The last time that I created a playground for experimenting with machine learning using Apache Spark and an InterSystems data platform, see Machine Learning with Spark and Caché, I installed and configured everything directly on my laptop: Caché, Python, Apache Spark, Java, some Hadoop libraries, to name a few. It required some effort, but eventually it worked.
CREATE TRIGGER "SQLUser".SYM_ON_U_FOR_SYM_CHNNL_125123 AFTER
UPDATE
ON "SQLUser"."sym_channel" REFERENCING OLD ROW AS OLD NEW ROW AS NEW FOR EACH ROW
DECLARE @var_row_data VARCHAR(16336);
WHEN
(
1 = 1
and 1 = 1
)
-- I will use OLD/NEW alias here set @var_row_data = ''; LANGUAGE sql insert into mytable (id, data) values(1, @var_row_data);
I want to have a generic PercentOfAll measure that can be used for any dimension of the cube. This PercentOfAll should act like Count, but instead of showing the number of rows in a cell, it will show the percentage (100*number of rows for that member/all rows) .It should not matter what dimension is being viewed. Ex
There are a total of 100 rows in the source class.
Is there a way to have a context property be available across 2 or 3 different business processes? For example if I am setting a value in a context property within one business process and want to use the same context property in a second business process, is this possible? Is this something that can be done with the "Context Superclass"?
https://www.youtube.com/embed/z9O0F1ovBUY [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
I have been playing around with the deployment tools from Ensemble (mostly based on this article https://docs.intersystems.com/latest/csp/docbook/DocBook.UI.Page.cls?KEY...) , but so far the one thing I still couldn't figure out is how to release the fact that I deleted files/components from my source production and that will get reflected in my target production.
When I try to send a hl7 message via TCP / IP with SSL certificate, I get the following error:
ERROR <Ens>ErrTCPTerminatedReadTimeoutExpired: Tiempo de espera de lectura TCP (600) superado mientras se esperaba el terminador SegTerminatorAscii=13, on |TCP|35000|6596, datos recibidos =''
I have increased the response time to 10 minutes but still, I do not get to receive the message.
Hi! I'm Andreas, the developer of Caché Monitor. Caché Monitor is a database tool with special feature for InterSystems Caché. Since today also with support for InterSystems IRIS! A public pre release is available for download: https://www.cachemonitor.de/cache-monitor-beta-releases/
I believe my current villain is “incremental compiling.” I was able to compare early compile output with later compile output:
Compilation started on 04/10/2019 10:46:32 with qualifiers 'cuk'
Compiling class UnitTests.Task3496 incremental compile detected
Compiling routine UnitTests.Task3496.2
Compilation finished successfully in 0.800s.
For the first time, InterSystems will be part of the WeAreDevelopers World Congress in Berlin, Germany, which brings together developers, IT experts and digital innovators to discuss and shape the future of application development.
From 6 to 7 June, we’re ready to welcome you at our booth #A5 and show you how InterSystems technologies enable intelligent interoperability and accelerate the creation of powerful, data-driven applications.
Schedule your individual meeting with InterSystems @ WeAreDevelopers in Berlin by quickly filing out the form on our event page [https://dach.intersystems.de/WeAreDevelopers2019] – the first three applicants will receive a FREE ticket!
I am attempting to use Node.js to perform a simple query against the database using some identifying information to get the ID. This is what I am attempting to do:
var irisobj = require('iris'); var myData = new irisobj.IRIS(); var result = myData.open({ path:"C:/InterSystems/IRIS/Mgr", username: "USERNAME", password: "PASSWORD", namespace: "YOURNAMESPACE" } );
We're pleased to announce that that InterSystems IRIS Community Edition is available on the Docker Store! InterSystems IRIS Community Edition is the no-cost developer edition designed to lower the barriers to entry to get started with IRIS. Now that it is listed on the Docker Store, running an IRIS Community instance is as easy as -
Hi! I've been fiddling with linked tables to get data from other servers, and I encountered a problem that I'm curious about. Maybe I'm not using these tools as intended or there's more going on, so I'm asking here.
I'm running a query on linked table A, something simple like this:
I was trying to create a query that can be exposed as a stored procedure (function actually) that would return a resultset with a random number of columns.
Unfortunately, it seems that unless I specify the ROWSPEC annotation on the Query method, I won't get any columns exposed. I was hoping to implement QueryNameGetInfo method and specify the names and number of columns I would be returning dynamically. But it seems that GetInfo information is simply ignored.
Unix: I'm trying to switch to using the $ZF(-100 function in place of $ZF(-1. I have it working for a simple, single, cmd, but can't figure out how to pass a piped sequence. Maybe this just isn't allowed?
For example:
S sc=$ZF(-1,"ps -Af | grep username")
works, but I can't figure out how to format this for the $ZF(-100 function.
So i'm having this problem with the task manager, the tasks simply stopped running. I had a problem with queued massages and trying to figure out what to do i'm afraid I may messed up something else, can someone help me ?