Hey Developers,
Here is a digest of the Developer Community videos on InterSystems Developers YouTube Channel in November 2020:
Hey Developers,
Here is a digest of the Developer Community videos on InterSystems Developers YouTube Channel in November 2020:
The aim of my question is to achieve this :

You can notice that some information of the PID segment are obfuscated.
I do have a solution but it involve a modification of EnsLib.HL7.Message. It's not a best practice.
Do you have idea or a simple solution that doesn't involve overloading every HL7 classes ?
Here is my solution :
https://github.com/grongierisc/hl7-msg-anonymizer
This is a modification of EnsLib.HL7.Message to not display Patient demography information in visual trace.
This feature is toggle by a resource named
Does anyone know how to convert UTC to Eastern time in the DTL? Input format "YYYYMMDDHHMMSS".
Can I use ConvertDateTime() function? If yes, what value do I use in these fields informat, outformat and outf?
Thank you.
Hi everyone,
I need to call a SOAP service using IRIS Interoperability and this SOAP service returns a Microsoft DataSet with a DiffGrams payload.
Do you know how to handle this kind of Objects ?
SOAP Payload Response :
<?xml version="1.0" ?>
<GetPatientsByClinicResponse>
<GetPatientsByClinicResult>
<xs:schema id="NewDataSet">
<xs:element name="NewDataSet" msdata:IsDataSet="true" msdata:MainDataTable="Patients" msdata:UseCurrentLocale="true">
<xs:complexType>
<xs:choice minOccurs="0" maxOccurs="unbounded">Hi Community,
We're pleased to invite you to the online meetup with the winners of the InterSystems Interoperability Contest!
Date & Time: Friday, November 27, 2020 – 10:00 EDT
What awaits you at this virtual Meetup?

.png)
This XML file does not appear to have any style information associated with it. The document tree is shown below.
WSDL
<definitions xmlns="http://schemas.xmlsoap.org/wsdl/" xmlns:SOAP-ENC="http://schemas.xmlsoap.org/soap/encoding/" xmlns:mime="http://schemas.xmlsoap.org/wsdl/mime/" xmlns:s="http://www.w3.org/2001/XMLSchema" xmlns:s0="http://herenit.com" xmlns:soap="http://schemas.xmlsoap.org/wsdl/soap/" xmlns:wsdl="http://schemas.xmlsoap.org/wsdl/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" targetNamespace="http://herenit.com">
<types>
<s:schemaHello,
In Caché Management Portal yo can see the SQL privileges per table.
For example : http://localhost:57772/csp/sys/sec/%25CSP.UI.Portal.User.zen?PID=P#3 (Page "SQL Tables").
Is it possible to get them by objectscript or SQL ?
Regards.
I am getting the date 20201121090000 in the HL7 message, How do I convert it to 2020-11-21 09:00:00 in a easy way?
I am currently doing it by extracting the first 7 values and splitting as date and time and then adding a hyphen using substring.
Is there an easier way by using $ZDATE? or something like that?
Hi!
If you see this page you are in the InterSystems Developer Community!
You are very welcome!
This is the place where you can read about and discuss InterSystems products and technologies: InterSystems IRIS, Caché, Ensemble, HealthShare, DeepSee, and iKnow.
What type of content can I find here?
We have three types of content on Developer Community(DC): articles, questions, announcements. And answers to questions. And videos.
The articles are about the best practices and experience with InterSystems technology and products.
Does anyone know about any documentation or sample code to create SFTP Business service in HealthShare?
Thank you for your help.
I'm running Windows 10 x64 Pro 20H2 with Intersystems Cache ODBC driver v2018.01.00.184. I've setup a System DSN using the 64-bit of ODBC Administrator.
I've been getting inconsistent results using my regular application (Microsoft Power BI) which I use through ODBC to query my hosted TrakCare T2017 instance.
Using Microsoft ODBC Test Tool (part of MDAC 2.8 SDK) I can verify I get the same errors as in Power BI. They seem to be ODBC driver related but I can't pin it down.
ODBC Test Tool GUI shows:
|
Full Connect(Default) Successfully connected to DSN 'Test DB'. |
I am looking for any pointers on how Intersystems IRIS Health can monitor a filesystem/Folder that user/s /applications can drop in CSV files via FTP and load the file to the IRIS DB . I understand that I will need create a record map for the CSV files, I am looking for any configuration references on how how to process files using file inbound adapters with the intent to pick up the CSV file as they are dropped in the target location and pass it to a Business process and ingest into the IRIS database
Any help would be greatly appreciated ...
Hi!
I need to solve a combinatorial optimization problem, like the Knapsack Problem.
I found some implementations, but they're memory inneficient. I also found libraries like this one from Google.
However, first I'd like to ask you guys for your experiences doing this kind of task in Caché/IRIS platform. Do you guys have implemented your own code or used a third party library?
Thanks, José
Hi
I have a table where i want to create join query like
(select * from patients where facility=abc) qry1
left join (select * from patients where facility=xyz) qry2 on qry1.patientid = qry2.patientid
Hi Community,
See how to use System Alerting & Monitoring (SAM) to view and customize metrics for a cluster of systems running InterSystems IRIS:
I have an application with maps that render and edit geographic data in a some layers using KML - geo data represented into XML file with layers, polygnons, points and metadata associated with it. Some maps are in GeoJSON too.
Is it DocDB the better option?
Hi Team,
What is the difference between Intersystems Cache DB and Intersystems IRIS ? I am looking for Cache DB installation details, but getting IRIS only everywhere. we are looking for a POC to test it.
Thanks,
Kranthi.
How InterSystems solutions handle C10k connections? ( https://en.wikipedia.org/wiki/C10k_problem )
For example, I want to create a social network on InterSystems platform.
In the scenario client (browser) => CSP Gateway => Cache can the Cache handle a large number of clients at the same time?
Quick analysis shows that CSP Gateway for each request opens a new TCP connection to Cache SuperServer (port 1972) allocating CSP session and license slot.
SuperServer has no queues or pools, and TCP connections are expensive (in terms of system resources).
Is there any real-world projects or tests/results?
Hi team
We are migrating the Code set of VMS to Unix.
In this process we are also taking care of all I/O operations of file.
In VMS, whenever we create a file, it generates a Version # associated with the file. Best part is we have a common library which we use to OPEN/USE/CLOSE/DELETE/RENAME/COPY the file.
Now, when we talk about VMS on cache, if we two processes creates a file "ABC.TXT", for 1st process it will create the version #1 and for second it will create #2.
This versioning cache handles internally, like which cache process is referring to #1 and which one for #2
Can someone help me to
I have a project to only filter certain pathology results into a downstream system.
Within a HL7 router and business I was planning on using a lookup table and either the exists() or Lookup(), but am having issues when using it with repeating fields or segments.
For example if I perform teh analysis per stated segment usign {} brackets this will work, as each stated repeat is assessed:
(HL7.{PIDgrpgrp(1).PIDgrp.PID:PatientIdentifierList(1).ID} DoesNotStartWith
As we all well know, InterSystems IRIS has an extensive range of tools for improving the scalability of application systems. In particular, much has been done to facilitate the parallel processing of data, including the use of parallelism in SQL query processing and the most attention-grabbing feature of IRIS: sharding. However, many mature developments that started back in Caché and have been carried over into IRIS actively use the multi-model features of this DBMS, which are understood as allowing the coexistence of different data models within a single database. For example, the HIS qMS database contains both semantic relational (electronic medical records) as well as traditional relational (interaction with PACS) and hierarchical data models (laboratory data and integration with other systems). Most of the listed models are implemented using SP.ARM's qWORD tool (a mini-DBMS that is based on direct access to globals). Therefore, unfortunately, it is not possible to use the new capabilities of parallel query processing for scaling, since these queries do not use IRIS SQL access.
Meanwhile, as the size of the database grows, most of the problems inherent to large relational databases become right for non-relational ones. So, this is a major reason why we are interested in parallel data processing as one of the tools that can be used for scaling.
In this article, I would like to discuss those aspects of parallel data processing that I have been dealing with over the years when solving tasks that are rarely mentioned in discussions of Big Data. I am going to be focusing on the technological transformation of databases, or, rather, technologies for transforming databases.
Developing a Full-Stack JavaScript web app with Caché requires you to bring together the right building blocks. Previously, I outlined the basic steps to install and connect Node.js to Caché and make it's powerful multi-model database capabilites available for use with Node.js. You can use Caché as a NoSQL-, document- (with unique key-level access!), SQL- and object-database with Node.js. When developing JavaScript applications, you'll see how powerful this combination is and makes Caché a perfect fit for Node.js.
In the first part of this article series I will show how to get started with the React framework, one of the most popular frameworks currently taking over front-end development. In the next parts you'll learn how to connect a basic web app to a Caché back-end.
You'll see, it's very easy to get started with this technology - you can even compare the amount of basic knowledge you need to COS because you only need to know a few basic concepts to start!
We have some code written by earlier team that read flat files data and create SDAs and push them to ECR Input Service. Now for every test cycle we have to load files to SDAs, check error log, fix the error, clear the ECR, and try new fixed flat files. I’d like to know if there is a method to validate SDA msgs prior to pushing to EFR Input operations ?
so everything we simple load flat files, create SDAs and validate them. If files all looks good with no SDA errros then we actually load them to ECR only once.
Hi,
I seem to be able to execute sqlproc using the convention defined here https://docs.intersystems.com/latest/csp/docbook/Doc.View.cls?KEY=GSQL_… provided the package/folder is a single level. As, soon I have a nested folder structure I get an error while trying to execute this.
For instance,
select id, Utils.Users_getRole(id) roles from users.users
works fine while,
select id, com.xyz.utils.Users_getRole(id) from users.users
does not work and gives, "User defined SQL function ... does not exist"
I am on Cache 2017. Is there a workaround to making this work?
Hi,
I am working with XML and I would like to be able to access parts of the XML document and return them as a string.
Hi Community!
We are pleased to invite you to the upcoming webinar in Spanish: "Improving Kubernetesperformance", on December 9, at 4:00 PM (CET).