Is there a way to generate an MD5 checksum of a file by performing a checksum against the file's parts? We potentially have to process large files and therefore cannot keep the entire file in memory. From what I understand checksum related logic is typically stateful, allowing file data to be passed through sequentially, however it appears cache script offers only stateless functions offering a simple method of generating a single checksum hash from a single string (or stream), correct me if I'm wrong?
We have a use case whereby very large HL7 MDM messages, which contain many OBX segments, need to be processed by a custom business process. Part of the process is to iterate over each OBX segment and set a value in the same field for each segment.
We are constantly running into issues where there are billions of Orphaned messages in our system that cause problems, and we have to manually run a cleanup to fix performance issues.
Could any one tell me, how the index works in cache db, Consider for example, I have a table called "Employee" and the fields as "EMPID, EMPNAME, EMPAGE" and I am having index for EMPID as IDX_EMPID and trying to get the record using the following query as
I have an HL7 message that contains NTE segment with line break inside.
when I try to transform it (with DTL) it stops reading the message after the line break inside the NTE because it treat that as "End of Message".
Is there a way to solve this in the DTL or I need to run a pre process on the HL7 message that identify the line break inside the NTE segments and replace it with an escape char?
Have you ever imported an XML schema from an XSD file? You might want to look at the original file again sometime later, but forgot where you put it.
Do not worry, that information is being kept as part of the import process.
The whole imported XSD schema is being kept in the ^EnsEDI.XML.Schema global. That global holds all the imported XSDs in your namespace. The first subscript is the name of the schema that you see in the portal.
We have an HTTP business operation where the messages gets stuck and we had to recycle the operation to resume the processing. We are trying to work around this (for time being until we find the root cause) using below code which should recycle the component.
Data for InterSystems products (table row data, object instance data) is stored in global variables. The data size of each global can be obtained by clicking the properties of the global you want to view from the Management Portal > System > Configuration > Local Database > Globals page, and then clicking the Calculate Size button on the Global Attributes page that appears. To display the data sizes of globals in a namespace, you can call ^%GSIZE utility on the terminal.
From time to time we develop an Ensemble Production with simple SQL Inbound data from external databases, we need to develop a few new classes. There are at least:
If you want to run an OS executable file, command, or a program created within an InterSystems product when the InterSystems product starts, write the processing in the SYSTEM^%ZSTART routine. (The %ZSTART routine is created in the %SYS namespace).
Before you write any code in SYSTEM^%ZSTART, make sure that it works properly under all conditions.
Not really familiar with how to secure URLs and using SSLs, but we currently looking to change our url from http:/www to https:/www so my understanding is once we get a certificate its a matter of creating a new SSL/TLS configuration and assign the certificate to it, then what how to apply it in out URL, or do we have to do this in IIS?
I'm having problems with an HL7 message that is not properly parsed by IRIS (any version, including Ensemble).
The message is a OML_O21 and is valid, in fact is similar of what is described in the document "IHE Laboratory Technical Framework, Volume 2a (LAB TF-2a) version 6.0". This is the sample from the IHE document (paragraph 2315):