I'm dealing with a situation that leaves in doubt what I understand about CACHE.WIJ (C:\Roche\CobasInfinity\HealthShare\mgr).. Journal files store records already written to the base.. and CACHE.WIJ records not yet written to the base.. theoretically, the data contained in CACHE.WIJ is temporary... until the record is written to the base (consequently generating journal).
We have developed a REST Operation. We wonder how could we face that String do have a character limit.
We just need to convert httpRequest.HttpResponse , which is a stream , because $isobject(response.Data) returns 1; which has a JSON inside it, and we need to convert it to a Ens.Response.
This response is structured as same as the JSON's properties:
I am having issues trying to obtain a JDBC driver which is backwards compatible with java 6 for a solution which will call my IRIS instance, I have already logged a WRC ticket which is looking unlikely that there will be any development to create a driver for such legacy tech.
The Java 6 app is end of life, however won't be replaced until after my project goes live, hence why I ask the question.
First of all, thanks for your help, and time, reading this questions, and answering to this doubt.
We have the following use case: to convert a HL7 ER7 ORU_R01 v2.5 message, to a Ens.Request Message which represent it with a custom data structure (each segment is a group of Properties).
To be more precise, the target Ens.Request Message's structure looks like:
I have a problem attempting to store a base 64 encode stream into the OBX:5 field of a HL7 message, by using the StoreFieldStreamBase64 method.
I'm using a transformer to input the contents of a pdf within a stream container (Ens.StreamContainer). The output is a HL7 message (EnsLib.HL7.Message 2.5:ORU_R01).
The transformer is being used in a business process.
This is the code used in the transformer to encode the stream to base 64:
First of all; thanks for your time reading this post, and thanks for your answers and help. I am really grateful.
We have the following need: Send a DICOM Find Document to get the complete list of a patient's studies from an Outbound System, which we simulate with the tool called "dcm4che" specifically with these two commands:
1º.Initialize a DICOM database in simulator. We will use this database to run queries using DICOM C-FIND commands:
Hello! I just had a quick question for anyone out there with more experience on deploying Ensemble productions.
I'm currently trying to export my Ensemble production WITH the Business Partners so I don't have to rebuild or add to that table after I import my Ensemble Production. I know that this is a small thing and doesn't actually affect anything in the production in terms of performance, but I like to have it for better documentation.
I think there have been articles published here before but I'm struggling to find them:
I'd like to be able to create my own customised version of the official IRIS Community Edition Container, eg with Node.js and a number of modules pre-installed. So, for example, can I create my own Dockerfile that begins with something like:
FROM containers.intersystems.com/intersystems/iris-community-arm64:2023.2.0.227.0
I am looking for an implementation strategy to create/update organizations in a local ISC based FHIR server. Thoase organizations do have a certain dependancy to one another. For Organization resources the "partOf" property symbolizes that. It is planned, that a Interoperability Production (supporting FHIR Server) receives a bundle which contains a batch export of all organizations of a customer (Hospital). The task is to bring those data into the FHIR server create and/or update organization.
Is there a difference in outcome between the two screengrabs below?
In both cases, when certain conditions are met, a transformation is called and the output sent on to two targets. In the first case we surmise the transformation is called twice, and the output of the first run sent to the first target, the output of the second run to the second target. In the second case we surmise the transformation is called once, and the output duplicated and sent to the two targets.
Hi all, I am trying to compile my swagger class file with swagger JSON which is a valid JSON as I created it and verified from swagger.io 2.0 spec. But getting below error. I am not very clear with the error message. Please help me resolve the same.
ObjectScript error: <CLASS DOES NOT EXIST>removeprj+7^%occClass *SwaggerClass
> ERROR #5091: An error has occurred while removing projection SwaggerClass:Reference.
> ERROR #5030: An error occurred while compiling class 'SwaggerClass'
1. The %ObjectSelectMode is working perfectly in the DynamicSQL for the object property in direct query. However this is not supported for class query. Is there any way to resolve this / Get the object value for class query result.
I'm trying to configure a specific process which dynamically sends messages to different endpoints based on datalookup keys, I've configured this aspect. What I'd like is to be able to visually see these connections without hardcoding them so is there a way to dynamically link them, i'll share what I tried below.
I have created a FHIR endpoint and send the FHIR resource to FHIR Interoperability production class which is HS.FHIRServer.Interop.Service through the the created endpoint(/r4). I can able to see/get the HS.SDA3.QuickStream(It use CacheTemp.HS.Stream temporary global) on the fly(In between the process). I'm unable to get the FHIR resource from the QuickStream once the process completed. Is this HS.SDA3.QuickStream is wiped out from the system once process completed?
Currently I have a program that is going to be used to compare global nodes between namespaces. Using nested FOR statements I am stepping through global nodes and comparing between the namespaces, if the global values are different I then parse the node comparing each delimited segment.
This is working.
I want to now be able to set what global I am in as a variable to expand on this concept.
I'm at a loss with configuring the IRIS ODBC driver (v02.10) on my Amazon Linux machine, connecting to a REMOTE IRIS backend. It should be simplified by not requiring a DSN (the app uses a DSN-less connection and generates the connection string). I have
We have some custom tools that do project based deployment from environment to environment. In VS Code, we're struggling to find the best way to export PRJ files that contain the contents of a project the way we were able to in Studio. We can edit existing projects, and in server side editing we can create projects, but we can't seem to extract the project files (Not the classes, the wrapper) to store in our source control solution for deployment purposes.
Hello community, I need to solve a complex but trivial issue.
Given:
Two tables "Comment" and "Post". Each one contains an "Author" field, which is essentially a user ID. In these tables, each user ID represents an author. The goal is to count all participants together and then group them by month, language, and other metrics.
The question is how to do this within the IRIS ecosystem. Is it even possible to take two tables, get distinct data from them and then combine into one cube?
I've an fileupload and button that executes Upload() clientmethod to upload files in my database and everything works fine when I run the application in the server where it is located, but if I run it from a client machine the %OnSubmit class method doesn't get invoked when form.submit() get executed so how can I fix this?