InterSystems Developer Community is a community of 18,183 amazing developers
We're a place where InterSystems IRIS programmers learn and share, stay up-to-date, grow together and have fun!

Receiving HL7 messages from our EMR, and processing to send out to downstream system. These are SIU_S12 messages, with a custom ZOR segment added by the EMR to include order information. The purpose of including this segment is to allow us to only send messages to the vendor that contain a procedure ID that is included on the list of procedures desired by the vendor. These values are in a LUT with the procedure ID in the key field, and a value of 1.

0 2
0 25

I would like to send bundle object to a router and then
perform a DTL with in the routing rule ( this BundleObject as a source to my DTL).
during this process , I get below exception

reqStream = json content in a stream

Set BundleObject=##class(HS.FHIR.DTL.vR4.Model.Resource.Bundle).FromJSON(reqStream,"vR4")

Set tSC = BusinessService.SendRequestSync("SAMPLE_ROUTER", BundleObj)

expection :

0 3
0 22

Hi guys,

Does anyone know what this error message mean, we are trying to save records to My.Package.List
class and sometimes we get this error message at first but when I try to save the same record again from the Terminal it does save fine?

0 1
0 23

Good morning, 🙂

I would like to ask a question, which has to do with how to manage %GlobalCharacterStream representing JSONS.

Thank you for reading this question, thank you for your help, and thank you for your time and attention.

Specifically, in a certain Process, we were querying 2 Operations, whose response we were converting to a Property called "informesAutorizadosRangoFechas" (reportsAuthorizedInRangeDates) which is %GlobalCharacterStream whose content is a JSON with the same structure.

0 4
0 74

Hi Community,

Play the new video on InterSystems Developers YouTube:

Clinica Universidad de Navarra - Low on Budget High on Innovation @ Global Summit 2023

https://www.youtube.com/embed/Yyb_I80oJF0
[This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]

0 0
0 26
InterSystems Official
· May 21
InterSystems Platforms Update Q2-2024

Here’s our Q2’2024 quarterly platforms update. If you’re new to these updates, welcome! This update aims to share recent changes as well as our best current knowledge on upcoming changes, but predicting the future is tricky business and this shouldn’t be considered a committed roadmap.

With that said, on to the update…

1 0
0 89

I want to customized retry frequency of transaction so that 503 error can be prevented while pushing FHIR bundle to the LFS.

How do I achieve this?

Reference the following algorithm to increase the retry interval of IRIS transaction

An exponential backoff algorithm retries requests exponentially, increasing the waiting time between retries up to a maximum backoff time. The following algorithm implements truncated exponential backoff with jitter:

0 2
0 33

I created %ZSTART, having just learned about it. I included some minimal code to test it, as follows:
%ZSTART
SYSTEM
WRITE "SYSTEM.",!
QUIT
JOB
WRITE "JOB.",!
QUIT
LOGIN
WRITE "LOGIN.",!
QUIT
CALLIN
WRITE "CALLIN.",!
QUIT

Now Terminal still works as normal, and sure enough I get "LOGIN" displayed, but when I start Studio I get an error message such as the following. I wonder what Caché ODBC has to do with it.

0 3
0 68

Continuing with the series of articles on voice file management, we are going to see how we can convert text into audio and receive the file with the chosen voice.
We will also explore how a service from OpenAI can help us analyze a text and determine the mood expressed in it.
Let's analyze how you can create your own voice file and how it can “read” your feelings.

2 0
2 27

Hello community members!

I am trying to connect to Caché database from Azure data factory using ODBC connection. The aim is to read data from Caché and write it in Azure blob storage using copy activity. The ODBC connection is successful and i can see the Caché tables, but i'm facing an error while reading the table contents.

I'm getting the below error :

0 4
0 246