I am trying to track down a problem we saw this morning with our TEST environment. We had a momentary issue where InterSystems HealthShare Health Connect could not connect correctly to LDAP. When we tried to login and could not connect to LDAP, the system would Delete our users.
the Test LDAP function would return a "Can't contact LDAP server". I went through the Certificates, made sure they had the correct permissions and were not expired.
Since version 1.93 of VS Code, the “import and compile” process has been very slow (more than an hour). Have you encountered this issue and do you know how to resolve it?
I am trying to use IRIS for Health as a Facade for an external FHIR Server, where IRIS provides the proper authentication.
The client authenticates using a bearer token obtained from the IRIS OAuth2 server via a jwt client assertion.
The IRIS endpoint, however, returns a 401 as soon as I remove the Unauthenticated access
Is there w way to make this work through configuration?
Hello, I try to develop a REST interface where I need to interact with legacy MUMPS routines. How can I pass in input to a Read without modifying the legacy code?
I think in linux I can execute command < inputfile to read from file, but how does it work in ObjectScript?
I am still trying to understand and come up with a use case to use FHIR as an enhancement to our existing HL7 messages. Does anyone have an example of taking a FHIR response and transforming it into a SDA?
How does the workflow go? Does it go through 1 BP?
Hi, I'm working on a large extraction from a database and need to parallelize the processes during the extraction.
Here's what happens:
The user starts the extraction from.
Several jobs are started using the Job <Method> instruction.
At the end, the user expects to find a document containing the results of all the extractions. What I'd like to do is start a new job that checks whether the previously started jobs have finished or are still working and consequently produce the document.
I have a theory that I could use an HL7 Data Transformer in a rule without specifying a specific DocType. I want to use the same DTL for a variety of different DocTypes. Here is my very simplistic DTL so far:
Is there a difference in outcome between the two screengrabs below?
In both cases, when certain conditions are met, a transformation is called and the output sent on to two targets. In the first case we surmise the transformation is called twice, and the output of the first run sent to the first target, the output of the second run to the second target. In the second case we surmise the transformation is called once, and the output duplicated and sent to the two targets.
When developing productions there are many places that we input endpoint information (IP/port) especially when there are dozens of operations going to the same environment. What later happens is that the IP changes and we have to then go into each operation and update the IP.
we converted from Ensemble 2014 to 2018 and now getting the attached error which looks like Manage Method don't exist anymore, has it been replaced by a new one?
How do you iterate over the array of namespaces returned in the result byref from the call below?
DO ##class(%SYS.Namespace).ListAll(.result)
Running "zw result" displays the namespaces (as individual subscripts in result) in the Terminal but I want to be able to programmatically iterate through each namespace name.
I have an SQL table as below (columns - ID and Value)
I will get an ID number (first column) as an input from FTP inbound , and after that I have to query table taking ID as an input to get all the values for the ID
as -
if ID = 11 ; QRY output = aaa,bbb
if ID = 22 ; QRY output = xxx,yyy,zzz
can someone help me with query
ID Value ---------- 11 aaa 22 xxx 22 yyy 11 bbb 22 zzz -----------
Is there a way to search for a string or use of a function in all namespaces in one go.
Studio option "Search in files"/Ctrl+shift+F search all files in the current namespace. We have 13 namespaces and wanted to know if there is an easy way to search in all namespaces in one go.
Hello every one, here in our company we are trying to develop a tool to help us in our "Code Review" that today is completely done for another developer.
So I need to develope a tool that reads a class/routine (Already done) and identify if in the current line there is some abbreviated command, that is against our policy of codification, for example:
s variable = "test" d ..SomeMethod()
So in this cases I want to warn the original developer that the code has parts that need to change the "s" to "Set" and de "d" to "Do".
Hi there, I'm wondering if anyone has run into an issue with <FILEFULL> when building an image from the ISC image? Specifically what's happening in our build is we are pre-loading our codebase into the image to make deployments faster and setting up source control, etc. When loading our libraries however we get hit with a <FILEFULL>. The resource limits on docker are pretty beefy and when observing resources on both the machine and container level we don't hit any issues. Oddly, this only happens when using the ARM64 version.
during an upgrade, a customer wants to load custom schemas in a particular order. For this, they renamed a file EPIC_MDM.HL7 to 1EPIC_MDM.HL7 Upon importing this file into their server using VSCode, the custom schema was renamed to 1EPIC_MDM.HL7 inside the <Category>
This is impacting the upgrade and they are looking for a way to import a custom schema without renaming the schema itself.
An IRIS.DAT file was removed, as it was not needed anymore. But the database was mirrored, so it still shows up in the mirror monitor and database list. How can this be fixed? There is no backup of the .DAT file so it cannot be restored.
I'm a bit confused about the number of elements in an XML object. I have an object of the class EnsLib.EDI.XML.Document The xsd structure ( DocType) is an SDA Object
I'm getting an unexpected behavior when using pandas function to_sql(), which uses sqlalchemy-iris. After the first execution, a transaction seems to be opened and all rows inserted are lost after closing the connection: