I am testing asynchronous Bulk Data retrieve with a large EHR vendor. When I retrieve the bulk data, they are returning ndjson (CONTENT-TYPE of "application/fhir+ndjson; charset=utf-8"). If I attempt to process that content using JSON functions - I get
ERROR #5035: General exception Name 'Parsing error' Code '3' Data ''
For example, the following generates the error
s objArray={}.%FromJSON(quickStreamOut)
The error makes sense in that ndjson is not a valid json format (I assume at least).
Hi, We have a business process that receives back a character stream that has csv content from a SOAP Operation call. I was hoping to make use of the record mapper to map the content to a record map to be able to process the transactions.
I only see examples/documentation of how to use a File or ftp business operation/service to map the stream to a record map.
Is this at all possible to get the csv stream into a record map batch object.
Artificial Intelligence (AI) is getting a lot of attention lately because it can change many areas of our lives. Better computer power and more data have helped AI do amazing things, like improving medical tests and making self-driving cars. AI can also help businesses make better decisions and work more efficiently, which is why it's becoming more popular and widely used. How can one integrate the OpenAI API calls into an existing IRIS Interoperability application?
For some time I have been planning to do some type of proof of concept with the Workflow functionality, which, like so many other functionalities present in IRIS, tends to go quite unnoticed by our clients (and for which I say mea culpa). That's why I decided a few days ago to develop an example of how to configure and exploit this functionality by connecting it with a user interface developed in Angular.
A simple production that enables FHIR transaction bundles to be loaded into InterSystems® FHIR® Server via Box and Dropbox. Using the included MFT Connection Components and a 14 liner Custom Business Process, this production will process your transaction bundles to FHIR Resources for immediate consumption with Harry Potter like wizardry. Great for Hackathons, Research and FHIR® Cocktail parties.
Maybe you have to implement scenarios that don't require a FHIR repository, but forwarding FHIR requests, manage the responses and maybe run transformations or extract some values in between. Here you will find some examples that can be implemented using InterSystems IRIS For Health o HealthShare Health Connect.
I'm trying to write a method that runs an OS command with $zf(-100,"/ASYNC") and then waits for it to finish before returning, but I can't figure out how to check if the child process (in $zchild) is still running. $System.Process.State($zchild) always returns "RUN" even once the child process is gone. Is there some better way to check this, short of running another OS-specific command and processing the output (which is just really annoying and unelegant)?
Apache Zeppelin it's a Multi-purpose notebook that allow you:
Data Ingestion
Data Discovery
Data Analytics
Data Visualization and Collaboration.
Apache Zeppelin interpreter concept allows any language/data-processing-backend to be plugged into Zeppelin. Currently Apache Zeppelin supports many interpreters such as Apache Spark, Apache Flink, Python, R, JDBC, Markdown and Shell.
IRIS for Health — is the world’s first and only data platform engineered specifically for healthcare. It empowers you to rapidly create and scale the industry’s next breakthrough applications.
I'm looking for information, suggestions, video (courses) on make a lot more use of the documatic system already built into Cache,
Probably talked of before within this forum, but after an hour or two of searching, can't find the articles that I mmay have seen before.
I want to do a lot more documenting of my installation, we use and actively comment classes and methods, but I want to take it further by actually documenting the "why" and high level "how" my system works. Documatic seems a good place to put that knowledge.
We are Upgrading the Application from Ensemble 2009 to 2016.1. We have imported the user, Roles and Resources.
But the Sql Tables and Sql Procedures are not being visible in Management portal -> Security Management -> User.
Please give me a solution or basics steps for importing those.
In Ensemble 2016.3 ‘Java Business Hosts’ provides the means to create new Business Services and Business Operations using Java without need for any ObjectScript coding. There are two major reasons to use this capability:
In general Global Streams are data objects embedded in Classes / Tables. Using and viewing them with SQL is normally a part of the access to the containing tables.
SO WHAT?
During debugging or searching for strange or unexpected behavior there could be the need to get closer to the stored stream. No big problem with direct access to Globals with SMP or Terminal. But with SQL you are lost. So my tool provides dynamic access to Global Streams wherever you may need this Special thanks to @Oliver Wilms for the inspiration for this tool.
I search how to send telemetry data from IRIS to azure. It is possible in other languages with openTelemetry SDK. But there are not SDK for interSystems platform. We would like something totaly automated/integrated.
(We already have telemetry data with prometheus and grafana).
Working on integrating with O365 Sharepoint REST API. I would want to know if anyone can share their experience with integration with Sharepoint REST API and how they implemented security?
I'm working in an application that uses %SIMILARITY to find matches among a set of documents that vary greatly in length. It's generally good but I've noticed issues with ranking short partially-matching documents over longer documents that match the search string entirely.
Hi All, I am implementing REST API's in IRIS using the below link as a guide. I chose this approach so i could have a single entry point for requests regardless of namespace
a sample form the HIHLib.REST.Server class i am using is below. my question is does IRIS and CSP cache responses for requests to the CSP.REST class? and if so what parameters or properties do i need to enable?
We have been storing raw messages in a MySQL database for DR and ad hoc purposes. We are thinking of using an Ensemble instance as our data lake instead. We could segregate the source data by namespace or by global. But either way we'll want a custom global to index the data for data retrieval performance purposes.