A business service is part of InterSystems Ensemble interoperability production which is responsible for accepting requests from external applications.
While using custom business service i'm not getting the Acknowledgement . But i tried with standard business service im getting acknowledgement . can you please help me to figure out the issue. Thanks in advance!!!
We have a simple BS that Extends Ens.BusinessService with an ADAPTER = "EnsLib.File.InboundAdapter"; The "incoming" folder has more than 4M files, so we get the following error: ERREUR <Ens>ErrException: <STORE>zFileSetExecute+38 ^%Library.File.1
I've a Service utilising the Adapter EnsLib.SQL.InboundAdapter, which uses a Credentials item set with the details of a local SQL account. This currently works, however, we're looking to use the credentials of an AD domain account.
The domain account is a member of an AD security group, which has the required permissions on the source SQL database. I've checked that access is possible with this account via SQL studio.
I have a file passthrough service that should check for new new files every 10 minutes, process them and then wait again for 10 minutes, even if there are new files waiting in the source directory. The "CallInterval" value feels like a wrong tool for this need since it would keep the service "locked" for a maximun of 10 minutes before releasin resources. Is there a better way than CallInterval to schedule the service to poll for files every 10 minutes?
In the Inbound SQL Adapter settings, is it possible to specify more than 1 field as the Key Field Name?
Because of the way the Query is being index in Ensemble by the Key Field Name, sometimes transactions get missed and I would like to see if we can add an additional key to the mix to ensure all the transactions are picked up. In this case the InterfaceTrigger is an ID that is auto generated by the table, and I would like to use that as well to ensure we don't miss transactions, and it does not throw any warning messages when it executes the Delete Query.
The Telegram Adapter for InterSystems IRIS serves as a bridge between the popular Telegram messaging platform and InterSystems IRIS, facilitating seamless communication and data exchange.
I have a production with one Business Host - a Business Service which I need to scale automatically to consume ~80% of CPU time. Business Service pulls data from a (non-FIFO) queue so that I can adjust pool size without any issues.
So far, I'm planning a different BS running every X seconds and sampling CPU with $system.Process.GetCPUTime() and scaling the pool size of the main BS up/down based on that metric.
Has anyone tried something similar? Any advice/code samples would be appreciated.
The Complex Recordmapper specify a map for headers, body, and trailers, but it expects all three to have fixed Leading Data to identify the record type. But what if the Header and Trailer have Leading Data, but not the body records? I can't seem to find a way to do this. For example:
My team is working on redesigning an Interoperability solution that currently runs on a HealthShare server that is part of a mirror. Most of the messages are delimited records processed using Complex Record Mapping. We were told to utilize cloud services available in AWS, use containers, autoscaling…
I am trying to interface LAB instruments with LAB System using ASTM over TCP. I read the documentation here . But I can't understand how to connect instrument and LAB system in same port, please see the screenshot of interface below. Please help.
I want to deploy Iris Interoperability Production in multiple containers . The production has File Services that process files from EFS location. All the files being processed by multiple containers are in the same directory. The standard Adapter class does not have Check Complete option to lock a file so that other containers File Services looking in the same file path cannot process the same file.
I updated the Inbound File Adapter class to offer another Check Complete option LOCK that will lock the file until it is processed and archived.
Sometimes we need to convert FHIR message to HL7 V2, e.g. to register a patient to the PACS system. In this article, I will explain the steps to achieve the desired by using IRIS FHIR Server production.
Below are the steps we need to follow:
Make sure FHIRServer production is started.
Register Business Service with FHIRServer endpoint.
Define Business Processes to convert FHIR message to SDA and then Convert SDA to HL7 v2.
Post JSON resource to FHIRServer endpoint and get HL7 V2 response.
Let's review the steps in detail.
Step 1. Make sure FHIRServer production is started
Open the production page and make sure Production is started. In the next step, we need to make sure business service HS.FHIRServer.Interop.Service is registered with FHIRServer
We are getting more and more request wondering if we could send/receive data via HTTPS to the outside world from within our Hospital Network. As you can imagine our Ensemble/Cache productions are not exposed to the DMZ or has access outside of the network. We only communicate with external vendors through a VPN, so communicating not using a VPN is rather new to us.
Currently there is a project to get rid of using Proxy, and instead of through a Load Balancer that can use rules to filter out traffic, which adds another layer of complexity.
I am trying to transform CSV file into SDA3 format by using Studio code. I am getting "No Stream contained in StreamContainer Request" Error at the business operation after converting CSV file to SDA3 Format but not able to generate the file in business operation
We have 3 env. using different ports schemes Is there a way to create a rule if msg comes on ports starting 22 then route it to devoperation, if it starts with 33 then route to testoperation, if its 44 route to stageoperation?
I like to know if we need to have the message in a file to process a Record Map?
I am working with Interoperability Production that processes files /messages using Record Maps. My team was asked to redesign the solution for deployment in AWS. We use containers. We had problems with having multiple containers processing files from the same directory. We are considering Amazon Simple Queue Service instead of having files on a shared file system.
We have a vendor that every couple of days will just stop transmitting messages, but still hold the TCP/IP connection open. No matter how many times we troubleshoot and talk with them, they don't seem to think its an issue with system. Normally if I just restart the service it will get the data flowing again.
I have a system that is sending Line Breaks\Carriage Returns within the text of a field that is breaking the parsing of the message. Does anyone have a way to get around this?
I have a Business Service which retrieves data via a SQL adapter and writes the data in text file. My production is doing the job, but instead of the Business Operation output being one file, I am getting many files, with 1 , 2 or 3 rows only.
I am having an issue with a file not being deleted from the FTP server and returning an error to this effect. The file, however, was fully processed. This causes the file to process again.
My question is I seem to recall a flag that can be set to enable more detailed tracing to be done on the FTP service/adapter. However, I have not been able to locate this. Does anyone recall such a logging setting or was I just dreaming?