I need a way to securely store credentials in IRIS that are used to access third-party APIs. These credentials include different clientId/secret pairs that I don't want to expose. While I know that it's possible to define Production Credentials in Ens.Config.Credentials, I understand that they are stored in plain text.
I'm trying to use LOAD DATA to insert 11k (11,377) rows of data. LOAD BULK DATA is not available for the version of IRIS I am using.
After calling LOAD DATA it says only 5,500 rows has been inserted. The LOAD DATA docs says any error rows are skipped and a count of skipped rows can be found in %SQL_Diag.Result however there are no results here. There are no errors in the xDBC error log either.
This link (https://github.com/CassiaNetworks/CassiaSDKGuide/blob/master/node_exampl...) uses Javascript request to open a a steam and collect data in an array (screenshot below) and I'm trying to replicate this with Cache script with HTTP request which is connecting Ok but don't know how can I collect data because the code is hanging at Set tSc=Httprequest.Get(HttpURL) and cannot execute S BLEConnect=Httprequest.HttpResponse.StatusLine or any line after it.
I need a help in the following error I am getting while running the SQL Query. I am running the SQL Query from different table using inner join. I run this everyday for last 90 days but for few days while extracting the data through this query I am getting this error. Please assist.
SQLException was thrown: Incorrect list format: 0 >= 0_SQLSTATE: S1000
SQLException was thrown: Incorrect list format: 0 >= 0_SQLSTATE: S1000
Does anyone have experience compacting and truncating IRIS datasets that are greater than 10 TB in size? How long did it take and what was the size? Did you run into any issues?
how can i use git for versioning while working in IRIS. I am having directory that i created while working in iris. How can i manage version control using git in this case.
Is it possible with vscode terminal or with a normal terminal in mac.
My understanding is that SQL gateway connection is a way where you can have two ensemble instances where in instance A you can have a class and then you create a connection from instance B and you can run queries for that table from instance A into B?
But as I checked the doc, it says that SQL gateways is a way to just import data from an instance to another?
I'm receiving in one of our interface a XML file as a %Stream.GlobalBinary, and in the source Application that produces this XML file, the user sometimes add a special character that force the XML file to have the header encoding to UTF-8, instead of ISO-8859-1 :
I am working on Converting Non HL7 message (Using record maps) into HL7 message.
1.Can anyone share few details how to save Non HL7 message into SQL table and the converted HL7 message into SQL table
2.In Message Viewer is there any SQL tables are linked to the session id or where the information regarding the message will be stored? Will the message trace details are stored in globals or in SQL table,If yes can anyone share the details in which tables or globals will it be stored?
Can %Service_telnet be enabled programmatically? I have the need to enable this feature so that I can automate a script I have for Cache, but unless the telnet service is on, the session cant connect.
We have been investigating how to activate a Server to Generate Tokens and an associated Resource Server to Validate the Token. This step, actually we have found out with the invaluable support of @Alberto Fuentes from Intersystems.
I have a test version of Cache installed on my laptop, but I cannot get logged into the database. I think my ID may be locked due to too many password failure attempts. I have not changed the Admin password or SuperUser password. Does anyone know what the default password would be for the Admin and/or SuperUser Id's? Or if someone knows how I can get in and unlock/reset my password, it would be much appreciated??
Curious if anyone has modified the HL7 TCP/IP Service and adapter class to receive an HL7 message, ACK it, and just dump it straight to a file? I don't want the messages in the engine at all - just receive and dump. Has anyone done this before and have code you're willing to share?