How can I effectively unit test a Business Process Language(BPL) that incorporates three code activities calling methods from distinct ObjectScript classes?.
As Intersystems Cache for Windows comes with a custom Apache version (2.4.52) and has some security risks, so not advised to be used for Production servers/environments. I want to install the new version of Apache (2.4.58) and replace it with the built-in version of it. Is there an easy and convenient way to perform this upgrade? My target is to completely replace the built-in version and replace it with a new version so that it passes the security scan.
Here is the link to the security risk mentioned by Intersystems:
First of all thanks for your help reading and answering this question, and thanks for your time.
We have been investigating how to handle the following need:
Our client which has the system which reveices our ESB's responses needs a SOAP response as follows, where all the ACK is inside a top XML tag called <ACK>:
I have ROUTINE for Get Data end of day. Routine Name is "getTicket.mac" . previously i execute at studio using this command to get data
w $$getTicket^production.etl.getTicket() and the result success as we want, in this case i want execute routine using BS ( Business Service ) and i want this routine execute every night at 12.00 AM
I want to create a task in the task manager to manage the member details. so what i have done let me first brief you :-
1) Created a persistent class to store the custom csv details into a sql table.
2) created a business service to take the custom csv file from a inbound folder and use this class (EnsLib.RecordMap.Service.FileService) to use RecordMap.
First i have created CSV Wizard then RecordMap then DTL and map the fields with csv values.
I've created an Object Gateway via the System Management Portal, and installed the following Java classes to allow connectivity between the IRIS server and RabbitMQ ...
RabbitMQ.jar
amqp-client-5.18.0.jar
slf4j-api-1.7.36.jar
slf4j-simple-1.7.36.jar
I've used the Java Gateway wizard via Studio and compiled the java classes that this would refer to, and a quick test worked on Day 1. However, on Day 2, on trying to connect, I got this error :
I'm facing issues uninstalling Ensemble software from my system. I've tried using the command first ensemble-2018.1.0.184.0-win_x64.exe /instance LOCALENSEMBLE /qn REMOVE=ALL. Additionally, when I try to uninstall it through Settings > Apps and Features, it doesn't get removed and lists the app again there.
Has anyone encountered a similar problem or could provide guidance on the proper uninstallation process for Ensemble? Any help would be greatly appreciated.
When running a sql insert query with CONVERT('datetime', '2023-12-12 11:27:00',120) in an insert statement into a table this works on one server and not the other.
Im using an ORU_R01 2.3.1 message for converting to pdf. For the stylesheet, i need all the OBX segments under each OBR segment with custom table headers. Can anyone help me with the xslt?
I am new to Deepsee (BI intelligence) concepts in IRIS. I am looking for a solution of below question
Question: We have a server where we are currently storing archive PDF reports of our application.
We want to add and view these archive reports (PDF format only) in dashboards. Is there any way we can do this? Or any article which can help us performing this task.
I am attempting to create a new table by executing a SELECT statement that involves joining multiple tables. However, I encountered an error during execution: '( expected, AS found^Combined AS.' I would also like to create a cube based on this SQL table. However, during the cube creation process, I am prompted to specify a source class, and I'm unsure which class to use as it requires an existing class. Could you please help me identify the issue with the table creation, and provide guidance on selecting the appropriate source class for the cube creation?"
[InterSystems IRIS SEVERE ERROR HSFLIVE] [Utility.Event] ISCLOG: WorkMgr CountWorkers: Found dead job ns=%SYS rtn=%SYS.WorkQueueMgr data=$lb("job","3209241","jobnum",67,"group","Run","category","SYS","breadcrumb","","jobtype",59)
01/10/24-09:00:54:225 (14316) 2 [Utility.Event] ISCLOG: HttpRequest Error while reading from web server $ZE=<EXTERNAL INTERRUPT>zRead+28^%Net.HttpRequest.1 ns=BDHEALTH rtn=%Net.HttpRequest.1 01/10/24-09:00:58:210 (5096) 2 [Utility.Event] ISCLOG: HttpRequest Error while reading from web server $ZE=<EXTERNAL INTERRUPT>zRead+28^%Net.HttpRequest.1 ns=BDHEALTH rtn=%Net.HttpRequest.1
I am trying to get data out of a cache backup. I am completely new to Intersystems Cache.
I got a docker image of Cache from this post:
I want to get the data from my backup into this docker image. First, I need access to a gui and a cli, then I need some instructions or documentation for actually pulling in the backup data.
Thanks! This is all new, so pardon my lack of knowledge here. If I can provide any more information let me know.
Hi all, Does anyone have experience with obtaining logs from Caché databases and parsing them now? Please leave me a message.
I ask because my project involves some hospitals that used Caché when creating databases over 10 years ago. Currently, we cannot replace the database, but we need to check and parse the database logs from the Caché.
I hope this message finds you well. I am reaching out because I have encountered an issue with a new dimension I created, named "Region," and I could use your assistance in resolving it.
The problem is that when I open the cube analyzer, the "Region" dimension does not populate data as expected. Instead, it merely displays the text "sourceRegion," which I specified as an expression in the architect.
I'm trying exported a project from version '2012.5' to '2018.1.4' but is returning ERROR #7602. Please, could you help me. How can I do it? Below is steps that I did:
We have a service account that is running queries to pull some data. It has read only access to the tables, but, unfortunately, the tool is starting a transaction before it reads the database. The query takes more than 20 minutes (21 - 23 minutes), so we're getting the console message that the transaction is open longer than 20 mins. Is there a way to prevent a user/service account from being able to start a transaction since it only has read access to all of the tables in the database?