I'm new to Zen reports and I'm wondering if what I'm trying to do is possible. I have a classmethod that accepts three parameters to generate an xml file from a global. I've set up my Report Definition to call the ClassMethod that generates the .xml file; however, I'm trying to figure out how to pass the parameters to that ClassMethod from the Report Definition. The example in the documentation shows parameter passing using SQL (which I'm not using), so I'm curious if this will work.
we are having a problem with the Cache.dat file is not expanding anymore since it has reached 2.2TB, we still have extra 260Gb of free disk space in our drive but for some reason it's raising a FILEFULL which is a result or the database not able to expand.
I am trying to import the SalesForce Enterprise WSDL in InterSystems Ensemble Studio using the SOAP Add-In. However, when I do so I receive a class dependency loop, Error #5316 during compilation. The WSDL is the standard WSDL provided with SalesForce and works fine in SoapUI. In the SalesForce WSDL it is allowed for an object A to include an object B as an element, while at the same time object B is allowed object A as an element. I think this is what causes the class dependency loop for InterSystems. Does anyone has any suggestion how I can circumvent this error?
I am Brazilian and I am starting to work with the latest version of CACHE, and I would like to know where I can see an example of a TRIGGER, I would like to validate the fields before writing the actual data in the database.
in advance, thank you very much! and sorry for any mistake in concordance in english
I work with IRIS for Health 2020.1 and I attempted to set up Enterprise Systems to populate Message Bank. I see entries in Message Bank Event Log from the defined Enterprise Systems. I can browse Messages going back before I set up Enterprise Systems using Enterprise Message Viewer. However, I find no messages in Message Bank Viewer. Do I need anything extra to archive messages?
tl;dr I have discovered that using $THIS in a very specific way will make persistent objects go stale.
I found that it only happens when using $get on an array that contains the OID reference of $THIS. I assume that this is an unwanted feature and have raised a WRC. The problem can be recreated following the steps below, I have also posted the workaround that I am currently testing and looks to be working fine.
Is there any way to change the sql Shell interface to output a <TAB> between the fields instead of outputting one or more spaces between columns
I needed to run an sql ( that took a while to run) and then cut and paste the generated output into excel, trouble was the paste into excel was not very good for columns when there's spaces used between the columns
Can anyone please help me how to use regular expression/wild character on a sql query for a pattern like C00.0 to C00.9, C01.0 to C01.9 and so on until C77.0 to C77.9 .
We are facing what seems to be a network problem while transferring HL7 messages from Ensemble/Healthshare to a distant target through TCP/IP.
Here is the version of the system in any case it could be useful: Cache for Windows (x86-64) 2017.2.1 (Build 801U) Wed Dec 6 2017 09:07:51 EST [HealthShare Modules:Core:14.02.2415 + Linkage Engine:15.03.9901]
I don't understand how to use this open exchange item for auto creating deleting items. I think it assumes knowledge in the steps of how to use it which i don't get.
Added the code into dev on SRFT.Utility.DeleteHelper.OnDeleteSuper and SRFT.Utility.DeleteHelper.AddHelper
Does anyone know if I can create an API integration, from Ensemble to Salesforce? What I would like to do is trigger a Salesforce case to be opened when for example a Queue reaches a predefined threshold or an interface connection goes from green to non-green.
Recently i've been using Restforms2 to create a CRUD API for a project. But it lacks some advanced functionality that we need, so we have created a production with a REST WS which handles those advanced methods. That works great but there's a drawback, it does not have authentication.
I would want to use the same authentication method as Restforms2 which is a basic auth using IRIS users and passwords.
Im trying to use an API running in AWS API Gateway. This API is over https and i am using the SSL/TLS config of Caché.
set httpRequest = ##class(%Net.HttpRequest).%New()
set httpRequest.Server = server
set httpRequest.Https=1
set httpRequest.SSLConfiguration = "SSLPadraoAdapcon"
do httpRequest.SetHeader("Content-Type","application/json")
do httpRequest.EntityBody.Write(json)
do httpRequest.Post("/dev/router")
I have been using a command pipe to open/run UNIX scripts for months without issue. We recently installed the 2016 Caché Upgrade and now I'm getting the following error when opening the pipe:
<PROTECT> *OPEN[Q](|CPIPE|)
I assume some security or permission setting is to blame, but I'm not sure exactly what. Here's the snippet of code...
s cpipename="|CPIPE|"
s filepath="/path/script"
o cpipename:(filepath:"R"):10
When server-side source control such as Deltanji is being used with Atelier we sometimes want to re-sync the local Atelier copy of a class or routine by pulling the copy from the server.
I have a project to only filter certain pathology results into a downstream system.
Within a HL7 router and business I was planning on using a lookup table and either the exists() or Lookup(), but am having issues when using it with repeating fields or segments.
For example if I perform teh analysis per stated segment usign {} brackets this will work, as each stated repeat is assessed:
I have a compactJWE that I want to decrypt using a key. I read the key from a .pem file and create a JWK with "RSA-OAEP" algorithm. I have the code below in a routine (.mac) file.
Hello! We want an Integration that should be moving files from a FTP server in a DMZ zone into another FTP server on our local network. I tried using EnsLib.FTP.PassthroughService(EnsLib.FTP.InboundAdapter, EnsLib.FTP.OutboundAdapter) Using this approach ensamble write data to the database, causing the CACHE.DAT to grow for every file that is moved. Looks like the entire file is written to the database, is this the case? We are not really interested in storing any file content information in ensamble in this particular case.