If anybody could give me some insight on creating the %All Namespace programmatically I would appreciate it. There are quite a few posts I found that reference its creation using the UI, but I cant seem to get passed some validations with any form of the below:
We have recently received the two following errors as part of a PDF extraction process and I am trying to figure what the difference is between the two errors?
Hello guys! I need some assistance with these problems I have.
The first one is that I need to execute a Python file, which is at /XY path. This PY is supposed to read an excel file and create a CSV version in /XZ path.
I am quite sure that the PY file is correct, but when using this function I get nothing but this error:
Is there a straightforward workflow in VS Code for moving a class from one namespace to another on the same server? With Caché/IRIS Studio, it was a (sort of) simple export/change namespace in file menu/import operation. But in VS Code it's not obvious (to me, at least).
I can add multiple namespaces to a workspace, but when using isfs:// paths I see only a "Download" option and no "Upload" or "Import/Compile." When I'm navigating the local file tree, I see Import options, but no way to select the namespace into which I want to import.
Could be enabled a service as "%service_telnet" from cmd? I need to open a cache terminal from cmd but without this service enabled it is not possible, I mean avoiding the use of manage portal of course.
I was wondering if there was a certain procedure or documentation on securing (Https://) the Web Portal into IRIS/Ensemble?
Currently we are using LDAP Delegated Authentication to access the Web Portal using LDAP. However as more and more emphasis is put on securing applications within networks, I can see Management/Security asking us to make sure that the web portal is more secure.
I have created a Class method to list employees that are subordinated to a certain boss. But I don´t know how to relate the data in the way it was given to me.
We got requirement to Migrate whole interfaces from prod instance (cache2017.1 ) to new instance iris2020.1.
Currently around 100+ integrations (health share and health connect) up and running in prod instance (cache2017.1 ) , we need to migrate this live server to brand new server with iris2020.1.
Could you please suggest me guidelines and best practises while doing this migration.
Please ping me any document reference for this kind of Lift and Shift Migration.
After migration done, we will use new Live (iris2020.1 )instance only.
Is it supported to be able to use member functions to define a set, in particular a set defined with a member range in the form:
set = member1:member2?
The docs only show hardcoded values being used for member1 and member2 and I can get it to work if I use hardcoded values. However if I try to use member functions it throws an error:
"ERROR #5001: Both parts of a range must be members"
I am having a scenario where I need to check difference between current date and a past date should be less than 31 in Routing Rule. Can anybody help me how to code that?
Hi All,
We have been using DeepSee which has been the integrated Analytic Dashboard built over Cache Cubes. It works fine but it's visual capabilities are limited and most probably is getting phased out.
If I am not wrong, Tableau is the suggested alternative to DeepSee . It is expensive and a big and considerable shift from existing technology
I wanted to get an opinion of the community as to these few key points
- What other BI tools others have been using with IRIS and what have been their experience
- For which tools are the best inbuilt support / api's avaialble in IRIS
I have an SDA feed from an Edge server that eventually is fed into HSHI / Analytics. This edge server is loading up patient demographics (in the Patient object), which feeds the HSAA.Patient table in HSHI / Analytics.
However, we have other edge servers also feeding into the same HSHI database, and these other edge servers have better demographic information.
I've two Ensemble instances running in a 2016 server but only one of them showing in startup programs.
the ensemble cubes doesn't showup in the programs list from the Start button as well so how can I add the cube so I can access it's facilities (Studio, MP,...etc)
I want to uniquely identify the installation of cache, and from that create a meaningful unique folder names to store files.
so to generate \192-168-1-100\KEVDEV\testfolder \ and \192-168-1-5\LIVE\testfolder\
so in this case, I could- use the IPaddress and thesecond part of $SYSTEM (cache instance name) - at least this way, I can instantly see where it came from with little risk of overwriting between cache instance.
I'm required to do a transformation around patient physical addresses. If (for whatever reason) a patient has multiple addresses, I want to move the home address to the first iteration of PID:11 and remove all other iterations. I've accomplished the first bit using DTL, but for patients who have multiple addresses they are left with empty trailing iterations. Here is an example I mocked up:
(I wasn't able to find this in the docs or the Community, so feel free to point me to a reference that I missed)
How can I determine the effective User and Group that will be used when an InterSystems IRIS process is doing file I/O on the file system for UNIX? Bonus points if you can tell me how to do it from within InterSystems IRIS as well as from the host OS.
What is the SQL table name for it? How can I obtain it via ObjectScript?
A quick search doesn't show any methods and properties. Documentation is a bit "wrong" here saying that the SQL table name is the same. It will be at least 'x_y.z'.
I'm trying to write a method that runs an OS command with $zf(-100,"/ASYNC") and then waits for it to finish before returning, but I can't figure out how to check if the child process (in $zchild) is still running. $System.Process.State($zchild) always returns "RUN" even once the child process is gone. Is there some better way to check this, short of running another OS-specific command and processing the output (which is just really annoying and unelegant)?
I have a client with cache instance that is growing by 10-30G/day. In order to determine what is consuming space I ran "DO ^%GSIZE" to generate a report. From the report the bulk of the space consumption is from a couple of globals which appear to be built in globals not user defined. I would like to find documentation that details the function of these globals to help me understand how and why space is being consumed at such a rapid pace.
Here FileContent field in json will contain converted base64 string of very large pdf(10pages). I need to submit this json as a payload to REST endpoint.
Hi, I am trying to work with visual studio with the extensions offered by InterSystems, but I can't find any option to work locally, is there any way to do it?