I'm running a query, joining nine tables together. When I run it with a SELECT COUNT(*) I get about 200,000 rows. When I run the SELECT with the columns I want, I get about 2,700 rows. Is there something inherent with the queries that limits the result set size, or anything I can do to make sure I'm getting back all the results? Rather, than trying to run it about 100 times limiting all previously gotten accounts.
What table would I find the MRN in? I'm trying to write a crystal report and pull that field, but I cannot seem to find it in any of the tables, to be able to add to the report.
Hi! I am trying to export Cache data to my SQL database and am trying to export the schema as well.
The Data Export wizard on the Management Portal only allows me to export the data but not the schema. Is there anyway to get the schema exported as well?
I'm calling the expandAll() method to expand the whole tree structure but for some reason its only expanding the root folders but not all elements under those folders ?
var tree = zenPage.getComponentById( "MyTree"); tree.expandAll(true);
I'm experimenting with FHIR bulk data load using NDJSONs, so far the import is running smoothly, but when I'm trying to perform a request of kind /Patient or /Procedure I'm getting back the following error:
Is there any document regarding the embedded sourceControl in HealthShare management portal GUI?
It has Status, Settings,GitWebUI,AddtoSC,Sync,NewBranch,SwitchBranch,Import, ImportForce buttons, I am pretty new to this and I am wondering what these buttons really do behind the scene and when I need to use them.
I would like to visulaize the tables and data on Power BI, so I try to connect the romote server to Power BI to access the tables and data of remote server. I can able to connect that remote server to our Studio also able to view class files. But I couldn't able to connect the remote server to power BI. I'm getting the below error while click on the tables
"Expression.Error: This native database query isn't currently supported"
Is anyone using the IPM client (e.g., running commands like zpm "install somepackagename") on an IRIS version earlier than 2022.1?
We're thinking about raising the minimum supported version so we can use Embedded Python in IPM. I'm curious if this would impact anyone. Of course, you'd be able to continue to use an earlier version of IPM.
How do you configure it to have source control on a namespace accessed via the InterSystems Server Manager extension?
An example from the video explains only how it will work with a cloned git repo on a local machine (PC/Mac), but so far, I can't get it to work on a remote development server (replacing existing remote Studio via RDP with local VS Code).
I built a monitoring system in Grafana using the IRIS API /api/monitor/metrics (reading with Prometheus) but I noticed that the RAM usage shown was below that shown by the operating system. I installed the Zabbix agent and the usage values were higher, but with a line with the same highs and lows but shifted.
So I know we can run routines using both XECUTE or @ However both of them run slower than just running the routine without it. (that is because as far as i understand, XECUTE command for example will re-compile before running?)
the question is: how can I run a routine by name - and keeping the performance as good as possible?
As the title suggests, I would like to programmatically export each HL7 schema category as XML, either as a stream object or to a file. How would I go about doing this?
i am trying to embed a dynamic SQL into Objectscript code , but the %NEXT() will return 0 , however when i copy the same query and run it in the SQL bit it will give me the result i want the code :
It's not supported at the moment but is it possible to implement the Sonatype Nexus (or JFrog Artifactory) support based on the current (or upcoming) IPM version?
In repo command help I see there is a support for filesystem repositories (which I have not yet tried) and my current (hopefully not far-fetched) interpretation is the IPM is designed to be able to support different repositories:
As the title suggests, I would like to programmatically export each individual data lookup table as XML, either as a stream object or to a file. How would I go about doing this?
https://www.youtube.com/embed/HZn4jazdowY [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
I am just writing something to share what I encountered last night, which is the IRIS 2024.3 does not comes with Python by default any more!!!
Which means that I need to install it by myself!!😅 The pros is, I can select my python version😁🤭💃 The trouble is.... at the first place.... I don't know what I should do😥. By going through the community (I am much more prefer than the official document, sorry InterSystems document team😓 ), I found the following piece
I'm using $system.obj.load() to synchronize a folder contains many classes (CLS) in UDL format. I use "-d" as flag (compilation is done later one). It's already done in parallel to make it as fast as possible but I was wondering if there was way to make this even faster, for example using some tradeoffs (eg: by turning off some features before loading them all (that would be re-enabled later on)). It's currently taking 250 seconds to import 3000 classes. MAC file and GBL import is much faster.