trying importing classDefinition to Iris via LoadStream() fails with zLoadStream+1^%SYSTEM.OBJ.1 ----> InterSystems.Data.IRISClient.IRISException : Exception thrown on server (code = 1192)...
Code is here; use any valid exported ClassDefinition as File to test that:
I was wondering if there is a way to maybe temporarily alter the [Private] tag dynamically or any other way to run a Unit test against class method, which has been labeled as [Private]?
I mean, I understand there is a workaround to create a non private class method inside this class, which will in turn call this private method and run a Unit test on the non private method, but I am looking for a way to do this without altering the original class.
I have a <tablePane> element with OnCreateResultSet and OnExecuteResultSet methods; autoExecute is "false" but the OnCreateResultSet and OnExecuteResultSet methods are nevertheless called on page load.
I want the user to be able to press a button to submit the parameters, then have this button call tablePane.executeQuery() to execute the query.
autoExecute is simply igonored. We're on 2014.1.3 with plans to upgrade to 2016 soon.
Are there more settings I need to set? Or is this attribute simply ignored?
I require an operation that will get a list of hosts every X minutes and call them, using values from my database.
The only issue is that I can't find any sort of scheduling/ timed event functionality in Ensemble. The documentation isn't particularly helpful so I thought I'd give the community a shot.
Is there any existing functionality in Ensemble that allows for scheduled operation calls?
https://www.youtube.com/embed/hMHRUYNl5Bk [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
I am accessing IRIS databases with JDBC (or ODBC) using Python. I want to fetch the data into a pandas dataframe to manipulate the data and create charts from it. I ran into a problem with string handling while using JDBC. This post is to help if anyone else has the same issues. Or, if there is an easier way to solve this, let me know in the comments!
I am using OSX, so I am unsure how unique my problem is. I am using Jupyter Notebooks, although the code would generally be the same if you used any other Python program or framework.
As an AI language model, ChatGPT is capable of performing a variety of tasks like language translation, writing songs, answering research questions, and even generating computer code. With its impressive abilities, ChatGPT has quickly become a popular tool for various applications, from chatbots to content creation. But despite its advanced capabilities, ChatGPT is not able to access your personal data. So we need to build a custom ChatGPT AI by using LangChain Framework:
Below are the steps to build a custom ChatGPT:
Step 1: Load the document
Step 2: Splitting the document into chunks
Step 3: Use Embedding against Chunks Data and convert to vectors
Step 4: Save data to the Vector database
Step 5: Take data (question) from the user and get the embedding
Step 6: Connect to VectorDB and do a semantic search
Step 7: Retrieve relevant responses based on user queries and send them to LLM(ChatGPT)
Step 8: Get an answer from LLM and send it back to the user
a customer engineer asked me how to use Http basic authentication on a Ensemble SOAP service. Anyone has done this before?
When secure SOAP service from Web Application setting, by select "password" for "Allowed Authentication Methods", it expected WSS, username and password in SOAP head. The customer hospital would like to add user authentication on sopa service, but SOAP client team prefer to add it on http header, instead of on SOAP header.
Also, I am very confused with adding password on URL for SOAP service, like
Watch this video to learn about InterSystems FHIR Server, the turnkey FHIR data solution that empowers FHIR application developers to focus on building life-changing healthcare applications:
https://www.youtube.com/embed/YyZc8dK9_jM [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
I'm trying to connect to a MySQL database through a SQL Gateway (type JDBC) on a MacBook M1. Can't get the connection working, anybody any experience on this topic?
Is there a group in or outside of this community for discussion of DTM issues? I am working on a conversion to Caché for a client but in the meantime, he is still live with his DTM system and a system management issue has arisen. The convo:
Client: Do you know how to increase the page limit in MUMPS DTM? One version of my database I use for one customer is telling me I have reached my page limit. I have increased the page limit in the database you are reviewing; so I figure there is a setting I am missing.
Can we make any property immutable once it value is set? For example, I may have a property called RecordCreatedTime to track when the record is created in IRIS, and once it is set, I don't want anyone to modify the value.
I am creating a WS as a server, but when I ask for the WSDL it is giving me an error because it cannot find the class.
I have added the following instructions:
set ^SYS("Security","CSP","AllowClass","MiProyecto.MiClaseWS","%SOAP.WebServiceInfo")=1
set ^SYS("Security","CSP","AllowClass","MiProyecto.MiClaseWS","%SOAP.WebServiceInvoke")=1
I have created an entry in the WS security configuration
Is there an elegant way to define a dynamic "Archive path" for EnsLib.File.InboundAdapter ?
What i mean is to have a sub-folders structure (under the root archive path) with a name similar to: yyyymmdd This way the adapter "could" store the files on those sub-folders (and create a new one for the 1st file for a specific date).
When a production is processing hundreds of thousands of files every day, we might end up with an archive with huge number of files (many millions) which is then hard to handle (read, search) especially on Windows O/S.