I have an above error when purging record map batches and was wondering if anyone out there has ever experienced this and if they have please any advice
Failed to purge body for header 9747192, BodyClassname='******.Batch':ERROR #5823: Cannot delete object, referenced by '*****.Record.%ParentBatch'
I would like to allow some departmental user to view the ensemble portal. I want to make sure they are not allowed to do any changes (like stop and start interfaces from portal)
I have created one userbut limited with SQL privilages. But using this account, the portal view is not accessible.
It would be appreciated if anyone can adice me on this. I know this may be a silly question.
Is %CSP.BinaryStream compatible with %StreamGlobalBinary because reviewing the classes this doesn't look to be the case?
What I am trying to do is zip large documents before sending over a REST interface. I compress the file and add that to an Ens.StreamContainer which is then passed to the operation to build the REST payload. We are using %Net.MimePart class to build the payload and use the BodySet method to add the zipped Ens.StreamContainer.
The issue is when the REST interface receives the POST, it doesn't recognize the data as a zip compressed item.
Working on integrating with O365 Sharepoint REST API. I would want to know if anyone can share their experience with integration with Sharepoint REST API and how they implemented security?
I am able to transmit a small batch of billing files: B/W 5 and 20 without issue- Put stream 11@%IO.LibraryStream to /SFTP/unmhsmcsa/OUT/TEST_021920201426.tst
But when the size of the batch increases to over 20 the operation hangs and I receive the error below: The file is created in the SFTP server with 0 data.
This article is aimed at developers implementing DICOM productions, specifically for cases with third-party endpoints that cannot handle the DIMSE timeout themselves.
We have created a utility to allow users to create a System task for a list of items the same way it’s done via SMP (attached below), the problem is that there may items and those systems tasks will grow a maybe exceed a couple of thousands so :
I know there is the ability to enable Rule Sets within a Business Rule by dates, but what about Sub-Rules or just Rules in general? We have an upcoming go-live that is at Midnight and I was wondering if there was a way I could script out enabling rules/sub-rules within a Business Rule to save my team some time and headaches. Or is the best practice is just to create another rule for the go-live, then move it over once we are live...
I have a requirement where in I had to read a HL7 message string ( which is a String ) from a SQL TBL to my Business Service. and I need to convert that string to EnsLib.HL7.Message format and traverse through the PID segment.
I've the below databinding on a combobox and the issue that the binding is coming with the display value of my levels so eg. I'm getting normal instead of 1 which mean the combobox in my form is not showing, so how get the binding to bring the VALUELIST rather then DISPLAYLIST ?
I have a business process that adds data to a global variable on receipt of an HL7 message, and a scheduled task that executes a class method defined within the same business process that removes data from the same global variable. With this in mind it makes sense to consider concurrency and therefore make use of the LOCK command.
My first question is whether this is actually necessary?
how do I know that my service is running in a production, I've this receiving service and the as blow that says lost TCP then closing TCP connection, then last one says ConfigItem starting job, and in the Jobs tab it says Listing.
Not everyone knows that InterSystems Caché has a built-in tool for code profiling called Caché Monitor.
Its main purpose (obviously) is the collection of statistics for programs running in Caché. It can provide statistics by program, as well as detailed Line-by-Line statistics for each program.
Using Caché Monitor
Let’s take a look at a potential use case for Caché Monitor and its key features. So, in order to start the profiler, you need to go to the terminal and switch to the namespace that you want to monitor, then launch the %SYS.MONLBL system routine:
A beginner’s guide to Exception Handling in RESTful web services. The article gives an example how the various error conditions during processing a service request can be handled.
We expect our client – server communication working in a flawless operational condition, running error free software. But we are prepared to handle exceptions. Are we? So far in the examples of the previous sessions were not. We did not care about exceptions. The result? In any error incident it took ages to figure out what the problem is and more importantly how to fix it.
I'm pretty new to Ensemble and Cache, and I ran into an error today that I haven't seen before: "ErrNoResponseClass: No Response Classname is assigned for Request class <<someNumber>>". I haven't found much documentation about this beyond the Ensemble Error Codes document, which is essentially just a duplicate of the information I can see in the error message when this problem occurs. Is there any further information about this error and the conditions that cause it? Or perhaps what the <<someNumber>> value represents?
Using the Rule Editor, I can open a routing rule and easily define an 'effectiveBegin' and 'effectiveEnd' date. I want to utilize this date feature for a "sub-rule" (i.e. rule #1) but not the entire routing rule therefore impacting all other rules. I also do not want to split and create a brand new routine rule just to send the message to the target business process. It would be ideal to keep it contained within this single routing class that was created.
Looking for some suggestion on what the community has been setting Confirm Complete to? The testing I did when this was set to size is way too slow at around 1 message for 2 seconds. Is it safe to set this value to None? Would I get half messages if the sending system didn't compete writing the message.