Software testing is an investigation conducted to provide stakeholders with information about the quality of the software product or service under test.
A few years ago, I was teaching the basics of our %UnitTest framework during Caché Foundations class (now called Developing Using InterSystems Objects and SQL). A student asked if it was possible to collect performance statistics while running unit tests. A few weeks later, I added some additional code to the %UnitTest examples to answer this question. I’m finally sharing it on the Community.
When we write unit test cases for cache object script code using %UnitTest.TestCase, what is the best way to write code to identify code coverage?
So, let say my unit test case hit all 10 lines of code of a method for a given class. So, unit test coverage should be 100% for that. But, using line-by-line coverage [(%Monitor.System.LineByLine] getting wrong percentage, because it also includes code comment/documentation as part of code. So, practically we can not ever achieve 100% of code coverage by using this API.
How can I effectively unit test a Business Process Language(BPL) that incorporates three code activities calling methods from distinct ObjectScript classes?.
We have one user that is having issues in Ensemble via browser. The two issues she reported are:
- If she opens a DTL, she has to Save or Compile the DTL before she is able to edit anything in the Actions tab. Typically, you can simply open the DTL, click on a line, and edit the Actions.
- If she opens the Tools tab and clicks the Test button, she does not get a response and it eventually times out on her. Typically, clicking the Test button will open a new window where you can paste a message to test your DTL with.
Continuing on with providing some examples of various storage technologies and their performance profiles, this time we looked at the growing trend of leveraging internal commodity-based server storage, specifically the new HPE Cloudline 3150 Gen10 AMD processor-based single socket servers with two 3.2TB Samsung PM1725a NVMe drives.
The Mockable.io (https://www.mockable.io/) is an online service to deploy REST API or SOAP services in seconds. This is useful to test the consumption of an API or SOAP service in your production or objectscript class without having to implement a real service, including https option.
Hi All, I was creating a unit test directory with the below script. But it not creating any directory as mentioned in code. I am working on MacBook with irisusr User as standard installation process.
if '$data(^UnitTestRoot) {
set mainFolder = "UnitTest\NewDir\"
set st = ##class(%File).CreateDirectoryChain(mainFolder _ testSuite)
if $$$ISOK(st) { set ^UnitTestRoot = mainFolder }
else {
write !, "UnitTestRoot folder could not be created. Consult with instructor."
I want to test automatically that HL7 business operation works correctly in error conditions. One is to test CE acknowledgement. I have planned to implement test production which includes business services for different situations (AA, AE, CA, CE, timeout, late response etc).
How should I implement HL7 business service that always returns CE (commit error)? I have tried but it keeps returning "AA".
I found a package on OEX for a Sharding Demo If Sharding is NOT included in the Community License I can not use the Community Distribution but require a different one. And have to add ZPM manually.
Is there a method I can utilise to check if a particular method has been called and perhaps log the args it was called with?
We are currently trying to expand out Cache Unit Testing processes and were wondering if there is an existing method that would capture this or how such a method could be structured.
Hello everyone, I work for a company that currently uses Studio on version 2016.1.2, and will be looking at options to upgrade to 2016.2 and beyond. This would also include the potential to switch from Studio to Atelier, however there are a lot of workflow changes and new features to test.
Does Atelier have a simple way to evaluate what it has to offer without needing to setup an entire separate environment? How did your organization demo its features before committing?
IRIS Interoperability Productions formerly known as Ensemble are fun to work with. Yes, I really think my work is fun. I have seen File Passthrough Services and File Passthrough Operations come in handy. At one point we placed test messages in files, then we utilized a File Passthrough Service with Inbound File Adapter to send the contents of the file as a Stream to a File Passthrough Operation with Outbound TCP Adapter.
Someone may have asked this before but is there a way to return the TRACE statements that we have placed in the Data transformation within the DTL Editor Testing Tool (EnsPortal.Dialog.TestTransform.cls)? Is this something we can get in an Enhancement request in for?
It would save sometime in troubleshooting where an error might be in the DTL.
We recently uploaded to OpenExchange a small application that I developed a while ago (and that @Jose-Tomas Salvador improved and refined) that I often use when I need to generate large volumes of HL7 messaging.
The application I am working on for Iris Contest needs to send an email. Is anybody aware of a configuration for SMTP server, port, credentials that I can use for the contest or any demo application?
Getting INVALID OREF error while unit testing. Also ResultDyanmicObj.Headers.%Size(),DummyResponse.Headers.%Size() these are working fine getting 7 value from both. but still it is not working.
do $$$AssertEquals(ResultDyanmicObj.Headers.%Size(),DummyResponse.Headers.%Size(), "Validation for Headers")
This is yet another short post that is intended to simplify developers' life. Now we'll talk about how to make GitHub run unit tests with every push to the repository by adding just one file to the repo. For free. On Github Cloud. Sounds great, isn't it?
It is possible and very easy to do. Credit goes to @Dmitry Maslennikov (and his repo), ZPM Package Manager, and GitHub Actions. Let's see how this all works!
Sometimes we need to know for sure if the current environment has sufficient cores, memory, and bandwidth to support the planned number of users and such SLAs as latency, response time, and availability. This is true for databases and backends. This is why it is mandatory for critical applications and databases to simulate the users simultaneous/concurrent requests and collect metrics about performance and availability.