Containerization is a lightweight alternative to full machine virtualization that involves encapsulating an application in a container with its own operating environment.
https://www.youtube.com/embed/DBpCPwLclUY [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
Such most popular nowadays CI platforms as GitHub and Gitlab, offers the ability to run any docker image as a service, mostly useful for integration tests.
And I could define a GitHub workflow like this for instance, for some of my Python project which, requires the connection to IRIS
Does anyone have experience with installing the Arbiter Container using Podman instead of Docker in a Red Hat environment? I was able to pull down the docker image, but unsure what are the next steps as I am confused on how to start the container using Podman and ensure the parameters are set appropriately? Does anyone have the steps that I should take? Should I go through the WRC? Does the WRC have experience using Podman?
Or should I just install the ISC Agent instead of using the Container?
My team works on implementing an Interoperability solution utilizing InterSystems Kubernetes Operator on Red Hat OpenShift container platform.
We are trying to determine how many messages we can process in any given time. We have a Feeder app running in 10 containers sending 50k messages each to a load balancer all beginning at the same time.
Messages are received via HTTPS protocol by webgateway containers.
Interoperability production runs in compute pods with persistent data, journals, and WIJ volumes.
Watch this video to get a brief overview of the near- and long-term plans for HealthShare containerization and Kubernetes adoption, as well as a preview/demo of our current progress.
https://www.youtube.com/embed/ZXXXnG_6SBc [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
https://www.youtube.com/embed/I0v18w9Y4X0 [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
Docker 20.10.14 (released March 23, 2022) changes the Linux capabilities given to containers in a manner that is incompatible with the Linux capability checker in InterSystems IRIS 2021.1 (and up) containers.
Users running Docker 20.10.14 on Linux will find that IRIS 2021.1+ containers will fail to start and the logs will incorrectly report that required Linux capabilities are missing. For example:
For quite some time InterSystems IRIS supports such thing as Merging CPF. So, with help of this it should be possible to define only desired changes in configuration. And get them applied even with vanilla Docker image.
And I though it could be useful when used with Dockerfile. Use this way to configure IRIS during docker build instead of using Installer manifest.
InterSystems Developer Community analytics. Project made with InterSystems IRIS BI (DeepSee), Power BI and Logi Report Designer to visualize and analyze members, articles, questions, answers, views and other pieces of content and activity on InterSystems Developer Community.
You can see your own activity, articles and questions. Track how your contribution changes developer community.
I am new to using containers. I am working on trying to specify some of the SAM settings and wondered how to access the terminal while IRIS is running in a Container. Can someone help me out?
I'll start with an apology as I am trying to wrap my head around the architecture of how InterSystems IRIS database management works. I am attempting to connect to the platform remotely through say a JDBC or ODBC connection in order to run queries, searches (through SQL statements) on my laptop and was trying to understand whether this would be possible? It is possible to setup an inbound client connection and wanted to better understand the architecture of how the database association works for IRIS database management. Does it use it's own internal SQL database or are we able to connect to our own database and which databases are certified to run against the platform?
https://www.youtube.com/embed/ZbzP7m0HDkU [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
The InterSystems Iris Fhirserver running on a Raspberry Pi Raspberry running as a FHIRserver
Raspberry running as FHIRserver
About a year ago I wrote some articles about the installation of the HAPI FHIRserver on a Raspberry Pi. At that time, I only knew the basics of the FHIR standard, little about the technology behind FHIR-servers and not much more about the Raspberry. By trying, failing, giving up and trying again I learned a lot.
The InterSystems Kubernetes Operation (IKO) version 3.3 is now available via the WRC download page and the InterSystems Container Registry.
IKO simplifies working with InterSystems IRIS or InterSystems IRIS for Health in Kubernetes by providing an easy-to-use irisCluster resource definition. See the documentation for a full list of features, including easy sharding, mirroring, and configuration of ECP.
IKO 3.3 Highlights:
Support for 2021.2 and 2022.1 editions of InterSystems IRIS & IRIS for Health
Support for Kuberentes 1.21
Deploy common System Alerting and Monitoring (SAM) configurations as part of your irisCluster
InterSystems API Manager (IAM) can now also be deployed and managed as part of the irisCluster
Automatic tagging of mirror pair active side, so a service can always point to the active mirror member.
The title says it all. I’m building an IRIS image with docker-compose using a separate Dockerfile. Pretty straightforward procedure: I import a Installer script inside the container containing a Installer Manifest I defined. Within the manifest, I create a namespace with code and data databases in separate locations. My intention is to keep the code database inside the container, so whenever I build the container, the imported code is replaced. The data, however, should be persistent.
We are developing some containarized cloud application level iris instances and using CPF Merge to do a lot of the initial buildout for the iris instance (i.e. create databases, namespaces, map globals/routines, ecp setup, etc...)
I am trying to figure out how to get package mappings into a namespace config, via cpf merge if possible... ?
This is the document I am working from to develop the cpf merge file -
If you're deploying to more than one environment/region/cloud/customer, you will inevitably encounter the issue of configuration management.
While all (or just several) of your deployments can share the same source code, some parts, such as configuration (settings, passwords) differ from deployment to deployment and must be managed somehow.
In this article, I will try to offer several tips on that topic. This article talks mainly about container deployments.
https://www.youtube.com/embed/u5ccd1kifwQ [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
Hi All! For those of you who attended experience labs at the 2021 Virtual Summit, you may recall that one of the lab sessions was around Kubernetes. We've now converted that lab to be fully on-demand. You can launch a small cluster of VMs and follow the exercise to manage your Kubernetes cluster, deploy InterSystems IRIS containers to it, and watch its self-healing nature when destroying a pod.
For those that, at some point, need to test what means that of ECP for horizontal escalability (computing power and/or users and processes concurrency), but they're lazy o have no much time to build the environment, configure the server nodes, etc..., I've just published in Open Exchange the app/sample OPNEx-ECP Deployment .
I have just created a new Global Master Topic, "IRIS Cheatsheets". IRIS has introduced a lot of new functionality, especially in scripting languages, FHIR R4 support, enhanced Interoperability Tools, and IRIS Analytics. Having spent 35 years working on Windows-based PC's and Laptops, I have surprisingly little knowledge of Linux, Docker and Git. Furthermore, I have written almost every application and Interface in ObjectScript with splatterings of SQL, .Net, and Java Gateways and the most basic knowledge of WinSCP, Putty, SSH. All that changed when I received my first Raspberry Pi.
I work as an Integration Engineer for United States Department of Veterans Affairs (VA). I work on a Health Connect production which processes many RecordMap files. I do not fully understand RecordMaps and I wanted to develop an application for the Interoperability contest where I could learn more about working with RecordMaps. I browsed InterSystems documentation for inspiration on how to start. I was happy to find CSV Record Wizard.
https://www.youtube.com/embed/PvlDp6xLQ5U [This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]
Is there an InterSystems supported dotnet core library or community contributed repo on the horizon? At this time we are exploring installing the ODBC driver in our containers but would rather use more robust solution.