Software deployment is all of the activities that make a software system available for use. The general deployment process consists of several interrelated activities with possible transitions between them.
I just deployed my production from test to acceptance but I found that the deployment misses some Soap Webclient classes which are used by my business operation components. I have used the management portal to create the deployment (i.e. production settings -> Export) and I expected that all classes used by the production were automatically included. Apparantly, that is not the case. Is this default behaviour for Ensemble? And can I somehow force Ensemble to automatically include these classes?
I'm deploying IRIS for Health on a Google Cloud VM, and I note that the SuperServer port is 1971, rather than good old 1972. Is this as it's supposed to be? I can't see that it's documented anywhere.
I'm trying to write an installer manifest that can create a namespace, resources (%DB_namespace) and a role (with the resource, above), based on the namespace. So you could pass in "ABC", or "XYZ", and it would create the %DB_ABC resource and the ABC role with %DB_ABC:RW permissions; or it will create the %DB_XYZ resource and the XYZ role with %DB_XYZ:RW permissions, accordingly.
we have Angular solution and Cache server. We need to have separate users and sessions on same browser (laptop, table etc) for every user and for one user with many connections.
I love the frequent mentions of versioning the CPF merge files in Source control. One thing that is not yet clear to me is how this would work for rolling back a change:
I am looking for a way to deploy production changes (code and/or production configs) via script or some other systematic way other than the Ensemble/Deploy page in the Management Portal. My goal is to store these configurations in Subversion, then have a piece that can pull the XML from source control and deploy it to a given server. Is there some kind of script that I can run on a given server to perform this deployment? If so, I can tie it up with Subversion using another tool that I have access to.
The Installer Manifest has the option to modify the production level settings for AutoStart but is there a way to change settings such as ActorPoolSize and other settings? What would the format be to change such a setting to change the ActorPoolSize to 2?
Do you know how to create workflow users and roles programmatically? I use Docker for test deploy and I need to set up IRIS Interoperability using the install script.
Maybe do you know how to import/export already existing workflow users and roles?
Using the Windows Subsystem for Linux (WSL2), is it possible to install IRIS and run it from there? I am wanting to test IAM, but unable to run Docker Desktop on my VM, and do not have access to a Linux machine to install and test with.
I would like to know if anyone has had some experience in building systems for multiple end-users.
To keep things simple, in a hypothetical example, say an Event Booking System, where a Venue could sign up to use such a system and define it's Venue, costs, calendars, etc.. and then invite their customers to book the Venue on different days/times.
If I wish to offer such a system to many Venues, there are a number of options available.
I work on deploying IRIS inside Docker container. I really like %Installer class can automate many steps. I want to establish an ECP connection to a mirror database and then define a remote database on the application server. I have already seen we can create local database and namespace in %Installer. What code is needed to establish ECP connection?
I am new to Intersystems, in our project we are directly connecting to the server (environment) using Intersystems VSCode extensions and publishing our changes from local machine. This is not the way we usually do as development process.
Is it possible to implement continuous integration ? So that developers can check-in their code in GIT Hub and can integrate Jenkins and automate the deployment?
I have been playing around with the Management Portal deployment tool, which involves: Ensemble > Manage > Deployment Changes > Deploy and Production Settings > Actions > Export Production Settings > Actions > Re-Export
Everything was going fine , until I came across this:
I'm using Git with DeepSee and when I need to do a commit to the git repo I'm exporting ALL the pivots and dashboards from the namespace. But I can forget to do that) And it can take time for a large system.
What is the way to manage automatical export of DeepSee artefacts which we are editing in UI (Cubes, Pivots, Dashboards, Pivot Variables, Term lists, Shared Measures) into files every time when I push Save button?
I am trying to copy an xml file generated on an Apache server into the Jenkins workspace post-build. I was thinking to use a 'send files over ssh' post-build script, but have not done this before and do not know how to refer to the file location on Apache server vs on Jenkins.
For example, if I want to copy from Apache's location of: "classes/UnitTest/Results.xml" into Jenkin's workspace: "/ReportFiles/Results.xml",
How does the script differentiate between whether the location address refers to the Apache server or Jenkins workspace?
We are in process of moving our first Production from development to test. We exported our project from Studio. On test server, we cannot Open the Production because it is not a valid production because nothing exists in Ens. Config.Production class. What are we missing? How do we add our Production in Ens.Config.Production?
I'm trying to build my project on a Linux machine using Docker.
In my development environment, I use Windows 10 Pro with Docker Desktop version 2.3.0.5. Everything works fine, and the docker-compose build runs flawlessly.
But, when I tried to run the same project in a Linux. Ubuntu 18.04.5 LTS (GNU/Linux 5.4.0-1025-azure x86_64), docker --version Docker version 19.03.6, build 369ce74a3c