If you're deploying to more than one environment/region/cloud/customer, you will inevitably encounter the issue of configuration management.

While all (or just several) of your deployments can share the same source code, some parts, such as configuration (settings, passwords) differ from deployment to deployment and must be managed somehow.

In this article, I will try to offer several tips on that topic. This article talks mainly about container deployments.

0 4 34

Requirement: Transform source XML message to target JSON.

step1: First create json equivalent xml from any online tool.

step2: use add in  studio utility create persistent classes from json equivalent xml and compile it , now target persistent classes are ready

step3: Do the above step for source xml or xsd and generate persistent classes for source and compile it

step4: complete the DTL by selecting root node from source, target persistent classes and complete the mappings and compile it

0 0 25
Robert Cemper · Jan 24 1m read

I have created a package that offers a utility to load a Global into JSON object and reverse     
to create a Global from this type of JSON object. Compact refers to the structure created.
Globals nodes are included with data for a fast data load.   
But also the related code is quite compact.

0 0 36
Robert Cemper · Jan 24 1m read

I have created a package that offers a utility to load a Global into JSON object and reverse     
to create a Global from this type of JSON object. Academic refers to the structure created.
Each logical node of the Global is presented separately with all its descendants.
Even if they don't contain any stored data.

0 0 31

Hi! For the Opendataset contest I've build a docker container app stack that use InterSystems IRIS & Openflights Dataset in a container and second container with Apache Zeppelin. You can found details here: https://github.com/andreas5588/openflights_demo

With that you can query the Opendflights Dataset from Apache Zepplin with zero configuration. The containers are on hub.docker so you can use it very easily.

0 0 25

Hi everyone,

I want to talk about our project and use the dataset theme for this contest.

Our intention never was to be a data curator, especially because sometimes my precious data means a lot for me, but not for the rest of the world.

My Precious

We want to go a step further and empower the user to find the perfect dataset for their needs.

Our project is a bridge between the data science community and the developer's community using InterSystems IRIS to achieve this mission.

0 0 29

Previously I had published The Article about the dataset from a real webserver, which can demonstrate how can activity and load of the Apache webserver depends on day of week, search engines indexing and some network noise.

Now I want to describe one useful function for most of webmasters and system administrators who are interested in obtaining of exactly information about visitors, hardware usage, and also about errors that gaing to their clients.

Here it is

0 0 32

With the release of InterSystems IRIS 2021.2 Preview and all-new LOAD DATA functionality dataset can by added with Objectscript Package Manager (ZPM) 

Medical Datasets contains following 12 datasets. For dataset tables and data details please visit ONLINE DEMO by using SuperUser | SYS 

0 0 34
Evgeniy Potapov · Jan 14 1m read
Real Webserver Logs Dataset

I'm happy to share with the community a web server log dataset from our longtime customer, an operating company.

Their webserver operates on Apache webserver and contains data which can be useful to analyse a load and search engines activity.

After installing the project, you will get the data for a few months that can show a typical load and activity of clients, robots and also you can see how it depends on day of week, holidays and time of a day.

The Cube is also included in package.

0 0 43

Hi community,

Prediction is a critical to the Maternal healthcare. The Health Dataset Application (https://openexchange.intersystems.com/package/Health-Dataset) has 10 real health datasets to predict the most important diseases and health problems, including Maternal Risk.

This article detail the steps to predict Maternal Risk using the InterSystems IRIS IntegratedML. This is a technology of InterSystems to do predictions using SQL Commnands! Great!

Follow these steps:

0 0 35

Hey community! How are you doing?

I hope to find everyone well, and a happy 2022 to all of you!

Over the years, I've been working on a lot of different projects, and I've been able to find a lot of interesting data.

But, most of the time, the dataset that I used to work with was the customer data. When I started to join the contest in the past couple of years, I began to look for specific web datasets.

I've curated a few data by myself, but I was thinking, "This dataset is enough to help others?"

0 3 84

The JSON is a data document free of types and validation rules. However, in some scenarios it is important that the JSON document has type and business rules validation, especially in interoperability scenarios. This article demonstrates how you can leverage a market-defined JSONSchema technology that is open for everyone to use and do advanced validations.

1 0 107

In this article I will explain the usage of %SQL_Diag.Result and %SQL_Diag.Message table along with all-new LOAD DATA functionality.

It is recommended to go through LOAD DATA documentation first. 

After successful operation LOAD DATA insert one record in %SQL_Diag.Result table and details are inserted in %SQL_Diag.Message table

Below is the basic command when table is already created and source file does not contain header row. 

LOAD DATA FROM FILE 'C://TEMP/mydata.txt' 
INTO MyTable

The file name must include a .txt or .csv (comma-separated values) suffix and both source and target have the same sequence of data columns.


Loading from File Source: Header

0 0 34

Hi folks!

Sometimes we need the docker image of the InterSystems IRIS solution we build to be published on some docker registry. The cases could be:

  1. Deploy it then in Kubernetes cluster
  2. Let your pal run the image of your public repo without building it locally.

You can push the image to Docker Hub Registry or Github Registry.

In this very short article, I provide a way how to do it automatically on every push to your GitHub repository.

1 0 32