Hey Community,
We invite you to join the next French stream on InterSystems Technologies #8, hosted by @Guillaume Rongier!
Date & Time: February 3rd, 12:00 Paris time.
👉 Direct link to join: https://youtu.be/2PFgnuF8bO8
.jpeg)
InterSystems IRIS is a Complete Data Platform
InterSystems IRIS gives you everything you need to capture, share, understand, and act upon your organization’s most valuable asset – your data.
As a complete platform, InterSystems IRIS eliminates the need to integrate multiple development technologies. Applications require less code, fewer system resources, and less maintenance.
Hey Community,
We invite you to join the next French stream on InterSystems Technologies #8, hosted by @Guillaume Rongier!
Date & Time: February 3rd, 12:00 Paris time.
👉 Direct link to join: https://youtu.be/2PFgnuF8bO8
.jpeg)
If you're deploying to more than one environment/region/cloud/customer, you will inevitably encounter the issue of configuration management.
While all (or just several) of your deployments can share the same source code, some parts, such as configuration (settings, passwords) differ from deployment to deployment and must be managed somehow.
In this article, I will try to offer several tips on that topic. This article talks mainly about container deployments.
Hey Community,
Learn about the changes we've made to InterSystems IRIS Containers, including security updates and the new web gateway container:
Sometimes it is necessary to transfer or migrate data and data schema from Postgres to IRIS. There are currently a few options for doing this, but the two most popular options are using DBeaver (https://openexchange.intersystems.com/package/DBeaver) or SQLGateway. The first will be demonstrated in this article and the second is presented in an excellent article by Robert Cemper, DB Migration using SQLgateway (https://community.intersystems.com/post/db-migration-using-sqlgateway), see in this article how to perform this migration using DBeaver:
In the
I'm using this Dockerfile to build .Net Gateway with my PEX BS. It works fine. If I change the image reference here to store/intersystems/iris-community:2021.2.0.617.0 in line 8 (so update from 2020.2 to 2021.2) it fails to build with:
KafkaConsumer.cs(4,20): error CS0234: The type or namespace name 'EnsLib' does not exist in the namespace 'InterSystems' (are you missing an assembly reference?) [/source/KafkaConsumer.csproj]
KafkaConsumer.cs(9,34): error CS0246: The type or namespace name 'BusinessService' could not be found (are you missing a using directive or an assembly reference?)Hello everyone,
I am having a challenge with the IRIS community version in Docker with wsl2. Host machine is Win 11 and the IRIS/Ubuntu image is Ubuntu-20.04 from running 'wsl -l -v.
I am not able to reach the WebPortal in IRIS nor the Super User port. docker ps -a = containers.intersystems.com/intersystems/iris-community:2021.1.0.215.3 "/tini -- /iris-main" 3 weeks ago Up 26 hours (healthy) 0.0.0.0:1972->1972/tcp, 2188/tcp, 53773/tcp, 0.0.0.0:52773->52773/tcp, 54773/tcp iris
Now, I was under the impression that the ports forwarded through the Host to the Image so I should be able to
I am looking for a way to find all properties defined in any given class through coding, including those inherited from parent class(s). In case of inherited properties, I would like to know from which class etc.
A related question: XML adaptor provides a way to export xml schema for a class. Is there a similar method for JSON adaptor to generate a JSON schema for a class?
Thanks in advance for your help.
Hello Community,
to return data from the OnPage() method of a %CSP.Page I used to simply write out the value of a %DynamicArray e.g. via
write tResponse.%ToJSON()
and return $$$OK afterwards.
Large objects bigger than 4MB result in an exception when trying to create return values using this way. What is best practice for returning values as JSON, especially if values exceeding the capcity of the %String class are to be expected? I saw that it is possible to write data to a file with e.g.:
set tmp = tResponse.%ToJSON("/opt/home/tmp/test_stream.txt")
Am I supposed to do this and afterwards
When I installed the IRIS 2021.2 Community Edition, I found some Globals, which seem to be set for the Intersystems Ensemble and DeepSee products. For me that is an annoying design error in those products and only confusing my developers.
As I do not want to use these, I tried to delete them, because they interfere in the list of globals mixed with our own Globals.
But for several of the EnsEdi Globals I get an error message telling me, they can't be deleted, because they are wrong.
Is there a way to get rid of them? Or do I have to wait for a new IRIS version?
Hi Community,
Let's meet together at the online meetup with the winners of the InterSystems Datasets contest!
You'll have an opportunity to have a discussion with InterSystems Product Managers on new LOAD DATA functionality, Q&A on datasets, etc. Our winners' demo included!
Date & Time: Thursday, January 20, 2022 – 11:00 EDT
I started some development with the IRIS 2020.4 version and did not continue it for some months.
Now I just wanted to continue, but saw the license is not valid anymore.
Is there still a possibility to upgrade to 2021.2 community edition and to recover the globals from the old edition without having a valid license key?
How to execute it from the command line? I am able to use SQL in Management Portal, but I need to do it in Unix
select * from Ens_Util.Log where ConfigName='SFTPTo.HQ.SAFETRACE.DFT' AND Text LIKE '%Put Stream%' and TimeLogged LIKE '%2021-10-16%'
Hey Developers,
Watch as the presenter uses IKO to deploy the WebGateway, SAM, and InterSystems API Manager with the latest version of the InterSystems Kubernetes Operator:
Hey Developers,
It's time to announce the winners of the InterSystems Datasets contest! Are you ready?
A storm of applause goes to these developers and their applications:
Hi all,
I am testing asynchronous Bulk Data retrieve with a large EHR vendor. When I retrieve the bulk data, they are returning ndjson (CONTENT-TYPE of "application/fhir+ndjson; charset=utf-8"). If I attempt to process that content using JSON functions - I get
ERROR #5035: General exception Name 'Parsing error' Code '3' Data ''
For example, the following generates the error
s objArray={}.%FromJSON(quickStreamOut)
The error makes sense in that ndjson is not a valid json format (I assume at least).
But how then should ndjson be parsed in COS?
For some years I missed being able to offer, to everybody interested in ObjectScript, a tutorial more or less complete, to start with ObjectScript. Something that could help more and make things easier to those new developers that come to our technology... something intermediate, halfway between the common "Hello World!", that doesn't really get you further, and the "Advanced Training", that is unaffordable because of lack of time,etc.
If there were something truly helpful not only as an introduction to the ecosystem, but as a starting point, as a boost, to really start to walk into ObjectScript and move forward by yourself... wouldn't that be awesome?
Does anyone know of a relatively quick and straightforward way of converting code written in the old dot scoping syntax with argumentless do (see here for reference: https://docs.intersystems.com/latest/csp/docbook/DocBook.UI.Page.cls?KEY=RCOS_cdo_legacy) to the modern parentheses scoping syntax? It's not too bad to do it by hand, but it's also easy to make a mistake and leave a "quit" in an if statement by accident for example.
E.g.
do:someCondition
.set value = do ##class(ExampleClass).ExampleFunctionThatReturnsValue() quit:'value
.do ##class(ExampleClass).ExampleOtherFunction()
Should
Hi !
I am new to the language and I am trying to read some json data from a government site (it is public data).
The address is: https://apisidra.ibge.gov.br/values/t/1736/n1/all/v/all/p/all/d/v44%202…
Here is some data sample:
[{"NC":"Nível Territorial (Código)","NN":"Nível Territorial","MC":"Unidade de Medida (Código)","MN":"Unidade de Medida","V":"Valor","D1C":"Brasil (Código)","D1N":"Brasil","D2C":"Variável (Código)","D2N":"Variável","D3C":"Mês
Application to import 12 Datasets along with 43 tables dynamically by using LOAD DATA command which loads data from a source into an IRIS SQL table.
List of Datasets
Hi contestants!
We've introduced a set of bonuses for the projects for the Datasets Contest!
Here are the projects that scored it:
Hi! For the Opendataset contest I've build a docker container app stack that use InterSystems IRIS & Openflights Dataset in a container and second container with Apache Zeppelin. You can found details here: https://github.com/andreas5588/openflights_demo
With that you can query the Opendflights Dataset from Apache Zepplin with zero configuration. The containers are on hub.docker so you can use it very easily.
If we look at the Docker Dashboard, we’ll see that there is a group named openflights_demo:
Apache Zeppelin is available via
http://localhost:8080
As you see on the screenshot the database
Hello,
Does anyone have experience sending files using %Net.FtpSession? I am not seeing any methods in the class for sending files and setting up an Ensemble production is not an option. Any advice would be appreciated. Thank you.
Hi everyone,
I want to talk about our project and use the dataset theme for this contest.
Our intention never was to be a data curator, especially because sometimes my precious data means a lot for me, but not for the rest of the world.

We want to go a step further and empower the user to find the perfect dataset for their needs.
Our project is a bridge between the data science community and the developer's community using InterSystems IRIS to achieve this mission.
Hey Developers,
Let the voting week begin! It's time to cast your votes for the best applications in the Datasets Contest!
🔥 You decide: VOTE HERE 🔥
How to vote? Details below.
Hi Developers!
Here're the technology bonuses for the InterSystems IRIS Datasets Contest 2021 that will give you extra points in the voting:
See the details below.

Previously I had published The Article about the dataset from a real webserver, which can demonstrate how can activity and load of the Apache webserver depends on day of week, search engines indexing and some network noise.
Now I want to describe one useful function for most of webmasters and system administrators who are interested in obtaining of exactly information about visitors, hardware usage, and also about errors that gaing to their clients.
Here it is
My project also contains a graphical UI that allows you to upload logfiles from your own webserver, load data to the IRIS and watch whats
With the release of InterSystems IRIS 2021.2 Preview and all-new LOAD DATA functionality dataset can by added with Objectscript Package Manager (ZPM)
Medical Datasets contains following 12 datasets. For dataset tables and data details please visit ONLINE DEMO by using SuperUser | SYS
| ID | Dataset Name | Tables | Licence |
|---|---|---|---|
| 1 | Synthetic Medical Data | 11 | Public Domain |
| 2 | Health Care Analytics - 1 | 8 | Public Domain |
| 3 | Global Suicide Data | 7 | Public Domain |
| 4 | COVID-19 Polls1 | 7 | CC-BY |
| 5 | Cancer Rates by U.S. State | 2 | Public Domain |
| 6 | Opioid Overdose Deaths | 2 | Public Domain |
| 7 | Heart Disease Prediction | 1 | Public Domain |
| 8 | Yellowpages medical services dataset | 1 | Public Domain |
| 9 | Hospital ratings | 1 | Public Domain |
| 10 | Heart Failure Prediction | 1 | (CC BY 4.0) |
| 11 | Diabetes Dataset | 1 | Public Domain |
| 12 | Chronic Kidney Disease | 1 | Public Domain |
According to the WHO, The top global causes of death, in order of total number of lives lost, are associated with three broad topics (source: https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death):
I created an application that's provides real data (without personal data) for some of these top 10
I'd like to share an example on how the new Embedded Python feature in IRIS helped me in my daily routines.
While I’m participating in the iris-kaggle-socrata-generator project with Henrique Dias, I got to unzip datasets from Kaggle in order to import them.
Such a task was easily achieved by using the zipfile lib in Python (this code was copied from this stackoverflow):
Method UnZip(pZipFileName As %String, pExtractToDir As %String) As %DynamicArray [ Language = python ]
{
import zipfile
import iris
with zipfile.ZipFile(pZipFileName, 'r') as zip_ref:I'm happy to share with the community a web server log dataset from our longtime customer, an operating company.
Their webserver operates on Apache webserver and contains data which can be useful to analyse a load and search engines activity.
After installing the project, you will get the data for a few months that can show a typical load and activity of clients, robots and also you can see how it depends on day of week, holidays and time of a day.
The Cube is also included in package.
So you can use my previous project (Promjet-Stats) and see dataset in dashboards.
To make that you shoud install