In this article I will explain the usage of %SQL_Diag.Result and %SQL_Diag.Message table along with all-new LOAD DATA functionality.

It is recommended to go through LOAD DATA documentation first.

After successful operation LOAD DATA insert one record in %SQL_Diag.Result table and details are inserted in %SQL_Diag.Message table


Below is the basic command when table is already created and source file does not contain header row.

LOAD DATA FROM FILE 'C://TEMP/mydata.txt' 
INTO MyTable

The file name must include a .txt or .csv (comma-separated values) suffix and both source and target have the same sequence of data columns.

Loading from File Source: Header

3 1
0 227

Astronomers’ tools

5 years ago, on December 19, 2013, the ESA launched an orbital telescope called Gaia. Learn more about the Gaia mission on the official website of the European Space Agency or in the article by Vitaly Egorov (Billion pixels for a billion stars).

However, few people know what technology the agency chose for storing and processing the data collected by Gaia. Two years before the launch, in 2011, the developers were considering a number of candidates (see “Astrostatistics and Data Mining” by Luis Manuel Sarro, Laurent Eyer, William O’Mullane, Joris De Ridder, pp. 111-112):

Comparing the technologies side-by-side produced the following results (source):

Technology Time
DB2 13min55s
PostgreSQL 8 14min50s
PostgreSQL 9 6min50s
Hadoop 3min37s
Cassandra 3min37s
Caché 2min25s

The first four will probably sound familiar even to schoolchildren. But what is Caché XEP?

10 9
1 968

Hi Developers!

Here're the technology bonuses for the InterSystems IRIS Datasets Contest 2021 that will give you extra points in the voting:

  • Dataset Usage Demo Repository - 4
  • LOAD DATA Usage - 3
  • Questionnaire - 2
  • Unique Real Dataset - 4
  • Docker container usage - 2
  • ZPM Package deployment - 3
  • Online Demo - 2
  • Code Quality pass - 1
  • First Article on Developer Community - 2
  • Second Article On DC - 1
  • Video on YouTube - 3

See the details below.<--break->

0 4
0 325
Article
· Dec 20, 2021 1m read
Holiday Reading: What Lies Beneath!

For those of you who might be new to IRIS, and even those who have used Cache or IRIS for some time but want to explore beyond its usually-assumed boundaries and practices, you might want to dive into this detailed exploration of the database engine that is at its heart, and discover just what you can really do with it, going way beyond what InterSystems have done with it for you.

10 5
4 437

Hi Community,

We are pleased to invite all the developers to the upcoming InterSystems Datasets Contest Kick-off Webinar! The topic of this webinar is dedicated to the Datasets contest.

In this webinar, we’ll do a quick tour of the new LOAD DATA feature, also chime in on packaging global data or file data with ZPM, and run a data-generation script as part of a method in the zpm install.

As always, our experts will answer the questions on how to develop, build, and deploy datasets using InterSystems IRIS.

Date & Time: Tuesday, December 28 – 10:00 AM EDT

Speakers:
🗣 @Benjamin De Boe, Product Manager, InterSystems
🗣 @Evgeny Shvarov, InterSystems Developer Ecosystem Manager

https://www.youtube.com/embed/jm7bDK0FoiI
[This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]

5 3
0 360
Discussion
· Dec 28, 2021
On Datasets Licensing

Hi developers!

We launched the datasets contest.

And one of the important questions that need to be covered is dataset licensing.

There are two general cases:

a) a dataset you take from another place in public Internet or private network/person.

b) a dataset you create by yourself or own for any other reason.

We decided to follow the principles and considerations that data.world site introduces for datasets' licensing.

0 2
0 207

Hi Community,

New video is already on InterSystems Developers YouTube:

Updates on Security: OpenSSL and a New "Security" Database

https://www.youtube.com/embed/Eb5kPw8-l08
[This is an embedded link, but you cannot view embedded content directly on the site because you have declined the cookies necessary to access it. To view embedded content, you would need to accept all cookies in your Cookies Settings]

3 0
0 257

I've been accessing Cache tables from a developer/reporting side, but am now involved in a project to create a data warehouse for our application. I'm trying to find a query I can use to return the sizes of all the tables in the database, so we can identify the largest tables and handle those individually. Can someone give me a query I can run against our Cache database to return the sizes of all the tables from largest to smallest?

Thanks for the help

0 8
0 650
Article
· Apr 9, 2019 3m read
IRIS/Ensemble as an ETL

IRIS and Ensemble are designed to act as an ESB/EAI. This mean they are build to process lots of small messages.

But some times, in real life we have to use them as ETL. The down side is not that they can't do so, but it can take a long time to process millions of row at once.

To improve performance, I have created a new SQLOutboundAdaptor who only works with JDBC.

BatchSqlOutboundAdapter

Extend EnsLib.SQL.OutboundAdapter to add batch batch and fetch support on JDBC connection.

3 7
2 1.4K

I am relatively new to Intersystems cache database management. As I understand it, a database is set with a maximum size or unlimited and the size of each space allocation for the database is defined. My understanding is that the cache database will then manage the allotment of space to the database as the current space allocation is consumed. Though I do not know why one would want to do this, I have been told that the space allocation to a database can be set to be handled manually. This is something, if true, I was not aware of.

0 6
0 213

I have a workstation with a CACHE instance up and running.

On that same workstation there is also an instance of IRIS (fresh install). I would like to migrate manually the CACHE database to IRIS (ideally, all globals, routines and classes).

What I tried is to copy C:\InterSystems\Cache\mgr\CACHE.DAT to C:\InterSystems\IRIS\mgr\IRIS.DAT (after shutting down both instances) but it does not work.

I got the following message : (112) The service for the IRIS instance did not start.

0 6
2 913

Hi Community,

Join the upcoming InterSystems Interoperability Contest Kick-off Webinar dedicated to the Interoperability Contest.

In this webinar, you'll get an overview of the interoperability capabilities of InterSystems IRIS, and we'll show you how to use the PEX. Also, we’ll discuss and answer your questions on how to build interoperability solutions using InterSystems IRIS and IRIS for Health.

Date & Time: Monday, October 4 — 12:00 AM EDT

Speakers:
🗣 @Stefan Wittmann, InterSystems Product Manager
🗣 @Bob Kuszewski, InterSystems Product Manager
🗣 @Evgeny Shvarov, InterSystems Developer Ecosystem Manager

5 5
0 327

I would like to know if an encrypted caché database can run significantly slower than a normal "unencrypted" database, in a way that is noticeable to the end user (e.g. slower response time for most pages, especially the ones that rely on read/writing to globals).

I searched in Intersystems knowledge base and couldn't find anything related. I'm looking for possible before/after benchmarks.

0 3
0 265

How to find a global's original namespace ? Potentially mapped from a different namespace .

I have a global ^Custom that exists in multiple namespaces but it could mapped from namespace Drone(A) to Launch(B)

Without access to Cache management portal how to find where is my global located using cache code ?

Like if ^Custom == ^[Drone]Custom ??

0 5
0 307

Hello gentlemen,

I hope you can help me because this is a very serious that I can't wait to resolve... Our back-up is currently taking 3 up to 4 Gb per day and we don't know why. We had 270 Gb yesterday and we are today at 274.3 Gb. We keep a computer history of the messages for only 50 days. Here is a picture of what's currently taking the most data place in our healtshare system:

(don't forget that our DB is also connected to the back-up, maybe it is related, I don't know)

0 1
0 153