InterSystems Caché database is a file where all the data, application scripts, and users, roles and security configurations are stored. Typically the name of the file is cache.dat.
I've been accessing Cache tables from a developer/reporting side, but am now involved in a project to create a data warehouse for our application. I'm trying to find a query I can use to return the sizes of all the tables in the database, so we can identify the largest tables and handle those individually. Can someone give me a query I can run against our Cache database to return the sizes of all the tables from largest to smallest?
I am relatively new to Intersystems cache database management. As I understand it, a database is set with a maximum size or unlimited and the size of each space allocation for the database is defined. My understanding is that the cache database will then manage the allotment of space to the database as the current space allocation is consumed. Though I do not know why one would want to do this, I have been told that the space allocation to a database can be set to be handled manually. This is something, if true, I was not aware of.
I have a workstation with a CACHE instance up and running.
On that same workstation there is also an instance of IRIS (fresh install). I would like to migrate manually the CACHE database to IRIS (ideally, all globals, routines and classes).
What I tried is to copy C:\InterSystems\Cache\mgr\CACHE.DAT to C:\InterSystems\IRIS\mgr\IRIS.DAT (after shutting down both instances) but it does not work.
I got the following message :(112) The service for the IRIS instance did not start.
I would like to know if an encrypted caché database can run significantly slower than a normal "unencrypted" database, in a way that is noticeable to the end user (e.g. slower response time for most pages, especially the ones that rely on read/writing to globals).
I searched in Intersystems knowledge base and couldn't find anything related. I'm looking for possible before/after benchmarks.
I hope you can help me because this is a very serious that I can't wait to resolve... Our back-up is currently taking 3 up to 4 Gb per day and we don't know why. We had 270 Gb yesterday and we are today at 274.3 Gb. We keep a computer history of the messages for only 50 days. Here is a picture of what's currently taking the most data place in our healtshare system:
(don't forget that our DB is also connected to the back-up, maybe it is related, I don't know)
Upon creating namespace I selected same database for both Globals and Routines. How can I separate both the databases. Please note that I already have data and code saved in database.
We have started to see Journal Daemon inactive and DBLatency warnings in the Console log of our Healthshare server. OS is Windows Server 2008 running in a VM. See below
How can you find referenced entries in a table that have been deleted with the drop column? I have to find all fields in the database afterwards and then delete them properly. Has anyone had this problem yet? Can the information be found in the class or storage definitions?
In a typical database, changing the type of a column from some kind of date/time to a string would require a degree of migration effort for existing values. In Caché, my suspicion is that this is not the case, and I could simply change a property definition from %Timestamp to %String (MAXLEN = "30") and recompile, without needing to worry about damaging the existing data stored for this property. Is this correct? Or would I find the existing data in this property being lost or corrupted?
I have a DR member and in this cache server the database "cachetemp" start to getting bigger without any reason (50GB that was all the free disk space we have)
In the members of the mirrors the cachetemp its ok and the size is 31MB.
I restarted the server because I read that the cachetemp database purge when restarting, but didnt happend.
Any recommendation to clean this database? can I just deleate the CACHE.DAT from this database?
We have a need to track Database changes over time - down to the SQL level of granularity if possible. Such as: User xyz runs routine ^abc and we get something similar to a changelog that tells us: table A had this value updated, insert, update etc....
Is that possible using IRIS level tools (Audit Log, Journal File, etc...) , is there a way to convert the global sets and kills from the journals into SQL level changes?
I want to create a Sample Database for Cache InterSystems and want to access it via Cache Entity Framework Provider. But, I cannot find a particular engine of Management Portal to create a database.
Can you guide how to install Management Portal in Windows?
After running an extensive block of daily statistics my IRISTEMP has expanded dramatically. But FreeSpace shows 97% unused and FREE space. How can I shrink IRISTEMP in runtime without shutdown and manual intervention (which was the traditional approach)
I have a case where I need to look up the NPI provider against an External MS SQL database to retrieve our Provider Identifier to send to a downstream system. In the past I would use a Business Process (BPL) to connect to the outside MS SQL via JDBC and get that information for me. But I was thinking instead of creating a BPL process to do this, would it be better just to link to the outside MS SQL database table to retrieve the information in a SQL statement within a DTL?
I have an application with maps that render and edit geographic data in a some layers using KML - geo data represented into XML file with layers, polygnons, points and metadata associated with it. Some maps are in GeoJSON too.
We have Edge Production where we uploading flat files clinical data, and during testing we do this for several rounds load data, test them and reload them. So everytime we have to generate differnet patient data or change MRNs etc. So request is is there a command to delete patient records that will remove from the tables, and delete reference in the registery as well.
Any guidance that you follow for testing backload data and delete loaded data, pls share
Hello everyone, First sorry for the ignorance, but, as I am new in the InterSystem world, I came to the following doubt: Thinking about a database migration.Is it possible, in a non-traumatic way, to take an Oracle database and migrate it completely to Caché?
I am new to InterSystems Cache and want to explore the database features for one of the use cases we have.
I am trying to find the community version of it from in https://download.InterSystems.com but only found InterSystems IRIS and Intersystems IRIS health community versions.
Hi Cache team, I am in the need of listing all the user defined schemas that are present my Cache db and also the user defined tables and views and Columns of those tables and views through Queries. So that I can write some JDBC code to run the queries and fetch the above metadata. Any help is appreciated.