MDM Thoughts
Master Data Management (MDM) is a discipline designed to make data more dependable, sharable and accessible. Here are some of IRI’s philosophies around MDM:
Most developers believe that data are, or should be, application-independent.Master Data Management (MDM) is a discipline designed to make data more dependable, sharable and accessible. Here are some of IRI’s philosophies around MDM:
Most developers believe that data are, or should be, application-independent.Editors note: CoSort Version 10 was released in mid-2018. Please see this article for links to its features and upgrade details.
What’s New in CoSort 9.5.3
Along with the new website, there is a new release of IRI’s flagship CoSort package for data management and data protection. Read More
Full disclosure: As this article is authored by an ETL-centric company with its strong suit in manipulating big data outside of databases, what follows will not seem objective to many. Read More
Data profiling, or data discovery, refers to the process of obtaining information from, and descriptive statistics about, various sources of data. The purpose of data profiling is to get a better understanding of the content of data, as well as its structure, relationships, and current levels of accuracy and integrity. Read More
Note: This article was originally drafted in 2015, but was updated in 2019 to reflect new integration between IRI Voracity and Knime (for Konstanz Information Miner), now the most powerful open source data mining platform available. Read More
IRI’s data management tools share a familiar and self-documenting metadata language called SortCL. All these tools — including CoSort, FieldShield, NextForm, and RowGen — require data definition file (DDF) layouts with /FIELD specifications for each data source so you can map your data and manage your metadata. Read More
A test data generator is an important part of the setup process for DevOps and data architects prototyping database and data warehouse operations, testing applications, benchmarking different platforms, and outsourcing work formats. Read More
VSE, short for Virtual Storage Extended, is an operating system for IBM mainframe computers. Programs scripted in its job control language (JCL) instruct the system in how to run batch jobs or start subsystems. Read More
One of the best ways to speed up big data processing operations is to not process so much data in the first place; i.e. to eliminate unnecessary data ahead of time. Read More
MVS, short for Multiple Virtual Storage, is the original operating system for IBM mainframe computers that is now z/OS. Its shell scripting or job control language (JCL), instructs the system how to run batch jobs or start subsystems. Read More
The IRI data management platform Voracity, as well as its constituent tools, can perform and speed big data warehouse extract, transform, load (ETL) operations, delaying the need for new hardware or expensive proprietary appliances: http://www.iri.com/blog/data-transformation2/a-big-data-quandary-hardware-or-software-appliances-or-cosort/ Read More