Database management
Terms related to databases, including definitions about relational databases and words and phrases about database management.- 99.999 (Five nines or Five 9s) - In computers, 99.
- active directory - Active Directory (AD) is Microsoft's proprietary directory service.
- Amazon RDS (Relational Database Service) - Amazon Relational Database Service (RDS) is a managed database service provided by Amazon Web Services (AWS).
- artifact (software development) - An artifact is a byproduct of software development that helps describe the architecture, design and function of software.
- Azure Data Studio (formerly SQL Operations Studio) - Azure Data Studio is a Microsoft tool, originally named SQL Operations Studio, for managing SQL Server databases and cloud-based Azure SQL Database and Azure SQL Data Warehouse systems.
- Basic Assembler Language (BAL) - BAL (Basic Assembler Language) is a version of IBM's assembler language (sometimes called assembly language) for its System/360 and System/370 mainframe operating system.
- C++ - C++ is an object-oriented programming (OOP) language that is viewed by many as the best language for creating large-scale applications.
- CICS (Customer Information Control System) - CICS (Customer Information Control System) is middleware that sits between the z/OS IBM mainframe operating system and business applications.
- cold backup (offline backup) - A cold backup is a backup of an offline database.
- columnar database - A columnar database (column-oriented) is a database management system (DBMS) that stores data on disk in columns instead of rows.
- conformed dimension - In data warehousing, a conformed dimension is a dimension that has the same meaning to every fact with which it relates.
- CRM (customer relationship management) analytics - CRM (customer relationship management) analytics comprises all of the programming that analyzes data about customers and presents it to an organization to help facilitate and streamline better business decisions.
- cryptographic nonce - A nonce is a random or semi-random number that is generated for a specific use.
- customer data integration (CDI) - Customer data integration (CDI) is the process of defining, consolidating and managing customer information across an organization's business units and systems to achieve a "single version of the truth" for customer data.
- data abstraction - Data abstraction is the reduction of a particular body of data to a simplified representation of the whole.
- data analytics (DA) - Data analytics (DA) is the process of examining data sets to find trends and draw conclusions about the information they contain.
- data availability - Data availability is a term used by computer storage manufacturers and storage service providers to describe how data should be available at a required level of performance in situations ranging from normal through disastrous.
- data center infrastructure management (DCIM) - Data center infrastructure management (DCIM) is the convergence of IT and building facilities functions within an organization.
- data cleansing (data cleaning, data scrubbing) - Data cleansing, also referred to as data cleaning or data scrubbing, is the process of fixing incorrect, incomplete, duplicate or otherwise erroneous data in a data set.
- Data Definition Language (DDL) - Data Definition Language (DDL) is used to create and modify the structure of objects in a database using predefined commands and a specific syntax.
- data engineer - A data engineer is an IT professional whose primary job is to prepare data for analytical or operational uses.
- data integrity - Data integrity is the assurance that digital information is uncorrupted and can only be accessed or modified by those authorized to do so.
- data mining - Data mining is the process of sorting through large data sets to identify patterns and relationships that can help solve business problems through data analysis.
- data modeling - Data modeling is the process of creating a simplified visual diagram of a software system and the data elements it contains, using text and symbols to represent the data and how it flows.
- data preprocessing - Data preprocessing, a component of data preparation, describes any type of processing performed on raw data to prepare it for another data processing procedure.
- data profiling - Data profiling refers to the process of examining, analyzing, reviewing and summarizing data sets to gain insight into the quality of data.
- data quality - Data quality is a measure of a data set's condition based on factors such as accuracy, completeness, consistency, reliability and validity.
- data set - A data set, also spelled 'dataset,' is a collection of related data that's usually organized in a standardized format.
- data source name (DSN) - A data source name (DSN) is a data structure containing information about a specific database to which an Open Database Connectivity (ODBC) driver needs to connect.
- data splitting - Data splitting is when data is divided into two or more subsets.
- data structure - A data structure is a specialized format for organizing, processing, retrieving and storing data.
- data warehouse - A data warehouse is a repository of data from an organization's operational systems and other sources that supports analytics applications to help drive business decision-making.
- database (DB) - A database is a collection of information that is organized so that it can be easily accessed, managed and updated.
- database as a service (DBaaS) - Database as a service (DBaaS) is a cloud computing managed service offering that provides access to a database without requiring the setup of physical hardware, the installation of software or the need to configure the database.
- database automation - Database automation is the use of unattended processes and self-updating procedures for administrative tasks in a database.
- database management system (DBMS) - A database management system (DBMS) is a software system for creating and managing databases.
- database marketing - Database marketing is a systematic approach to the gathering, consolidation and processing of consumer data.
- database replication - Database replication is the frequent electronic copying of data from a database in one computer or server to a database in another -- so that all users share the same level of information.
- Db2 - Db2 is a family of database management system (DBMS) products from IBM that serve a number of different operating system (OS) platforms.
- deep analytics - Deep analytics is the application of sophisticated data processing techniques to yield information from large and typically multi-source data sets comprised of both unstructured and semi-structured data.
- dimension - In data warehousing, a dimension is a collection of reference information that supports a measurable event, such as a customer transaction.
- distributed database - A distributed database is a database that consists of two or more files located in different sites either on the same network or on entirely different networks.
- distributed ledger technology (DLT) - Distributed ledger technology (DLT) is a digital system for recording the transaction of assets in which the transactions and their details are recorded in multiple places at the same time.
- document-oriented database - A document-oriented database is a type of NoSQL database in which data is stored in binary document files.
- Dublin Core - Dublin Core is an international metadata standard formally known as the Dublin Core Metadata Element Set and includes 15 metadata (data that describes data) terms.
- ebXML (Electronic Business XML) - EbXML (Electronic Business XML or e-business XML) is a project to use the Extensible Markup Language (XML) to standardize the secure exchange of business data.
- Eclipse (Eclipse Foundation) - Eclipse is a free, Java-based development platform known for its plugins that allow developers to develop and test code written in other programming languages.
- encoding and decoding - Encoding and decoding are used in many forms of communications, including computing, data communications, programming, digital electronics and human communications.
- encryption key management - Encryption key management is the practice of generating, organizing, protecting, storing, backing up and distributing encryption keys.
- ensemble modeling - Ensemble modeling is the process of running two or more related but different analytical models and then synthesizing the results into a single score or spread.
- entity relationship diagram (ERD) - An entity relationship diagram (ERD), also known as an 'entity relationship model,' is a graphical representation that depicts relationships among people, objects, places, concepts or events in an information technology (IT) system.
- Excel - Excel is a spreadsheet program from Microsoft and a component of its Office product group for business applications.
- extension - An extension typically refers to a file name extension.
- failover - Failover is a backup operational mode in which the functions of a system component are assumed by a secondary component when the primary becomes unavailable.
- field - A field is an area in a fixed or known location in a unit of data such as a record, message header, or computer instruction that has a purpose and usually a fixed size.
- file extension (file format) - In computing, a file extension is a suffix added to the name of a file to indicate the file's layout, in terms of how the data within the file is organized.
- flat file - A flat file is a collection of data stored in a two-dimensional database in which similar yet discrete strings of information are stored as records in a table.
- foreign key - A foreign key is a column or columns of data in one table that refers to the unique data values -- often the primary key data -- in another table.
- Google BigQuery - Google BigQuery is a cloud-based big data analytics web service for processing very large read-only data sets.
- Google Bigtable - Google Bigtable is a distributed, column-oriented data store created by Google Inc.
- graph analytics - Graph analytics is a category of software tools and data mining techniques that help an analyst understand the relationship between entries in a graph database.
- graph database - A graph database, also referred to as a semantic database, is a software application designed to store, query and modify network graphs.
- Hadoop Distributed File System (HDFS) - The Hadoop Distributed File System (HDFS) is the primary data storage system Hadoop applications use.
- hashing - Hashing is the process of transforming any given key or a string of characters into another value.
- IBM IMS (Information Management System) - IBM IMS (Information Management System) is a database and transaction management system that was first introduced by IBM in 1968.
- in-database analytics - In-database analytics is a scheme for processing data within the database, avoiding the data movement that slows response time.
- in-memory database - An in-memory database is a type of analytic database designed to streamline the work involved in processing queries.
- information - Information is the output that results from analyzing, contextualizing, structuring, interpreting or in other ways processing data.
- ISAM (Indexed Sequential Access Method) - ISAM (Indexed Sequential Access Method) is a file management system developed at IBM that enables records to be accessed either sequentially, in the order they were entered or randomly with an index.
- Java Database Connectivity (JDBC) - Java Database Connectivity (JDBC) is an API packaged with the Java SE edition that makes it possible to connect from a Java Runtime Environment (JRE) to external, relational database systems.
- job scheduler - A job scheduler is a computer program that enables an enterprise to schedule and, in some cases, monitor computer 'batch' jobs (units of work).
- job step - In certain computer operating systems, a job step is part of a job, a unit of work that a computer operator (or a program called a job scheduler) gives to the operating system.
- JOLAP (Java Online Analytical Processing) - JOLAP (Java Online Analytical Processing) is a Java application-programming interface (API) for the Java 2 Platform, Enterprise Edition (J2EE) environment that supports the creation, storage, access, and management of data in an online analytical processing (OLAP) application.
- key-value pair (KVP) - A key-value pair (KVP) is a set of two linked data items: a key, which is a unique identifier for some item of data, and the value, which is either the data that is identified or a pointer to the location of that data.
- knowledge base - In general, a knowledge base is a centralized repository of information.
- knowledge management (KM) - Knowledge management is the process an enterprise uses to gather, organize, share and analyze its knowledge in a way that's easily accessible to employees.
- Lisp (programming language) - Lisp, an acronym for list processing, is a functional programming language that was designed for easy manipulation of data strings.
- MariaDB - MariaDB is an open source relational database management system (DBMS) that is a compatible drop-in replacement for the widely used MySQL database technology.
- Microsoft Office SharePoint Server (MOSS) - Microsoft Office SharePoint Server (MOSS) is the full version of a portal-based platform for collaboratively creating, managing and sharing documents and Web services.
- Microsoft SQL Server - Microsoft SQL Server is a relational database management system (RDBMS) that supports a wide variety of transaction processing, business intelligence (BI) and data analytics applications in corporate IT environments.
- Microsoft SQL Server Management Studio (SSMS) - Microsoft SQL Server Management Studio (SSMS) is an integrated environment to manage a SQL Server infrastructure.
- Microsoft SSIS (SQL Server Integration Services) - Microsoft SSIS (SQL Server Integration Services) is an enterprise data integration, data transformation and data migration tool built into Microsoft's SQL Server database.
- middleware - Middleware is software that bridges the gap between applications and operating systems by providing a method for communication and data management.
- MongoDB - MongoDB is an open source NoSQL database management program.
- MPP database (massively parallel processing database) - An MPP database is a database that is optimized to be processed in parallel for many operations to be performed by many processing units at a time.
- multidimensional database (MDB) - A multidimensional database (MDB) is a type of database that is optimized for data warehouse and online analytical processing (OLAP) applications.
- national identity card - A national identity card is a portable document, typically a plasticized card with digitally embedded information, that is used to verify aspects of a person's identity.
- object-oriented database management system (OODBMS) - An object-oriented database management system (OODBMS), sometimes shortened to ODBMS for object database management system, is a database management system (DBMS) that supports the modelling and creation of data as objects.
- OLAP (online analytical processing) - OLAP (online analytical processing) is a computing method that enables users to easily and selectively extract and query data in order to analyze it from different points of view.
- OPAC (Online Public Access Catalog) - An OPAC (Online Public Access Catalog) is an online bibliography of a library collection that is available to the public.
- Open Database Connectivity (ODBC) - Open Database Connectivity (ODBC) is an open standard application programming interface (API) that allows application programmers to easily access data stored in a database.
- operational data store (ODS) - An operational data store (ODS) is a type of database that's often used as an interim logical area for a data warehouse.
- Oracle RMAN (Oracle Recovery Manager) - Oracle RMAN (Oracle Recovery Manager) is a utility built into Oracle databases to automate backup and recovery; it includes features not available in third-party backup tools.
- personal health record (PHR) - A personal health record (PHR) is an electronic summary of health information that a patient maintains control of themselves, as opposed to their healthcare provider.
- PL/SQL (procedural language extension to Structured Query Language) - In Oracle database management, PL/SQL is a procedural language extension to Structured Query Language (SQL).
- predictive modeling - Predictive modeling is a mathematical process used to predict future events or outcomes by analyzing patterns in a given set of input data.
- primary key (primary keyword) - A primary key, also called a primary keyword, is a column in a relational database table that's distinctive for each record.
- query - A query is a question or a request for information expressed in a formal manner.
- raw data (source data or atomic data) - Raw data is the data originally generated by a system, device or operation, and has not been processed or changed in any way.
- RDBMS (relational database management system) - A relational database management system (RDBMS) is a collection of programs and capabilities that enable IT teams and others to create, update, administer and otherwise interact with a relational database.