70 relations: ACID, Asynchronous I/O, Atomic broadcast, Backup, Change data capture, Cloud computing, Cluster manager, Clustered file system, Computer cluster, Computer hardware, Computer network, Computing, Consensus (computer science), Data storage, Database, Deterministic finite automaton, Device driver, Device file, Disaster recovery, Disk array controller, Disk mirroring, Distributed computing, Distributed concurrency control, Distributed lock manager, Distributed shared memory, Dynamic data, Failover, Fault tolerance, Fault-tolerant computer system, File system, Filter driver, Hard disk drive, High-availability cluster, Jim Gray (computer scientist), Journaling file system, Keyspace (distributed data store), Lag, Latency (engineering), Linearizability, List of file systems, Load balancing (computing), Log shipping, Master/slave (technology), Microsoft, Microsoft Cluster Server, Mirror website, Multi-master replication, Operating system, Optimistic replication, Paxos (computer science), ..., Process group, Quality of service, Recovery point objective, Replica, Rsync, Serializability, Snapshot (computer storage), Software, Software transactional memory, Speed of light, Spread Toolkit, State machine replication, Synchronization, System Center Data Protection Manager, Transaction log, Transparency (human–computer interaction), Virtual synchrony, WAN optimization, WANdisco, Wide area network. Expand index (20 more) » « Shrink index
In computer science, ACID (Atomicity, Consistency, Isolation, Durability) is a set of properties of database transactions intended to guarantee validity even in the event of errors, power failures, etc.
In computer science, asynchronous I/O (also non-sequential I/O) is a form of input/output processing that permits other processing to continue before the transmission has finished.
In fault-tolerant distributed computing, an atomic broadcast or total order broadcast is a broadcast where all correct processes in a system of multiple processes receive the same set of messages in the same order; that is, the same sequence of messages.
In information technology, a backup, or the process of backing up, refers to the copying into an archive file of computer data so it may be used to restore the original after a data loss event.
In databases, change data capture (CDC) is a set of software design patterns used to determine (and track) the data that has changed so that action can be taken using the changed data.
Cloud computing is an information technology (IT) paradigm that enables ubiquitous access to shared pools of configurable system resources and higher-level services that can be rapidly provisioned with minimal management effort, often over the Internet.
A cluster manager usually is a backend graphical user interface (GUI) or command-line software that runs on one or all cluster nodes (in some cases it runs on a different server or cluster of management servers.) The cluster manager works together with a cluster management agent.
A clustered file system is a file system which is shared by being simultaneously mounted on multiple servers.
A computer cluster is a set of loosely or tightly connected computers that work together so that, in many respects, they can be viewed as a single system.
Computer hardware includes the physical parts or components of a computer, such as the central processing unit, monitor, keyboard, computer data storage, graphic card, sound card and motherboard.
A computer network, or data network, is a digital telecommunications network which allows nodes to share resources.
Computing is any goal-oriented activity requiring, benefiting from, or creating computers.
A fundamental problem in distributed computing and multi-agent systems is to achieve overall system reliability in the presence of a number of faulty processes.
Data storage is the recording (storing) of information (data) in a storage medium.
A database is an organized collection of data, stored and accessed electronically.
In the theory of computation, a branch of theoretical computer science, a deterministic finite automaton (DFA)—also known as a deterministic finite acceptor (DFA) and a deterministic finite state machine (DFSM) or a deterministic finite state automaton (DFSA)—is a finite-state machine that accepts or rejects strings of symbols and only produces a unique computation (or run) of the automaton for each input string.
In computing, a device driver is a computer program that operates or controls a particular type of device that is attached to a computer.
In Unix-like operating systems, a device file or special file is an interface to a device driver that appears in a file system as if it were an ordinary file.
Disaster recovery (DR) involves a set of policies, tools and procedures to enable the recovery or continuation of vital technology infrastructure and systems following a natural or human-induced disaster.
A disk array controller is a device which manages the physical disk drives and presents them to the computer as logical units.
In data storage, disk mirroring is the replication of logical disk volumes onto separate physical hard disks in real time to ensure continuous availability.
Distributed computing is a field of computer science that studies distributed systems.
Distributed concurrency control is the concurrency control of a system distributed over a computer network (Bernstein et al. 1987, Weikum and Vossen 2001).
Operating systems use lock managers to organise and serialise the access to resources.
In computer science, distributed shared memory (DSM) is a form of memory architecture where physically separated memories can be addressed as one logically shared address space.
In data management, the time scale of the data determines how it is processed and stored.
In computing and related technologies such as networking, failover is switching to a redundant or standby computer server, system, hardware component or network upon the failure or abnormal termination of the previously active application, server, system, hardware component, or network.
Fault tolerance is the property that enables a system to continue operating properly in the event of the failure (or one or more faults within) some of its components.
Fault-tolerant computer systems are systems designed around the concepts of fault tolerance.
In computing, a file system or filesystem controls how data is stored and retrieved.
A filter driver is a Microsoft Windows driver that extends or modifies the function of peripheral devices or supports a specialized device in the personal computer.
A hard disk drive (HDD), hard disk, hard drive or fixed disk is an electromechanical data storage device that uses magnetic storage to store and retrieve digital information using one or more rigid rapidly rotating disks (platters) coated with magnetic material.
High-availability clusters (also known as HA clusters or fail-over clusters) are groups of computers that support server applications that can be reliably utilized with a minimum amount of down-time.
James Nicholas Gray (19442007) was an American computer scientist who received the Turing Award in 1998 "for seminal contributions to database and transaction processing research and technical leadership in system implementation".
A journaling file system is a file system that keeps track of changes not yet committed to the file system's main part by recording the intentions of such changes in a data structure known as a "journal", which is usually a circular log.
A keyspace (or key space) in a NoSQL data store is an object that holds together all column families of a design.
In online gaming, lag is a noticeable delay between the action of players and the reaction of the server in a video game.
Latency is a time interval between the stimulation and response, or, from a more general point of view, a time delay between the cause and the effect of some physical change in the system being observed.
In concurrent programming, an operation (or set of operations) is atomic, linearizable, indivisible or uninterruptible if it appears to the rest of the system to occur at once without being interrupted.
The following lists identify, characterize, and link to more thorough information on computer file systems.
In computing, load balancing improves the distribution of workloads across multiple computing resources, such as computers, a computer cluster, network links, central processing units, or disk drives.
Log shipping is the process of automating the backup of a transaction log files on a primary (production) database server, and then restoring them onto a standby server.
Master/slave or primary/replica is a model of communication where one device or process has unidirectional control over one or more other devices.
Microsoft Corporation (abbreviated as MS) is an American multinational technology company with headquarters in Redmond, Washington.
Microsoft Cluster Server (MSCS) is a computer program that allows server computers to work together as a computer cluster, to provide failover and increased availability of applications, or parallel calculating power in case of high-performance computing (HPC) clusters (as in supercomputing).
Mirror websites or mirrors are replicas of other websites.
Multi-master replication is a method of database replication which allows data to be stored by a group of computers, and updated by any member of the group.
An operating system (OS) is system software that manages computer hardware and software resources and provides common services for computer programs.
Optimistic replication (also known as lazy replication) is a strategy for replication in which replicas are allowed to diverge.
Paxos is a family of protocols for solving consensus in a network of unreliable processors.
In a POSIX-conformant operating system, a process group denotes a collection of one or more processes.
Quality of service (QoS) is the description or measurement of the overall performance of a service, such as a telephony or computer network or a cloud computing service, particularly the performance seen by the users of the network.
A recovery point objective (RPO) is defined by business continuity planning.
A replica is an exact reproduction, such as of a painting, as it was executed by the original artist or a copy or reproduction, especially one on a scale smaller than the original.
rsync is a utility for efficiently transferring and synchronizing files across computer systems, by checking the timestamp and size of files.
In concurrency control of databases,Philip A. Bernstein, Vassos Hadzilacos, Nathan Goodman (1987): (free PDF download), Addison Wesley Publishing Company, Gerhard Weikum, Gottfried Vossen (2001):, Elsevier, transaction processing (transaction management), and various transactional applications (e.g., transactional memoryMaurice Herlihy and J. Eliot B. Moss. Transactional memory: architectural support for lock-free data structures. Proceedings of the 20th annual international symposium on Computer architecture (ISCA '93). Volume 21, Issue 2, May 1993. and software transactional memory), both centralized and distributed, a transaction schedule is serializable if its outcome (e.g., the resulting database state) is equal to the outcome of its transactions executed serially, i.e. without overlapping in time.
In computer systems, a snapshot is the state of a system at a particular point in time.
Computer software, or simply software, is a generic term that refers to a collection of data or computer instructions that tell the computer how to work, in contrast to the physical hardware from which the system is built, that actually performs the work.
In computer science, software transactional memory (STM) is a concurrency control mechanism analogous to database transactions for controlling access to shared memory in concurrent computing.
The speed of light in vacuum, commonly denoted, is a universal physical constant important in many areas of physics.
The Spread Toolkit is a computer software package that provides a high performance group communication system that is resilient to faults across local and wide area networks.
In computer science, state machine replication or state machine approach is a general method for implementing a fault-tolerant service by replicating servers and coordinating client interactions with server replicas.
Synchronization is the coordination of events to operate a system in unison.
System Center Data Protection Manager (DPM) is a software product from Microsoft that provides near-continuous data protection and data recovery in a Microsoft Windows environment.
In the field of databases in computer science, a transaction log (also transaction journal, database log, binary log or audit trail) is a history of actions executed by a database management system used to guarantee ACID properties over crashes or hardware failures.
Any change in a computing system, such as a new feature or new component, is transparent if the system after change adheres to previous external interface as much as possible while changing its internal behaviour.
Virtual synchrony is an interprocess message passing (sometimes called ordered, reliable multicast) technology.
WAN optimization is a collection of techniques for increasing data transfer efficiencies across wide-area networks (WANs).
WANdisco, plc., dually headquartered in Sheffield, England and San Ramon, California in the US, is a public software company specialized in the area of distributed computing.
A wide area network (WAN) is a telecommunications network or computer network that extends over a large geographical distance/place.
Asynchronous replication, Copy Services, Copy services, Data Redundancy, Data Reliability, Data reliability, Data replication, Database replication, Heterogeneous Replication, Master election, Master-slave replication, Replag, Replicated database, Replication (computer science), Secondary database server, Storage replication, Synchronous replication.