Transaction processing system

Last updated

A transaction processing system (TPS) is a software system, or software/hardware combination, that supports transaction processing.

Contents

History

The first transaction processing system was SABRE, made by IBM for American Airlines, which became operational in 1964. [1] Designed to process up to 83,000 transactions a day, the system ran on two IBM 7090 computers. SABRE was migrated to IBM System/360 computers in 1972, and became an IBM product first as Airline control Program (ACP) and later as Transaction Processing Facility (TPF) . In addition to airlines, TPF is used by large banks, credit card companies, and hotel chains.

The Hewlett Packard Enterprise NonStop system (formerly Tandem NonStop) is a hardware and software system designed for Online Transaction Processing (OLTP) introduced in 1976. [2] The system provides an extreme level of availability and data integrity.

List of transaction processing systems

Processing types

Transaction processing is distinct from and can be contrasted with other computer processing models, such as batch processing, time-sharing, and real-time processing. [10]

Batch processing

Batch processing is execution of a series of programs (jobs) on a computer without manual intervention. Several transactions, called a batch are collected and processed at the same time. The results of each transaction are not immediately available when the transaction is being entered; there is a time delay.

Real-time processing

"Real time systems attempt to guarantee an appropriate response to a stimulus or request quickly enough to affect the conditions that caused the stimulus." [10] Each transaction in realtime processing is unique; it is not part of a group of transactions.

Transaction processing

A Transaction Processing System (TPS) is an information system that collects, stores, modifies, and retrieves the data transactions of an enterprise. Transaction processing systems also attempt to provide predictable response times to requests, although this is not as critical as real-time systems. Rather than allowing the user to run arbitrary programs as time-sharing, transaction processing allows only predefined, structured transactions. Each transaction is usually short, and each transaction's processing activity is programmed in advance. It is an MIS model.

Transaction processing system features

The following features are considered important in evaluating transaction processing systems. [10]

Performance

Fast performance with a rapid response time is critical. Transaction processing systems are usually measured by the number of transactions they can process in a given period of time.

Continuous availability

The system must be available during the time period when the users are entering transactions. Many organizations rely heavily on their TPS; a breakdown will disrupt operations or even stop the business.

Data integrity

The system must be able to handle hardware or software problems without corrupting data. Multiple users must be protected from attempting to change the same piece of data at the same time, for example two operators cannot sell the same seat on an airplane.

Ease of use

Often users of transaction processing systems are casual users. The system should be simple for them to understand, protect them from data-entry errors as much as possible, and allow them to easily correct their errors.

Modular growth

The system should be capable of growth at incremental costs, rather than requiring a complete replacement. It should be possible to add, replace, or update hardware and software components without shutting down the system.

Types of transaction processing

Processing in a batch

Transactions may be collected and processed as in batch processing. Transactions will be collected and later updated as a batch when it is convenient or economical to process them. Historically, this was the most common method as the information technology did not yet exist to allow real-time processing.

Processing in real-time

This is the immediate processing of data. It provides instant confirmation of a transaction. It may involve a large number of users who are simultaneously performing transactions which change data. Because of advances in technology (such as the increase in the speed of data transmission and larger systems and networking bandwidth), real-time updating is possible.

Databases for transaction processing

A database is an organized collection of data. Databases offer fast retrieval times for non-structured requests as in a typical transaction processing application.

Databases for transaction processing may be constructed using hierarchical, network, or relational structures.

A hierarchical structure Hierachical-Diagram.PNG
A hierarchical structure
A network structure Network-Diagram.PNG
A network structure
A relational structure Relational-Diagram.PNG
A relational structure

The following features are desirable in a database system used in transaction processing systems:

Backup procedures

A Dataflow Diagram of backup and recovery procedures Backup-DFD.png
A Dataflow Diagram of backup and recovery procedures

Since business organizations have become very dependent on transaction processing, a breakdown may disrupt the business' regular routine and stop its operation for a certain amount of time. In order to prevent data loss and minimize disruptions well-designed backup and recovery procedures must exist and be adhered to. The recovery process can rebuild the system when it goes down.

Types of back-up procedures

There are two main types of back-up procedures: grandfather-father-son and partial backups:

Grandfather-father-son

This procedure involves taking complete backups of all data at regular intervals daily, weekly, monthly, or whatever is appropriate. Multiple generations of backup are retained, often three which gives rise to the name. The most recent backup is the son, the previous the father, and the oldest backup is the grandfather. This method is commonly used for a batch transaction processing system with a magnetic tape drive. If the system fails during a batch run, the master file is recreated by restoring the son backup and then restarting the batch. However, if the son backup fails, is corrupted or destroyed, then the previous generation of backup (the father) is used. Likewise, if that fails, then the generation of backup previous to the father (i.e. the grandfather) is required. Of course the older the generation, the more the data may be out of date.

Partial backups

Partial backups generally contain only records that have changed. For example, a full backup could be performed weekly, and then partial backups taken nightly. Recovery using this scheme involves restoring the last full backup and then restoring all partial backups in order to produce an up-to-date database. This process is quicker than taking only complete backups, at the expense of longer recovery time.

Advantages

  • Batch or real-time processing available
  • Reduction in processing time, lead time and order cycle time
  • Reduction in inventory, personnel and ordering costs
  • Increase in productivity and customer satisfaction

See also

Related Research Articles

<span class="mw-page-title-main">Database</span> Organized collection of data in computing

In computing, a database is an organized collection of data or a type of data store based on the use of a database management system (DBMS), the software that interacts with end users, applications, and the database itself to capture and analyze the data. The DBMS additionally encompasses the core facilities provided to administer the database. The sum total of the database, the DBMS and the associated applications can be referred to as a database system. Often the term "database" is also used loosely to refer to any of the DBMS, the database system or an application associated with the database.

<span class="mw-page-title-main">Mainframe computer</span> Large computer

A mainframe computer, informally called a mainframe or big iron, is a computer used primarily by large organizations for critical applications like bulk data processing for tasks such as censuses, industry and consumer statistics, enterprise resource planning, and large-scale transaction processing. A mainframe computer is large but not as large as a supercomputer and has more processing power than some other classes of computers, such as minicomputers, servers, workstations, and personal computers. Most large-scale computer-system architectures were established in the 1960s, but they continue to evolve. Mainframe computers are often used as servers.

<span class="mw-page-title-main">Transaction Processing Facility</span> IBM real-time operating system

Transaction Processing Facility (TPF) is an IBM real-time operating system for mainframe computers descended from the IBM System/360 family, including zSeries and System z9.

Computerized batch processing is a method of running software programs called jobs in batches automatically. While users are required to submit the jobs, no other interaction by the user is required to process the batch. Batches may automatically be run at scheduled times as well as being run contingent on the availability of computer resources.

In computer science, ACID is a set of properties of database transactions intended to guarantee data validity despite errors, power failures, and other mishaps. In the context of databases, a sequence of database operations that satisfies the ACID properties is called a transaction. For example, a transfer of funds from one bank account to another, even involving multiple changes such as debiting one account and crediting another, is a single transaction.

Systems Network Architecture (SNA) is IBM's proprietary networking architecture, created in 1974. It is a complete protocol stack for interconnecting computers and their resources. SNA describes formats and protocols but, in itself, is not a piece of software. The implementation of SNA takes the form of various communications packages, most notably Virtual Telecommunications Access Method (VTAM), the mainframe software package for SNA communications.

<span class="mw-page-title-main">Ingres (database)</span>

Ingres Database is a proprietary SQL relational database management system intended to support large commercial and government applications.

<span class="mw-page-title-main">IBM Db2</span> Relational model database server

Db2 is a family of data management products, including database servers, developed by IBM. It initially supported the relational model, but was extended to support object–relational features and non-relational structures like JSON and XML. The brand name was originally styled as DB/2, then DB2 until 2017 and finally changed to its present form.

<span class="mw-page-title-main">CICS</span> IBM mainframe transaction monitor

IBM CICS is a family of mixed-language application servers that provide online transaction management and connectivity for applications on IBM mainframe systems under z/OS and z/VSE.

NonStop is a series of server computers introduced to market in 1976 by Tandem Computers Inc., beginning with the NonStop product line. It was followed by the Tandem Integrity NonStop line of lock-step fault tolerant computers, now defunct. The original NonStop product line is currently offered by Hewlett Packard Enterprise since Hewlett-Packard Company's split in 2015. Because NonStop systems are based on an integrated hardware/software stack, Tandem and later HPE also developed the NonStop OS operating system for them.

In computing, the Distributed Computing Environment (DCE) software system was developed in the early 1990s from the work of the Open Software Foundation (OSF), a consortium that included Apollo Computer, IBM, Digital Equipment Corporation, and others. The DCE supplies a framework and a toolkit for developing client/server applications. The framework includes:

In computing, a Parallel Sysplex is a cluster of IBM mainframes acting together as a single system image with z/OS. Used for disaster recovery, Parallel Sysplex combines data sharing and parallel computing to allow a cluster of up to 32 systems to share a workload for high performance and high availability.

Replication in computing involves sharing information so as to ensure consistency between redundant resources, such as software or hardware components, to improve reliability, fault-tolerance, or accessibility.

Transarc Corporation was a private Pittsburgh-based software company founded in 1989 by Jeffrey Eppinger, Michael L. Kazar, Alfred Spector, and Dean Thompson of Carnegie Mellon University.

The history of IBM mainframe operating systems is significant within the history of mainframe operating systems, because of IBM's long-standing position as the world's largest hardware supplier of mainframe computers. IBM mainframes run operating systems supplied by IBM and by third parties.

A teleprocessing monitor is a control program that monitors the transfer of data between multiple local and remote terminals to ensure that the transaction processes completely or, if an error occurs, to take appropriate actions.

OS 2200 is the operating system for the Unisys ClearPath Dorado family of mainframe systems. The operating system kernel of OS 2200 is a lineal descendant of Exec 8 for the UNIVAC 1108. Documentation and other information on current and past Unisys systems can be found on the Unisys public support website.

<span class="mw-page-title-main">Telecommunication transaction processing systems</span>

Telecommunication networks can generate a vast amount of transactions where each transaction contains information about a particular subscriber's activity. Telecommunication network consist of various interacting devices and platforms. Any transaction carried out by a subscriber is often recorded in multiple devices as it passes through the network. Telecommunication organizations generally need to be able to extract transaction information from these various network elements in order to correctly bill subscribers for the usage on the network. Transaction processing system is a subset of information systems, and in the telecommunications industry, forms an integral part of the management information system. TPS can be regarded as the link between the various network elements and platforms and the information management uses to drive the business.

The following is provided as an overview of and topical guide to databases:

Encina was a DCE-based transaction processing system developed by Transarc, who were later acquired by IBM.

References

  1. "The Sabre Story" (PDF). Sabre.
  2. "History of TANDEM COMPUTERS, INC. – FundingUniverse". www.fundinguniverse.com. Retrieved 2023-03-01.
  3. "Terminals Help Manage Aluminum Firm's Production". Computerworld. July 26, 1976. Retrieved November 14, 2012.
  4. UNISYS Corporation (2012). Transaction Server for ClearPath MCP Configuration Guide (PDF).
  5. Digital Equipment Corporation (1989). VAX ACMS Guide to Creating Transaction Processing Applications.
  6. Bell, Gordon. "Digital Computing Timeline (1985)" . Retrieved November 15, 2012.
  7. Van Vleck, Thomas. "Multics Glossary -T-" . Retrieved November 15, 2012.
  8. Transarc. "Corporate Overview". Archived from the original on February 3, 1999. Retrieved November 16, 2012.
  9. IBM Corporation. "TXSeries for Multiplatforms" . Retrieved November 16, 2012.
  10. 1 2 3 Schuster, Stewart A. (June 15, 1981). "In Depth: Relational Data Base Management". Computerworld. Retrieved November 16, 2012.

Further reading