OpenPDC

Last updated

The openPDC is a complete set of applications for processing streaming time-series data in real-time. The name stands for "open source phasor data concentrator" and was originally designed for the concentration and management of real-time streaming synchrophasors. Due to the system's modular design, the openPDC can be classified as a generic event stream processor.

Contents

History

The openPDC is based on the SuperPDC which was developed by the Tennessee Valley Authority starting in 2004. The openPDC officially launched on October 7, 2009. [1] The first stable release of the openPDC was made available on January 28, 2010. [2] Then on February 23, 2010, AREVA T&D announced that they would be providing commercial support for the openPDC. Finally, on April 1, 2010, the Grid Protection Alliance inherited development of the openPDC and entered into a contract with the North American Electric Reliability Corporation in an effort to further enable its use as a distributed system.

Operation

The phasor data concentrator runs as a Windows service. The service is responsible for managing the life cycle of adapters that create and process the streaming phasor measurements. Adapters are split into three layers: [3] [4]

  1. The input adapter layer is typically responsible for receiving data from an outside source such as a phasor measurement unit. That data is used to create measurements which are sent to other adapters to be processed or archived.
  2. The action adapter layer is typically responsible for concentration and processing of the input measurements. Adapters in this layer can also introduce new measurements to the system, just like adapters in the input adapter layer.
  3. The output adapter layer is typically responsible for archival of measurements received from the input adapter layer and the action adapter layer.

These adapters can be configured using any one of three supported database systems: Microsoft SQL Server, MySQL, and Microsoft Access. [5]

Features

The openPDC implements a number of standard phasor protocols which can be used to receive data from devices. The supported protocols are IEEE C37.118 (v1.0 and Draft 6), IEEE 1344, BPA PDCstream, FNET, SEL Fast Message, and Macrodyne. [3] There is also a built-in concentration engine which sorts the real-time data into frames based on the timestamp associated with each measurement. The sorted measurements can then be streamed to other applications using either IEEE C37.118 or BPA PDCstream. [6] The system is capable of concentrating over one billion measurements per day coming from over 100 phasor measurement units across the Eastern Interconnection. [7] [8]

The project includes a historian for data archival. The files produced by the historian can later be analyzed using Hadoop. [3] The historian is also designed to be used in distributed systems, having a single node dedicated to archival and several others sending concentrated measurements to the central archive. This design even allows for its use in any distributed system enabling one PDC to filter measurements to another that is running a resource intensive adapter. [9] The local historian provides web services which can be used to monitor the data being archived in real-time. [10] [11]

The data quality monitoring adapters provide information about measurements entering the concentrator that have flat-lined, gone out of a specified range, or entered the concentrator with a bad timestamp. [12]

Applications

Applications included in the openPDC project:

See also

Related Research Articles

<span class="mw-page-title-main">IBM 3270</span> Family of block-oriented display terminals and printers made by IBM

The IBM 3270 is a family of block oriented display and printer computer terminals introduced by IBM in 1971 and normally used to communicate with IBM mainframes. The 3270 was the successor to the IBM 2260 display terminal. Due to the text color on the original models, these terminals are informally known as green screen terminals. Unlike a character-oriented terminal, the 3270 minimizes the number of I/O interrupts required by transferring large blocks of data known as data streams, and uses a high speed proprietary communications interface, using coaxial cable.

<span class="mw-page-title-main">Software testing</span> Checking software against a standard

Software testing is the act of checking whether software satisfies expectations.

<span class="mw-page-title-main">Multiplexing</span> Method of combining multiple signals into one signal over a shared medium

In telecommunications and computer networking, multiplexing is a method by which multiple analog or digital signals are combined into one signal over a shared medium. The aim is to share a scarce resource – a physical transmission medium. For example, in telecommunications, several telephone calls may be carried using one wire. Multiplexing originated in telegraphy in the 1870s, and is now widely applied in communications. In telephony, George Owen Squier is credited with the development of telephone carrier multiplexing in 1910.

<span class="mw-page-title-main">Network interface controller</span> Hardware component that connects a computer to a network

A network interface controller is a computer hardware component that connects a computer to a computer network.

<span class="mw-page-title-main">Electronic test equipment</span> Testing appliance for electronics systems

Electronic test equipment is used to create signals and capture responses from electronic devices under test (DUTs). In this way, the proper operation of the DUT can be proven or faults in the device can be traced. Use of electronic test equipment is essential to any serious work on electronics systems.

In software project management, software testing, and software engineering, verification and validation is the process of checking that a software engineer system meets specifications and requirements so that it fulfills its intended purpose. It may also be referred to as software quality control. It is normally the responsibility of software testers as part of the software development lifecycle. In simple terms, software verification is: "Assuming we should build X, does our software achieve its goals without any bugs or gaps?" On the other hand, software validation is: "Was X what we should have built? Does X meet the high-level requirements?"

An application firewall is a form of firewall that controls input/output or system calls of an application or service. It operates by monitoring and blocking communications based on a configured policy, generally with predefined rule sets to choose from. The two primary categories of application firewalls are network-based and host-based.

Tuxedo is a middleware platform used to manage distributed transaction processing in distributed computing environments. Tuxedo is a transaction processing system or transaction-oriented middleware, or enterprise application server for a variety of systems and programming languages. Developed by AT&T in the 1980s, it became a software product of Oracle Corporation in 2008 when they acquired BEA Systems. Tuxedo is now part of the Oracle Fusion Middleware.

A penetration test, colloquially known as a pentest, is an authorized simulated cyberattack on a computer system, performed to evaluate the security of the system; this is not to be confused with a vulnerability assessment. The test is performed to identify weaknesses, including the potential for unauthorized parties to gain access to the system's features and data, as well as strengths, enabling a full risk assessment to be completed.

<span class="mw-page-title-main">Automatic test equipment</span> Apparatus used in hardware testing that carries out a series of tests automatically

Automatic test equipment or automated test equipment (ATE) is any apparatus that performs tests on a device, known as the device under test (DUT), equipment under test (EUT) or unit under test (UUT), using automation to quickly perform measurements and evaluate the test results. An ATE can be a simple computer-controlled digital multimeter, or a complicated system containing dozens of complex test instruments capable of automatically testing and diagnosing faults in sophisticated electronic packaged parts or on wafer testing, including system on chips and integrated circuits.

Power-system automation is the act of automatically controlling the power system via instrumentation and control devices. Substation automation refers to using data from Intelligent electronic devices (IED), control and automation capabilities within the substation, and control commands from remote users to control power-system devices.

The international standard IEC 61499, addressing the topic of function blocks for industrial process measurement and control systems, was initially published by the International Electrotechnical Commission (IEC) in 2005. The specification of IEC 61499 defines a generic model for distributed control systems and is based on the IEC 61131 standard. The concepts of IEC 61499 are also explained by Lewis and Zoitl as well as Vyatkin.

<span class="mw-page-title-main">Phasor measurement unit</span> Device measuring electrical waves on a power grid

A phasor measurement unit (PMU) is a device used to estimate the magnitude and phase angle of an electrical phasor quantity in the electricity grid using a common time source for synchronization. Time synchronization is usually provided by GPS or IEEE 1588 Precision Time Protocol, which allows synchronized real-time measurements of multiple remote points on the grid. PMUs are capable of capturing samples from a waveform in quick succession and reconstructing the phasor quantity, made up of an angle measurement and a magnitude measurement. The resulting measurement is known as a synchrophasor. These time synchronized measurements are important because if the grid’s supply and demand are not perfectly matched, frequency imbalances can cause stress on the grid, which is a potential cause for power outages.

In electric power systems and industrial automation, ANSI Device Numbers can be used to identify equipment and devices in a system such as relays, circuit breakers, or instruments. The device numbers are enumerated in ANSI/IEEE Standard C37.2 Standard for Electrical Power System Device Function Numbers, Acronyms, and Contact Designations.

In software engineering, graphical user interface testing is the process of testing a product's graphical user interface (GUI) to ensure it meets its specifications. This is normally done through the use of a variety of test cases.

<span class="mw-page-title-main">UniPro</span> High-speed interface technology

UniPro is a high-speed interface technology for interconnecting integrated circuits in mobile and mobile-influenced electronics. The various versions of the UniPro protocol are created within the MIPI Alliance, an organization that defines specifications targeting mobile and mobile-influenced applications.

<span class="mw-page-title-main">Storage area network</span> Network which provides access to consolidated, block-level data storage

A storage area network (SAN) or storage network is a computer network which provides access to consolidated, block-level data storage. SANs are primarily used to access data storage devices, such as disk arrays and tape libraries from servers so that the devices appear to the operating system as direct-attached storage. A SAN typically is a dedicated network of storage devices not accessible through the local area network (LAN).

<span class="mw-page-title-main">FNET</span>

FNET is a wide-area power system frequency measurement system. Using a type of phasor measurement unit (PMU) known as a frequency disturbance recorder (FDR), FNET/GridEye is able to measure the power system frequency, voltage, and angle very accurately. These measurements can then be used to study various power system phenomena, and may play an important role in the development of future smart grid technologies. The FNET/GridEye system is currently operated by the Power Information Technology Laboratory at the University of Tennessee (UTK) in Knoxville, Tennessee, and Oak Ridge National Laboratory (ORNL) in Oak Ridge, Tennessee.

RTP-MIDI is a protocol to transport MIDI messages within Real-time Transport Protocol (RTP) packets over Ethernet and WiFi networks. It is completely open and free, and is compatible both with LAN and WAN application fields. Compared to MIDI 1.0, RTP-MIDI includes new features like session management, device synchronization and detection of lost packets, with automatic regeneration of lost data. RTP-MIDI is compatible with real-time applications, and supports sample-accurate synchronization for each MIDI message.

<span class="mw-page-title-main">Apache Flink</span> Framework and distributed processing engine

Apache Flink is an open-source, unified stream-processing and batch-processing framework developed by the Apache Software Foundation. The core of Apache Flink is a distributed streaming data-flow engine written in Java and Scala. Flink executes arbitrary dataflow programs in a data-parallel and pipelined manner. Flink's pipelined runtime system enables the execution of bulk/batch and stream processing programs. Furthermore, Flink's runtime supports the execution of iterative algorithms natively.

References

  1. Jim Allen (7 October 2009). TVA Opens Data Collection Software for Industry Use. Retrieved 8 June 2010.
  2. openPDC v1.0 Release. Retrieved 8 June 2010.
  3. 1 2 3 openPDC Single Page Overview. Retrieved 8 June 2010.
  4. How to Create a Custom Adapter. Retrieved 9 June 2010.
  5. openPDC Getting Started: Set up the database. Retrieved 9 June 2010.
  6. How to Use the openPDC Manager: Configuring outputs. Retrieved 10 June 2010.
  7. James Ritchie Carroll (May 2007). TVA SPDC Update Archived 2011-07-27 at the Wayback Machine , page 8. Retrieved 15 June 2010.
  8. Josh Patterson (2 June 2009). The Smart Grid: Hadoop at the Tennessee Valley Authority (TVA). Retrieved 15 June 2010.
  9. James Ritchie Carroll. Updates for Simplification of Scale-Out. Retrieved 11 June 2010.
  10. openPDC Getting Started: Using the in-process historian adapter. Retrieved 10 June 2010.
  11. How to Use the openPDC Manager: Viewing real-time data. Retrieved 10 June 2010.
  12. Data Quality Monitoring Adapters. Retrieved 10 June 2010.
  13. openPDC Getting Started: Using the openPDC Console. Retrieved 9 June 2010.
  14. 1 2 3 Introducing the openPDC Management System. Retrieved 9 June 2010.
  15. openPDC Getting Started: Running the PMU Connection Tester. Retrieved 9 June 2010.
  16. PMU Connection Tester. Retrieved 9 June 2010.
  17. Board Technology Committee (23 October 2008). NERC Conference Call Agenda Archived 2011-06-08 at the Wayback Machine . Retrieved 7 July 2010.
  18. June 2010 openPDC v1.1 Release. Retrieved 7 July 2010.
  19. NASPI PMU Registry. Retrieved 14 June 2010.
  20. openPDC Getting Started: Encrypt the configuration settings. Retrieved 9 June 2010.