Architecture Design and Assessment System

Last updated

The Architecture Design and Assessment System (ADAS) was a set of software programs offered by the Research Triangle Institute from the mid-1980s until the early 1990s. [1]

Contents

A petri net-like graph model of a system was graphically created. The hierarchical graphs were simulated to determine resource utilization and throughput. Functional simulation of the model could be realized by attaching C or Ada code to the nodes. This enabled dynamic resource assignment, timing, and priority.

Simulation Model

An ADAS model consisted of nodes connected by directed arcs.

For abstract simulation a node represents a process (systems engineering) in the system that is being modeled. The readiness for execution or firing of this process requires that: its inputs are satisfied, space is available for its outputs, and its shared resource/hardware is available. During execution the node consumes its inputs, uses the resource for the prescribed duration, then produces its outputs.

The inputs and outputs of a process are represented by discrete tokens. These tokens flow along the arcs in the graph. If the maximum arc size is greater than one, then an arc would represent a buffer between system processes.

To refine the model of the system, a subgraph could be placed below the node to refine the behavior of that process.

Physical and behavioral properties were attached to nodes and arcs in the form of attributes. The attribute definition language allowed the computation of attributes from ancestor attributes and global values.

Related Research Articles

Distributed computing is a field of computer science that studies distributed systems. A distributed system is a system whose components are located on different networked computers, which communicate and coordinate their actions by passing messages to one another from any system. The components interact with one another in order to achieve a common goal. Three significant characteristics of distributed systems are: concurrency of components, lack of a global clock, and independent failure of components. It deals with a central challenge that, when components of a system fails, it doesn't imply the entire system fails. Examples of distributed systems vary from SOA-based systems to massively multiplayer online games to peer-to-peer applications.

VHDL

The VHSIC Hardware Description Language (VHDL) is a hardware description language (HDL) that can model the behavior and structure of digital systems at multiple levels of abstraction, ranging from the system level down to that of logic gates, for design entry, documentation, and verification purposes. Since 1987, VHDL has been standardized by the Institute of Electrical and Electronics Engineers (IEEE) as IEEE Std 1076; the latest version of which is IEEE Std 1076-2019. To model analog and mixed-signal systems, an IEEE-standardized HDL based on VHDL called VHDL-AMS has been developed.

In computer engineering, a hardware description language (HDL) is a specialized computer language used to describe the structure and behavior of electronic circuits, and most commonly, digital logic circuits.

Parallel computing Programming paradigm in which many processes are executed simultaneously

Parallel computing is a type of computation in which many calculations or processes are carried out simultaneously. Large problems can often be divided into smaller ones, which can then be solved at the same time. There are several different forms of parallel computing: bit-level, instruction-level, data, and task parallelism. Parallelism has long been employed in high-performance computing, but has gained broader interest due to the physical constraints preventing frequency scaling. As power consumption by computers has become a concern in recent years, parallel computing has become the dominant paradigm in computer architecture, mainly in the form of multi-core processors.

Visual programming language Programming language written graphically by a user

In computing, a visual programming language (VPL) is any programming language that lets users create programs by manipulating program elements graphically rather than by specifying them textually. A VPL allows programming with visual expressions, spatial arrangements of text and graphic symbols, used either as elements of syntax or secondary notation. For example, many VPLs are based on the idea of "boxes and arrows", where boxes or other screen objects are treated as entities, connected by arrows, lines or arcs which represent relations.

System of systems is a collection of task-oriented or dedicated systems that pool their resources and capabilities together to create a new, more complex system which offers more functionality and performance than simply the sum of the constituent systems. Currently, systems of systems is a critical research discipline for which frames of reference, thought processes, quantitative analysis, tools, and design methods are incomplete. The methodology for defining, abstracting, modeling, and analyzing system of systems problems is typically referred to as system of systems engineering.

A cognitive architecture refers to both a theory about the structure of the human mind and to a computational instantiation of such a theory used in the fields of artificial intelligence (AI) and computational cognitive science. The formalized models can be used to further refine a comprehensive theory of cognition and as a useful artificial intelligence program. Successful cognitive architectures include ACT-R and SOAR. The research on cognitive architectures as software instantiation of cognitive theories was initiated by Allen Newell in 1990.

Electric (software)

The Electric VLSI Design System is an EDA tool written in the early 1980s by Steven M. Rubin. Electric is used to construct logic wire schematics and to perform analysis of integrated circuit layout. It can also handle hardware description languages such as VHDL and Verilog. The system has many analysis and synthesis tools, including Design rule checking, Simulation, Routing, Layout vs. Schematic, Logical Effort, and more.

NEC V60 CISC microprocessor

The NEC V60 is a CISC microprocessor manufactured by NEC starting in 1986. Several improved versions were introduced with the same instruction set architecture (ISA), the V70 in 1987, and the V80 and AFPP in 1989. They were succeeded by the V800 product families, which is currently produced by Renesas.

Randal Bryant American computer scientist (born 1952)

Randal E. Bryant is an American computer scientist and academic noted for his research on formally verifying digital hardware and software. Bryant has been a faculty member at Carnegie Mellon University since 1984. He served as the Dean of the School of Computer Science (SCS) at Carnegie Mellon from 2004 to 2014. Dr. Bryant retired and became a Founders University Professor Emeritus on June 30th, 2020.

Neil J. Gunther

Neil Gunther is a computer information systems researcher best known internationally for developing the open-source performance modeling software Pretty Damn Quick and developing the Guerrilla approach to computer capacity planning and performance analysis. He has also been cited for his contributions to the theory of large transients in computer systems and packet networks, and his universal law of computational scalability.

MULTICUBE is a Seventh Framework Programme (FP7) project aimed to define innovative methods for the design optimization of computer architectures for the embedded system domain.

MLDesigner is an integrated modeling and simulation tool for the design and analysis of complex embedded and networked systems. MLDesigner speeds up modeling, simulation and analysis of discrete event, discrete time and continuous time systems concerning architecture, function and performance. The tools is based on ideas of the "Ptolemy Project", done at the University if California Berkeley. MLDesigner is developed by MLDesign Technologies Inc. Palo Alto, CA, USA in collaboration with Mission Level Design GmbH, Ilmenau, Germany.

SpiNNaker

SpiNNaker is a massively parallel, manycore supercomputer architecture designed by the Advanced Processor Technologies Research Group (APT) at the Department of Computer Science, University of Manchester. It is composed of 57,600 processing nodes, each with 18 ARM9 processors and 128 MB of mobile DDR SDRAM, totalling 1,036,800 cores and over 7 TB of RAM. The computing platform is based on spiking neural networks, useful in simulating the human brain.

Jerzy W. Rozenblit

Jerzy W. Rozenblit is a University Distinguished Professor and Raymond J. Oglethorpe Endowed Chair in the Department of Electrical and Computer Engineering (ECE) at the University of Arizona, in Tucson, Arizona. He also holds a joint appointment as Professor of Surgery in the University of Arizona College of Medicine. From 2003 to 2011 he served as the ECE Department Head. During his tenure at the University of Arizona, he established the Model-Based Design Laboratory with major projects in complex systems design hardware software codesign, modeling, and computer-aided minimally-invasive surgical training. He presently serves as Director of the Life-Critical Computing Systems Initiative, a research enterprise intended to improve the reliability and safety of technology in life-critical applications.

A high-level language computer architecture (HLLCA) is a computer architecture designed to be targeted by a specific high-level programming language (HLL), rather than the architecture being dictated by hardware considerations. It is accordingly also termed language-directed computer design, coined in McKeeman (1967) and primarily used in the 1960s and 1970s. HLLCAs were popular in the 1960s and 1970s, but largely disappeared in the 1980s. This followed the dramatic failure of the Intel 432 (1981) and the emergence of optimizing compilers and reduced instruction set computer (RISC) architectures and RISC-like complex instruction set computer (CISC) architectures, and the later development of just-in-time compilation (JIT) for HLLs. A detailed survey and critique can be found in Ditzel & Patterson (1980).

Charles Albert Zukowski is a professor and former chair of the Department of Electrical Engineering at Columbia University. Zukowski was born in Buffalo, New York. While still a student at MIT, from 1979 to 1982, Zukowski worked at the Thomas J. Watson Research Center. He received his BS in Electrical Engineering from the Massachusetts Institute of Technology in 1982. He received the IBM PhD Fellowship from 1982 to 1985; in 1985 he earned his PhD in Electrical Engineering with a thesis entitled "Design and measurement of a reconfigurable multi-microprocessor machine". The same year, he joined the faculty of Columbia University as Assistant Professor, and was awarded tenure in 1993. Zukowski is an active member of IEEE and was made an IEEE Fellow in 2000. Zukowski's present research focuses on VLSI circuits and integrated circuit (IC) optimization, though in the past he has published in the fields of systems biology and computer architecture.

Prabhat Mishra

Prabhat Mishra is a Professor in the Department of Computer and Information Science and Engineering and a UF Research Foundation Professor at the University of Florida. Prof. Mishra's research interests are in hardware security, quantum computing, embedded systems, system-on-chip validation, formal verification, and post-silicon debug.

Rajat Moona is the Director of Indian Institute of Technology Bhilai. He was also a professor of Computer Science and Engineering at IIT Kanpur and Director General of Centre for Development of Advanced Computing.

Signal Transition Graphs (STGs) are typically used in electronic engineering and computer engineering to describe dynamic behaviour of asynchronous circuits, for the purposes of their analysis or synthesis.

References

  1. G.A. Frank, D.L. Franke, and W.F. Ingogly, "An Architecture Design and Assessment System,"VLSI Design, Vol. 6, No. 8, Aug. 1985, pp. 30-50

Further reading