This article includes a list of general references, but it remains largely unverified because it lacks sufficient corresponding inline citations .(January 2021) |
BORO (Business Objects Reference Ontology) is an approach to developing ontological or semantic models for large complex operational applications that consists of a top ontology as well as a process for constructing the ontology. It was originally developed as a method for mining ontologies from multiple legacy systems – as the first stage in an architectural transformation [1] or software modernization. It has also been used to enable semantic interoperability between legacy systems. It is described in detail in (Partridge 1996, 2005). It is the analysis method used in the development and maintenance of the U.S. Department of Defense Architecture Framework (DoDAF) Meta Model (DM2), where a data modeling working group of over 350 members was able to systematically resolve a broad spectrum of knowledge representation issues.
The approach was developed in the late 1980s and early 1990s by a team of KPMG consultants led by Chris Partridge. The team was working on a complex legacy systems re-engineering project and needed a new approach. The prime challenge of the re-engineering work was to clarify the underlying ontology of the systems and the work focussed on developing a process for mining ontologies and a top ontology that formed the foundation for the analysis. The top ontology was tailored to meet the needs of the re-engineering. Early work established that a key factor was to make a series of clear metaphysical choices to provide a solid (metaphysical) foundation.
A key choice was for an extensional (and hence, four-dimensional) ontology which provided neat Criterion of identity. Using this top ontology as a basis, a systematic process for re-engineering legacy systems was developed. From a software engineering perspective, a key feature of this process was the identification of common general patterns, under which the legacy system was subsumed. It has been substantially developed since then.
Much of the approach and the associated tools are proprietary, but some aspects have been delivered to the public domain and elements of it have appeared in a number of standards. For example, the ISO standard, ISO 15926 – Industrial automation systems and integration – was heavily influenced by an early version. [2] The IDEAS (International Defence Enterprise Architecture Specification for exchange) standard is based upon BORO, which in turn was used to develop DODAF 2.0. From 2003 to 2008, the start-up company 42 Objects, funded by private equity company 3i worked on developing systems based upon BORO.
The BORO approach is designed to be a simple, repeatable process for developing formal ontologies. The method takes an extensional approach to ontology development. The method aims to be grounded in physical reality so that if followed to the letter the method should consistently produce the same ontology given the same inputs. It can then be used for comparing multiple data-sources for semantic matches/mismatches and for re-engineering multiple legacy systems into a coherent whole (either as a new monolithic system, or as a method for designing federation of existing systems).
BORO's purpose is to improve the quality of information and information models, to integrate multiple information sources and extract hidden semantics. The purpose of the method is to re-engineer disparate data sources into a common model. It is meant to be focused on semantic analysis – establishing whether two concepts are the same, if they overlap, or if they are unrelated. This is based on using resources from higher order logic, mathematics and philosophy.
For example in the case of Criterion of identity, the method adopts an extensional approach. As an example, take “Waterloo Bridge” as a term. The first thing we ask is “does it refer to an object that has a spatial and temporal extent ?”. It has spatial extent; it spans the River Thames. However, when we examine the temporal extent we realise there have been two bridges at that site. The first, built in 1817 (two years after the battle of Waterloo) was demolished in 1920. The bridge that stands there now was built in 1942. This analysis has immediately highlighted a problem with a name-based approach – there are two bridges of that name, which one are we referring to? At this point, the analyst can add one or both of the bridges to the ontology, then apply the appropriate names to each.
The process also works for types of things. Take “bridges” as a concept. It doesn’t have spatiotemporal extent, so we go to the next question “does it have members ?”. It does – the members are all the bridges in the world. We then identify some exemplar members – e.g. Waterloo Bridge. At this stage, it is advisable to identify exemplars that are “on the edge” of the set – e.g. things that may or may not be bridges – e.g. pontoons, bridging vehicles, etc. so as to accurately identify the extent of the type.
The final concept covered by the process is the tuple. A tuple is a relationship between things. If the concept under analysis is neither a type nor an individual, then it must be a tuple. We identify the things at the end of the tuple then add it to the ontology.
Traditional methods of data analysis tend to be linguistic; comparison of concepts is based on the names these concepts have. More modern methods have introduced a semantic approach, where the analyst will tend to analyse the underlying senses of the word (meaning). A lot of it depends on the analyst’s domain knowledge and linguistic interpretation. Although BORO produces an ontology (information science) in the very strictest sense of the term, it is not intended to produce the type of ontology (information science) that computer scientists would use for reasoning and inference. BORO is different to many other data analysis techniques in that treats the names of things as a secondary concern. With BORO, the analyst is forced to identify individual concepts by their extent.
The BORO methodology is best summarised as a flowchart:
The method has been presented several times, including a tutorial at the Integrated Enterprise Architecture Conference in London in 2008. It was also presented at the UK Ministry of Defence's EKIG conference in October 2009.
In computer science and information science, an ontology encompasses a representation, formal naming and definition of the categories, properties and relations between the concepts, data and entities that substantiate one, many, or all domains of discourse. More simply, an ontology is a way of showing the properties of a subject area and how they are related, by defining a set of concepts and categories that represent the subject.
A data model is an abstract model that organizes elements of data and standardizes how they relate to one another and to the properties of real-world entities. For instance, a data model may specify that the data element representing a car be composed of a number of other elements which, in turn, represent the color and size of the car and define its owner.
Software development is the process of conceiving, specifying, designing, programming, documenting, testing, and bug fixing involved in creating and maintaining applications, frameworks, or other software components. Software development involves writing and maintaining the source code, but in a broader sense, it includes all processes from the conception of the desired software through to the final manifestation of the software, typically in a planned and structured process. Software development also includes research, new development, prototyping, modification, reuse, re-engineering, maintenance, or any other activities that result in software products.
A modeling language is any artificial language that can be used to express information or knowledge or systems in a structure that is defined by a consistent set of rules. The rules are used for interpretation of the meaning of components in the structure.
IDEF, initially an abbreviation of ICAM Definition and renamed in 1999 as Integration Definition, is a family of modeling languages in the field of systems and software engineering. They cover a wide range of uses from functional modeling to data, simulation, object-oriented analysis and design, and knowledge acquisition. These definition languages were developed under funding from U.S. Air Force and, although still most commonly used by them and other military and United States Department of Defense (DoD) agencies, are in the public domain.
Data modeling in software engineering is the process of creating a data model for an information system by applying certain formal techniques.
Object-role modeling (ORM) is used to model the semantics of a universe of discourse. ORM is often used for data modeling and software engineering.
A metamodel or surrogate model is a model of a model, and metamodeling is the process of generating such metamodels. Thus metamodeling or meta-modeling is the analysis, construction and development of the frames, rules, constraints, models and theories applicable and useful for modeling a predefined class of problems. As its name implies, this concept applies the notions of meta- and modeling in software engineering and systems engineering. Metamodels are of many types and have diverse applications.
A conceptual model is a representation of a system. It consists of concepts used to help people know, understand, or simulate a subject the model represents. It is also a set of concepts. In contrast, physical models are physical objects, such as a toy model that may be assembled and made to work like the object it represents.
The Shlaer–Mellor method, also known as Object-Oriented Systems Analysis (OOSA) or Object-Oriented Analysis (OOA) is an object-oriented software development methodology introduced by Sally Shlaer and Stephen Mellor in 1988. The method makes the documented analysis so precise that it is possible to implement the analysis model directly by translation to the target architecture, rather than by elaborating model changes through a series of more platform-specific models. In the new millennium the Shlaer–Mellor method has migrated to the UML notation, becoming Executable UML.
In information science, an upper ontology is an ontology which consists of very general terms that are common across all domains. An important function of an upper ontology is to support broad semantic interoperability among a large number of domain-specific ontologies by providing a common starting point for the formulation of definitions. Terms in the domain ontology are ranked under the terms in the upper ontology, e.g., the upper ontology classes are superclasses or supersets of all the classes in the domain ontologies.
The IDEAS Group is the International Defence Enterprise Architecture Specification for exchange Group. The deliverable of the project is a data exchange format for military Enterprise Architectures. The scope is four nation and covers MODAF (UK), DoDAF (USA), DNDAF (Canada) and the Australian Defence Architecture Framework (AUSDAF). The initial scope for exchange is the architectural data required to support coalition operations planning -
Resources, events, agents (REA) is a model of how an accounting system can be re-engineered for the computer age. REA was originally proposed in 1982 by William E. McCarthy as a generalized accounting model, and contained the concepts of resources, events and agents.
Enterprise modelling is the abstract representation, description and definition of the structure, processes, information and resources of an identifiable business, government body, or other large organization.
DOGMA, short for Developing Ontology-Grounded Methods and Applications, is the name of research project in progress at Vrije Universiteit Brussel's STARLab, Semantics Technology and Applications Research Laboratory. It is an internally funded project, concerned with the more general aspects of extracting, storing, representing and browsing information.
IDEF5 is a software engineering method to develop and maintain usable, accurate domain ontologies. This standard is part of the IDEF family of modeling languages in the field of software engineering.
Enterprise engineering is the body of knowledge, principles, and practices used to design all or part of an enterprise. An enterprise is a complex socio-technical system that comprises people, information, and technology that interact with each other and their environment in support of a common mission. One definition is: "an enterprise life-cycle oriented discipline for the identification, design, and implementation of enterprises and their continuous evolution", supported by enterprise modelling. The discipline examines each aspect of the enterprise, including business processes, information flows, material flows, and organizational structure. Enterprise engineering may focus on the design of the enterprise as a whole, or on the design and integration of certain business components.
Business Semantics Management (BSM) encompasses the technology, methodology, organization, and culture that brings business stakeholders together to collaboratively realize the reconciliation of their heterogeneous metadata; and consequently the application of the derived business semantics patterns to establish semantic alignment between the underlying data structures.
In computer science, information science and systems engineering, ontology engineering is a field which studies the methods and methodologies for building ontologies, which encompasses a representation, formal naming and definition of the categories, properties and relations between the concepts, data and entities. In a broader sense, this field also includes a knowledge construction of the domain using formal ontology representations such as OWL/RDF. A large-scale representation of abstract concepts such as actions, time, physical objects and beliefs would be an example of ontological engineering. Ontology engineering is one of the areas of applied ontology, and can be seen as an application of philosophical ontology. Core ideas and objectives of ontology engineering are also central in conceptual modeling.
Model Driven Interoperability (MDI) is a methodological framework, which provides a conceptual and technical support to make interoperable enterprises using ontologies and semantic annotations, following model driven development (MDD) principles.