OpenCorporates

Last updated

OpenCorporates
OpenCorporates Logo.png
Type of site
Public records database
Available inEnglish
OwnerOpenCorporates Ltd
URL opencorporates.com OOjs UI icon edit-ltr-progressive.svg
RegistrationOptional [lower-alpha 1]
Launched20 December 2010;13 years ago (2010-12-20)
Current statusActive
Content license
Open Database Licence

OpenCorporates is a website that shares data on corporations under the copyleft Open Database License. The company, OpenCorporates Ltd, [lower-alpha 2] [3] was incorporated on 18 December 2010 [2] by Chris Taggart and Rob McKinnon, and the website was officially launched on 20th. [4]

Contents

Data is sourced from national business registries in 140 jurisdictions, and presented in a standardised form. Collected data comprises the name of the entity, date of incorporation, registered addresses, and the names of directors. Some data, such as the ownership structure, is contributed by users. [5] [6]

Recognition

Co-founder Chris Taggart Chris Taggart 2013.JPG
Co-founder Chris Taggart

In 2011, the site won third place in the Open Data Challenge. [7] Vice President of the European Commission Neelie Kroes said the site "is the kind of resource the (Digital) Single Market needs and it is encouraging to see that it is being built." [8] The project was represented on the European Union's Core Vocabularies Working Group's Core Business Task Force. [9]

In early 2012, the project was appointed to the Financial Stability Board's advisory panel on a Legal Entity Identification for Financial Contracts. [10]

In July 2015, OpenCorporates was a finalist in both the Business and Publisher categories at the Open Data Institute Awards. [11] It was announced as the winner of the Open Data Business Award due to work with promoting data transparency in the corporate sector. [12]

Usage

The service has been used to study public procurement data, [13] online hiring market, [14] to visualize and analyze company data [15] [16] [17] to analyze tax havens, illicit activities of companies. [18]

See also

Notes

  1. Not required for all features, but required for certain information, such as directors and incorporation data. [1]
  2. Known as Chrinon Ltd until 2018. [2]

Related Research Articles

<span class="mw-page-title-main">Semantic Web</span> Extension of the Web to facilitate data exchange

The Semantic Web, sometimes known as Web 3.0, is an extension of the World Wide Web through standards set by the World Wide Web Consortium (W3C). The goal of the Semantic Web is to make Internet data machine-readable.

In information science, an ontology encompasses a representation, formal naming, and definitions of the categories, properties, and relations between the concepts, data, or entities that pertain to one, many, or all domains of discourse. More simply, an ontology is a way of showing the properties of a subject area and how they are related, by defining a set of terms and relational expressions that represent the entities in that subject area. The field which studies ontologies so conceived is sometimes referred to as applied ontology.

An annotation is extra information associated with a particular point in a document or other piece of information. It can be a note that includes a comment or explanation. Annotations are sometimes presented in the margin of book pages. For annotations of different digital media, see web annotation and text annotation.

<span class="mw-page-title-main">Neelie Kroes</span> Dutch politician

Neelie Kroes is a retired Dutch politician of the People's Party for Freedom and Democracy (VVD) and businessperson who served as European Commissioner from 22 November 2004 to 1 November 2014.

Semantic similarity is a metric defined over a set of documents or terms, where the idea of distance between items is based on the likeness of their meaning or semantic content as opposed to lexicographical similarity. These are mathematical tools used to estimate the strength of the semantic relationship between units of language, concepts or instances, through a numerical description obtained according to the comparison of information supporting their meaning or describing their nature. The term semantic similarity is often confused with semantic relatedness. Semantic relatedness includes any relation between two terms, while semantic similarity only includes "is a" relations. For example, "car" is similar to "bus", but is also related to "road" and "driving".

In information science, an upper ontology is an ontology that consists of very general terms that are common across all domains. An important function of an upper ontology is to support broad semantic interoperability among a large number of domain-specific ontologies by providing a common starting point for the formulation of definitions. Terms in the domain ontology are ranked under the terms in the upper ontology, e.g., the upper ontology classes are superclasses or supersets of all the classes in the domain ontologies.

The CIDOC Conceptual Reference Model (CRM) provides an extensible ontology for concepts and information in cultural heritage and museum documentation. It is the international standard (ISO 21127:2023) for the controlled exchange of cultural heritage information. Galleries, libraries, archives, museums (GLAMs), and other cultural institutions are encouraged to use the CIDOC CRM to enhance accessibility to museum-related information and knowledge.

The concept of the Social Semantic Web subsumes developments in which social interactions on the Web lead to the creation of explicit and semantically rich knowledge representations. The Social Semantic Web can be seen as a Web of collective knowledge systems, which are able to provide useful information based on human contributions and which get better as more people participate. The Social Semantic Web combines technologies, strategies and methodologies from the Semantic Web, social software and the Web 2.0.

<span class="mw-page-title-main">Linked data</span> Structured data and method for its publication

In computing, linked data is structured data which is interlinked with other data so it becomes more useful through semantic queries. It builds upon standard Web technologies such as HTTP, RDF and URIs, but rather than using them to serve web pages only for human readers, it extends them to share information in a way that can be read automatically by computers. Part of the vision of linked data is for the Internet to become a global database.

<span class="mw-page-title-main">DBpedia</span> Online database project

DBpedia is a project aiming to extract structured content from the information created in the Wikipedia project. This structured information is made available on the World Wide Web using OpenLink Virtuoso. DBpedia allows users to semantically query relationships and properties of Wikipedia resources, including links to other related datasets.

<span class="mw-page-title-main">Nigel Shadbolt</span> Principal of Jesus College, Oxford

Sir Nigel Richard Shadbolt is Principal of Jesus College, Oxford, and Professorial Research Fellow in the Department of Computer Science, University of Oxford. He is chairman of the Open Data Institute which he co-founded with Tim Berners-Lee. He is also a visiting professor in the School of Electronics and Computer Science at the University of Southampton. Shadbolt is an interdisciplinary researcher, policy expert and commentator. His research focuses on understanding how intelligent behaviour is embodied and emerges in humans, machines and, most recently, on the Web, and has made contributions to the fields of Psychology, Cognitive science, Computational neuroscience, Artificial Intelligence (AI), Computer science and the emerging field of Web science.

Amit Sheth is a computer scientist at University of South Carolina in Columbia, South Carolina. He is the founding Director of the Artificial Intelligence Institute, and a Professor of Computer Science and Engineering. From 2007 to June 2019, he was the Lexis Nexis Ohio Eminent Scholar, director of the Ohio Center of Excellence in Knowledge-enabled Computing, and a Professor of Computer Science at Wright State University. Sheth's work has been cited by over 48,800 publications. He has an h-index of 117, which puts him among the top 100 computer scientists with the highest h-index. Prior to founding the Kno.e.sis Center, he served as the director of the Large Scale Distributed Information Systems Lab at the University of Georgia in Athens, Georgia.

<span class="mw-page-title-main">Ontology engineering</span> Field that studies the methods and methodologies for building ontologies

In computer science, information science and systems engineering, ontology engineering is a field which studies the methods and methodologies for building ontologies, which encompasses a representation, formal naming and definition of the categories, properties and relations between the concepts, data and entities of a given domain of interest. In a broader sense, this field also includes a knowledge construction of the domain using formal ontology representations such as OWL/RDF. A large-scale representation of abstract concepts such as actions, time, physical objects and beliefs would be an example of ontological engineering. Ontology engineering is one of the areas of applied ontology, and can be seen as an application of philosophical ontology. Core ideas and objectives of ontology engineering are also central in conceptual modeling.

<span class="mw-page-title-main">YAGO (database)</span> Open-source information repository

YAGO is an open source knowledge base developed at the Max Planck Institute for Informatics in Saarbrücken. It is automatically extracted from Wikidata and Schema.org.

Knowledge extraction is the creation of knowledge from structured and unstructured sources. The resulting knowledge needs to be in a machine-readable and machine-interpretable format and must represent knowledge in a manner that facilitates inferencing. Although it is methodically similar to information extraction (NLP) and ETL, the main criterion is that the extraction result goes beyond the creation of structured information or the transformation into a relational schema. It requires either the reuse of existing formal knowledge or the generation of a schema based on the source data.

<span class="mw-page-title-main">Apache Stanbol</span>

Apache Stanbol is an open source modular software stack and reusable set of components for semantic content management. Apache Stanbol components are meant to be accessed over RESTful interfaces to provide semantic services for content management. Thus, one application is to extend traditional content management systems with semantic services.

<span class="mw-page-title-main">UMBEL</span>

UMBEL is a logically organized knowledge graph of 34,000 concepts and entity types that can be used in information science for relating information from disparate sources to one another. It was retired at the end of 2019. UMBEL was first released in July 2008. Version 1.00 was released in February 2011. Its current release is version 1.50.

In natural language processing, linguistics, and neighboring fields, Linguistic Linked Open Data (LLOD) describes a method and an interdisciplinary community concerned with creating, sharing, and (re-)using language resources in accordance with Linked Data principles. The Linguistic Linked Open Data Cloud was conceived and is being maintained by the Open Linguistics Working Group (OWLG) of the Open Knowledge Foundation, but has been a point of focal activity for several W3C community groups, research projects, and infrastructure efforts since then.

The Bahama Leaks are 1.3 million internal files from the company register of the Bahamas. After the release of the Panama Papers in 2016, an unknown source handed over internal data from the national corporate registry of the Bahamas to Frederik Obermaier and Bastian Obermayer, who analyzed them with the help of the International Consortium of Investigative Journalists (ICIJ). At the same time, ICIJ, Süddeutsche Zeitung and other media partners published detailed reporting before they published an online database of offshore entities.

<span class="mw-page-title-main">Knowledge graph</span> Type of knowledge base

In knowledge representation and reasoning, a knowledge graph is a knowledge base that uses a graph-structured data model or topology to represent and operate on data. Knowledge graphs are often used to store interlinked descriptions of entities – objects, events, situations or abstract concepts – while also encoding the free-form semantics or relationships underlying these entities.

References

  1. "Important changes to our website service". OpenCorporates Knowledge Base. 26 September 2023. Retrieved 18 January 2024.
  2. 1 2 "OPENCORPORATES LTD overview - Find and update company information - GOV.UK". Companies House . 18 November 2010. Retrieved 18 January 2024.
  3. "Public records privacy policy". OpenCorporates. Retrieved 18 January 2024.
  4. "OpenCorporates launches". OpenCorporates (Press release). 20 December 2010. Retrieved 4 January 2021.
  5. "Researching Corporations and Their Owners". Global Investigative Journalism Network. Retrieved 1 August 2023.
  6. "Finding Company Information". The Institute of Chartered Accountants in England and Wales. Retrieved 1 August 2023.
  7. "Open Data Challenge winners". Archived from the original on 16 August 2011. Retrieved 2 July 2011.
  8. Kroes, Neelie. "Getting out the Data". Europa . Retrieved 2 July 2011.
  9. "Core Vocabularies Working Group Members". Europa . 15 February 2012. Retrieved 6 April 2012.
  10. "FSB Legal Entity Identifier (LEI) Initiative Industry Advisory Panel – Membership" (PDF). Financial Stability Board. 2012. Retrieved 6 April 2012.
  11. "ODI Awards 2015 Finalists". Open Data Award 2015 (Press release). Retrieved 13 August 2015.
  12. "Celebrating Generation Open – ODI awards network thinkers who are changing the world". open data institute (Press release). Retrieved 13 August 2015.
  13. Simperl, Elena; Corcho, Oscar; Grobelnik, Marko; Roman, Dumitru; Soylu, Ahmet; Ruíz, María Jesús Fernández; Gatti, Stefano; Taggart, Chris; Klima, Urška Skok (2019), "Towards a Knowledge Graph Based Platform for Public Procurement", Metadata and Semantic Research, Cham: Springer International Publishing, pp. 317–323, doi:10.1007/978-3-030-14401-2_29, hdl: 11250/2628715 , ISBN   978-3-030-14400-5, S2CID   67866805
  14. Mezaour, Amar-Djalil (2005), "Filtering Web Documents for a Thematic Warehouse Case Study: eDot a Food Risk Data Warehouse (extended)", Intelligent Information Processing and Web Mining, Advances in Soft Computing, Berlin/Heidelberg: Springer-Verlag, pp. 269–278, doi:10.1007/3-540-32392-9_28, ISBN   3-540-25056-5
  15. Mane, U. V.; Gurav, P. N.; Deshmukh, A. M.; Govindwar, S. P. (2008). "Degradation of textile dye reactive navy – blue Rx (Reactive blue–59) by an isolated Actinomycete Streptomyces krainskii SUK – 5". Malaysian Journal of Microbiology. 4 (2). doi: 10.21161/10.21161/mjm.106717 . ISSN   2231-7538.
  16. Roman, Dumitru; Alexiev, Vladimir; Paniagua, Javier; Elvesæter, Brian; von Zernichow, Bjørn Marius; Soylu, Ahmet; Simeonov, Boyan; Taggart, Chris (25 November 2021). "The euBusinessGraph ontology: A lightweight ontology for harmonizing basic company information". Semantic Web. 13 (1): 41–68. doi: 10.3233/sw-210424 . hdl: 11250/2980609 . ISSN   2210-4968.
  17. Berthelé, Emmanuel (19 January 2018), "Using Big Data in Insurance", Big Data for Insurance Companies, Hoboken, NJ, US: John Wiley & Sons, Inc., pp. 131–161, doi:10.1002/9781119489368.ch5, ISBN   9781119489368 , retrieved 23 February 2022
  18. Tiwari, Milind; Gepp, Adrian; Kumar, Kuldeep (20 October 2021). "Global money laundering appeal index: application of principal component analysis". Journal of Money Laundering Control. 26: 205–211. doi:10.1108/jmlc-10-2021-0108. ISSN   1368-5201. S2CID   244618995.