C3 superclass linearization is an algorithm used primarily to obtain the order in which methods should be inherited in the presence of multiple inheritance. In other words, the output of C3 superclass linearization is a deterministic Method Resolution Order (MRO).
In object-oriented systems with multiple inheritance, some mechanism must be used for resolving conflicts when inheriting different definitions of the same property from multiple superclasses. [1]
C3 superclass linearization is called C3 because it "is consistent with three properties": [1]
It was first published at the 1996 OOPSLA conference, in a paper entitled "A Monotonic Superclass Linearization for Dylan". [1] It was adapted to the Open Dylan implementation in January 2012 [2] following an enhancement proposal. [3] It has been chosen as the default algorithm for method resolution in Python 2.3 (and newer), [4] [5] Raku, [6] Parrot, [7] Solidity, and PGF/TikZ's Object-Oriented Programming module. [8] It is also available as an alternative, non-default MRO in the core of Perl 5 starting with version 5.10.0. [9] An extension implementation for earlier versions of Perl 5 named Class::C3
exists on CPAN. [10]
Python's Guido van Rossum summarizes C3 superclass linearization thus: [11]
Basically, the idea behind C3 is that if you write down all of the ordering rules imposed by inheritance relationships in a complex class hierarchy, the algorithm will determine a monotonic ordering of the classes that satisfies all of them. If such an ordering can not be determined, the algorithm will fail.
![]() | This article may be confusing or unclear to readers.(April 2018) |
The C3 superclass linearization of a class is the sum of the class plus a unique merge of the linearizations of its parents and a list of the parents itself. The list of parents as the last argument to the merge process preserves the local precedence order of direct parent classes.
The merge of parents' linearizations and parents list is done by selecting the first head of the lists which does not appear in the tail (all elements of a list except the first) of any of the lists. Note, that a good head may appear as the first element in multiple lists at the same time, but it is forbidden to appear anywhere else. The selected element is removed from all the lists where it appears as a head and appended to the output list. The process of selecting and removing a good head to extend the output list is repeated until all remaining lists are exhausted. If at some point no good head can be selected, because the heads of all remaining lists appear in any one tail of the lists, then the merge is impossible to compute due to inconsistent orderings of dependencies in the inheritance hierarchy and no linearization of the original class exists.
A naive divide-and-conquer approach to computing the linearization of a class may invoke the algorithm recursively to find the linearizations of parent classes for the merge-subroutine. However, this will result in an infinitely looping recursion in the presence of a cyclic class hierarchy. To detect such a cycle and to break the infinite recursion (and to reuse the results of previous computations as an optimization), the recursive invocation should be shielded against re-entrance of a previous argument by means of a cache or memoization.
This algorithm is similar to finding a topological ordering.
Given
<nowiki/>classOclassAextendsOclassBextendsOclassCextendsOclassDextendsOclassEextendsOclassK1extendsC,A,BclassK3extendsA,DclassK2extendsB,D,EclassZextendsK1,K3,K2
the linearization of Z is computed as
<nowiki/>L(O):=[O]// the linearization of O is trivially the singleton list [O], because O has no parentsL(A):=[A]+merge(L(O),[O])// the linearization of A is A plus the merge of its parents' linearizations with the list of parents...=[A]+merge([O],[O])=[A,O]// ...which simply prepends A to its single parent's linearizationL(B):=[B,O]// linearizations of B, C, D and E are computed similar to that of AL(C):=[C,O]L(D):=[D,O]L(E):=[E,O]L(K1):=[K1]+merge(L(C),L(B),L(A),[C,A,B])// first, find the linearization of K1's parents, L(C), L(B), and L(A), and merge them with the parent list [C, A, B]=[K1]+merge([C,O],[B,O],[A,O],[C,A,B])// class C is a good candidate for the first merge step, because it only appears as the head of the first and last lists=[K1,C]+merge([O],[B,O],[A,O],[A,B])// class O is not a good candidate for the next merge step, because it also appears in the tails of list 2 and 3, class B is also not good; but class A is a good candidate=[K1,C,A]+merge([O],[B,O],[O],[B])// class B is a good candidate; class O still appears in the tail of list 2=[K1,C,A,B]+merge([O],[O],[O])// finally, class O is a valid candidate, which also exhausts all remaining lists=[K1,C,A,B,O]L(K3):=[K3]+merge(L(A),L(D),[A,D])=[K3]+merge([A,O],[D,O],[A,D])// select A=[K3,A]+merge([O],[D,O],[D])// fail O, select D=[K3,A,D]+merge([O],[O])// select O=[K3,A,D,O]L(K2):=[K2]+merge(L(B),L(D),L(E),[B,D,E])=[K2]+merge([B,O],[D,O],[E,O],[B,D,E])// select B=[K2,B]+merge([O],[D,O],[E,O],[D,E])// fail O, select D=[K2,B,D]+merge([O],[O],[E,O],[E])// fail O, select E=[K2,B,D,E]+merge([O],[O],[O])// select O=[K2,B,D,E,O]L(Z):=[Z]+merge(L(K1),L(K3),L(K2),[K1,K3,K2])=[Z]+merge([K1,C,A,B,O],[K3,A,D,O],[K2,B,D,E,O],[K1,K3,K2])// select K1=[Z,K1]+merge([C,A,B,O],[K3,A,D,O],[K2,B,D,E,O],[K3,K2])// select C=[Z,K1,C]+merge([A,B,O],[K3,A,D,O],[K2,B,D,E,O],[K3,K2])// fail A, select K3=[Z,K1,C,K3]+merge([A,B,O],[A,D,O],[K2,B,D,E,O],[K2])// select A=[Z,K1,C,K3,A]+merge([B,O],[D,O],[K2,B,D,E,O],[K2])// fail B, fail D, select K2=[Z,K1,C,K3,A,K2]+merge([B,O],[D,O],[B,D,E,O])// select B=[Z,K1,C,K3,A,K2,B]+merge([O],[D,O],[D,E,O])// fail O, select D=[Z,K1,C,K3,A,K2,B,D]+merge([O],[O],[E,O])// fail O, select E=[Z,K1,C,K3,A,K2,B,D,E]+merge([O],[O],[O])// select O=[Z,K1,C,K3,A,K2,B,D,E,O]// done.
First, a metaclass to enable a short representation of the objects by name instead of the default class REPR value:
classType(type):def__repr__(cls):# Will make it so that repr(O) is "O" instead of "<__main__.O>",# and the same for any subclasses of O.returncls.__name__classO(object,metaclass=Type):pass
Next, we define our base classes:
classA(O):passclassB(O):passclassC(O):passclassD(O):passclassE(O):pass
Then we construct the inheritance tree:
classK1(C,A,B):passclassK3(A,D):passclassK2(B,D,E):passclassZ(K1,K3,K2):pass
And now:
>>> Z.mro()[Z, K1, C, K3, A, K2, B, D, E, O, <class 'object'>]
Raku uses C3 linearization for classes by default:
classA {} classB {} classC {} classD {} classE {} classK1isCisAisB {} classK3isAisD {} classK2isBisDisE {} classZisK1isK3isK2 {} sayZ.^mro; # OUTPUT: ((Z) (K1) (C) (K3) (A) (K2) (B) (D) (E) (Any) (Mu))
(the Any
and Mu
are the types all Raku objects inherit from, so Any
stands in place of O)
In computer science, a heap is a tree-based data structure that satisfies the heap property: In a max heap, for any given node C, if P is the parent node of C, then the key of P is greater than or equal to the key of C. In a min heap, the key of P is less than or equal to the key of C. The node at the "top" of the heap is called the root node.
Multiple inheritance is a feature of some object-oriented computer programming languages in which an object or class can inherit features from more than one parent object or parent class. It is distinct from single inheritance, where an object or class may only inherit from one particular object or class.
In computer science, merge sort is an efficient, general-purpose, and comparison-based sorting algorithm. Most implementations produce a stable sort, which means that the relative order of equal elements is the same in the input and output. Merge sort is a divide-and-conquer algorithm that was invented by John von Neumann in 1945. A detailed description and analysis of bottom-up merge sort appeared in a report by Goldstine and von Neumann as early as 1948.
A regular expression, sometimes referred to as rational expression, is a sequence of characters that specifies a match pattern in text. Usually such patterns are used by string-searching algorithms for "find" or "find and replace" operations on strings, or for input validation. Regular expression techniques are developed in theoretical computer science and formal language theory.
In computer science, a sorting algorithm is an algorithm that puts elements of a list into an order. The most frequently used orders are numerical order and lexicographical order, and either ascending or descending. Efficient sorting is important for optimizing the efficiency of other algorithms that require input data to be in sorted lists. Sorting is also often useful for canonicalizing data and for producing human-readable output.
In cryptography, Triple DES, officially the Triple Data Encryption Algorithm, is a symmetric-key block cipher, which applies the DES cipher algorithm three times to each data block. The 56-bit key of the Data Encryption Standard (DES) is no longer considered adequate in the face of modern cryptanalytic techniques and supercomputing power; Triple DES increases the effective security to 112 bits. A CVE released in 2016, CVE-2016-2183, disclosed a major security vulnerability in the DES and 3DES encryption algorithms. This CVE, combined with the inadequate key size of 3DES, led to NIST deprecating 3DES in 2019 and disallowing all uses by the end of 2023. It has been replaced with the more secure, more robust AES.
A multiplication algorithm is an algorithm to multiply two numbers. Depending on the size of the numbers, different algorithms are more efficient than others. Numerous algorithms are known and there has been much research into the topic.
In time series analysis, dynamic time warping (DTW) is an algorithm for measuring similarity between two temporal sequences, which may vary in speed. For instance, similarities in walking could be detected using DTW, even if one person was walking faster than the other, or if there were accelerations and decelerations during the course of an observation. DTW has been applied to temporal sequences of video, audio, and graphics data — indeed, any data that can be turned into a one-dimensional sequence can be analyzed with DTW. A well-known application has been automatic speech recognition, to cope with different speaking speeds. Other applications include speaker recognition and online signature recognition. It can also be used in partial shape matching applications.
MRO or Mro may refer to:
In physics, a ferromagnetic material is said to have magnetocrystalline anisotropy if it takes more energy to magnetize it in certain directions than in others. These directions are usually related to the principal axes of its crystal lattice. It is a special case of magnetic anisotropy. In other words, the excess energy required to magnetize a specimen in a particular direction over that required to magnetize it along the easy direction is called crystalline anisotropy energy.
In cryptography, the Tiny Encryption Algorithm (TEA) is a block cipher notable for its simplicity of description and implementation, typically a few lines of code. It was designed by David Wheeler and Roger Needham of the Cambridge Computer Laboratory; it was first presented at the Fast Software Encryption workshop in Leuven in 1994, and first published in the proceedings of that workshop.
One-key MAC (OMAC) is a family of message authentication codes constructed from a block cipher much like the CBC-MAC algorithm. It may be used to provide assurance of the authenticity and, hence, the integrity of data. Two versions are defined:
In mathematics, a complex analytic K3 surface is a compact connected complex manifold of dimension 2 with а trivial canonical bundle and irregularity zero. An (algebraic) K3 surface over any field means a smooth proper geometrically connected algebraic surface that satisfies the same conditions. In the Enriques–Kodaira classification of surfaces, K3 surfaces form one of the four classes of minimal surfaces of Kodaira dimension zero. A simple example is the Fermat quartic surface
qrpff is a Perl script created by Keith Winstein and Marc Horowitz of the MIT SIPB. It performs DeCSS in six or seven lines. The name itself is an encoding of "decss" in rot-13. The algorithm was rewritten 77 times to condense it down to six lines.
Extendible hashing is a type of hash system which treats a hash as a bit string and uses a trie for bucket lookup. Because of the hierarchical nature of the system, re-hashing is an incremental operation. This means that time-sensitive applications are less affected by table growth than by standard full-table rehashes.
In geometry, a polygon P in the plane is called monotone with respect to a straight line L, if every line orthogonal to L intersects the boundary of P at most twice.
In computer science, a skew binomial heap is a data structure for priority queue operations. It is a variant of the binomial heap that supports constant-time insertion operations in the worst case, rather than amortized time.
In computer graphics, the centripetal Catmull–Rom spline is a variant form of the Catmull–Rom spline, originally formulated by Edwin Catmull and Raphael Rom, which can be evaluated using a recursive algorithm proposed by Barry and Goldman. It is a type of interpolating spline defined by four control points , with the curve drawn only from to .
Phragmén's voting rules are rules for multiwinner voting. They allow voters to vote for individual candidates rather than parties, but still guarantee proportional representation. They were published by Lars Edvard Phragmén in French and Swedish between 1893 and 1899, and translated to English by Svante Janson in 2016.
{{cite conference}}
: CS1 maint: multiple names: authors list (link)