Bastiaan Quast

Last updated
Bastiaan Quast
Bastiaan-Quast-Stanford.jpg
Bastiaan Quast at Stanford
CitizenshipNetherlands
Known for Rnn (software)
Scientific career
Institutions United Nations (ITU)
Doctoral advisor Richard Baldwin, Jean-Louis Arcand
Website bastiaanquast.com

Bastiaan Quast is a Dutch Machine learning researcher. He is the author and lead maintainer of the open-source rnn and transformer deep-learning frameworks in the R programming language, and the datasets.load GUI package, as well as R packages on Global Value Chain decomposition & WIOD and on Regression Discontinuity Design. [1] Quast is a great-great-grandson of the Nobel Peace Prize laureate Tobias Asser. [2]

Contents

Early life and education

Bastiaan Quast graduated from University of Groningen with a bachelor's degree in Economics and bachelor's degree in Theoretical philosophy. He holds a master's degree in Econometrics from the University of St. Gallen He obtained his Ph.D from the Graduate Institute of International and Development Studies with advisors Richard Baldwin and Jean-Louis Arcand, his work on local languages and internet usage was discussed at the 2017 G20 meeting in Germany. [3] [4] [5]

Career

A screenshot of the datasets.load GUI Datasets.load.png
A screenshot of the datasets.load GUI

Quast is an functionary of the United Nations at the International Telecommunication Union, as Secretary of the ITU-WHO Focus Group on Artificial Intelligence for Health and AI for Good.

Bastiaan Quast created the popular machine learning framework rnn in R, which allows native implementations of recurrent neural network architectures, such as LSTM and GRU (>100,000 downloads [6] ). While working at UNCTAD, Quast developed the popular package datasets.load, which is part of the top 10% of most downloaded R packages (>100,000 [7] ). The R packages decompr and wiod have been downloaded >20,000 times. [8] [9]

Bibliography

Kummritz, Victor; Quast, Bastiaan (2017). Global value chains in developing economies. London, United Kingdom: VoxEU.

Related Research Articles

<span class="mw-page-title-main">Neural network (machine learning)</span> Computational model used in machine learning, based on connected, hierarchical functions

In machine learning, a neural network is a model inspired by the structure and function of biological neural networks in animal brains.

<span class="mw-page-title-main">Overfitting</span> Flaw in mathematical modelling

In mathematical modeling, overfitting is "the production of an analysis that corresponds too closely or exactly to a particular set of data, and may therefore fail to fit to additional data or predict future observations reliably". An overfitted model is a mathematical model that contains more parameters than can be justified by the data. In a mathematical sense, these parameters represent the degree of a polynomial. The essence of overfitting is to have unknowingly extracted some of the residual variation as if that variation represented underlying model structure.

A recurrent neural network (RNN) is one of the two broad types of artificial neural network, characterized by direction of the flow of information between its layers. In contrast to the uni-directional feedforward neural network, it is a bi-directional artificial neural network, meaning that it allows the output from some nodes to affect subsequent input to the same nodes. Their ability to use internal state (memory) to process arbitrary sequences of inputs makes them applicable to tasks such as unsegmented, connected handwriting recognition or speech recognition. The term "recurrent neural network" is used to refer to the class of networks with an infinite impulse response, whereas "convolutional neural network" refers to the class of finite impulse response. Both classes of networks exhibit temporal dynamic behavior. A finite impulse recurrent network is a directed acyclic graph that can be unrolled and replaced with a strictly feedforward neural network, while an infinite impulse recurrent network is a directed cyclic graph that cannot be unrolled.

<span class="mw-page-title-main">Orange (software)</span> Open-source data analysis software

Orange is an open-source data visualization, machine learning and data mining toolkit. It features a visual programming front-end for explorative qualitative data analysis and interactive data visualization.

In data analysis, anomaly detection is generally understood to be the identification of rare items, events or observations which deviate significantly from the majority of the data and do not conform to a well defined notion of normal behavior. Such examples may arouse suspicions of being generated by a different mechanism, or appear inconsistent with the remainder of that set of data.

<span class="mw-page-title-main">Echo state network</span> Type of reservoir computer

An echo state network (ESN) is a type of reservoir computer that uses a recurrent neural network with a sparsely connected hidden layer. The connectivity and weights of hidden neurons are fixed and randomly assigned. The weights of output neurons can be learned so that the network can produce or reproduce specific temporal patterns. The main interest of this network is that although its behavior is non-linear, the only weights that are modified during training are for the synapses that connect the hidden neurons to output neurons. Thus, the error function is quadratic with respect to the parameter vector and can be differentiated easily to a linear system.

Reservoir computing is a framework for computation derived from recurrent neural network theory that maps input signals into higher dimensional computational spaces through the dynamics of a fixed, non-linear system called a reservoir. After the input signal is fed into the reservoir, which is treated as a "black box," a simple readout mechanism is trained to read the state of the reservoir and map it to the desired output. The first key benefit of this framework is that training is performed only at the readout stage, as the reservoir dynamics are fixed. The second is that the computational power of naturally available systems, both classical and quantum mechanical, can be used to reduce the effective computational cost.

<span class="mw-page-title-main">Long short-term memory</span> Artificial recurrent neural network architecture used in deep learning

Long short-term memory (LSTM) is a type of recurrent neural network (RNN) aimed at dealing with the vanishing gradient problem present in traditional RNNs. Its relative insensitivity to gap length is its advantage over other RNNs, hidden Markov models and other sequence learning methods. It aims to provide a short-term memory for RNN that can last thousands of timesteps, thus "long short-term memory". It is applicable to classification, processing and predicting data based on time series, such as in handwriting, speech recognition, machine translation, speech activity detection, robot control, video games, and healthcare.

Sparse principal component analysis is a technique used in statistical analysis and, in particular, in the analysis of multivariate data sets. It extends the classic method of principal component analysis (PCA) for the reduction of dimensionality of data by introducing sparsity structures to the input variables.

In statistics and machine learning, ensemble methods use multiple learning algorithms to obtain better predictive performance than could be obtained from any of the constituent learning algorithms alone. Unlike a statistical ensemble in statistical mechanics, which is usually infinite, a machine learning ensemble consists of only a concrete finite set of alternative models, but typically allows for much more flexible structure to exist among those alternatives.

There are many types of artificial neural networks (ANN).

<span class="mw-page-title-main">Deep learning</span> Branch of machine learning

Deep learning is the subset of machine learning methods based on neural networks with representation learning. The adjective "deep" refers to the use of multiple layers in the network. Methods used can be either supervised, semi-supervised or unsupervised.

A convolutional neural network (CNN) is a regularized type of feed-forward neural network that learns features by itself via filter optimization. Vanishing gradients and exploding gradients, seen during backpropagation in earlier neural networks, are prevented by using regularized weights over fewer connections. For example, for each neuron in the fully-connected layer, 10,000 weights would be required for processing an image sized 100 × 100 pixels. However, applying cascaded convolution kernels, only 25 neurons are required to process 5x5-sized tiles. Higher-layer features are extracted from wider context windows, compared to lower-layer features.

<span class="mw-page-title-main">Apache Spark</span> Open-source data analytics cluster computing framework

Apache Spark is an open-source unified analytics engine for large-scale data processing. Spark provides an interface for programming clusters with implicit data parallelism and fault tolerance. Originally developed at the University of California, Berkeley's AMPLab, the Spark codebase was later donated to the Apache Software Foundation, which has maintained it since.

A recursive neural network is a kind of deep neural network created by applying the same set of weights recursively over a structured input, to produce a structured prediction over variable-size input structures, or a scalar prediction on it, by traversing a given structure in topological order. Recursive neural networks, sometimes abbreviated as RvNNs, have been successful, for instance, in learning sequence and tree structures in natural language processing, mainly phrase and sentence continuous representations based on word embedding. RvNNs have first been introduced to learn distributed representations of structure, such as logical terms. Models and general frameworks have been developed in further works since the 1990s.

oneAPI Data Analytics Library, is a library of optimized algorithmic building blocks for data analysis stages most commonly associated with solving Big Data problems.

Neural machine translation (NMT) is an approach to machine translation that uses an artificial neural network to predict the likelihood of a sequence of words, typically modeling entire sentences in a single integrated model.

<span class="mw-page-title-main">ML.NET</span> Machine learning library

ML.NET is a free software machine learning library for the C# and F# programming languages. It also supports Python models when used together with NimbusML. The preview release of ML.NET included transforms for feature engineering like n-gram creation, and learners to handle binary classification, multi-class classification, and regression tasks. Additional ML tasks like anomaly detection and recommendation systems have since been added, and other approaches like deep learning will be included in future versions.

rnn (software) Machine Learning framework written in the R language

rnn is an open-source machine learning framework that implements recurrent neural network architectures, such as LSTM and GRU, natively in the R programming language, that has been downloaded over 100,000 times.

Emotion recognition in conversation (ERC) is a sub-field of emotion recognition, that focuses on mining human emotions from conversations or dialogues having two or more interlocutors. The datasets in this field are usually derived from social platforms that allow free and plenty of samples, often containing multimodal data. Self- and inter-personal influences play critical role in identifying some basic emotions, such as, fear, anger, joy, surprise, etc. The more fine grained the emotion labels are the harder it is to detect the correct emotion. ERC poses a number of challenges, such as, conversational-context modeling, speaker-state modeling, presence of sarcasm in conversation, emotion shift across consecutive utterances of the same interlocutor.

References

  1. Quast, Bastiaan (2019-06-12), Regression Discontinuity in R. , retrieved 2019-09-19
  2. "Steven Guy Quast". www.dutchjewry.org. Retrieved 2019-10-06.
  3. Taylor, Emily (February 16, 2017). "Bridging the Digital Divide: infrastructure skills and women's empowerment" (PDF). G20 Insights.
  4. "Les motifs de la lenteur d'Internet en Afrique | Info Afrique". Info Afrique (in French). 2016-08-30. Retrieved 2018-09-24.
  5. Manager, African (2016-08-30). "Internet Society se penche sur l'Afrique". African Manager (in French). Retrieved 2018-09-24.
  6. Quast, Bastiaan (2019-08-30), Recurrent Neural Networks in R. , retrieved 2019-09-19
  7. Quast, Bastiaan (2019-03-26), Tools for view datasets in RVisual interface for loading datasets in RStudio from all installed (unloaded) packages.: bquast/datasets.load , retrieved 2019-09-19
  8. Quast, Bastiaan (2019-05-16), GVC decomposition in R. , retrieved 2019-09-19
  9. Quast, Bastiaan (2019-09-12), Data sets from the World Input Output database, for the years 1995-2011: bquast/wiod , retrieved 2019-09-19