Juyang (John) Weng | |
---|---|
Nationality | Chinese-American |
Occupation(s) | Computer engineer, neuroscientist, author, and academic |
Academic background | |
Education | BSc., Computer Science MSc., Computer Science PhD., Computer Science |
Alma mater | Fudan University University of Illinois at Urbana-Champaign |
Thesis | (1989) |
Doctoral advisor | Thomas S. Huang Narendra Ahuja |
Academic work | |
Institutions | Brain-Mind Institute GENISAMA Michigan State University |
Juyang (John) Weng is a Chinese-American computer engineer,neuroscientist,author,and academic. He is a former professor at the Department of Computer Science and Engineering at Michigan State University and the President of Brain-Mind Institute and GENISAMA. [1]
Weng has conducted research on grounded machine learning by intersecting computer science and engineering,with brain and cognitive science. In collaborative research efforts with his coworkers,he has explored mental architectures and computational models for autonomous development across various domains such as vision,audition,touch,behaviors,and motivational systems,both in biological and engineered systems. He has authored two books,Natural and Artificial Intelligence:Introduction to Computational Brain-Mind and Motion and Structure from Image Sequences,and is the editor of the book series 'New Frontiers in Robotics.' In addition,he has published over 300 articles.
Weng is a Life Fellow of the Institute of Electrical and Electronics Engineers (IEEE) and the Founder and President of the Brain-Mind Institute,and the startup GENISAMA. He is also the Founder and Editor-in-chief of the International Journal of Humanoid Robotics and the Brain-Mind Magazine and an Associate Editor of the IEEE Transactions on Autonomous Mental Development (now Cognitive and Developmental Systems). [2] Additionally,he served as a Guest editor for five special issues,including What AI and Neuroscience Can Learn from Each Other:Open Problems in Models and Theories,Cognitive Computation, [3] The Special Issue on Brain Imaging-informed Multimodal Analysis,IEEE Transactions on Autonomous Mental Development, [4] and The Special Issue on Autonomous Mental Development,International Journal of Humanoid Robotics. [5]
Weng obtained his BS degree from Fudan University in 1982,followed by earning his M.Sc. and Ph.D. degrees in computer science from the University of Illinois at Urbana-Champaign in 1985 and 1989,respectively. [6]
Following his Ph.D.,in 1990,Weng began his academic career as a visiting assistant research professor at the Beckman Institute of the University of Illinois,Urbana. From 1992 to 1998,he served as an assistant professor at Michigan State University,becoming associate professor in 1998 and professor in 2003. Retired now [7]
Weng's research revolves around grounded machine learning,spanning vision,audition,natural language understanding,planning,and real-time hardware implementations. He is also involved in technology transfer through his startup,GENISAMA,which focuses on grounded,emergent,natural,incremental,skull-closed,attentive,motivated,and abstract systems. His theoretical contributions include mathematically proving that Developmental Networks (DNs) he developed can learn any universal turing machines and establishing a theory on Autonomous Programming For General Purposes (APFGP),supporting Conscious Machine Learning. [8] [9]
Weng has worked on developmental networks from Cresceptron to DN3 to achieve the first-ever conscious learning algorithm which is free from "deep learning" misconduct. [10] His research has been featured on 'Discovery Channel , Enel ,and BBC . [11]
From 1983 to 1989,Weng's research work during his master's degree and Ph.D. degree focused on the analysis of the motion of objects and estimating 3D structures from motion. [12] He realized that such model-based approaches can provide with piecemeal insights but are too restrictive for understanding how animal brains learn vision and other brain skills. Soon after his Ph.D. degree work,he started Cresceptron. [13]
Cresceptron represented a direction that Weng later termed Autonomous Mental Development (AMD). In 1992,he and his collaborators pioneered the development of a framework titled Cresceptron for segmenting and recognizing real-world 3D objects from their images through automated learning. [13] The framework was tested for visual recognition,specifically recognizing 3-D objects from 2-D images and segmenting them from cluttered backgrounds without the need for handcrafted 3D models. It employed techniques such as stochastic distortion modeling,view-based interpolation,and a combination of individual and class-based learning approaches. Cresceptron achieved seven significant accomplishments,including the development of techniques like learning large-scale 3D objects with a deep convolutional neural network (CNN) and feature-independent learning for extensive datasets,among others. It was also established that Cresceptron significantly differs from later "deep learning" networks due to its approach of developing a sole network using Hebbian learning (i.e.,unsupervised in all hidden layers). [14] [15]
Weng introduced another framework named SHOSLIF which provided a unified theory and methodology for comprehensive sensor-actuator learning. [16] It addressed single sensory problems as well as critical issues that Cresceptron faces,such as the automated selection of the most valuable features,the automatic organization of sensory and control information through a coarse-to-fine space partition tree,resulting in a remarkably low,logarithmic time complexity for content-based retrieval from extensive visual knowledge bases. [17] It also deals with handling invariance through learning,enabling online incremental learning,and facilitating autonomous learning,among other objectives. [18] [19]
From 1998 to 2010,Weng developed SAIL [20] and Dav robots [21] using sensory mapping models including self-aware self-effecting (SASE),staggered hierarchical mapping (SHM),and incremental hierarchical discriminant regression (IHDR) methods. It has been applied to the recognition of occluded objects, [22] speech recognition, [23] vision-guided navigation, [24] and range-based collision avoidance. [25]
Since 2005,Weng and his team have been working on the development of brain-like and cortex-like Developmental Networks (DNs) and their embodiments Where-What Networks (WWNs) [26] using brain-like architecture,including modeling pathways,laminar 6-layer cortex,and brain areas. [27] [28] In addition,they have analyzed how the brain deals with modulation,time,and space and have created three versions (DN1 through DN3) by 2023. A significant enhancement introduced in the transition from DN-2 to DN-3 involves initiating the brain-size network from a single-cell zygote. This means a fully autonomous process for brain patterning from a single cell. The key mechanisms of patterning include the Lobe Component Analysis (LCA) [29] and Synaptic Maintenance, [30] which automatically maintain the global smoothness of brain representation and local refinements of area representations. This approach enabled the developmental algorithm to progressively develop sensors,a complex brain,and motor functions in a sequential and self-organizing manner,ensuring that the wiring and pattern formation processes occur automatically from the initial conception stages throughout the entire life of the system. [31]
These Developmental Networks (DNs) and Where-What Networks (WWNs 1–9) have been developed for versatile visual learning in complex environments. [32] DNs can recognize objects and autonomously determine where and what to focus on using self-generated task context. Furthermore,these WWNs and DNs have been applied to general-purpose vision, [33] temporal visual event recognition, [34] vision-guided navigation, [35] learning audition while learning to speak, [36] and language acquisition as brain's responses to text temporal events. [37]
Weng is the first to formally raise that robotic consciousness is necessary for AI,consciousness can and should be learned (i.e.,developed),and proposed a fully implementable algorithm to do so. He proposed DN3 [31] as the engine to conduct conscious learning [38] where a robot is able to become increasingly conscious,like an infant and then a child,through its 'living' experience in the physical world which typically include human parents and teachers. However,there is no central controller within DN3's skull,emphasizing that consciousness should not be statically handcrafted and must encompass elements beyond a programmer's design. [31]
Since 2016,Weng has alleged instances of plagiarism and post-selection misconduct on a worldwide scale,but the implicated institutions have not yet acknowledged his allegations.
Weng alleged that many deep learning networks that use images of 3D objects copied their key idea from Cresceptron [13] but almost all later deep learning publications did not cite Cresceptron. He highlighted that the Cresceptron (for 3D) is very different from the Neocognitron [39] (for 2D) because the Cresceptron is a fundamental departure from Neocognitron. Cresceptron enables a neural network to grow incrementally from a zero-neuron hierarchy and learn 3D objects from their 2D images in cluttered scenes. This is different from the aspect graphs of the 1990s and all other methods that had an inside-skull human teacher as a central controller. [40] This alleged plagiarism includes HMAX at MIT [41] and the ACM Turing Award 2018. [42] Without internal weight supervision like human manual selections [39] [41] and error-backprop, [42] feature learning and sharing in hidden areas of Cresceptron are based on (unsupervised) Hebbian mechanisms. [43]
Weng raised the issue of Post-Selection in AI and argued that it constitutes misconduct. He addressed that many AI methods require two steps in their training stage. The first step consists of training multiple systems by randomly fitting a fit data set. The second step consists of Post-Model Selection (Post-Selection). The Post-Selection chooses a few luckiest trained systems or relies on human manual parameter-tuning based on the systems’errors on a validation data set. He alleged that Post-Selection in AI contains two types of misconduct:(1) cheating in the absence of a test,because the Post-Selection step belongs to the training stage;(2) hiding bad-looking data,because less lucky systems were not reported. [10]
Weng further alleged that more categories of AI methods suffered from their Post-Selection steps,such as Neocognitron,HMAX,Deep Learning,Long Short-Term Memories,Extreme Learning Machines,Evolving Networks,Reservoir Computing,Transformers,Large Language Models,ChatGPT,and Bard,as long as they contain the Post-Selection step,which is either automatic or requires human manual tuning. He mathematically reasoned that the luckiest system on a validation set gives only an expected performance on a future test set that is only near the average performance of all trained systems on the validation set. [10]
Weng has sued institutions to address the issue of alleged misconduct outside of academia including Alphabet,in the United States District Court for the Western District of Michigan (Civil Action No. 1:22-cv-998) [44] and the US Court of Appeal 6th Circuit (Civil Action No,23–1567). [45]
Neural networks are a branch of machine learning models inspired by the neuronal organization found in the biological neural networks in animal brains.
Jürgen Schmidhuber is a German computer scientist noted for his work in the field of artificial intelligence,specifically artificial neural networks. He is a scientific director of the Dalle Molle Institute for Artificial Intelligence Research in Switzerland. He is also director of the Artificial Intelligence Initiative and professor of the Computer Science program in the Computer,Electrical,and Mathematical Sciences and Engineering (CEMSE) division at the King Abdullah University of Science and Technology (KAUST) in Saudi Arabia.
Neuromorphic computing is an approach to computing that is inspired by the structure and function of the human brain. A neuromorphic computer/chip is any device that uses physical artificial neurons to do computations. In recent times,the term neuromorphic has been used to describe analog,digital,mixed-mode analog/digital VLSI,and software systems that implement models of neural systems. The implementation of neuromorphic computing on the hardware level can be realized by oxide-based memristors,spintronic memories,threshold switches,transistors,among others. Training software-based neuromorphic systems of spiking neural networks can be achieved using error backpropagation,e.g.,using Python based frameworks such as snnTorch,or using canonical learning rules from the biological learning literature,e.g.,using BindsNet.
Developmental robotics (DevRob),sometimes called epigenetic robotics,is a scientific field which aims at studying the developmental mechanisms,architectures and constraints that allow lifelong and open-ended learning of new skills and new knowledge in embodied machines. As in human children,learning is expected to be cumulative and of progressively increasing complexity,and to result from self-exploration of the world in combination with social interaction. The typical methodological approach consists in starting from theories of human and animal development elaborated in fields such as developmental psychology,neuroscience,developmental and evolutionary biology,and linguistics,then to formalize and implement them in robots,sometimes exploring extensions or variants of them. The experimentation of those models in robots allows researchers to confront them with reality,and as a consequence,developmental robotics also provides feedback and novel hypotheses on theories of human and animal development.
The expression computational intelligence (CI) usually refers to the ability of a computer to learn a specific task from data or experimental observation. Even though it is commonly considered a synonym of soft computing,there is still no commonly accepted definition of computational intelligence.
For holographic data storage,holographic associative memory (HAM) is an information storage and retrieval system based on the principles of holography. Holograms are made by using two beams of light,called a "reference beam" and an "object beam". They produce a pattern on the film that contains them both. Afterwards,by reproducing the reference beam,the hologram recreates a visual image of the original object. In theory,one could use the object beam to do the same thing:reproduce the original reference beam. In HAM,the pieces of information act like the two beams. Each can be used to retrieve the other from the pattern. It can be thought of as an artificial neural network which mimics the way the brain uses information. The information is presented in abstract form by a complex vector which may be expressed directly by a waveform possessing frequency and magnitude. This waveform is analogous to electrochemical impulses believed to transmit information between biological neuron cells.
The neocognitron is a hierarchical,multilayered artificial neural network proposed by Kunihiko Fukushima in 1979. It has been used for Japanese handwritten character recognition and other pattern recognition tasks,and served as the inspiration for convolutional neural networks.
Computational neurogenetic modeling (CNGM) is concerned with the study and development of dynamic neuronal models for modeling brain functions with respect to genes and dynamic interactions between genes. These include neural network models and their integration with gene network models. This area brings together knowledge from various scientific disciplines,such as computer and information science,neuroscience and cognitive science,genetics and molecular biology,as well as engineering.
Spiking neural networks (SNNs) are artificial neural networks (ANN) that more closely mimic natural neural networks. In addition to neuronal and synaptic state,SNNs incorporate the concept of time into their operating model. The idea is that neurons in the SNN do not transmit information at each propagation cycle,but rather transmit information only when a membrane potential—an intrinsic quality of the neuron related to its membrane electrical charge—reaches a specific value,called the threshold. When the membrane potential reaches the threshold,the neuron fires,and generates a signal that travels to other neurons which,in turn,increase or decrease their potentials in response to this signal. A neuron model that fires at the moment of threshold crossing is also called a spiking neuron model.
Neurorobotics is the combined study of neuroscience,robotics,and artificial intelligence. It is the science and technology of embodied autonomous neural systems. Neural systems include brain-inspired algorithms,computational models of biological neural networks and actual biological systems. Such neural systems can be embodied in machines with mechanic or any other forms of physical actuation. This includes robots,prosthetic or wearable systems but also,at smaller scale,micro-machines and,at the larger scales,furniture and infrastructures.
An area of computer vision is active vision,sometimes also called active computer vision. An active vision system is one that can manipulate the viewpoint of the camera(s) in order to investigate the environment and get better information from it.
Kunihiko Fukushima is a Japanese computer scientist,most noted for his work on artificial neural networks and deep learning. He is currently working part-time as a senior research scientist at the Fuzzy Logic Systems Institute in Fukuoka,Japan.
Morphogenetic robotics generally refers to the methodologies that address challenges in robotics inspired by biological morphogenesis.
Fusion adaptive resonance theory (fusion ART) is a generalization of self-organizing neural networks known as the original Adaptive Resonance Theory models for learning recognition categories across multiple pattern channels. There is a separate stream of work on fusion ARTMAP,that extends fuzzy ARTMAP consisting of two fuzzy ART modules connected by an inter-ART map field to an extended architecture consisting of multiple ART modules.
AlexNet is the name of a convolutional neural network (CNN) architecture,designed by Alex Krizhevsky in collaboration with Ilya Sutskever and Geoffrey Hinton,who was Krizhevsky's Ph.D. advisor at the University of Toronto.
Artificial neural networks (ANNs) are models created using machine learning to perform a number of tasks. Their creation was inspired by neural circuitry. While some of the computational implementations ANNs relate to earlier discoveries in mathematics,the first implementation of ANNs was by psychologist Frank Rosenblatt,who developed the perceptron. Little research was conducted on ANNs in the 1970s and 1980s,with the AAAI calling that period an "AI winter".
Silvia Ferrari is an Italian-American aerospace engineer. She is John Brancaccio Professor at the Sibley School of Mechanical and Aerospace Engineering at Cornell University and also the director of the Laboratory for Intelligent Systems and Control (LISC) at the same university.
Aude G. Billard is a Swiss physicist in the fields of machine learning and human-robot interactions. As a full professor at the School of Engineering at Swiss Federal Institute of Technology in Lausanne (EPFL),Billard’s research focuses on applying machine learning to support robot learning through human guidance. Billard’s work on human-robot interactions has been recognized numerous times by the Institute of Electrical and Electronics Engineers (IEEE) and she currently holds a leadership position on the executive committee of the IEEE Robotics and Automation Society (RAS) as the vice president of publication activities.
Javier Andreu-Perez is a British computer scientist and a Senior Lecturer and Chair in Smart Health Technologies at the University of Essex. He is also associate editor-in-chief of Neurocomputing for the area of Deep Learning and Machine Learning. Andreu-Perez research is mainly focused on Human-Centered Artificial Intelligence (HCAI). He also chairs a interdisciplinary lab in this area,HCAI-Essex.
Small object detection is a particular case of object detection where various techniques are employed to detect small objects in digital images and videos. "Small objects" are objects having a small pixel footprint in the input image. In areas such as aerial imagery,state-of-the-art object detection techniques under performed because of small objects.