Robotics

Last updated
The Shadow robot hand system Shadow Hand Bulb large.jpg
The Shadow robot hand system

Robotics is an interdisciplinary branch of computer science and engineering. [1] Robotics involves the design, construction, operation, and use of robots. The goal of robotics is to design machines that can help and assist humans. Robotics integrates fields of mechanical engineering, electrical engineering, information engineering, mechatronics, electronics, bioengineering, computer engineering, control engineering, software engineering, mathematics, etc.

Contents

Robotics develops machines that can substitute for humans and replicate human actions. Robots can be used in many situations for many purposes, but today many are used in dangerous environments (including inspection of radioactive materials, bomb detection and deactivation), manufacturing processes, or where humans cannot survive (e.g., in space, underwater, in high heat, and clean up and containment of hazardous materials and radiation). Robots can take any form, but some are made to resemble humans in appearance. This is claimed to help in the acceptance of robots in certain replicative behaviors which are usually performed by people. Such robots attempt to replicate walking, lifting, speech, cognition, or any other human activity. Many of today's robots are inspired by nature, contributing to the field of bio-inspired robotics.

Certain robots require user input to operate, while other robots function autonomously. The concept of creating robots that can operate autonomously dates back to classical times, but research into the functionality and potential uses of robots did not grow substantially until the 20th century. Throughout history, it has been frequently assumed by various scholars, inventors, engineers, and technicians that robots will one day be able to mimic human behavior and manage tasks in a human-like fashion. Today, robotics is a rapidly growing field, as technological advances continue; researching, designing, and building new robots serve various practical purposes, whether domestically, commercially, or militarily. Many robots are built to do jobs that are hazardous to people, such as defusing bombs, finding survivors in unstable ruins, and exploring mines and shipwrecks. Robotics is also used in STEM (science, technology, engineering, and mathematics) as a teaching aid. [2]

Etymology

The word robotics was derived from the word robot, which was introduced to the public by Czech writer Karel Čapek in his play R.U.R. (Rossum's Universal Robots) , which was published in 1920. [3] The word robot comes from the Slavic word robota, which means work/job. The play begins in a factory that makes artificial people called robots, creatures who can be mistaken for humans – very similar to the modern ideas of androids. Karel Čapek himself did not coin the word. He wrote a short letter in reference to an etymology in the Oxford English Dictionary in which he named his brother Josef Čapek as its actual originator. [3]

According to the Oxford English Dictionary, the word robotics was first used in print by Isaac Asimov, in his science fiction short story "Liar!" , published in May 1941 in Astounding Science Fiction . Asimov was unaware that he was coining the term; since the science and technology of electrical devices is electronics, he assumed robotics already referred to the science and technology of robots. In some of Asimov's other works, he states that the first use of the word robotics was in his short story Runaround (Astounding Science Fiction, March 1942), [4] [5] where he introduced his concept of The Three Laws of Robotics. However, the original publication of "Liar!" predates that of "Runaround" by ten months, so the former is generally cited as the word's origin.

History

In 1948, Norbert Wiener formulated the principles of cybernetics, the basis of practical robotics.

Fully autonomous robots only appeared in the second half of the 20th century. The first digitally operated and programmable robot, the Unimate, was installed in 1961 to lift hot pieces of metal from a die casting machine and stack them. Commercial and industrial robots are widespread today and used to perform jobs more cheaply, more accurately, and more reliably than humans. They are also employed in some jobs which are too dirty, dangerous, or dull to be suitable for humans. Robots are widely used in manufacturing, assembly, packing and packaging, mining, transport, earth and space exploration, surgery, [6] weaponry, laboratory research, safety, and the mass production of consumer and industrial goods. [7]

DateSignificanceRobot nameInventor
Third century B.C. and earlierOne of the earliest descriptions of automata appears in the Lie Zi text, on a much earlier encounter between King Mu of Zhou (1023–957 BC) and a mechanical engineer known as Yan Shi, an 'artificer'. The latter allegedly presented the king with a life-size, human-shaped figure of his mechanical handiwork. [8] Yan Shi (Chinese:偃师)
First century A.D. and earlierDescriptions of more than 100 machines and automata, including a fire engine, a wind organ, a coin-operated machine, and a steam-powered engine, in Pneumatica and Automata by Heron of Alexandria Ctesibius, Philo of Byzantium, Heron of Alexandria, and others
c. 420 B.CA wooden, steam-propelled bird, which was able to flyFlying pigeonArchytas of Tarentum
1206Created early humanoid automata, programmable automaton band [9] Robot band, hand-washing automaton, [10] automated moving peacocks [11] Al-Jazari
1495Designs for a humanoid robot Mechanical Knight Leonardo da Vinci
1560sClockwork Prayer that had machinal feet built under its robes that imitated walking. The robot's eyes, lips and head all move in lifelike gestures.Clockwork Prayer[ citation needed ] Gianello della Torre
1738Mechanical duck that was able to eat, flap its wings, and excrete Digesting Duck Jacques de Vaucanson
1898Nikola Tesla demonstrates the first radio-controlled vessel.Teleautomaton Nikola Tesla
1903Leonardo Torres y Quevedo presented the Telekino at the Paris Academy of Science, which consisted of a robot that executed commands transmitted by electromagnetic waves. [12] [13] Telekino Leonardo Torres y Quevedo
1912Leonardo Torres y Quevedo builds the first truly autonomous machine capable of playing chess. As opposed to the human-operated The Turk and Ajeeb, El Ajedrecista was an automaton that played chess without human guidance. It only played an endgame with three chess pieces, automatically moving a white king and a rook to checkmate the black king moved by a human opponent. [14] [15] El Ajedrecista Leonardo Torres y Quevedo
1914In his paper Essays on Automatics published in 1914, Leonardo Torres y Quevedo proposed a machine that makes "judgments" using sensors that capture information from the outside, parts that manipulate the outside world like arms, power sources such as batteries, and air pressure, and the most important, captured information and past information. It is defined as a part that can control the reaction of a living thing according to external information and adapt to changes in the environment to change its behavior. [16] [17] [18] Essays on Automatics Leonardo Torres y Quevedo
1921First fictional automatons called "robots" appear in the play R.U.R. Rossum's Universal Robots Karel Čapek
1930sHumanoid robot exhibited at the 1939 and 1940 World's Fairs Elektro Westinghouse Electric Corporation
1946First general-purpose digital computer Whirlwind Multiple people
1948Simple robots exhibiting biological behaviors [19] Elsie and Elmer William Grey Walter
1956First commercial robot, from the Unimation company founded by George Devol and Joseph Engelberger, based on Devol's patents [20] Unimate George Devol
1961First installed industrial robot. Unimate George Devol
1967 to 1972First full-scale humanoid intelligent robot, [21] [22] and first android. Its limb control system allowed it to walk with the lower limbs, and to grip and transport objects with its hands, using tactile sensors. Its vision system allowed it to measure distances and directions to objects using external receptors, artificial eyes, and ears. And its conversation system allowed it to communicate with a person in Japanese, with an artificial mouth. [23] [24] [25] WABOT-1 Waseda University
1973First industrial robot with six electromechanically driven axes [26] [27] Famulus KUKA Robot Group
1974The world's first microcomputer controlled electric industrial robot, IRB 6 from ASEA, was delivered to a small mechanical engineering company in southern Sweden. The design of this robot had been patented already in 1972.IRB 6 ABB Robot Group
1975Programmable universal manipulation arm, a Unimation product PUMA Victor Scheinman
1978First object-level robot programming language, allowing robots to handle variations in object position, shape, and sensor noise. Freddy I and II, RAPT robot programming language Patricia Ambler and Robin Popplestone
1983First multitasking, the parallel programming language used for robot control. It was the Event Driven Language (EDL) on the IBM/Series/1 process computer, with the implementation of both inter-process communication (WAIT/POST) and mutual exclusion (ENQ/DEQ) mechanisms for robot control. [28] ADRIEL IStevo Bozinovski and Mihail Sestakov

Robotic aspects

Mechanical construction Type 95 wheel and treads detail.JPG
Mechanical construction
Electrical aspect Computer Circuit Board MOD 45153624.jpg
Electrical aspect
A level of programming Dev win32.png
A level of programming

There are many types of robots; they are used in many different environments and for many different uses. Although being very diverse in application and form, they all share three basic similarities when it comes to their construction:

  1. Robots all have some kind of mechanical construction, a frame, form or shape designed to achieve a particular task. For example, a robot designed to travel across heavy dirt or mud might use caterpillar tracks. The mechanical aspect is mostly the creator's solution to completing the assigned task and dealing with the physics of the environment around it. Form follows function.
  2. Robots have electrical components that power and control the machinery. For example, the robot with caterpillar tracks would need some kind of power to move the tracker treads. That power comes in the form of electricity, which will have to travel through a wire and originate from a battery, a basic electrical circuit. Even petrol-powered machines that get their power mainly from petrol still require an electric current to start the combustion process which is why most petrol-powered machines like cars, have batteries. The electrical aspect of robots is used for movement (through motors), sensing (where electrical signals are used to measure things like heat, sound, position, and energy status), and operation (robots need some level of electrical energy supplied to their motors and sensors in order to activate and perform basic operations)
  3. All robots contain some level of computer programming code. A program is how a robot decides when or how to do something. In the caterpillar track example, a robot that needs to move across a muddy road may have the correct mechanical construction and receive the correct amount of power from its battery, but would not go anywhere without a program telling it to move. Programs are the core essence of a robot, it could have excellent mechanical and electrical construction, but if its program is poorly constructed its performance will be very poor (or it may not perform at all). There are three different types of robotic programs: remote control, artificial intelligence, and hybrid. A robot with remote control programming has a preexisting set of commands that it will only perform if and when it receives a signal from a control source, typically a human being with remote control. It is perhaps more appropriate to view devices controlled primarily by human commands as falling in the discipline of automation rather than robotics. Robots that use artificial intelligence interact with their environment on their own without a control source, and can determine reactions to objects and problems they encounter using their preexisting programming. A hybrid is a form of programming that incorporates both AI and RC functions in them.

Applications

As more and more robots are designed for specific tasks, this method of classification becomes more relevant. For example, many robots are designed for assembly work, which may not be readily adaptable for other applications. They are termed "assembly robots". For seam welding, some suppliers provide complete welding systems with the robot i.e. the welding equipment along with other material handling facilities like turntables, etc. as an integrated unit. Such an integrated robotic system is called a "welding robot" even though its discrete manipulator unit could be adapted to a variety of tasks. Some robots are specifically designed for heavy load manipulation, and are labeled as "heavy-duty robots". [29]

Current and potential applications include:

Components

Power source

The InSight lander with solar panels deployed in a cleanroom PIA19664-MarsInSightLander-Assembly-20150430.jpg
The InSight lander with solar panels deployed in a cleanroom

At present, mostly (lead–acid) batteries are used as a power source. Many different types of batteries can be used as a power source for robots. They range from lead–acid batteries, which are safe and have relatively long shelf lives but are rather heavy compared to silver–cadmium batteries which are much smaller in volume and are currently much more expensive. Designing a battery-powered robot needs to take into account factors such as safety, cycle lifetime, and weight. Generators, often some type of internal combustion engine, can also be used. However, such designs are often mechanically complex and need fuel, require heat dissipation, and are relatively heavy. A tether connecting the robot to a power supply would remove the power supply from the robot entirely. This has the advantage of saving weight and space by moving all power generation and storage components elsewhere. However, this design does come with the drawback of constantly having a cable connected to the robot, which can be difficult to manage. [44] Potential power sources could be:

Actuation

A robotic leg powered by air muscles 2005-11-14 ShadowLeg Finished medium.jpg
A robotic leg powered by air muscles

Actuators are the "muscles" of a robot, the parts which convert stored energy into movement. [45] By far the most popular actuators are electric motors that rotate a wheel or gear, and linear actuators that control industrial robots in factories. There are some recent advances in alternative types of actuators, powered by electricity, chemicals, or compressed air.

Electric motors

The vast majority of robots use electric motors, often brushed and brushless DC motors in portable robots or AC motors in industrial robots and CNC machines. These motors are often preferred in systems with lighter loads, and where the predominant form of motion is rotational.

Linear actuators

Various types of linear actuators move in and out instead of by spinning, and often have quicker direction changes, particularly when very large forces are needed such as with industrial robotics. They are typically powered by compressed and oxidized air (pneumatic actuator) or an oil (hydraulic actuator) Linear actuators can also be powered by electricity which usually consists of a motor and a leadscrew. Another common type is a mechanical linear actuator that is turned by hand, such as a rack and pinion on a car.

Series elastic actuators

Series elastic actuation (SEA) relies on the idea of introducing intentional elasticity between the motor actuator and the load for robust force control. Due to the resultant lower reflected inertia, series elastic actuation improves safety when a robot interacts with the environment (e.g., humans or workpieces) or during collisions. [46] Furthermore, it also provides energy efficiency and shock absorption (mechanical filtering) while reducing excessive wear on the transmission and other mechanical components. This approach has successfully been employed in various robots, particularly advanced manufacturing robots [47] and walking humanoid robots. [48] [49]

The controller design of a series elastic actuator is most often performed within the passivity framework as it ensures the safety of interaction with unstructured environments. [50] Despite its remarkable stability and robustness, this framework suffers from the stringent limitations imposed on the controller which may trade-off performance. The reader is referred to the following survey which summarizes the common controller architectures for SEA along with the corresponding sufficient passivity conditions. [51] One recent study has derived the necessary and sufficient passivity conditions for one of the most common impedance control architectures, namely velocity-sourced SEA. [52] This work is of particular importance as it drives the non-conservative passivity bounds in an SEA scheme for the first time which allows a larger selection of control gains.

Air muscles

Pneumatic artificial muscles also known as air muscles, are special tubes that expand (typically up to 42%) when air is forced inside them. They are used in some robot applications. [53] [54] [55]

Wire muscles

Muscle wire, also known as shape memory alloy, Nitinol® or Flexinol® wire, is a material that contracts (under 5%) when electricity is applied. They have been used for some small robot applications. [56] [57]

Electroactive polymers

EAPs or EPAMs are a plastic material that can contract substantially (up to 380% activation strain) from electricity, and have been used in facial muscles and arms of humanoid robots, [58] and to enable new robots to float, [59] fly, swim or walk. [60]

Piezo motors

Recent alternatives to DC motors are piezo motors or ultrasonic motors. These work on a fundamentally different principle, whereby tiny piezoceramic elements, vibrating many thousands of times per second, cause linear or rotary motion. There are different mechanisms of operation; one type uses the vibration of the piezo elements to step the motor in a circle or a straight line. [61] Another type uses the piezo elements to cause a nut to vibrate or to drive a screw. The advantages of these motors are nanometer resolution, speed, and available force for their size. [62] These motors are already available commercially, and being used on some robots. [63] [64]

Elastic nanotubes

Elastic nanotubes are a promising artificial muscle technology in early-stage experimental development. The absence of defects in carbon nanotubes enables these filaments to deform elastically by several percent, with energy storage levels of perhaps 10  J/cm3 for metal nanotubes. Human biceps could be replaced with an 8 mm diameter wire of this material. Such compact "muscle" might allow future robots to outrun and outjump humans. [65]

Sensing

Sensors allow robots to receive information about a certain measurement of the environment, or internal components. This is essential for robots to perform their tasks, and act upon any changes in the environment to calculate the appropriate response. They are used for various forms of measurements, to give the robots warnings about safety or malfunctions, and to provide real-time information about the task it is performing.

Touch

Current robotic and prosthetic hands receive far less tactile information than the human hand. Recent research has developed a tactile sensor array that mimics the mechanical properties and touch receptors of human fingertips. [66] [67] The sensor array is constructed as a rigid core surrounded by conductive fluid contained by an elastomeric skin. Electrodes are mounted on the surface of the rigid core and are connected to an impedance-measuring device within the core. When the artificial skin touches an object the fluid path around the electrodes is deformed, producing impedance changes that map the forces received from the object. The researchers expect that an important function of such artificial fingertips will be adjusting the robotic grip on held objects.

Scientists from several European countries and Israel developed a prosthetic hand in 2009, called SmartHand, which functions like a real one—allowing patients to write with it, type on a keyboard, play piano and perform other fine movements. The prosthesis has sensors which enable the patient to sense real feeling in its fingertips. [68]

Vision

Computer vision is the science and technology of machines that see. As a scientific discipline, computer vision is concerned with the theory behind artificial systems that extract information from images. The image data can take many forms, such as video sequences and views from cameras.

In most practical computer vision applications, the computers are pre-programmed to solve a particular task, but methods based on learning are now becoming increasingly common.

Computer vision systems rely on image sensors that detect electromagnetic radiation which is typically in the form of either visible light or infra-red light. The sensors are designed using solid-state physics. The process by which light propagates and reflects off surfaces is explained using optics. Sophisticated image sensors even require quantum mechanics to provide a complete understanding of the image formation process. Robots can also be equipped with multiple vision sensors to be better able to compute the sense of depth in the environment. Like human eyes, robots' "eyes" must also be able to focus on a particular area of interest, and also adjust to variations in light intensities.

There is a subfield within computer vision where artificial systems are designed to mimic the processing and behavior of biological system, at different levels of complexity. Also, some of the learning-based methods developed within computer vision have a background in biology.

Other

Other common forms of sensing in robotics use lidar, radar, and sonar. [69] Lidar measures distance to a target by illuminating the target with laser light and measuring the reflected light with a sensor. Radar uses radio waves to determine the range, angle, or velocity of objects. Sonar uses sound propagation to navigate, communicate with or detect objects on or under the surface of the water.

Manipulation

KUKA industrial robot operating in a foundry Automation of foundry with robot.jpg
KUKA industrial robot operating in a foundry
Puma, one of the first industrial robots Puma Robotic Arm - GPN-2000-001817.jpg
Puma, one of the first industrial robots
Baxter, a modern and versatile industrial robot developed by Rodney Brooks Caught Coding (9690512888).jpg
Baxter, a modern and versatile industrial robot developed by Rodney Brooks
Lefty, first checker playing robot Lefty the Robot - First Checker Playing Robot 1983.png
Lefty, first checker playing robot

A definition of robotic manipulation has been provided by Matt Mason as: "manipulation refers to an agent’s control of its environment through selective contact”. [70]

Robots need to manipulate objects; pick up, modify, destroy, or otherwise have an effect. Thus the functional end of a robot arm intended to make the effect (whether a hand, or tool) are often referred to as end effectors , [71] while the "arm" is referred to as a manipulator. [72] Most robot arms have replaceable end-effectors, each allowing them to perform some small range of tasks. Some have a fixed manipulator that cannot be replaced, while a few have one very general-purpose manipulator, for example, a humanoid hand. [73]

Mechanical grippers

One of the most common types of end-effectors are "grippers". In its simplest manifestation, it consists of just two fingers that can open and close to pick up and let go of a range of small objects. Fingers can, for example, be made of a chain with a metal wire running through it. [74] Hands that resemble and work more like a human hand include the Shadow Hand and the Robonaut hand. [75] Hands that are of a mid-level complexity include the Delft hand. [76] [77] Mechanical grippers can come in various types, including friction and encompassing jaws. Friction jaws use all the force of the gripper to hold the object in place using friction. Encompassing jaws cradle the object in place, using less friction.

Suction end-effectors

Suction end-effectors, powered by vacuum generators, are very simple astrictive [78] devices that can hold very large loads provided the prehension surface is smooth enough to ensure suction.

Pick and place robots for electronic components and for large objects like car windscreens, often use very simple vacuum end-effectors.

Suction is a highly used type of end-effector in industry, in part because the natural compliance of soft suction end-effectors can enable a robot to be more robust in the presence of imperfect robotic perception. As an example: consider the case of a robot vision system that estimates the position of a water bottle but has 1 centimeter of error. While this may cause a rigid mechanical gripper to puncture the water bottle, the soft suction end-effector may just bend slightly and conform to the shape of the water bottle surface.

General purpose effectors

Some advanced robots are beginning to use fully humanoid hands, like the Shadow Hand, MANUS, [79] and the Schunk hand. [80] These are highly dexterous manipulators, with as many as 20 degrees of freedom and hundreds of tactile sensors. [81]

Locomotion

Rolling robots

Segway in the Robot museum in Nagoya Segway 01.JPG
Segway in the Robot museum in Nagoya

For simplicity, most mobile robots have four wheels or a number of continuous tracks. Some researchers have tried to create more complex wheeled robots with only one or two wheels. These can have certain advantages such as greater efficiency and reduced parts, as well as allowing a robot to navigate in confined places that a four-wheeled robot would not be able to.

Two-wheeled balancing robots

Balancing robots generally use a gyroscope to detect how much a robot is falling and then drive the wheels proportionally in the same direction, to counterbalance the fall at hundreds of times per second, based on the dynamics of an inverted pendulum. [82] Many different balancing robots have been designed. [83] While the Segway is not commonly thought of as a robot, it can be thought of as a component of a robot, when used as such Segway refer to them as RMP (Robotic Mobility Platform). An example of this use has been as NASA's Robonaut that has been mounted on a Segway. [84]

One-wheeled balancing robots

A one-wheeled balancing robot is an extension of a two-wheeled balancing robot so that it can move in any 2D direction using a round ball as its only wheel. Several one-wheeled balancing robots have been designed recently, such as Carnegie Mellon University's "Ballbot" which is the approximate height and width of a person, and Tohoku Gakuin University's "BallIP". [85] Because of the long, thin shape and ability to maneuver in tight spaces, they have the potential to function better than other robots in environments with people. [86]

Spherical orb robots

Several attempts have been made in robots that are completely inside a spherical ball, either by spinning a weight inside the ball, [87] [88] or by rotating the outer shells of the sphere. [89] [90] These have also been referred to as an orb bot [91] or a ball bot. [92] [93]

Six-wheeled robots

Using six wheels instead of four wheels can give better traction or grip in outdoor terrain such as on rocky dirt or grass.

Tracked robots
TALON military robots used by the United States Army Foster-Miller TALON SWORDS.jpg
TALON military robots used by the United States Army

Tank tracks provide even more traction than a six-wheeled robot. Tracked wheels behave as if they were made of hundreds of wheels, therefore are very common for outdoor and military robots, where the robot must drive on very rough terrain. However, they are difficult to use indoors such as on carpets and smooth floors. Examples include NASA's Urban Robot "Urbie". [94]

Walking applied to robots

Walking is a difficult and dynamic problem to solve. Several robots have been made which can walk reliably on two legs, however, none have yet been made which are as robust as a human. There has been much study on human-inspired walking, such as AMBER lab which was established in 2008 by the Mechanical Engineering Department at Texas A&M University. [95] Many other robots have been built that walk on more than two legs, due to these robots being significantly easier to construct. [96] [97] Walking robots can be used for uneven terrains, which would provide better mobility and energy efficiency than other locomotion methods. Typically, robots on two legs can walk well on flat floors and can occasionally walk up stairs. None can walk over rocky, uneven terrain. Some of the methods which have been tried are:

ZMP technique

The zero moment point (ZMP) is the algorithm used by robots such as Honda's ASIMO. The robot's onboard computer tries to keep the total inertial forces (the combination of Earth's gravity and the acceleration and deceleration of walking), exactly opposed by the floor reaction force (the force of the floor pushing back on the robot's foot). In this way, the two forces cancel out, leaving no moment (force causing the robot to rotate and fall over). [98] However, this is not exactly how a human walks, and the difference is obvious to human observers, some of whom have pointed out that ASIMO walks as if it needs the lavatory. [99] [100] [101] ASIMO's walking algorithm is not static, and some dynamic balancing is used (see below). However, it still requires a smooth surface to walk on.

Hopping

Several robots, built in the 1980s by Marc Raibert at the MIT Leg Laboratory, successfully demonstrated very dynamic walking. Initially, a robot with only one leg, and a very small foot could stay upright simply by hopping. The movement is the same as that of a person on a pogo stick. As the robot falls to one side, it would jump slightly in that direction, in order to catch itself. [102] Soon, the algorithm was generalised to two and four legs. A bipedal robot was demonstrated running and even performing somersaults. [103] A quadruped was also demonstrated which could trot, run, pace, and bound. [104] For a full list of these robots, see the MIT Leg Lab Robots page. [105]

Dynamic balancing (controlled falling)

A more advanced way for a robot to walk is by using a dynamic balancing algorithm, which is potentially more robust than the Zero Moment Point technique, as it constantly monitors the robot's motion, and places the feet in order to maintain stability. [106] This technique was recently demonstrated by Anybots' Dexter Robot, [107] which is so stable, it can even jump. [108] Another example is the TU Delft Flame.

Passive dynamics

Perhaps the most promising approach uses passive dynamics where the momentum of swinging limbs is used for greater efficiency. It has been shown that totally unpowered humanoid mechanisms can walk down a gentle slope, using only gravity to propel themselves. Using this technique, a robot need only supply a small amount of motor power to walk along a flat surface or a little more to walk up a hill. This technique promises to make walking robots at least ten times more efficient than ZMP walkers, like ASIMO. [109] [110]

Other methods of locomotion

Flying

A modern passenger airliner is essentially a flying robot, with two humans to manage it. The autopilot can control the plane for each stage of the journey, including takeoff, normal flight, and even landing. [111] Other flying robots are uninhabited and are known as unmanned aerial vehicles (UAVs). They can be smaller and lighter without a human pilot on board, and fly into dangerous territory for military surveillance missions. Some can even fire on targets under command. UAVs are also being developed which can fire on targets automatically, without the need for a command from a human. Other flying robots include cruise missiles, the Entomopter, and the Epson micro helicopter robot. Robots such as the Air Penguin, Air Ray, and Air Jelly have lighter-than-air bodies, are propelled by paddles, and are guided by sonar.

Snaking
Two robot snakes. The left one has 64 motors (with 2 degrees of freedom per segment), the right one 10. Robosnakes.jpg
Two robot snakes. The left one has 64 motors (with 2 degrees of freedom per segment), the right one 10.

Several snake robots have been successfully developed. Mimicking the way real snakes move, these robots can navigate very confined spaces, meaning they may one day be used to search for people trapped in collapsed buildings. [112] The Japanese ACM-R5 snake robot [113] can even navigate both on land and in water. [114]

Skating

A small number of skating robots have been developed, one of which is a multi-mode walking and skating device. It has four legs, with unpowered wheels, which can either step or roll. [115] Another robot, Plen, can use a miniature skateboard or roller-skates, and skate across a desktop. [116]

Capuchin, a climbing robot Capuchin Free Climbing Robot.jpg
Capuchin, a climbing robot
Climbing

Several different approaches have been used to develop robots that have the ability to climb vertical surfaces. One approach mimics the movements of a human climber on a wall with protrusions; adjusting the center of mass and moving each limb in turn to gain leverage. An example of this is Capuchin, [117] built by Dr. Ruixiang Zhang at Stanford University, California. Another approach uses the specialized toe pad method of wall-climbing geckoes, which can run on smooth surfaces such as vertical glass. Examples of this approach include Wallbot [118] and Stickybot. [119]

China's Technology Daily reported on 15 November 2008, that Dr. Li Hiu Yeung and his research group of New Concept Aircraft (Zhuhai) Co., Ltd. had successfully developed a bionic gecko robot named "Speedy Freelander". According to Dr. Yeung, the gecko robot could rapidly climb up and down a variety of building walls, navigate through ground and wall fissures, and walk upside-down on the ceiling. It was also able to adapt to the surfaces of smooth glass, rough, sticky or dusty walls as well as various types of metallic materials. It could also identify and circumvent obstacles automatically. Its flexibility and speed were comparable to a natural gecko. A third approach is to mimic the motion of a snake climbing a pole. [69]

Swimming (Piscine)

It is calculated that when swimming some fish can achieve a propulsive efficiency greater than 90%. [120] Furthermore, they can accelerate and maneuver far better than any man-made boat or submarine, and produce less noise and water disturbance. Therefore, many researchers studying underwater robots would like to copy this type of locomotion. [121] Notable examples are the Essex University Computer Science Robotic Fish G9, [122] and the Robot Tuna built by the Institute of Field Robotics, to analyze and mathematically model thunniform motion. [123] The Aqua Penguin, [124] designed and built by Festo of Germany, copies the streamlined shape and propulsion by front "flippers" of penguins. Festo have also built the Aqua Ray and Aqua Jelly, which emulate the locomotion of manta ray, and jellyfish, respectively.

Robotic Fish: iSplash-II ISplash Robotic Fish.jpg
Robotic Fish: iSplash-II

In 2014 iSplash-II was developed by PhD student Richard James Clapham and Prof. Huosheng Hu at Essex University. It was the first robotic fish capable of outperforming real carangiform fish in terms of average maximum velocity (measured in body lengths/ second) and endurance, the duration that top speed is maintained. [125] This build attained swimming speeds of 11.6BL/s (i.e. 3.7 m/s). [126] The first build, iSplash-I (2014) was the first robotic platform to apply a full-body length carangiform swimming motion which was found to increase swimming speed by 27% over the traditional approach of a posterior confined waveform. [127]

Sailing
The autonomous sailboat robot Vaimos Vaimosluc.jpg
The autonomous sailboat robot Vaimos

Sailboat robots have also been developed in order to make measurements at the surface of the ocean. A typical sailboat robot is Vaimos [128] built by IFREMER and ENSTA-Bretagne. Since the propulsion of sailboat robots uses the wind, the energy of the batteries is only used for the computer, for the communication and for the actuators (to tune the rudder and the sail). If the robot is equipped with solar panels, the robot could theoretically navigate forever. The two main competitions of sailboat robots are WRSC, which takes place every year in Europe, and Sailbot.

Environmental interaction and navigation

Radar, GPS, and lidar, are all combined to provide proper navigation and obstacle avoidance (vehicle developed for 2007 DARPA Urban Challenge). ElementBlack2.jpg
Radar, GPS, and lidar, are all combined to provide proper navigation and obstacle avoidance (vehicle developed for 2007 DARPA Urban Challenge).

Though a significant percentage of robots in commission today are either human controlled or operate in a static environment, there is an increasing interest in robots that can operate autonomously in a dynamic environment. These robots require some combination of navigation hardware and software in order to traverse their environment. In particular, unforeseen events (e.g. people and other obstacles that are not stationary) can cause problems or collisions. Some highly advanced robots such as ASIMO and Meinü robot have particularly good robot navigation hardware and software. Also, self-controlled cars, Ernst Dickmanns' driverless car, and the entries in the DARPA Grand Challenge, are capable of sensing the environment well and subsequently making navigational decisions based on this information, including by a swarm of autonomous robots. [41] Most of these robots employ a GPS navigation device with waypoints, along with radar, sometimes combined with other sensory data such as lidar, video cameras, and inertial guidance systems for better navigation between waypoints.

Human-robot interaction

Kismet can produce a range of facial expressions. Kismet-IMG 6007-gradient.jpg
Kismet can produce a range of facial expressions.

The state of the art in sensory intelligence for robots will have to progress through several orders of magnitude if we want the robots working in our homes to go beyond vacuum-cleaning the floors. If robots are to work effectively in homes and other non-industrial environments, the way they are instructed to perform their jobs, and especially how they will be told to stop will be of critical importance. The people who interact with them may have little or no training in robotics, and so any interface will need to be extremely intuitive. Science fiction authors also typically assume that robots will eventually be capable of communicating with humans through speech, gestures, and facial expressions, rather than a command-line interface. Although speech would be the most natural way for the human to communicate, it is unnatural for the robot. It will probably be a long time before robots interact as naturally as the fictional C-3PO, or Data of Star Trek, Next Generation. Even though the current state of robotics cannot meet the standards of these robots from science-fiction, robotic media characters (e.g., Wall-E, R2-D2) can elicit audience sympathies that increase people's willingness to accept actual robots in the future. [129] Acceptance of social robots is also likely to increase if people can meet a social robot under appropriate conditions. Studies have shown that interacting with a robot by looking at, touching, or even imagining interacting with the robot can reduce negative feelings that some people have about robots before interacting with them. [130] However, if pre-existing negative sentiments are especially strong, interacting with a robot can increase those negative feelings towards robots. [130]

Speech recognition

Interpreting the continuous flow of sounds coming from a human, in real time, is a difficult task for a computer, mostly because of the great variability of speech. [131] The same word, spoken by the same person may sound different depending on local acoustics, volume, the previous word, whether or not the speaker has a cold, etc.. It becomes even harder when the speaker has a different accent. [132] Nevertheless, great strides have been made in the field since Davis, Biddulph, and Balashek designed the first "voice input system" which recognized "ten digits spoken by a single user with 100% accuracy" in 1952. [133] Currently, the best systems can recognize continuous, natural speech, up to 160 words per minute, with an accuracy of 95%. [134] With the help of artificial intelligence, machines nowadays can use people's voice to identify their emotions such as satisfied or angry. [135]

Robotic voice

Other hurdles exist when allowing the robot to use voice for interacting with humans. For social reasons, synthetic voice proves suboptimal as a communication medium, [136] making it necessary to develop the emotional component of robotic voice through various techniques. [137] [138] An advantage of diphonic branching is the emotion that the robot is programmed to project, can be carried on the voice tape, or phoneme, already pre-programmed onto the voice media. One of the earliest examples is a teaching robot named Leachim developed in 1974 by Michael J. Freeman. [139] [140] Leachim was able to convert digital memory to rudimentary verbal speech on pre-recorded computer discs. [141] It was programmed to teach students in The Bronx, New York. [141]

Gestures

One can imagine, in the future, explaining to a robot chef how to make a pastry, or asking directions from a robot police officer. In both of these cases, making hand gestures would aid the verbal descriptions. In the first case, the robot would be recognizing gestures made by the human, and perhaps repeating them for confirmation. In the second case, the robot police officer would gesture to indicate "down the road, then turn right". It is likely that gestures will make up a part of the interaction between humans and robots. [142] A great many systems have been developed to recognize human hand gestures. [143]

Facial expression

Facial expressions can provide rapid feedback on the progress of a dialog between two humans, and soon may be able to do the same for humans and robots. Robotic faces have been constructed by Hanson Robotics using their elastic polymer called Frubber, allowing a large number of facial expressions due to the elasticity of the rubber facial coating and embedded subsurface motors (servos). [144] The coating and servos are built on a metal skull. A robot should know how to approach a human, judging by their facial expression and body language. Whether the person is happy, frightened, or crazy-looking affects the type of interaction expected of the robot. Likewise, robots like Kismet and the more recent addition, Nexi [145] can produce a range of facial expressions, allowing it to have meaningful social exchanges with humans. [146]

Artificial emotions

Artificial emotions can also be generated, composed of a sequence of facial expressions or gestures. As can be seen from the movie Final Fantasy: The Spirits Within , the programming of these artificial emotions is complex and requires a large amount of human observation. To simplify this programming in the movie, presets were created together with a special software program. This decreased the amount of time needed to make the film. These presets could possibly be transferred for use in real-life robots. An example of a robot with artificial emotions is Robin the Robot developed by an Armenian IT company Expper Technologies, which uses AI-based peer-to-peer interaction. Its main task is achieving emotional well-being, i.e. overcome stress and anxiety. Robin was trained to analyze facial expressions and use his face to display his emotions given the context. The robot has been tested by kids in US clinics, and observations show that Robin increased the appetite and cheerfulness of children after meeting and talking. [147]

Personality

Many of the robots of science fiction have a personality, something which may or may not be desirable in the commercial robots of the future. [148] Nevertheless, researchers are trying to create robots which appear to have a personality: [149] [150] i.e. they use sounds, facial expressions, and body language to try to convey an internal state, which may be joy, sadness, or fear. One commercial example is Pleo, a toy robot dinosaur, which can exhibit several apparent emotions. [151]

Social intelligence

The Socially Intelligent Machines Lab of the Georgia Institute of Technology researches new concepts of guided teaching interaction with robots. The aim of the projects is a social robot that learns task and goals from human demonstrations without prior knowledge of high-level concepts. These new concepts are grounded from low-level continuous sensor data through unsupervised learning, and task goals are subsequently learned using a Bayesian approach. These concepts can be used to transfer knowledge to future tasks, resulting in faster learning of those tasks. The results are demonstrated by the robot Curi who can scoop some pasta from a pot onto a plate and serve the sauce on top. [152]

Control

Puppet Magnus, a robot-manipulated marionette with complex control systems Magnus B. Egerstedt puppet.jpg
Puppet Magnus, a robot-manipulated marionette with complex control systems
Experimental planar robot arm and sensor-based, open-architecture robot controller developed at Sunderland University, UK in 2000 Sunderland University Experimental Robot Arm.jpg
Experimental planar robot arm and sensor-based, open-architecture robot controller developed at Sunderland University, UK in 2000
RuBot II can manually resolve Rubik's cubes. RuBot II.jpg
RuBot II can manually resolve Rubik's cubes.

The mechanical structure of a robot must be controlled to perform tasks. [153] The control of a robot involves three distinct phases – perception, processing, and action (robotic paradigms). [154] Sensors give information about the environment or the robot itself (e.g. the position of its joints or its end effector). This information is then processed to be stored or transmitted and to calculate the appropriate signals to the actuators (motors), which move the mechanical structure to achieve the required co-ordinated motion or force actions.

The processing phase can range in complexity. At a reactive level, it may translate raw sensor information directly into actuator commands (e.g. firing motor power electronic gates based directly upon encoder feedback signals to achieve the required torque/velocity of the shaft). Sensor fusion and internal models may first be used to estimate parameters of interest (e.g. the position of the robot's gripper) from noisy sensor data. An immediate task (such as moving the gripper in a certain direction until an object is detected with a proximity sensor) is sometimes inferred from these estimates. Techniques from control theory are generally used to convert the higher-level tasks into individual commands that drive the actuators, most often using kinematic and dynamic models of the mechanical structure. [153] [154] [155]

At longer time scales or with more sophisticated tasks, the robot may need to build and reason with a "cognitive" model. Cognitive models try to represent the robot, the world, and how the two interact. Pattern recognition and computer vision can be used to track objects. [153] Mapping techniques can be used to build maps of the world. Finally, motion planning and other artificial intelligence techniques may be used to figure out how to act. For example, a planner may figure out how to achieve a task without hitting obstacles, falling over, etc.

Modern commercial robotic control systems are highly complex, integrate multiple sensors and effectors, have many interacting degrees-of-freedom (DOF) and require operator interfaces, programming tools and real-time capabilities. [154] They are oftentimes interconnected to wider communication networks and in many cases are now both IoT-enabled and mobile. [156] Progress towards open architecture, layered, user-friendly and ‘intelligent’ sensor-based interconnected robots has emerged from earlier concepts related to Flexible Manufacturing Systems (FMS), and several 'open or 'hybrid' reference architectures exist which assist developers of robot control software and hardware to move beyond traditional, earlier notions of 'closed' robot control systems have been proposed. [155] Open architecture controllers are said to be better able to meet the growing requirements of a wide range of robot users, including system developers, end users and research scientists, and are better positioned to deliver the advanced robotic concepts related to Industry 4.0. [155] In addition to utilizing many established features of robot controllers, such as position, velocity and force control of end effectors, they also enable IoT interconnection and the implementation of more advanced sensor fusion and control techniques, including adaptive control, Fuzzy control and Artificial Neural Network (ANN)-based control. [155] When implemented in real-time, such techniques can potentially improve the stability and performance of robots operating in unknown or uncertain environments by enabling the control systems to learn and adapt to environmental changes. [157] There are several examples of reference architectures for robot controllers, and also examples of successful implementations of actual robot controllers developed from them. One example of a generic reference architecture and associated interconnected, open-architecture robot and controller implementation was developed by Michael Short and colleagues at the University of Sunderland in the UK in 2000 (pictured right). [155] The robot was used in a number of research and development studies, including prototype implementation of novel advanced and intelligent control and environment mapping methods in real-time. [157] [158]

Autonomy levels

TOPIO, a humanoid robot, played ping pong at Tokyo IREX 2009. TOPIO 3.jpg
TOPIO, a humanoid robot, played ping pong at Tokyo IREX 2009.

Control systems may also have varying levels of autonomy.

  1. Direct interaction is used for haptic or teleoperated devices, and the human has nearly complete control over the robot's motion.
  2. Operator-assist modes have the operator commanding medium-to-high-level tasks, with the robot automatically figuring out how to achieve them. [160]
  3. An autonomous robot may go without human interaction for extended periods of time . Higher levels of autonomy do not necessarily require more complex cognitive capabilities. For example, robots in assembly plants are completely autonomous but operate in a fixed pattern.

Another classification takes into account the interaction between human control and the machine motions.

  1. Teleoperation. A human controls each movement, each machine actuator change is specified by the operator.
  2. Supervisory. A human specifies general moves or position changes and the machine decides specific movements of its actuators.
  3. Task-level autonomy. The operator specifies only the task and the robot manages itself to complete it.
  4. Full autonomy. The machine will create and complete all its tasks without human interaction.

Research

Two Jet Propulsion Laboratory engineers stand with three vehicles, providing a size comparison of three generations of Mars rovers. Front and center is the flight spare for the first Mars rover, Sojourner, which landed on Mars in 1997 as part of the Mars Pathfinder Project. On the left is a Mars Exploration Rover (MER) test vehicle that is a working sibling to Spirit and Opportunity, which landed on Mars in 2004. On the right is a test rover for the Mars Science Laboratory, which landed Curiosity on Mars in 2012.
Sojourner is 65 cm (2.13 ft) long. The Mars Exploration Rovers (MER) are 1.6 m (5.2 ft) long. Curiosity on the right is 3 m (9.8 ft) long. PIA15279 3rovers-stand D2011 1215 D521.jpg
Two Jet Propulsion Laboratory engineers stand with three vehicles, providing a size comparison of three generations of Mars rovers. Front and center is the flight spare for the first Mars rover, Sojourner, which landed on Mars in 1997 as part of the Mars Pathfinder Project. On the left is a Mars Exploration Rover (MER) test vehicle that is a working sibling to Spirit and Opportunity, which landed on Mars in 2004. On the right is a test rover for the Mars Science Laboratory, which landed Curiosity on Mars in 2012.
Sojourner is 65 cm (2.13 ft) long. The Mars Exploration Rovers (MER) are 1.6 m (5.2 ft) long. Curiosity on the right is 3 m (9.8 ft) long.

Much of the research in robotics focuses not on specific industrial tasks, but on investigations into new types of robots, alternative ways to think about or design robots, and new ways to manufacture them. Other investigations, such as MIT's cyberflora project, are almost wholly academic.

A first particular new innovation in robot design is the open sourcing of robot-projects. To describe the level of advancement of a robot, the term "Generation Robots" can be used. This term is coined by Professor Hans Moravec, Principal Research Scientist at the Carnegie Mellon University Robotics Institute in describing the near future evolution of robot technology. First-generation robots, Moravec predicted in 1997, should have an intellectual capacity comparable to perhaps a lizard and should become available by 2010. Because the first generation robot would be incapable of learning, however, Moravec predicts that the second generation robot would be an improvement over the first and become available by 2020, with the intelligence maybe comparable to that of a mouse. The third generation robot should have intelligence comparable to that of a monkey. Though fourth generation robots, robots with human intelligence, professor Moravec predicts, would become possible, he does not predict this happening before around 2040 or 2050. [161]

The second is evolutionary robots. This is a methodology that uses evolutionary computation to help design robots, especially the body form, or motion and behavior controllers. In a similar way to natural evolution, a large population of robots is allowed to compete in some way, or their ability to perform a task is measured using a fitness function. Those that perform worst are removed from the population and replaced by a new set, which have new behaviors based on those of the winners. Over time the population improves, and eventually a satisfactory robot may appear. This happens without any direct programming of the robots by the researchers. Researchers use this method both to create better robots, [162] and to explore the nature of evolution. [163] Because the process often requires many generations of robots to be simulated, [164] this technique may be run entirely or mostly in simulation, using a robot simulator software package, then tested on real robots once the evolved algorithms are good enough. [165] Currently, there are about 10 million industrial robots toiling around the world, and Japan is the top country having high density of utilizing robots in its manufacturing industry.[ citation needed ]

Dynamics and kinematics

External video
Nuvola apps kaboodle.svg How the BB-8 Sphero Toy Works

The study of motion can be divided into kinematics and dynamics. [166] Direct kinematics or forward kinematics refers to the calculation of end effector position, orientation, velocity, and acceleration when the corresponding joint values are known. Inverse kinematics refers to the opposite case in which required joint values are calculated for given end effector values, as done in path planning. Some special aspects of kinematics include handling of redundancy (different possibilities of performing the same movement), collision avoidance, and singularity avoidance. Once all relevant positions, velocities, and accelerations have been calculated using kinematics, methods from the field of dynamics are used to study the effect of forces upon these movements. Direct dynamics refers to the calculation of accelerations in the robot once the applied forces are known. Direct dynamics is used in computer simulations of the robot. Inverse dynamics refers to the calculation of the actuator forces necessary to create a prescribed end-effector acceleration. This information can be used to improve the control algorithms of a robot.

In each area mentioned above, researchers strive to develop new concepts and strategies, improve existing ones, and improve the interaction between these areas. To do this, criteria for "optimal" performance and ways to optimize design, structure, and control of robots must be developed and implemented.

Bionics and biomimetics

Bionics and biomimetics apply the physiology and methods of locomotion of animals to the design of robots. For example, the design of BionicKangaroo was based on the way kangaroos jump.

Quantum computing

There has been some research into whether robotics algorithms can be run more quickly on quantum computers than they can be run on digital computers. This area has been referred to as quantum robotics. [167]

Education and training

The SCORBOT-ER 4u educational robot ER4u.jpg
The SCORBOT-ER 4u educational robot

Robotics engineers design robots, maintain them, develop new applications for them, and conduct research to expand the potential of robotics. [168] Robots have become a popular educational tool in some middle and high schools, particularly in parts of the USA, [169] as well as in numerous youth summer camps, raising interest in programming, artificial intelligence, and robotics among students.

Employment

A robot technician builds small all-terrain robots (courtesy: MobileRobots, Inc.). MobileRobotsPioneerAT.jpg
A robot technician builds small all-terrain robots (courtesy: MobileRobots, Inc.).

Robotics is an essential component in many modern manufacturing environments. As factories increase their use of robots, the number of robotics–related jobs grow and have been observed to be steadily rising. [170] The employment of robots in industries has increased productivity and efficiency savings and is typically seen as a long-term investment for benefactors. A paper by Michael Osborne and  Carl Benedikt Frey  found that 47 percent of US jobs are at risk to automation "over some unspecified number of years". [171] These claims have been criticized on the ground that social policy, not AI, causes unemployment. [172] In a 2016 article in The Guardian, Stephen Hawking stated "The automation of factories has already decimated jobs in traditional manufacturing, and the rise of artificial intelligence is likely to extend this job destruction deep into the middle classes, with only the most caring, creative or supervisory roles remaining". [173]

According to a GlobalData September 2021 report, the robotics industry was worth $45bn in 2020, and by 2030, it will have grown at a compound annual growth rate (CAGR) of 29% to $568bn, driving jobs in robotics and related industries. [174]

Occupational safety and health implications

A discussion paper drawn up by EU-OSHA highlights how the spread of robotics presents both opportunities and challenges for occupational safety and health (OSH). [175]

The greatest OSH benefits stemming from the wider use of robotics should be substitution for people working in unhealthy or dangerous environments. In space, defense, security, or the nuclear industry, but also in logistics, maintenance, and inspection, autonomous robots are particularly useful in replacing human workers performing dirty, dull or unsafe tasks, thus avoiding workers' exposures to hazardous agents and conditions and reducing physical, ergonomic and psychosocial risks. For example, robots are already used to perform repetitive and monotonous tasks, to handle radioactive material or to work in explosive atmospheres. In the future, many other highly repetitive, risky or unpleasant tasks will be performed by robots in a variety of sectors like agriculture, construction, transport, healthcare, firefighting or cleaning services. [176]

Moreover, there are certain skills to which humans will be better suited than machines for some time to come and the question is how to achieve the best combination of human and robot skills. The advantages of robotics include heavy-duty jobs with precision and repeatability, whereas the advantages of humans include creativity, decision-making, flexibility, and adaptability. This need to combine optimal skills has resulted in collaborative robots and humans sharing a common workspace more closely and led to the development of new approaches and standards to guarantee the safety of the "man-robot merger". Some European countries are including robotics in their national programs and trying to promote a safe and flexible cooperation between robots and operators to achieve better productivity. For example, the German Federal Institute for Occupational Safety and Health (BAuA) organises annual workshops on the topic "human-robot collaboration".

In the future, cooperation between robots and humans will be diversified, with robots increasing their autonomy and human-robot collaboration reaching completely new forms. Current approaches and technical standards [177] [178] aiming to protect employees from the risk of working with collaborative robots will have to be revised.

User experience

Great user experience predicts the needs, experiences, behaviors, language and cognitive abilities, and other factors of each user group. It then uses these insights to produce a product or solution that is ultimately useful and usable. For robots, user experience begins with an understanding of the robot's intended task and environment, while considering any possible social impact the robot may have on human operations and interactions with it. [179]

It defines that communication as the transmission of information through signals, which are elements perceived through touch, sound, smell and sight. [180] The author states that the signal connects the sender to the receiver and consists of three parts: the signal itself, what it refers to, and the interpreter. Body postures and gestures, facial expressions, hand and head movements are all part of nonverbal behavior and communication. Robots are no exception when it comes to human-robot interaction. Therefore, humans use their verbal and nonverbal behaviors to communicate their defining characteristics. Similarly, social robots need this coordination to perform human-like behaviors.

See also

Related Research Articles

Computer vision tasks include methods for acquiring, processing, analyzing and understanding digital images, and extraction of high-dimensional data from the real world in order to produce numerical or symbolic information, e.g. in the forms of decisions. Understanding in this context means the transformation of visual images into descriptions of the world that make sense to thought processes and can elicit appropriate action. This image understanding can be seen as the disentangling of symbolic information from image data using models constructed with the aid of geometry, physics, statistics, and learning theory.

<span class="mw-page-title-main">Control engineering</span> Engineering discipline that deals with control systems

Control engineering or control systems engineering is an engineering discipline that deals with control systems, applying control theory to design equipment and systems with desired behaviors in control environments. The discipline of controls overlaps and is usually taught along with electrical engineering and mechanical engineering at many institutions around the world.

<span class="mw-page-title-main">Robot</span> Machine capable of carrying out a complex series of actions automatically

A robot is a machine—especially one programmable by a computer—capable of carrying out a complex series of actions automatically. A robot can be guided by an external control device, or the control may be embedded within. Robots may be constructed to evoke human form, but most robots are task-performing machines, designed with an emphasis on stark functionality, rather than expressive aesthetics.

<span class="mw-page-title-main">Humanoid robot</span> Body shape similar to a human

A humanoid robot is a robot resembling the human body in shape. The design may be for functional purposes, such as interacting with human tools and environments, for experimental purposes, such as the study of bipedal locomotion, or for other purposes. In general, humanoid robots have a torso, a head, two arms, and two legs, though some humanoid robots may replicate only part of the body, for example, from the waist up. Some humanoid robots also have heads designed to replicate human facial features such as eyes and mouths. Androids are humanoid robots built to aesthetically resemble humans.

<span class="mw-page-title-main">Automation</span> Use of various control systems for operating equipment

Automation describes a wide range of technologies that reduce human intervention in processes, namely by predetermining decision criteria, subprocess relationships, and related actions, as well as embodying those predeterminations in machines. Automation has been achieved by various means including mechanical, hydraulic, pneumatic, electrical, electronic devices, and computers, usually in combination. Complicated systems, such as modern factories, airplanes, and ships typically use combinations of all of these techniques. The benefit of automation includes labor savings, reducing waste, savings in electricity costs, savings in material costs, and improvements to quality, accuracy, and precision.

An actuator is a component of a machine that is responsible for moving and controlling a mechanism or system, for example by opening a valve. In simple terms, it is a "mover".

<span class="mw-page-title-main">Haptic technology</span> Any form of interaction involving touch

Haptic technology is technology that can create an experience of touch by applying forces, vibrations, or motions to the user. These technologies can be used to create virtual objects in a computer simulation, to control virtual objects, and to enhance remote control of machines and devices (telerobotics). Haptic devices may incorporate tactile sensors that measure forces exerted by the user on the interface. The word haptic, from the Greek: ἁπτικός (haptikos), means "tactile, pertaining to the sense of touch". Simple haptic devices are common in the form of game controllers, joysticks, and steering wheels.

<span class="mw-page-title-main">Mechatronics engineering</span> Combination of electronics and mechanics

Mechatronics engineering also called mechatronics, is an interdisciplinary branch of engineering that focuses on the integration of mechanical, electrical and electronic engineering systems, and also includes a combination of robotics, electronics, computer science, telecommunications, systems, control, and product engineering.

<span class="mw-page-title-main">Electroactive polymer</span>

An electroactive polymer (EAP) is a polymer that exhibits a change in size or shape when stimulated by an electric field. The most common applications of this type of material are in actuators and sensors. A typical characteristic property of an EAP is that they will undergo a large amount of deformation while sustaining large forces.

Neurorobotics is the combined study of neuroscience, robotics, and artificial intelligence. It is the science and technology of embodied autonomous neural systems. Neural systems include brain-inspired algorithms, computational models of biological neural networks and actual biological systems. Such neural systems can be embodied in machines with mechanic or any other forms of physical actuation. This includes robots, prosthetic or wearable systems but also, at smaller scale, micro-machines and, at the larger scales, furniture and infrastructures.

<span class="mw-page-title-main">History of robots</span> First true automaton able to play chess

The history of robots has its origins in the ancient world. During the industrial revolution, humans developed the structural engineering capability to control electricity so that machines could be powered with small motors. In the early 20th century, the notion of a humanoid machine was developed.

<span class="mw-page-title-main">Daniela L. Rus</span> American computer scientist

Daniela L. Rus is a roboticist and computer scientist, Director of the MIT Computer Science and Artificial Intelligence Laboratory (CSAIL), and the Andrew and Erna Viterbi Professor in the Department of Electrical Engineering and Computer Science (EECS) at the Massachusetts Institute of Technology.

<span class="mw-page-title-main">Glossary of robotics</span> List of definitions of terms and concepts commonly used in the study of robotics

Robotics is the branch of technology that deals with the design, construction, operation, structural disposition, manufacture and application of robots. Robotics is related to the sciences of electronics, engineering, mechanics, and software.

The following outline is provided as an overview of and topical guide to robotics:

Robotic sensing is a subarea of robotics science intended to provide sensing capabilities to robots. Robotic sensing provides robots with the ability to sense their environments and is typically used as feedback to enable robots to adjust their behavior based on sensed input. Robot sensing includes the ability to see, touch, hear and move and associated algorithms to process and make use of environmental feedback and sensory data. Robot sensing is important in applications such as vehicular automation, robotic prosthetics, and for industrial, medical, entertainment and educational robots.

As humans move through their environment, they must change the stiffness of their joints in order to effectively interact with their surroundings. Stiffness is the degree to a which an object resists deformation when subjected to a known force. This idea is also referred to as impedance, however, sometimes the idea of deformation under a given load is discussed under the term "compliance" which is the opposite of stiffness . In order to effectively interact with their environment, humans must adjust the stiffness of their limbs. This is accomplished via the co-contraction of antagonistic muscle groups.

<span class="mw-page-title-main">Proportional myoelectric control</span>

Proportional myoelectric control can be used to activate robotic lower limb exoskeletons. A proportional myoelectric control system utilizes a microcontroller or computer that inputs electromyography (EMG) signals from sensors on the leg muscle(s) and then activates the corresponding joint actuator(s) proportionally to the EMG signal.

<span class="mw-page-title-main">Soft robotics</span> Subfield of robotics

Soft robotics is a subfield of robotics that concerns the design, control, and fabrication of robots composed of compliant materials, instead of rigid links. In contrast to rigid-bodied robots built from metals, ceramics and hard plastics, the compliance of soft robots can improve their safety when working in close contact with humans.

The term “soft robots” designs a broad class of robotic systems whose architecture includes soft elements, with much higher elasticity than traditional rigid robots. Articulated Soft Robots are robots with both soft and rigid parts, inspired to the muscle-skeletal system of vertebrate animals – from reptiles to birds to mammalians to humans. Compliance is typically concentrated in actuators, transmission and joints while structural stability is provided by rigid or semi-rigid links.

A continuum robot is a type of robot that is characterised by infinite degrees of freedom and number of joints. These characteristics allow continuum manipulators to adjust and modify their shape at any point along their length, granting them the possibility to work in confined spaces and complex environments where standard rigid-link robots cannot operate. In particular, we can define a continuum robot as an actuatable structure whose constitutive material forms curves with continuous tangent vectors. This is a fundamental definition that allows to distinguish between continuum robots and snake-arm robots or hyper-redundant manipulators: the presence of rigid links and joints allows them to only approximately perform curves with continuous tangent vectors.

References

  1. "German National Library". International classification system of the German National Library (GND).
  2. Nocks, Lisa (2007). The robot : the life story of a technology. Westport, CT: Greenwood Publishing Group.
  3. 1 2 Zunt, Dominik. "Who did actually invent the word "robot" and what does it mean?". The Karel Čapek website. Archived from the original on 23 January 2013. Retrieved 5 February 2017.
  4. Asimov, Isaac (1996) [1995]. "The Robot Chronicles". Gold. London: Voyager. pp. 224–225. ISBN   978-0-00-648202-4.
  5. Asimov, Isaac (1983). "4 The Word I Invented". Counting the Eons. Doubleday. Bibcode:1983coeo.book.....A. Robotics has become a sufficiently well-developed technology to warrant articles and books on its history and I have watched this in amazement, and in some disbelief because I invented … the word
  6. Svoboda, Elizabeth (25 September 2019). "Your robot surgeon will see you now". Nature. 573 (7775): S110–S111. Bibcode:2019Natur.573S.110S. doi: 10.1038/d41586-019-02874-0 . PMID   31554995.
  7. "Robotics: About the Exhibition". The Tech Museum of Innovation. Archived from the original on 13 September 2008. Retrieved 15 September 2008.
  8. Needham, Joseph (1991). Science and Civilisation in China: Volume 2, History of Scientific Thought. Cambridge University Press. ISBN   978-0-521-05800-1.
  9. Fowler, Charles B. (October 1967). "The Museum of Music: A History of Mechanical Instruments". Music Educators Journal. 54 (2): 45–49. doi:10.2307/3391092. JSTOR   3391092. S2CID   190524140.
  10. Rosheim, Mark E. (1994). Robot Evolution: The Development of Anthrobotics. Wiley-IEEE. pp.  9–10. ISBN   978-0-471-02622-8.
  11. al-Jazari (Islamic artist), Encyclopædia Britannica .
  12. Sarkar 2006, page 97[ full citation needed ]
  13. H. R. Everett, Unmanned Systems of World Wars I and II, MIT Press - 2015, pages 91-95
  14. Williams, Andrew (2017-03-16). History of Digital Games: Developments in Art, Design and Interaction. CRC Press. ISBN   9781317503811.
  15. Randell, Brian (October 1982). "From Analytical Engine to Electronic Digital Computer: The Contributions of Ludgate, Torres, and Bush". IEEE Annals of the History of Computing. 4 (4): 327–341. doi:10.1109/MAHC.1982.10042. S2CID   1737953.
  16. L. Torres Quevedo. Ensayos sobre Automática - Su definicion. Extension teórica de sus aplicaciones, Revista de la Academia de Ciencias Exacta, Revista 12, pp.391-418, 1913.
  17. L. Torres Quevedo. Essais sur l'Automatique - Sa définition. Etendue théorique de ses applications, Revue Génerale des Sciences Pures et Appliquées, vol.2, pp.601-611, 1915.
  18. B. Randell. Essays on Automatics, The Origins of Digital Computers, pp.89-107, 1982.
  19. PhD, Renato M.E. Sabbatini. "Sabbatini, RME: An Imitation of Life: The First Robots".
  20. Waurzyniak, Patrick (2006). "Masters of Manufacturing: Joseph F. Engelberger". Society of Manufacturing Engineers. 137 (1). Archived from the original on 9 November 2011.
  21. "Humanoid History -WABOT-". www.humanoid.waseda.ac.jp.
  22. Zeghloul, Saïd; Laribi, Med Amine; Gazeau, Jean-Pierre (21 September 2015). Robotics and Mechatronics: Proceedings of the 4th IFToMM International Symposium on Robotics and Mechatronics. Springer. ISBN   9783319223681 via Google Books.
  23. "Historical Android Projects". androidworld.com.
  24. Robots: From Science Fiction to Technological Revolution, page 130
  25. Duffy, Vincent G. (19 April 2016). Handbook of Digital Human Modeling: Research for Applied Ergonomics and Human Factors Engineering. CRC Press. ISBN   9781420063523 via Google Books.
  26. "KUKA Industrial Robot FAMULUS". Archived from the original on 20 February 2009. Retrieved 10 January 2008.
  27. "History of Industrial Robots" (PDF). Archived from the original (PDF) on 24 December 2012. Retrieved 27 October 2012.
  28. Bozinovski, S. (1994). "Parallel programming for mobile robot control: Agent-based approach". 14th International Conference on Distributed Computing Systems. pp. 202–208. doi:10.1109/ICDCS.1994.302412. ISBN   0-8186-5840-1. S2CID   27855786.
  29. Hunt, V. Daniel (1985). "Smart Robots". Smart Robots: A Handbook of Intelligent Robotic Systems. Chapman and Hall. p. 141. ISBN   978-1-4613-2533-8.
  30. "Robot density rises globally". Robotic Industries Association. 8 February 2018. Retrieved 3 December 2018.
  31. Pinto, Jim (1 October 2003). "Fully automated factories approach reality". Automation World . Archived from the original on 1 October 2011. Retrieved 3 December 2018.
  32. Dragani, Rachelle (8 November 2018). "Can a robot make you a 'superworker'?". Verizon Communications . Retrieved 3 December 2018.
  33. Pollock, Emily (7 June 2018). "Construction Robotics Industry Set to Double by 2023". engineering.com. Retrieved 3 December 2018.
  34. Grift, Tony E. (2004). "Agricultural Robotics". University of Illinois at Urbana–Champaign. Archived from the original on 4 May 2007. Retrieved 3 December 2018.
  35. Thomas, Jim (1 November 2017). "How corporate giants are automating the farm". New Internationalist . Retrieved 3 December 2018.
  36. "OUCL Robot Sheepdog Project". Department of Computer Science, University of Oxford. 3 July 2001. Retrieved 3 December 2018.
  37. Kolodny, Lora (4 July 2017). "Robots are coming to a burger joint near you". CNBC . Retrieved 3 December 2018.
  38. Corner, Stuart (23 November 2017). "AI-driven robot makes 'perfect' flatbread". iothub.com.au. Retrieved 3 December 2018.
  39. Eyre, Michael (12 September 2014). "'Boris' the robot can load up dishwasher". BBC News . Retrieved 3 December 2018.
  40. One database, developed by the United States Department of Energy contains information on almost 500 existing robotic technologies and can be found on the D&D Knowledge Management Information Tool.
  41. 1 2 Kagan, Eugene, and Irad Ben-Gal (2015). Search and foraging:individual motion and swarm dynamics. Chapman and Hall/CRC, 2015. ISBN   9781482242102.{{cite book}}: CS1 maint: multiple names: authors list (link)
  42. Fojtik, Rostislav (2017). "The Ozobot and education of programming". New Trends and Issues Proceedings on Humanities and Social Sciences. 4 (5). doi:10.18844/prosoc.v4i5.2666.
  43. Masril, Mardhiah; Hendrik, Billy; Theozard Fikri, Harry; Hazidar, Al Hamidy; Priambodo, Bagus; Naf'An, Emil; Handriani, Inge; Pratama Putra, Zico; Kudr Nseaf, Asama (2019). "The Effect of Lego Mindstorms as an Innovative Educational Tool to Develop Students' Creativity Skills for a Creative Society". Journal of Physics: Conference Series. 1339 (1): 012082. Bibcode:2019JPhCS1339a2082M. doi:10.1088/1742-6596/1339/1/012082. S2CID   213941566.
  44. Dowling, Kevin. "Power Sources for Small Robots" (PDF). Carnegie Mellon University. Retrieved 11 May 2012.
  45. Roozing, Wesley; Li, Zhibin; Tsagarakis, Nikos; Caldwell, Darwin (2016). "Design Optimisation and Control of Compliant Actuation Arrangements in Articulated Robots for Improved Energy Efficiency". IEEE Robotics and Automation Letters. 1 (2): 1110–1117. doi:10.1109/LRA.2016.2521926. S2CID   1940410.
  46. Pratt, G.A.; Williamson, M.M. (1995). "Series elastic actuators". Proceedings 1995 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human-Robot Interaction and Cooperative Robots. Vol. 1. pp. 399–406. doi:10.1109/IROS.1995.525827. hdl:1721.1/36966. ISBN   0-8186-7108-4. S2CID   17120394.
  47. Furnémont, Raphaël; Mathijssen, Glenn; Verstraten, Tom; Lefeber, Dirk; Vanderborght, Bram (27 January 2016). "Bi-directional series-parallel elastic actuator and overlap of the actuation layers" (PDF). Bioinspiration & Biomimetics. 11 (1): 016005. Bibcode:2016BiBi...11a6005F. doi:10.1088/1748-3190/11/1/016005. PMID   26813145. S2CID   37031990.
  48. Pratt, Jerry E.; Krupp, Benjamin T. (2004). "Series Elastic Actuators for legged robots". In Gerhart, Grant R; Shoemaker, Chuck M; Gage, Douglas W (eds.). Unmanned Ground Vehicle Technology VI. Vol. 5422. pp. 135–144. doi:10.1117/12.548000. S2CID   16586246.
  49. Li, Zhibin; Tsagarakis, Nikos; Caldwell, Darwin (2013). "Walking Pattern Generation for a Humanoid Robot with Compliant Joints". Autonomous Robots. 35 (1): 1–14. doi:10.1007/s10514-013-9330-7. S2CID   624563.
  50. Colgate, J. Edward (1988). The control of dynamically interacting systems (Thesis). hdl:1721.1/14380.
  51. Calanca, Andrea; Muradore, Riccardo; Fiorini, Paolo (November 2017). "Impedance control of series elastic actuators: Passivity and acceleration-based control". Mechatronics. 47: 37–48. doi:10.1016/j.mechatronics.2017.08.010.
  52. Tosun, Fatih Emre; Patoglu, Volkan (June 2020). "Necessary and Sufficient Conditions for the Passivity of Impedance Rendering With Velocity-Sourced Series Elastic Actuation". IEEE Transactions on Robotics. 36 (3): 757–772. doi:10.1109/TRO.2019.2962332. S2CID   212907787.
  53. www.imagesco.com, Images SI Inc -. "Air Muscle actuators, going further, page 6".
  54. "Air Muscles". Shadow Robot. Archived from the original on 27 September 2007.
  55. Tondu, Bertrand (2012). "Modelling of the McKibben artificial muscle: A review". Journal of Intelligent Material Systems and Structures. 23 (3): 225–253. doi:10.1177/1045389X11435435. S2CID   136854390.
  56. "TALKING ELECTRONICS Nitinol Page-1". Talkingelectronics.com. Retrieved 27 November 2010.
  57. "lf205, Hardware: Building a Linux-controlled walking robot". Ibiblio.org. 1 November 2001. Retrieved 27 November 2010.
  58. "WW-EAP and Artificial Muscles". Eap.jpl.nasa.gov. Retrieved 27 November 2010.
  59. "Empa – a117-2-eap". Empa.ch. Retrieved 27 November 2010.
  60. "Electroactive Polymers (EAP) as Artificial Muscles (EPAM) for Robot Applications". Hizook. Archived from the original on 6 August 2020. Retrieved 27 November 2010.
  61. "Piezo LEGS – -09-26". Archived from the original on 30 January 2008. Retrieved 28 October 2007.
  62. "Squiggle Motors: Overview" . Retrieved 8 October 2007.
  63. Nishibori; et al. (2003). "Robot Hand with Fingers Using Vibration-Type Ultrasonic Motors (Driving Characteristics)". Journal of Robotics and Mechatronics. 15 (6): 588–595. doi:10.20965/jrm.2003.p0588.
  64. Otake, Mihoko; Kagami, Yoshiharu; Ishikawa, Kohei; Inaba, Masayuki; Inoue, Hirochika (6 April 2001). "Shape design of gel robots made of electroactive polymer gel". Smart Materials. 4234: 194–202. Bibcode:2001SPIE.4234..194O. doi:10.1117/12.424407. S2CID   30357330.
  65. Madden, John D. (16 November 2007). "Mobile Robots: Motor Challenges and Materials Solutions". Science. 318 (5853): 1094–1097. Bibcode:2007Sci...318.1094M. CiteSeerX   10.1.1.395.4635 . doi:10.1126/science.1146351. PMID   18006737. S2CID   52827127.
  66. "Syntouch LLC: BioTac(R) Biomimetic Tactile Sensor Array". Archived from the original on 3 October 2009. Retrieved 10 August 2009.
  67. Wettels, Nicholas; Santos, Veronica J.; Johansson, Roland S.; Loeb, Gerald E. (January 2008). "Biomimetic Tactile Sensor Array". Advanced Robotics. 22 (8): 829–849. doi:10.1163/156855308X314533. S2CID   4594917.
  68. "What is The SmartHand?". SmartHand Project. Retrieved 4 February 2011.
  69. 1 2 Arreguin, Juan (2008). Automation and Robotics. Vienna, Austria: I-Tech and Publishing.
  70. Mason, Matthew T. (2001). Mechanics of Robotic Manipulation. doi:10.7551/mitpress/4527.001.0001. ISBN   9780262256629. S2CID   5260407.
  71. "What is a robotic end-effector?". ATI Industrial Automation. 2007. Retrieved 16 October 2007.
  72. Crane, Carl D.; Joseph Duffy (1998). Kinematic Analysis of Robot Manipulators. Cambridge University Press. ISBN   978-0-521-57063-3 . Retrieved 16 October 2007.
  73. G.J. Monkman, S. Hesse, R. Steinmann & H. Schunk (2007). Robot Grippers. Berlin: Wiley
  74. "Annotated Mythbusters: Episode 78: Ninja Myths – Walking on Water, Catching a Sword, Catching an Arrow". (Discovery Channel's Mythbusters making mechanical gripper from the chain and metal wire)
  75. Robonaut hand
  76. "Delft hand". TU Delft. Archived from the original on 3 February 2012. Retrieved 21 November 2011.
  77. M&C. "TU Delft ontwikkelt goedkope, voorzichtige robothand".
  78. "astrictive definition – English definition dictionary – Reverso".
  79. Tijsma, H.A.; Liefhebber, F.; Herder, J.L. (2005). "Evaluation of New User Interface Features for the MANUS Robot Arm". 9th International Conference on Rehabilitation Robotics, 2005. ICORR 2005. pp. 258–263. doi:10.1109/ICORR.2005.1501097. ISBN   0-7803-9003-2. S2CID   36445389.
  80. Allcock, Andrew (2006). "Anthropomorphic hand is almost human". Machinery. Archived from the original on 28 September 2007. Retrieved 17 October 2007.
  81. "Welcome".
  82. "T.O.B.B". Mtoussaint.de. Retrieved 27 November 2010.
  83. "nBot, a two wheel balancing robot". Geology.heroy.smu.edu. Retrieved 27 November 2010.
  84. "ROBONAUT Activity Report". NASA. 2004. Archived from the original on 20 August 2007. Retrieved 20 October 2007.
  85. Guizzo, Erico (29 April 2010). "A Robot That Balances on a Ball". IEEE Spectrum.
  86. "Carnegie Mellon Researchers Develop New Type of Mobile Robot That Balances and Moves on a Ball Instead of Legs or Wheels" (Press release). Carnegie Mellon. 9 August 2006. Archived from the original on 9 June 2007. Retrieved 20 October 2007.
  87. "Spherical Robot Can Climb Over Obstacles". BotJunkie. Retrieved 27 November 2010.
  88. "Rotundus". Rotundus.se. Archived from the original on 26 August 2011. Retrieved 27 November 2010.
  89. "OrbSwarm Gets A Brain". BotJunkie. 11 July 2007. Retrieved 27 November 2010.
  90. "Rolling Orbital Bluetooth Operated Thing". BotJunkie. Retrieved 27 November 2010.
  91. "Swarm". Orbswarm.com. Retrieved 27 November 2010.
  92. "The Ball Bot : Johnnytronic@Sun". Blogs.sun.com. Archived from the original on 24 August 2011. Retrieved 27 November 2010.
  93. "Senior Design Projects | College of Engineering & Applied Science| University of Colorado at Boulder". Engineering.colorado.edu. 30 April 2008. Archived from the original on 23 July 2011. Retrieved 27 November 2010.
  94. "JPL Robotics: System: Commercial Rovers". Archived from the original on 2006-06-15.
  95. "AMBER Lab".
  96. "Micromagic Systems Robotics Lab". Archived from the original on 2017-06-01. Retrieved 2009-04-29.
  97. "AMRU-5 hexapod robot" (PDF).
  98. "Achieving Stable Walking". Honda Worldwide. Retrieved 22 October 2007.
  99. "Funny Walk". Pooter Geek. 28 December 2004. Retrieved 22 October 2007.
  100. "ASIMO's Pimp Shuffle". Popular Science. 9 January 2007. Retrieved 22 October 2007.
  101. "The Temple of VTEC – Honda and Acura Enthusiasts Online Forums > Robot Shows Prime Minister How to Loosen Up > > A drunk robot?".
  102. "3D One-Leg Hopper (1983–1984)". MIT Leg Laboratory. Retrieved 22 October 2007.
  103. "3D Biped (1989–1995)". MIT Leg Laboratory.
  104. "Quadruped (1984–1987)". MIT Leg Laboratory.
  105. "MIT Leg Lab Robots- Main".
  106. "About the robots". Anybots. Archived from the original on 9 September 2007. Retrieved 23 October 2007.
  107. "Homepage". Anybots. Archived from the original on 16 May 2014. Retrieved 23 October 2007.
  108. "Dexter Jumps video". YouTube. 1 March 2007. Archived from the original on 2021-10-30. Retrieved 23 October 2007.
  109. Collins, Steve; Ruina, Andy; Tedrake, Russ; Wisse, Martijn (18 February 2005). "Efficient Bipedal Robots Based on Passive-Dynamic Walkers". Science. 307 (5712): 1082–1085. Bibcode:2005Sci...307.1082C. doi:10.1126/science.1107799. PMID   15718465. S2CID   1315227.
  110. Collins, S.H.; Ruina, A. (2005). "A Bipedal Walking Robot with Efficient and Human-Like Gait". Proceedings of the 2005 IEEE International Conference on Robotics and Automation. pp. 1983–1988. doi:10.1109/ROBOT.2005.1570404. ISBN   0-7803-8914-X. S2CID   15145353.
  111. "Testing the Limits" (PDF). Boeing. p. 29. Retrieved 9 April 2008.
  112. Miller, Gavin. "Introduction". snakerobots.com. Retrieved 22 October 2007.
  113. "ACM-R5". Archived from the original on 11 October 2011.
  114. "Swimming snake robot (commentary in Japanese)". Archived from the original on 2012-02-08. Retrieved 2007-10-28.
  115. "Commercialized Quadruped Walking Vehicle "TITAN VII"". Hirose Fukushima Robotics Lab. Archived from the original on 6 November 2007. Retrieved 23 October 2007.
  116. "Plen, the robot that skates across your desk". SCI FI Tech. 23 January 2007. Archived from the original on 11 October 2007. Retrieved 23 October 2007.
  117. Capuchin on YouTube
  118. Wallbot on YouTube
  119. Stanford University: Stickybot on YouTube
  120. Sfakiotakis, M.; Lane, D.M.; Davies, J.B.C. (April 1999). "Review of fish swimming modes for aquatic locomotion". IEEE Journal of Oceanic Engineering. 24 (2): 237–252. Bibcode:1999IJOE...24..237S. CiteSeerX   10.1.1.459.8614 . doi:10.1109/48.757275. S2CID   17226211.
  121. Richard Mason. "What is the market for robot fish?". Archived from the original on 4 July 2009.
  122. "Robotic fish powered by Gumstix PC and PIC". Human Centred Robotics Group at Essex University. Archived from the original on 14 August 2011. Retrieved 25 October 2007.
  123. Witoon Juwarahawong. "Fish Robot". Institute of Field Robotics. Archived from the original on 4 November 2007. Retrieved 25 October 2007.
  124. "YouTube". YouTube . Archived from the original on 2009-06-09.
  125. "High-Speed Robotic Fish | iSplash". isplash-robot. Retrieved 7 January 2017.
  126. "iSplash-II: Realizing Fast Carangiform Swimming to Outperform a Real Fish" (PDF). Robotics Group at Essex University. Archived from the original (PDF) on 30 September 2015. Retrieved 29 September 2015.
  127. "iSplash-I: High Performance Swimming Motion of a Carangiform Robotic Fish with Full-Body Coordination" (PDF). Robotics Group at Essex University. Archived from the original (PDF) on 30 September 2015. Retrieved 29 September 2015.
  128. Jaulin, Luc; Le Bars, Fabrice (February 2013). "An Interval Approach for Stability Analysis: Application to Sailboat Robotics". IEEE Transactions on Robotics. 29 (1): 282–287. CiteSeerX   10.1.1.711.7180 . doi:10.1109/TRO.2012.2217794. S2CID   4977937.
  129. Banks, Jaime (2020). "Optimus Primed: Media Cultivation of Robot Mental Models and Social Judgments". Frontiers in Robotics and AI. 7: 62. doi: 10.3389/frobt.2020.00062 . PMC   7805817 . PMID   33501230.
  130. 1 2 Wullenkord, Ricarda; Fraune, Marlena R.; Eyssel, Friederike; Sabanovic, Selma (2016). "Getting in Touch: How imagined, actual, and physical contact affect evaluations of robots". 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). pp. 980–985. doi:10.1109/ROMAN.2016.7745228. ISBN   978-1-5090-3929-6. S2CID   6305599.
  131. Norberto Pires, J. (December 2005). "Robot‐by‐voice: experiments on commanding an industrial robot using the human voice". Industrial Robot: An International Journal. 32 (6): 505–511. doi:10.1108/01439910510629244.
  132. "Survey of the State of the Art in Human Language Technology: 1.2: Speech Recognition". Archived from the original on 11 November 2007.
  133. Fournier, Randolph Scott; Schmidt, B. June (1995). "Voice input technology: Learning style and attitude toward its use". Delta Pi Epsilon Journal. 37 (1): 1–12. ProQuest   1297783046.
  134. "History of Speech & Voice Recognition and Transcription Software". Dragon Naturally Speaking. Retrieved 27 October 2007.
  135. Cheng Lin, Kuan; Huang, Tien‐Chi; Hung, Jason C.; Yen, Neil Y.; Ju Chen, Szu (7 June 2013). "Facial emotion recognition towards affective computing‐based learning". Library Hi Tech. 31 (2): 294–307. doi:10.1108/07378831311329068.
  136. Walters, M. L.; Syrdal, D. S.; Koay, K. L.; Dautenhahn, K.; Te Boekhorst, R. (2008). "Human approach distances to a mechanical-looking robot with different robot voice styles". RO-MAN 2008 - the 17th IEEE International Symposium on Robot and Human Interactive Communication. pp. 707–712. doi:10.1109/ROMAN.2008.4600750. ISBN   978-1-4244-2212-8. S2CID   8653718.
  137. Pauletto, Sandra; Bowles, Tristan (2010). "Designing the emotional content of a robotic speech signal". Proceedings of the 5th Audio Mostly Conference on a Conference on Interaction with Sound - AM '10. pp. 1–8. doi:10.1145/1859799.1859804. ISBN   978-1-4503-0046-9. S2CID   30423778.
  138. Bowles, Tristan; Pauletto, Sandra (2010). Emotions in the Voice: Humanising a Robotic Voice (PDF). Proceedings of the 7th Sound and Music Computing Conference. Barcelona.
  139. "World of 2-XL: Leachim". www.2xlrobot.com. Retrieved 28 May 2019.
  140. "The Boston Globe from Boston, Massachusetts on June 23, 1974 · 132". Newspapers.com. Retrieved 28 May 2019.
  141. 1 2 "cyberneticzoo.com - Page 135 of 194 - a history of cybernetic animals and early robots". cyberneticzoo.com. Retrieved 28 May 2019.
  142. Waldherr, Stefan; Romero, Roseli; Thrun, Sebastian (1 September 2000). "A Gesture Based Interface for Human-Robot Interaction". Autonomous Robots. 9 (2): 151–173. doi:10.1023/A:1008918401478. S2CID   1980239.
  143. Li, Ling Hua; Du, Ji Fang (December 2012). "Visual Based Hand Gesture Recognition Systems". Applied Mechanics and Materials. 263–266: 2422–2425. Bibcode:2012AMM...263.2422L. doi:10.4028/www.scientific.net/AMM.263-266.2422. S2CID   62744240.
  144. "Frubber facial expressions". Archived from the original on 7 February 2009.
  145. "Best Inventions of 2008 – TIME". Time. 29 October 2008. Archived from the original on November 2, 2008 via www.time.com.
  146. "Kismet: Robot at MIT's AI Lab Interacts With Humans". Sam Ogden. Archived from the original on 12 October 2007. Retrieved 28 October 2007.
  147. "Armenian Robin the Robot to comfort kids at U.S. clinics starting July". Public Radio of Armenia. Retrieved 2021-05-13.
  148. Park, S.; Sharlin, Ehud; Kitamura, Y.; Lau, E. (29 April 2005). Synthetic Personality in Robots and its Effect on Human-Robot Relationship (Report). doi:10.11575/PRISM/31041. hdl:1880/45619.
  149. "Robot Receptionist Dishes Directions and Attitude". NPR.org.
  150. "New Scientist: A good robot has personality but not looks" (PDF). Archived from the original (PDF) on 29 September 2006.
  151. "Playtime with Pleo, your robotic dinosaur friend". 25 September 2008.
  152. Jennifer Bogo (31 October 2014). "Meet a woman who trains robots for a living".
  153. 1 2 3 Corke, Peter (2017). "Robotics, Vision and Control". Springer Tracts in Advanced Robotics. 118. doi:10.1007/978-3-319-54413-7. ISBN   978-3-319-54412-0. ISSN   1610-7438.
  154. 1 2 3 Lee, K. S. Fu, Ralph Gonzalez, C S. G. (1987). Robotics: Control Sensing. Vis. McGraw-Hill. ISBN   978-0-07-026510-3.
  155. 1 2 3 4 5 Short, Michael; Burn, Kevin (2011-04-01). "A generic controller architecture for intelligent robotic systems". Robotics and Computer-Integrated Manufacturing. 27 (2): 292–305. doi:10.1016/j.rcim.2010.07.013. ISSN   0736-5845.
  156. Ray, Partha Pratim (2016). "Internet of Robotic Things: Concept, Technologies, and Challenges". IEEE Access. 4: 9489–9500. doi:10.1109/ACCESS.2017.2647747. ISSN   2169-3536. S2CID   9273802.
  157. 1 2 Burn, K.; Short, M.; Bicker, R. (July 2003). "Adaptive and Nonlinear Fuzzy Force Control Techniques Applied to Robots Operating in Uncertain Environments". Journal of Robotic Systems. 20 (7): 391–400. doi:10.1002/rob.10093. ISSN   0741-2223.
  158. Burn, Kevin; Home, Geoffrey (2008-05-01). "Environment classification using Kohonen self-organizing maps". Expert Systems. 25 (2): 98–114. doi:10.1111/j.1468-0394.2008.00441.x. ISSN   0266-4720. S2CID   33369232.
  159. "A Ping-Pong-Playing Terminator". Popular Science.
  160. "Synthiam Exosphere combines AI, human operators to train robots". The Robot Report.
  161. NOVA conversation with Professor Moravec, October 1997. NOVA Online
  162. Sandhana, Lakshmi (5 September 2002). "A Theory of Evolution, for Robots". Wired. Wired Magazine. Retrieved 28 October 2007.
  163. Experimental Evolution In Robots Probes The Emergence Of Biological Communication. Science Daily. 24 February 2007. Retrieved 28 October 2007.
  164. Žlajpah, Leon (15 December 2008). "Simulation in robotics". Mathematics and Computers in Simulation. 79 (4): 879–897. doi:10.1016/j.matcom.2008.02.017.
  165. News, Technology Research. "Evolution trains robot teams TRN 051904". www.trnmag.com.{{cite web}}: |last= has generic name (help)
  166. Agarwal, P.K. Elements of Physics XI. Rastogi Publications. p. 2. ISBN   978-81-7133-911-2.
  167. Tandon, Prateek (2017). Quantum Robotics. Morgan & Claypool Publishers. ISBN   978-1627059138.
  168. "Career: Robotics Engineer". Princeton Review. 2012. Retrieved 27 January 2012.
  169. Saad, Ashraf; Kroutil, Ryan (2012). Hands-on Learning of Programming Concepts Using Robotics for Middle and High School Students. Proceedings of the 50th Annual Southeast Regional Conference of the Association for Computer Machinery. ACM. pp. 361–362. doi:10.1145/2184512.2184605.
  170. Toy, Tommy (29 June 2011). "Outlook for robotics and Automation for 2011 and beyond are excellent says expert". PBT Consulting. Retrieved 27 January 2012.
  171. Frey, Carl Benedikt; Osborne, Michael A. (January 2017). "The future of employment: How susceptible are jobs to computerisation?". Technological Forecasting and Social Change. 114: 254–280. CiteSeerX   10.1.1.395.416 . doi:10.1016/j.techfore.2016.08.019.
  172. McGaughey, Ewan (16 October 2019). "Will robots automate your job away? Full employment, basic income, and economic democracy". doi:10.31228/osf.io/udbj8. S2CID   243172487. SSRN   3044448.{{cite journal}}: Cite journal requires |journal= (help)
  173. Hawking, Stephen (1 January 2016). "This is the most dangerous time for our planet". The Guardian. Retrieved 22 November 2019.
  174. "Robotics – Thematic Research". GlobalData. GlobalData. Retrieved 22 September 2021.
  175. "Focal Points Seminar on review articles in the future of work – Safety and health at work – EU-OSHA". osha.europa.eu. Retrieved 19 April 2016.
  176. "Robotics: Redefining crime prevention, public safety and security". SourceSecurity.com.
  177. "Draft Standard for Intelligent Assist Devices — Personnel Safety Requirements" (PDF).
  178. "ISO/TS 15066:2016 – Robots and robotic devices – Collaborative robots".
  179. Brogårdh, Torgny (January 2007). "Present and future robot control development—An industrial perspective". Annual Reviews in Control. 31 (1): 69–79. doi:10.1016/j.arcontrol.2007.01.002. ISSN   1367-5788.
  180. Wang, Tian-Miao; Tao, Yong; Liu, Hui (2018-04-17). "Current Researches and Future Development Trend of Intelligent Robot: A Review". International Journal of Automation and Computing. 15 (5): 525–546. doi:10.1007/s11633-018-1115-1. ISSN   1476-8186. S2CID   126037910.

Further reading