Machine vision

Last updated
Early Automatix (now part of Omron) machine vision system Autovision II from 1983 being demonstrated at a trade show. Camera on tripod is pointing down at a light table to produce backlit image shown on screen, which is then subjected to blob extraction. AutovisionIIatRDT.jpg
Early Automatix (now part of Omron) machine vision system Autovision II from 1983 being demonstrated at a trade show. Camera on tripod is pointing down at a light table to produce backlit image shown on screen, which is then subjected to blob extraction.

Machine vision (MV) is the technology and methods used to provide imaging-based automatic inspection and analysis for such applications as automatic inspection, process control, and robot guidance, usually in industry. Machine vision refers to many technologies, software and hardware products, integrated systems, actions, methods and expertise. Machine vision as a systems engineering discipline can be considered distinct from computer vision, a form of computer science. It attempts to integrate existing technologies in new ways and apply them to solve real world problems. The term is the prevalent one for these functions in industrial automation environments but is also used for these functions in other environment vehicle guidance.

Contents

The overall machine vision process includes planning the details of the requirements and project, and then creating a solution. During run-time, the process starts with imaging, followed by automated analysis of the image and extraction of the required information.

Definition

Definitions of the term "Machine vision" vary, but all include the technology and methods used to extract information from an image on an automated basis, as opposed to image processing, where the output is another image. The information extracted can be a simple good-part/bad-part signal, or more a complex set of data such as the identity, position and orientation of each object in an image. The information can be used for such applications as automatic inspection and robot and process guidance in industry, for security monitoring and vehicle guidance. [1] [2] [3] This field encompasses a large number of technologies, software and hardware products, integrated systems, actions, methods and expertise. [3] [4] Machine vision is practically the only term used for these functions in industrial automation applications; the term is less universal for these functions in other environments such as security and vehicle guidance. Machine vision as a systems engineering discipline can be considered distinct from computer vision, a form of basic computer science; machine vision attempts to integrate existing technologies in new ways and apply them to solve real world problems in a way that meets the requirements of industrial automation and similar application areas. [3] :5 [5] The term is also used in a broader sense by trade shows and trade groups such as the Automated Imaging Association and the European Machine Vision Association. This broader definition also encompasses products and applications most often associated with image processing. [4] The primary uses for machine vision are automatic inspection and industrial robot/process guidance. [6] [7] :6–10 [8] In more recent times the terms computer vision and machine vision have converged to a greater degree. [9] :13 See glossary of machine vision.

Imaging based automatic inspection and sorting

The primary uses for machine vision are imaging-based automatic inspection and sorting and robot guidance.; [6] [7] :6–10 in this section the former is abbreviated as "automatic inspection". The overall process includes planning the details of the requirements and project, and then creating a solution. [10] [11] This section describes the technical process that occurs during the operation of the solution.

Methods and sequence of operation

The first step in the automatic inspection sequence of operation is acquisition of an image, typically using cameras, lenses, and lighting that has been designed to provide the differentiation required by subsequent processing. [12] [13] MV software packages and programs developed in them then employ various digital image processing techniques to extract the required information, and often make decisions (such as pass/fail) based on the extracted information. [14]

Equipment

The components of an automatic inspection system usually include lighting, a camera or other imager, a processor, software, and output devices. [7] :11–13

Imaging

The imaging device (e.g. camera) can either be separate from the main image processing unit or combined with it in which case the combination is generally called a smart camera or smart sensor. [15] [16] Inclusion of the full processing function into the same enclosure as the camera is often referred to as embedded processing. [17] When separated, the connection may be made to specialized intermediate hardware, a custom processing appliance, or a frame grabber within a computer using either an analog or standardized digital interface (Camera Link, CoaXPress). [18] [19] [20] [21] MV implementations also use digital cameras capable of direct connections (without a framegrabber) to a computer via FireWire, USB or Gigabit Ethernet interfaces. [21] [22]

While conventional (2D visible light) imaging is most commonly used in MV, alternatives include multispectral imaging, hyperspectral imaging, imaging various infrared bands, [23] line scan imaging, 3D imaging of surfaces and X-ray imaging. [6] Key differentiations within MV 2D visible light imaging are monochromatic vs. color, frame rate, resolution, and whether or not the imaging process is simultaneous over the entire image, making it suitable for moving processes. [24]

Though the vast majority of machine vision applications are solved using two-dimensional imaging, machine vision applications utilizing 3D imaging are a growing niche within the industry. [25] [26] The most commonly used method for 3D imaging is scanning based triangulation which utilizes motion of the product or image during the imaging process. A laser is projected onto the surfaces of an object. In machine vision this is accomplished with a scanning motion, either by moving the workpiece, or by moving the camera & laser imaging system. The line is viewed by a camera from a different angle; the deviation of the line represents shape variations. Lines from multiple scans are assembled into a depth map or point cloud. [27] Stereoscopic vision is used in special cases involving unique features present in both views of a pair of cameras. [27] Other 3D methods used for machine vision are time of flight and grid based. [27] [25] One method is grid array based systems using pseudorandom structured light system as employed by the Microsoft Kinect system circa 2012. [28] [29]

Image processing

After an image is acquired, it is processed. [20] Central processing functions are generally done by a CPU, a GPU, a FPGA or a combination of these. [17] Deep learning training and inference impose higher processing performance requirements. [30] Multiple stages of processing are generally used in a sequence that ends up as a desired result. A typical sequence might start with tools such as filters which modify the image, followed by extraction of objects, then extraction (e.g. measurements, reading of codes) of data from those objects, followed by communicating that data, or comparing it against target values to create and communicate "pass/fail" results. Machine vision image processing methods include;

Outputs

A common output from automatic inspection systems is pass/fail decisions. [14] These decisions may in turn trigger mechanisms that reject failed items or sound an alarm. Other common outputs include object position and orientation information for robot guidance systems. [6] Additionally, output types include numerical measurement data, data read from codes and characters, counts and classification of objects, displays of the process or results, stored images, alarms from automated space monitoring MV systems, and process control signals. [10] [13] This also includes user interfaces, interfaces for the integration of multi-component systems and automated data interchange. [42]

Deep Learning

The term "Deep Learning" has variable meanings, most of which can be applied to techniques used in machine vision for over 20 years. However the usage of the term in Machine Vision began in the later 2010s with the advent of the capability to successfully apply such techniques to entire images in the industrial machine vision space. [43] Conventional machine vision usually requires the "physics" phase of a machine vision automatic inspection solution to create reliable simple differentiation of defects. An example of "simple" differentiation is that the defects are dark and the good parts of the product are light. A common reason why some applications were not doable was when it was impossible to achieve the "simple"; deep learning removes this requirement, in essence "seeing" the object more as a human does, making it now possible to accomplish those automatic applications. [43] The system learns from a large amount of images during a training phase and then executes the inspection during run-time use which is called "inference". [43]

Imaging based robot guidance

Machine vision commonly provides location and orientation information to a robot to allow the robot to properly grasp the product. This capability is also used to guide motion that is simpler than robots, such as a 1 or 2 axis motion controller. [6] The overall process includes planning the details of the requirements and project, and then creating a solution. This section describes the technical process that occurs during the operation of the solution. Many of the process steps are the same as with automatic inspection except with a focus on providing position and orientation information as the result. [6]

Market

As recently as 2006, one industry consultant reported that MV represented a $1.5 billion market in North America. [44] However, the editor-in-chief of an MV trade magazine asserted that "machine vision is not an industry per se" but rather "the integration of technologies and products that provide services or applications that benefit true industries such as automotive or consumer goods manufacturing, agriculture, and defense." [4]

See also

Related Research Articles

Computer vision tasks include methods for acquiring, processing, analyzing and understanding digital images, and extraction of high-dimensional data from the real world in order to produce numerical or symbolic information, e.g. in the forms of decisions. Understanding in this context means the transformation of visual images into descriptions of the world that make sense to thought processes and can elicit appropriate action. This image understanding can be seen as the disentangling of symbolic information from image data using models constructed with the aid of geometry, physics, statistics, and learning theory.

<span class="mw-page-title-main">Thermography</span> Infrared imaging used to reveal temperature

Infrared thermography (IRT), thermal video and/or thermal imaging, is a process where a thermal camera captures and creates an image of an object by using infrared radiation emitted from the object in a process, which are examples of infrared imaging science. Thermographic cameras usually detect radiation in the long-infrared range of the electromagnetic spectrum and produce images of that radiation, called thermograms. Since infrared radiation is emitted by all objects with a temperature above absolute zero according to the black body radiation law, thermography makes it possible to see one's environment with or without visible illumination. The amount of radiation emitted by an object increases with temperature; therefore, thermography allows one to see variations in temperature. When viewed through a thermal imaging camera, warm objects stand out well against cooler backgrounds; humans and other warm-blooded animals become easily visible against the environment, day or night. As a result, thermography is particularly useful to the military and other users of surveillance cameras.

<span class="mw-page-title-main">Gesture recognition</span> Topic in computer science and language technology

Gesture recognition is an area of research and development in computer science and language technology concerned with the recognition and interpretation of human gestures. A subdiscipline of computer vision, it employs mathematical algorithms to interpret gestures.

<span class="mw-page-title-main">Smart camera</span> Machine vision system

A smart camera (sensor) or intelligent camera (sensor) or (smart) vision sensor or intelligent vision sensor or smart optical sensor or intelligent optical sensor or smart visual sensor or intelligent visual sensor is a machine vision system which, in addition to image capture circuitry, is capable of extracting application-specific information from the captured images, along with generating event descriptions or making decisions that are used in an intelligent and automated system. A smart camera is a self-contained, standalone vision system with built-in image sensor in the housing of an industrial video camera. The vision system and the image sensor can be integrated into one single piece of hardware known as intelligent image sensor or smart image sensor. It contains all necessary communication interfaces, e.g. Ethernet, as well as industry-proof 24V I/O lines for connection to a PLC, actuators, relays or pneumatic valves, and can be either static or mobile. It is not necessarily larger than an industrial or surveillance camera. A capability in machine vision generally means a degree of development such that these capabilities are ready for use on individual applications. This architecture has the advantage of a more compact volume compared to PC-based vision systems and often achieves lower cost, at the expense of a somewhat simpler (or omitted) user interface. Smart cameras are also referred to by the more general term smart sensors.

<span class="mw-page-title-main">3D scanning</span> Scanning of an object or environment to collect data on its shape

3D scanning is the process of analyzing a real-world object or environment to collect three dimensional data of its shape and possibly its appearance. The collected data can then be used to construct digital 3D models.

The following are common definitions related to the machine vision field.

The following outline is provided as an overview of and topical guide to computer vision:

InspecVision Ltd. is a UK engineering company based in Mallusk, Northern Ireland, established in 2003. It is a manufacturing company that produces computer vision inspection systems. The company is one of several local companies created as spinoffs or inspired by research conducted at the Queen's University of Belfast.

Object recognition – technology in the field of computer vision for finding and identifying objects in an image or video sequence. Humans recognize a multitude of objects in images with little effort, despite the fact that the image of the objects may vary somewhat in different view points, in many different sizes and scales or even when they are translated or rotated. Objects can even be recognized when they are partially obstructed from view. This task is still a challenge for computer vision systems. Many approaches to the task have been implemented over multiple decades.

An area of computer vision is active vision, sometimes also called active computer vision. An active vision system is one that can manipulate the viewpoint of the camera(s) in order to investigate the environment and get better information from it.

<span class="mw-page-title-main">3D reconstruction</span> Process of capturing the shape and appearance of real objects

In computer vision and computer graphics, 3D reconstruction is the process of capturing the shape and appearance of real objects. This process can be accomplished either by active or passive methods. If the model is allowed to change its shape in time, this is referred to as non-rigid or spatio-temporal reconstruction.

A structured-light 3D scanner is a 3D scanning device for measuring the three-dimensional shape of an object using projected light patterns and a camera system.

Visual servoing, also known as vision-based robot control and abbreviated VS, is a technique which uses feedback information extracted from a vision sensor to control the motion of a robot. One of the earliest papers that talks about visual servoing was from the SRI International Labs in 1979.

<span class="mw-page-title-main">Visual odometry</span> Determining the position and orientation of a robot by analyzing associated camera images

In robotics and computer vision, visual odometry is the process of determining the position and orientation of a robot by analyzing the associated camera images. It has been used in a wide variety of robotic applications, such as on the Mars Exploration Rovers.

William Ward Armstrong is a Canadian mathematician and computer scientist. He earned his Ph.D. from the University of British Columbia in 1966 and is most known as the originator Armstrong's axioms of dependency in a Relational database.

<span class="mw-page-title-main">3D stereo view</span> Enables viewing of objects through any stereo pattern

A 3D stereo view is the viewing of objects through any stereo pattern.

Visual computing is a generic term for all computer science disciplines dealing with images and 3D models, such as computer graphics, image processing, visualization, computer vision, virtual and augmented reality and video processing. Visual computing also includes aspects of pattern recognition, human computer interaction, machine learning and digital libraries. The core challenges are the acquisition, processing, analysis and rendering of visual information. Application areas include industrial quality control, medical image processing and visualization, surveying, robotics, multimedia systems, virtual heritage, special effects in movies and television, and computer games.

<span class="mw-page-title-main">Air-Cobot</span> French research and development project (2013–)

Air-Cobot (Aircraft Inspection enhanced by smaRt & Collaborative rOBOT) is a French research and development project of a wheeled collaborative mobile robot able to inspect aircraft during maintenance operations. This multi-partner project involves research laboratories and industry. Research around this prototype was developed in three domains: autonomous navigation, human-robot collaboration and nondestructive testing.

<span class="mw-page-title-main">Objective vision</span>

Objective Vision is a project mainly aimed at real-time computer vision and simulation vision of living creatures. it has three sections contain of an open-source library of programming functions for using inside the projects, Virtual laboratory for scholars to check the application of functions directly and by command-line code for external and instant access, and the research section consists of paperwork and libraries to expand the scientific prove of works.

<span class="mw-page-title-main">Subhasis Chaudhuri</span>

Subhasis Chaudhuri is an Indian electrical engineer and the director at the Indian Institute of Technology, Bombay. He is a former K. N. Bajaj Chair Professor of the Department of Electrical Engineering of IIT Bombay. He is known for his pioneering studies on computer vision and is an elected fellow of all the three major Indian science academies viz. the National Academy of Sciences, India, Indian Academy of Sciences, and Indian National Science Academy. He is also a fellow of Institute of Electrical and Electronics Engineers, and the Indian National Academy of Engineering. The Council of Scientific and Industrial Research, the apex agency of the Government of India for scientific research, awarded him the Shanti Swarup Bhatnagar Prize for Science and Technology, one of the highest Indian science awards, in 2004 for his contributions to Engineering Sciences.

References

  1. Steger, Carsten; Markus Ulrich; Christian Wiedemann (2018). Machine Vision Algorithms and Applications (2nd ed.). Weinheim: Wiley-VCH. p. 1. ISBN   978-3-527-41365-2 . Retrieved 2018-01-30.
  2. Beyerer, Jürgen; Puente León, Fernando & Frese, Christian (2016). Machine Vision - Automated Visual Inspection: Theory, Practice and Applications. Berlin: Springer. doi:10.1007/978-3-662-47794-6. ISBN   978-3-662-47793-9 . Retrieved 2016-10-11.
  3. 1 2 3 Graves, Mark & Bruce G. Batchelor (2003). Machine Vision for the Inspection of Natural Products. Springer. p. 5. ISBN   978-1-85233-525-0 . Retrieved 2010-11-02.
  4. 1 2 3 Holton, W. Conard (October 2010). "By Any Other Name". Vision Systems Design. 15 (10). ISSN   1089-3709 . Retrieved 2013-03-05.
  5. Owen-Hill, Alex (July 21, 2016). "Robot Vision vs Computer Vision: What's the Difference?". Robotics Tomorrow.
  6. 1 2 3 4 5 6 7 Turek, Fred D. (June 2011). "Machine Vision Fundamentals, How to Make Robots See". NASA Tech Briefs . 35 (6): 60–62. Retrieved 2011-11-29.
  7. 1 2 3 Cognex (2016). "Introduction to Machine Vision" (PDF). Assembly Magazine. Retrieved 9 February 2017.
  8. Lückenhaus, Maximilian (May 1, 2016). "Machine Vision in IIoT". Quality Magazine.
  9. Computer Vision Principles, algorithms, Applications, Learning 5th Edition by E.R. Davies Academic Press, Elselvier 2018 ISBN 978-0-12-809284-2
  10. 1 2 West, Perry A Roadmap For Building A Machine Vision System Pages 1-35
  11. Dechow, David (January 2009). "Integration: Making it Work". Vision & Sensors: 16–20. Archived from the original on 2020-03-14. Retrieved 2012-05-12.
  12. Hornberg, Alexander (2006). Handbook of Machine Vision. Wiley-VCH. p. 427. ISBN   978-3-527-40584-8 . Retrieved 2010-11-05.
  13. 1 2 Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. ISBN   3-540-66410-6.[ page needed ]
  14. 1 2 Hornberg, Alexander (2006). Handbook of Machine Vision. Wiley-VCH. p. 429. ISBN   978-3-527-40584-8 . Retrieved 2010-11-05.
  15. Belbachir, Ahmed Nabil, ed. (2009). Smart Cameras. Springer. ISBN   978-1-4419-0952-7.[ page needed ]
  16. Dechow, David (February 2013). "Explore the Fundamentals of Machine Vision: Part 1". Vision Systems Design. 18 (2): 14–15. Retrieved 2013-03-05.
  17. 1 2 Critical Considerations for Embedded Vision Design by Dave Rice and Amber Thousand Photonics Spectra magazine published by Laurin Publishing Co. July 2019 issue Pages 60-64
  18. Wilson, Andrew (May 31, 2011). "CoaXPress standard gets camera, frame grabber support". Vision Systems Design. Retrieved 2012-11-28.
  19. Wilson, Dave (November 12, 2012). "Cameras certified as compliant with CoaXPress standard". Vision Systems Design. Retrieved 2013-03-05.
  20. 1 2 Davies, E.R. (1996). Machine Vision - Theory Algorithms Practicalities (2nd ed.). Harcourt & Company. ISBN   978-0-12-206092-2.[ page needed ].
  21. 1 2 Dinev, Petko (March 2008). "Digital or Analog? Selecting the Right Camera for an Application Depends on What the Machine Vision System is Trying to Achieve". Vision & Sensors: 10–14. Archived from the original on 2020-03-14. Retrieved 2012-05-12.
  22. Wilson, Andrew (December 2011). "Product Focus - Looking to the Future of Vision". Vision Systems Design. 16 (12). Retrieved 2013-03-05.
  23. Wilson, Andrew (April 2011). "The Infrared Choice". Vision Systems Design. 16 (4): 20–23. Retrieved 2013-03-05.
  24. West, Perry High Speed, Real-Time Machine Vision CyberOptics, pages 1-38
  25. 1 2 Murray, Charles J (February 2012). "3D Machine Vison Comes into Focus". Design News . Archived from the original on 2012-06-05. Retrieved 2012-05-12.
  26. Davies, E.R. (2012). Computer and Machine Vision: Theory, Algorithms, Practicalities (4th ed.). Academic Press. pp. 410–411. ISBN   9780123869081 . Retrieved 2012-05-13.
  27. 1 2 3 3-D Imaging: A practical Overview for Machine Vision By Fred Turek & Kim Jackson Quality Magazine, March 2014 issue, Volume 53/Number 3 Pages 6-8
  28. http://research.microsoft.com/en-us/people/fengwu/depth-icip-12.pdf HYBRID STRUCTURED LIGHT FOR SCALABLE DEPTH SENSING Yueyi Zhang, Zhiwei Xiong, Feng Wu University of Science and Technology of China, Hefei, China Microsoft Research Asia, Beijing, China
  29. R.Morano, C.Ozturk, R.Conn, S.Dubin, S.Zietz, J.Nissano, "Structured light using pseudorandom codes", IEEE Transactions on Pattern Analysis and Machine Intelligence 20 (3)(1998)322–327
  30. Finding the optimal hardware for deep learining inference in machine vision by Mike Fussell Vision Systems Design magazine September 2019 issue pages 8-9
  31. Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 39. ISBN   3-540-66410-6.
  32. Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 96. ISBN   3-540-66410-6.
  33. Linda G. Shapiro and George C. Stockman (2001): “Computer Vision”, pp 279-325, New Jersey, Prentice-Hall, ISBN   0-13-030796-3
  34. Lauren Barghout. Visual Taxometric approach Image Segmentation using Fuzzy-Spatial Taxon Cut Yields Contextually Relevant Regions. Information Processing and Management of Uncertainty in Knowledge-Based Systems. CCIS Springer-Verlag. 2014
  35. Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 108. ISBN   3-540-66410-6.
  36. Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 95. ISBN   3-540-66410-6.
  37. 1 2 Turek, Fred D. (March 2007). "Introduction to Neural Net Machine Vision". Vision Systems Design. 12 (3). Retrieved 2013-03-05.
  38. Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 111. ISBN   3-540-66410-6.
  39. 1 2 Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 125. ISBN   3-540-66410-6.
  40. Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 132. ISBN   3-540-66410-6.
  41. Demant C.; Streicher-Abel B. & Waszkewitz P. (1999). Industrial Image Processing: Visual Quality Control in Manufacturing. Springer-Verlag. p. 191. ISBN   3-540-66410-6.
  42. Hornberg, Alexander (2006). Handbook of Machine Vision. Wiley-VCH. p. 709. ISBN   978-3-527-40584-8 . Retrieved 2010-11-05.
  43. 1 2 3 The Place for Deep Learning in Machine Vision Quality Magazine May 2022 issue, Volume 61, Number 5 Published by BNP Media II
  44. Hapgood, Fred (December 15, 2006 – January 1, 2007). "Factories of the Future". CIO. 20 (6): 46. ISSN   0894-9301 . Retrieved 2010-10-28.