![]() | |
Company type | Product Line |
---|---|
Industry | Artificial intelligence, Camera, Depth perception, 3D reconstruction |
Founded | 2015 |
Headquarters | Santa Clara, California |
Products | Intel RealSense |
Parent | Intel |
Website | www |
![]() | |
Developer | Intel RealSense |
---|---|
Manufacturer | Intel |
Type | Cameras, Technology |
Intel RealSense Technology, formerly known as Intel Perceptual Computing, is a product range of depth and tracking technologies designed to give machines and devices depth perception capabilities. The technologies, owned by Intel are used in autonomous drones, robots, AR/VR, smart home devices amongst many others broad market products.
The RealSense products are made of Vision Processors, Depth and Tracking Modules, and Depth Cameras, supported by an open source, cross-platform SDK in an attempt to simplify supporting cameras for third party software developers, system integrators, ODMs and OEMs. [1]
Intel began producing hardware and software that utilized depth tracking, gestures, facial recognition, eye tracking, and other technologies under the branding Perceptual Computing in 2013. [2] [3] According to Intel, much of their research into the technologies is focused around "sensory inputs that make [computers] more human like". They initially hoped to begin including 3D cameras that could support their Perceptual Computing as opposed to traditional 2D cameras by late 2014. [4]
In 2013, Intel ran a competition among seven teams to create software highlighting the capabilities of its Perceptual Computing technology entitled "Intel Ultimate Coder Challenge: Going Perceptual". [3] [5]
In 2014, Intel rebranded their Perceptual Computing line of technology as Intel RealSense. [2] [6]
Intel RealSense Group supports multiple depth and tracking technologies including Coded Light Depth, Stereo Depth and Positional Tracking. [7]
To address the lack of applications built on the RealSense platform and to promote the platform among software developers, in 2014 Intel organized the "Intel RealSense App Challenge". The winners were awarded large sums of money. [8]
As of January 2018, new Intel RealSense D400 Product Family was launched with the Intel RealSense Vision Processor D4, Intel RealSense Depth Module D400 Series, and 2 ready to use depth cameras: Intel RealSense Depth Cameras D435 and D415.
The Intel RealSense Vision Processor D4 series are vision processors based on 28 nanometer (nm) process technology to compute real-time stereo depth data. They utilise a depth algorithm that enables more accurate and longer range depth perception than previously available. There are two products in this family: RealSense Vision processor D4 and RealSense Vision Processor D4m.[ citation needed ]
The Intel RealSense Depth Module D400 Series is designed for easy integration to bring 3D into devices and machines. Intel also released the D415 and D435 in 2018. Both cameras feature the RealSense Vision processor D4 and camera sensors. They are supported by the cross-platform and open source Intel RealSense SDK 2.0. The Intel D415 is designed for more precise measurements.[ citation needed ]
The Intel RealSense Depth Camera D435 is for capturing stereo images as well as depth (by using a Structured-light 3D scanner) for a variety of applications that need perceive the world in 3D. [9] [10] [11] [12] [13] [14] [15] [16] [17]
Previous generations of Intel RealSense depth cameras (F200, R200 and SR300) were implemented in multiple laptop and tablet computers by Asus, HP, Dell, Lenovo, and Acer. [18] Additionally, Razer and Creative offered consumer ready standalone webcams with the Intel RealSense camera built into the design.: [19] Razer Stargazer and the Creative BlasterX Senz3D. [20]
This is a stand-alone camera that can be attached to a desktop or laptop computer. [21] It is intended to be used for natural gesture-based interaction, face recognition, immersive, video conferencing and collaboration, gaming and learning and 3D scanning. [22] There was also version of this camera to be embedded into laptop computers. [18]
Snapshot is a camera system intended to be built into tablet computers and possibly smartphones. Its intended uses include taking photographs and performing after the fact refocusing, distance measurements, and applying depth photo filters. [23] The refocus feature differs from a plenoptic camera in that RealSense Snapshot takes pictures with large depth of field so that initially the whole picture is in focus and then in software it selectively blurs parts of the image depending on their distance. The Dell Venue 8 and 10 7000 Series Android tablets are equipped with this camera. [24]
Rear-mounted camera for Microsoft Surface or a similar tablet, like the HP Spectre X2. [25] This camera is intended for augmented reality applications, content creation, and object scanning. Its depth accuracy is on the order of millimeters and its range is up to 6.0 meters. The R200 is a stereo camera and is able to obtain accurate depth outdoors as well as indoors. [26]
In an early preview article in 2015, PC World's Mark Hachman concluded that RealSense is an enabling technology that will be largely defined by the software that will take advantage of its features. He noted that as of the time the article was written, the technology was new and there was no such software. [27]
Camera 3D uses Intel RealSense (Serie D400) and Microsoft Kinect sensors to create holographic memories, 3D models and Facebook 3D photos
Specifications: Intel RealSense Depth Camera D415, D435 and D455 [28] [29] [30] [31]
D415 | D435 | D455 | |
---|---|---|---|
Use Environment | Indoor/Outdoor | ||
Depth Technology | Active infrared (IR) stereo | ||
IR Projector & Left/Right Camera Type | Standard | Wide | |
Shutter Type | Rolling | Global | |
Image Sensor Module | OV2740 (OV02740-H34A-Z) [32] | OV9782 (OV09782-GA4A) [33] | |
Image Sensor Technology | PureCel HDR [32] | OmniPixel3-GS [33] | |
Image Sensor Size | 1⁄6 inch, 3855 μm × 2919 μm [32] | 1⁄4 inch, 3896 μm × 2453 μm [33] | |
Image Sensor Pixel Size | 1.4 μm × 1.4 μm [32] | 3 μm × 3 μm [33] | |
Vision Processor Board | RealSense Vision Processor D4 | ||
Depth Sensor Module | RealSense Module D415 | RealSense Module D430 + RGB Camera | RealSense Module D450 |
Depth Field of View for HD | H:65°±2 V:40°±1 D:72°±2 | H:87°±3 V:58°±1 D:95°±3 | |
Depth Field of View for VGA | H:50°±2 V:40°±1 D:61°±2 | H:75°±3 V:62°±1 D:89°±3 | |
Depth Resolution and Framerate | Up to 1280px × 720px @ 90fps | ||
Minimum Depth Distance at Maximum Resolution | 45 cm / 17.7" | 28 cm / 11.0" | 52 cm / 20.5" |
Depth Accuracy | <2% at 2.0m / 2.2yd | <2% at 4.0m / 4.4yd | |
Ideal and Maximum Range | 0.5m to 3m / 0.6yd to 3.3yd | 0.3m to 3m / 0.3yd to 3.3yd | 0.6m to 6m / 0.3yd to 6.6yd |
RGB Resolution, Framerate and Aspect Ratio | 1920px × 1080px @ 30fps (16:9) | 1280px × 800px @ 30fps (8:5) | |
RGB Field of View | H:69.4° V:42.5° D:77.0° | H:91.2° V:65.5° D:100.6° | |
RGB Lens Distortion | ≤1.5% | ||
Device Dimensions | 99mm × 20mm × 23mm | 90mm × 25mm × 25mm | 124mm × 26mm × 29mm |
Connector | USB Type-C 3.1 Gen 1 | ||
Mounting Mechanisms | One 1/4-20UNC thread mounting point | ||
Two M3 thread mounting points | Two M4 thread mounting points |
Specifications: Intel RealSense Vision Processor D4 Series [34]
(Not available separately as these are just the bare PCB Vision Processor boards, only used as basis for the RealSense Depth Camera series)
RealSense Vision Processor D4 | RealSense Vision Processor D4M | |
Depth Technology | Stereo | Stereo |
Form Factor | ASIC BGA | ASIC BGA |
Package Size | 6.4mm x 6.4mm x 1mm | 4.7mm x 3.8mm x 0.55mm |
Process Technology | 28 nm | 28 nm |
Depth Max Throughput | 36.6 MP/sec (848×480@90fps) | |
Depth Stream Output Resolution | Up to 1280×720 | Up to 720×720 |
Depth Stream Output Frame Rate | Up to 90fps | |
RGB Sensor Max Resolution & Max Frame Rate | 1920×1080, Up to 60fps | 720×720, Up to 30 fps |
IR Projector Controls | Yes | Yes |
Host Interface | USB 3.0 | 2x MIPI |
Multi Camera Support | Yes, up to 5 | Up to 30fps |
I/O | 5x MIPI CSI-2, 5x I2C, 1x SPI, GPIO, Timer | 2x MIPI 1x I2C, 1x SPI, GPIO, Timer |
Specifications: Intel Stereo DepthModule SKUs [35]
(Not available separately as these are just the bare PCB Depth Sensor Modules, only used as basis for the RealSense Depth Camera series)
D400 | D410 | D415 | D420 | D430 | |
Depth technology | Passive IR Stereo | Active IR Stereo | Active IR Stereo | Passive IR Stereo | Active IR Stereo |
Image Sensor technology | Rolling Shutter | Rolling Shutter | Rolling Shutter | Global Shutter | Global Shutter |
Depth FOV (HxV for HD 16:9) | 63.4degx 40.4deg | 63.4degx 40.4deg | 63.4degx 40.4deg | 85.2degx 58deg | 85.2degx 58deg |
RGB Frame Rate and Resolution | - | - | Up to 60FPS | - | - |
Depth Resolution | Up to 1280x720 | Up to 1280x720 | Up to 1280x720 | Up to 1280x720 | Up to 1280x720 |
Depth Frame Rate | Up to 90fps | Up to 90fps | Up to 90fps | Up to 90fps | Up to 90fps |
Range | 0.16-10m+ | 0.16-10m+ | 0.16-10m+ | 0.11-10m+ | 0.11-10m+ |
Gesture recognition is an area of research and development in computer science and language technology concerned with the recognition and interpretation of human gestures. A subdiscipline of computer vision, it employs mathematical algorithms to interpret gestures.
A stereo camera is a type of camera with two or more lenses with a separate image sensor or film frame for each lens. This allows the camera to simulate human binocular vision, and therefore gives it the ability to capture three-dimensional images, a process known as stereo photography. Stereo cameras may be used for making stereoviews and 3D pictures for movies, or for range imaging. The distance between the lenses in a typical stereo camera is about the distance between one's eyes and is about 6.35 cm, though a longer base line produces more extreme 3-dimensionality.
The stereo cameras approach is a method of distilling a noisy video signal into a coherent data set that a computer can begin to process into actionable symbolic objects, or abstractions. Stereo cameras is one of many approaches used in the broader fields of computer vision and machine vision.
Microsoft PixelSense was an interactive surface computing platform that allowed one or more people to use and touch real-world objects, and share digital content at the same time. The PixelSense platform consists of software and hardware products that combine vision based multitouch PC hardware, 360-degree multiuser application design, and Windows software to create a natural user interface (NUI).
An image processor, also known as an image processing engine, image processing unit (IPU), or image signal processor (ISP), is a type of media processor or specialized digital signal processor (DSP) used for image processing, in digital cameras or other devices. Image processors often employ parallel computing even with SIMD or MIMD technologies to increase speed and efficiency. The digital image processing engine can perform a range of tasks. To increase the system integration on embedded devices, often it is a system on a chip with multi-core processor architecture.
2D-plus-Depth is a stereoscopic video coding format that is used for 3D displays, such as Philips WOWvx. Philips discontinued work on the WOWvx line in 2009, citing "current market developments". Currently, this Philips technology is used by SeeCubic company, led by former key 3D engineers and scientists of Philips. They offer autostereoscopic 3D displays which use the 2D-plus-Depth format for 3D video input.
A time-of-flight camera, also known as time-of-flight sensor, is a range imaging camera system for measuring distances between the camera and the subject for each point of the image based on time-of-flight, the round trip time of an artificial light signal, as provided by a laser or an LED. Laser-based time-of-flight cameras are part of a broader class of scannerless LIDAR, in which the entire scene is captured with each laser pulse, as opposed to point-by-point with a laser beam such as in scanning LIDAR systems. Time-of-flight camera products for civil applications began to emerge around 2000, as the semiconductor processes allowed the production of components fast enough for such devices. The systems cover ranges of a few centimeters up to several kilometers.
Sony Depthsensing Solutions SA/NV, formerly known as SoftKinetic Systems, is a Belgian company originating from the merger of Optrima NV, founded by André Miodezky, Maarten Kuijk, Daniël Van Nieuwenhove, Ward Van der Tempel, Riemer Grootjans and Tomas Van den Hauwe and SoftKinetic SA founded by Eric Krzeslo, Thibaud Remacle, Gilles Pinault and Xavier Baele. Sony Depthsensing Solutions develops gesture recognition hardware and software for real-time range imaging (3D) cameras. SoftKinetic was founded in July 2007 providing gesture recognition solutions based on its technology to the interactive digital entertainment, consumer electronics, health & fitness, and serious game industries. SoftKinetic technology has been applied to interactive digital signage and advergaming, interactive television, and physical therapy.
Kinect is a discontinued line of motion sensing input devices produced by Microsoft and first released in 2010. The devices generally contain RGB cameras, and infrared projectors and detectors that map depth through either structured light or time of flight calculations, which can in turn be used to perform real-time gesture recognition and body skeletal detection, among other capabilities. They also contain microphones that can be used for speech recognition and voice control.
Rockchip is a Chinese fabless semiconductor company based in Fuzhou, Fujian province. It has offices in Shanghai, Beijing, Shenzhen, Hangzhou and Hong Kong. It designs system on a chip (SoC) products, using the ARM architecture licensed from ARM Holdings for the majority of its projects.
PrimeSense was an Israeli 3D sensing company based in Tel Aviv. PrimeSense had offices in Israel, North America, Japan, Singapore, Korea, China and Taiwan. PrimeSense was bought by Apple Inc. for $360 million on November 24, 2013.
The Lenovo IdeaCentre is a line of consumer-oriented desktop computers designed, developed and marketed by Lenovo. The first IdeaCentre desktop, the IdeaCentre K210, was announced by Lenovo on June 30, 2008 as a consumer-focussed alternative to the business-oriented ThinkCentre and ThinkStation brandnames acquired from IBM. While the IdeaCentre line consists entirely of desktops, they share a common design language with the IdeaPad line of laptops and hybrids. One such feature is Veriface facial recognition technology.
Omek Interactive was a venture-backed technology company developing advanced motion-sensing software for human-computer interaction in automobiles, casinos and arcades, consumer electronics, video games, health care and digital signage. Omek was co-founded in 2006 by Janine Kutliroff and Gershom Kutliroff.
Imageon was a series of media coprocessors and mobile chipsets produced by ATI in 2002–2008, providing graphics acceleration and other multimedia features for handheld devices such as mobile phones and Personal Digital Assistants (PDAs). AMD later sold the Imageon mobile handheld graphics division to Qualcomm in 2009, where it was used exclusively inside their Snapdragon SoC processors under the Adreno brand name.
Tango was an augmented reality computing platform, developed and authored by the Advanced Technology and Projects (ATAP), a skunkworks division of Google. It used computer vision to enable mobile devices, such as smartphones and tablets, to detect their position relative to the world around them without using GPS or other external signals. This allowed application developers to create user experiences that include indoor navigation, 3D mapping, physical space measurement, environmental recognition, augmented reality, and windows into a virtual world.
Movidius Ltd. was a company based in San Mateo, California, that designed low-power processor chips for computer vision. The company was acquired by Intel in September 2016, who continues to sell the company's products under the Movidius line.
OpenXR is an open-source, royalty-free standard for access to virtual reality and augmented reality platforms and devices. It is developed by a working group managed by the Khronos Group consortium. OpenXR was announced by the Khronos Group on February 27, 2017, during GDC 2017. A provisional version of the standard was released on March 18, 2019, to enable developers and implementers to provide feedback on it. On July 29, 2019, OpenXR 1.0 was released to the public by Khronos Group at SIGGRAPH 2019 and on April 15, 2024, OpenXR 1.1 was released by Khronos.
The Sony Xperia XZ2 Premium is an Android smartphone manufactured and marketed by Sony. Part of the Xperia X series, the device was announced to the public on April 16, 2018, featuring a 4K HDR display and a MotionEye™ Dual Camera.
Gérard G. Medioni is a computer scientist, author, academic and inventor. He is a vice president and distinguished scientist at Amazon and serves as emeritus professor of Computer Science at the University of Southern California.
{{cite web}}
: CS1 maint: numeric names: authors list (link){{cite AV media}}
: CS1 maint: bot: original URL status unknown (link): Mail incoming, trying out and tearing down the Intel Euclid @scanlime-in-progress. YouTube .{{cite web}}
: CS1 maint: location (link){{cite web}}
: CS1 maint: location (link)