DeepScale

Last updated
DeepScale, Inc.
Founded2015;9 years ago (2015)
Founder
DefunctOctober 1, 2019 (2019-10-01)
FateAcquired by Tesla, Inc.
Headquarters Mountain View, California, U.S
Key people
Forrest N. Iandola (CEO)
Website deepscale.ai

DeepScale, Inc. was an American technology company headquartered in Mountain View, California, that developed perceptual system technologies for automated vehicles. On October 1, 2019, the company was acquired by Tesla, Inc. [1]

Contents

History

DeepScale was co-founded in 2015 by Forrest Iandola and Kurt Keutzer. [2] [3] In 2018, DeepScale raised US$15 million in Series A funding. [4] In 2018, the firm announced strategic partnerships with automotive suppliers including Visteon and Hella Aglaia Mobile Vision GmbH. [5] [6] On October 1, 2019, the firm was purchased by Tesla, which works on autonomous vehicle technology. [1]

Technology

Prior to the founding of DeepScale, Forrest Iandola and Kurt Keutzer worked together at University of California, Berkeley, on making deep neural networks (DNNs) more efficient. [7] [8] In 2016, shortly after the founding of DeepScale, Iandola, Keutzer, and their collaborators released SqueezeNet, which is a small and energy-efficient DNN for computer vision. [9] [10] [11] By developing smaller DNNs, the firm has been able to run deep learning on scaled-down processing hardware such as smartphones and automotive-grade chips. [9] [12] [13] In 2018, the firm said that its engineering team had moved beyond SqueezeNet and that it had developed even faster and more accurate DNNs for use in commercial products. [14]

In recent years, neural architecture search (NAS) has begun to outperform humans at designing DNNs that produce high-accuracy results while running fast. [15] In 2019, DeepScale published a paper called SqueezeNAS, which used supernetwork-based NAS to design a family of fast and accurate DNNs for semantic segmentation of images. [16] The paper claimed that the SqueezeNAS neural networks outperform the speed-accuracy tradeoff curve of Google's MobileNetV3 family of neural network models. [17] While Google used thousands of GPU-days to search for the design of MobileNetV3, DeepScale used just tens of GPU-days to automatically design the DNNs presented in the SqueezeNAS paper. [18]

Product

The firm develops perceptual system software which uses deep neural networks to enable cars to interpret their environment. The software is designed for integration into an open platform, where a wide range of sensors and processors can be used. [5] The software is able to run on a variety of processors, ranging from NVIDIA GPUs to smaller ARM-based processing chips that are designed specifically for the automotive market. [5] [14]

In January 2019, the firm launched an automotive perception software product called "Carver" that uses deep neural networks to perform object detection, lane identification, and drivable area identification. To accomplish this, Carver uses three neural networks which run in parallel. While running in real-time, these three networks perform a total of 0.6 trillion operations per second ("tera-ops/sec"). [19] As a point of reference, each of the two redundant chips on the Tesla Full Self-Driving computer system board can perform 36 tera-ops/sec. [20] So 0.6 tera-ops/sec is only 2% of the capacity of each Tesla chip.

Acquisition by Tesla

On October 1, 2019, CNBC reported that Tesla had purchased DeepScale. [1] Fortune stated that "it's apparent that DeepScale's technology will be integrated into Tesla's Autopilot, the self-driving technology the company is currently working on." [13] Further, CNET reported that "DeepScale's approach to autonomy fits the bigger picture that [Tesla CEO Elon] Musk has promoted for a few years now. Rather than relying on LiDAR, Musk has consistently believed cameras, radar and ultrasonic sensors will make up a robust system without other hardware." [21]

Related Research Articles

<span class="mw-page-title-main">OpenCV</span> Computer vision library

OpenCV is a library of programming functions mainly for real-time computer vision. Originally developed by Intel, it was later supported by Willow Garage, then Itseez. The library is cross-platform and licensed as free and open-source software under Apache License 2. Starting in 2011, OpenCV features GPU acceleration for real-time operations.

<span class="mw-page-title-main">Cadence Design Systems</span> American multinational computational software company

Cadence Design Systems, Inc., headquartered in San Jose, California, is an American multinational computational software company, founded in 1988 by the merger of SDA Systems and ECAD, Inc. The company produces software, hardware, and silicon structures for designing integrated circuits, systems on chips (SoCs), and printed circuit boards.

No instruction set computing (NISC) is a computing architecture and compiler technology for designing highly efficient custom processors and hardware accelerators by allowing a compiler to have low-level control of hardware resources.

<span class="mw-page-title-main">Object detection</span> Computer technology related to computer vision and image processing

Object detection is a computer technology related to computer vision and image processing that deals with detecting instances of semantic objects of a certain class in digital images and videos. Well-researched domains of object detection include face detection and pedestrian detection. Object detection has applications in many areas of computer vision, including image retrieval and video surveillance.

<span class="mw-page-title-main">Deep learning</span> Branch of machine learning

Deep learning is the subset of machine learning methods based on artificial neural networks (ANNs) with representation learning. The adjective "deep" refers to the use of multiple layers in the network. Methods used can be either supervised, semi-supervised or unsupervised.

Convolutional neural network (CNN) is a regularized type of feed-forward neural network that learns feature engineering by itself via filters optimization. Vanishing gradients and exploding gradients, seen during backpropagation in earlier neural networks, are prevented by using regularized weights over fewer connections. For example, for each neuron in the fully-connected layer 10,000 weights would be required for processing an image sized 100 × 100 pixels. However, applying cascaded convolution kernels, only 25 neurons are required to process 5x5-sized tiles. Higher-layer features are extracted from wider context windows, compared to lower-layer features.

Google Brain was a deep learning artificial intelligence research team under the umbrella of Google AI, a research division at Google dedicated to artificial intelligence. Formed in 2011, Google Brain combined open-ended machine learning research with information systems and large-scale computing resources. The team has created tools such as TensorFlow, which allow for neural networks to be used by the public, with multiple internal AI research projects. The team aims to create research opportunities in machine learning and natural language processing. The team was merged into former Google sister company DeepMind to form Google DeepMind in April 2023.

Headquartered in Tel Aviv Cortica utilizes unsupervised learning methods to recognize and analyze digital images and video. The technology developed by the Cortica team is based on research of the function of the human brain.

An AI accelerator, deep learning processor, or neural processing unit is a class of specialized hardware accelerator or computer system designed to accelerate artificial intelligence and machine learning applications, including artificial neural networks and machine vision. Typical applications include algorithms for robotics, Internet of Things, and other data-intensive or sensor-driven tasks. They are often manycore designs and generally focus on low-precision arithmetic, novel dataflow architectures or in-memory computing capability. As of 2024, a typical AI integrated circuit chip contains tens of billions of MOSFET transistors.

Chainer is an open source deep learning framework written purely in Python on top of NumPy and CuPy Python libraries. The development is led by Japanese venture company Preferred Networks in partnership with IBM, Intel, Microsoft, and Nvidia.

Caffe is a deep learning framework, originally developed at University of California, Berkeley. It is open source, under a BSD license. It is written in C++, with a Python interface.

The CIFAR-10 dataset is a collection of images that are commonly used to train machine learning and computer vision algorithms. It is one of the most widely used datasets for machine learning research. The CIFAR-10 dataset contains 60,000 32x32 color images in 10 different classes. The 10 different classes represent airplanes, cars, birds, cats, deer, dogs, frogs, horses, ships, and trucks. There are 6,000 images of each class.

Neural architecture search (NAS) is a technique for automating the design of artificial neural networks (ANN), a widely used model in the field of machine learning. NAS has been used to design networks that are on par or outperform hand-designed architectures. Methods for NAS can be categorized according to the search space, search strategy and performance estimation strategy used:

In computer vision, SqueezeNet is the name of a deep neural network for image classification that was released in 2016. SqueezeNet was developed by researchers at DeepScale, University of California, Berkeley, and Stanford University. In designing SqueezeNet, the authors' goal was to create a smaller neural network with fewer parameters while achieving competitive accuracy.

Beijing Unisound Information Technology Co., Ltd., often shortened to Unisound, is a Chinese technology company based in Beijing. It is a unicorn startup specialising in speech recognition and artificial intelligence services applicable to a variety of industries.

Amazon SageMaker is a cloud based machine-learning platform that allows the creation, training, and deployment by developers of machine-learning (ML) models on the cloud. It can be used to deploy ML models on embedded systems and edge-devices. SageMaker was launched in November 2017.

Kurt Keutzer is an American computer scientist.

Forrest N. Iandola is an American computer scientist specializing in efficient AI.

Deep learning speech synthesis refers to the application of deep learning models to generate natural-sounding human speech from written text (text-to-speech) or spectrum (vocoder). Deep neural networks (DNN) are trained using a large amount of recorded speech and, in the case of a text-to-speech system, the associated labels and/or input text.

Meta AI is an artificial intelligence laboratory owned by Meta Platforms Inc. Meta AI develops various forms of artificial intelligence, developing augmented and artificial reality technologies. Meta AI is an academic research laboratory focused on generating knowledge for the AI community. This is in contrast to Facebook's Applied Machine Learning (AML) team, which focuses on practical applications of its products.

References

  1. 1 2 3 Kolodny, Lora (2019-10-01). "Tesla is buying computer vision start-up DeepScale in a quest to create truly driverless cars". CNBC. Retrieved 2019-10-02.
  2. Johnson, Khari (2020-06-23). "SqueezeBERT promises faster mobile NLP while maintaining BERT levels of accuracy". VentureBeat .
  3. "Are We Short of Deep Learning Experts?". EE Times . 2018-04-03.
  4. Marinova, Polina (2018-04-04). "Term Sheet". Fortune. Retrieved 2018-05-22.
  5. 1 2 3 Yoshida, Junko (2018-01-09). "Visteon Works with DNN Vanguard DeepScale". EE Times. Retrieved 2018-04-07.
  6. Yoshida, Junko (2018-04-03). "Are We Short of Deep Learning Experts?". EE Times. Retrieved 2018-04-07.
  7. Keutzer, Kurt. "Faculty Webpage". UC Berkeley. Retrieved 2018-05-22.
  8. Keutzer, Kurt. "Students". UC Berkeley. Retrieved 2018-05-22.
  9. 1 2 Yoshida, Junko (2017-09-21). "DeepScale on Robo-Car: Fuse Raw Data". EE Times. Retrieved 2018-05-22.
  10. Iandola, Forrest N; Han, Song; Moskewicz, Matthew W; Ashraf, Khalid; Dally, William J; Keutzer, Kurt (2016). "SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size". arXiv: 1602.07360 [cs.CV].
  11. Niedermeyer, Edward (2019-10-01). "Tesla Beefs Up Autonomy Effort With DeepScale Acqui-Hire". The Drive. Archived from the original on 2019-10-02. Retrieved 2019-11-10.
  12. Shazar, Jon (2018-04-05). "Steve Cohen Buys The Dip In Self-Driving Cars". Dealbreaker. Retrieved 2018-05-22.
  13. 1 2 Reisinger, Don (2019-10-02). "Why Tesla Quietly Acquired DeepScale, a Machine Learning Startup That's 'Squeezing' A.I." Fortune. Retrieved 2019-11-25.
  14. 1 2 "How to become a Full-Stack Deep Learning Engineer (time: 51:30)". Silicon Valley Deep Learning Group. Retrieved 2018-05-22.
  15. Zoph, Barret; Vasudevan, Vijay; Shlens, Jonathon; Le, Quoc V. (2017-07-21). "Learning Transferable Architectures for Scalable Image Recognition". arXiv: 1707.07012 [cs.CV].
  16. Shaw, Albert; Hunter, Daniel; Iandola, Forrest; Sidhu, Sammy (2019). "SqueezeNAS: Fast neural architecture search for faster semantic segmentation". arXiv: 1908.01748 [cs.CV].
  17. Howard, Andrew; Sandler, Mark; Chu, Grace; Chen, Liang-Chieh; Chen, Bo; Tan, Mingxing; Wang, Weijun; Zhu, Yukun; Pang, Ruoming; Vasudevan, Vijay; Le, Quoc V.; Adam, Hartwig (2019-05-06). "Searching for MobileNetV3". arXiv: 1905.02244 [cs.CV].
  18. Yoshida, Junko (2019-08-25). "Does Your AI Chip Have Its Own DNN?". EE Times. Retrieved 2019-09-26.
  19. Landen, Ben (2019-01-25). "DeepScale Announces Carver21: Modular Deep Learning Perception Software for Driver-Assistance". DeepScale Blog. Retrieved 2019-02-04.
  20. Hollister, Sean (2019-04-22). "Tesla's new self-driving chip is here, and this is your best look yet". The Verge. Retrieved 2020-08-24.
  21. Szymkowski, Sean (2019-10-02). "Tesla reportedly buys machine-learning startup DeepScale for self-driving car tech". CNET. Archived from the original on 2019-10-10. Retrieved 2019-11-10.