3 things you need to know
computer vision is a set of techniques for extracting information from images, videos, or point clouds. computer vision includes image recognition, object detection, activity recognition, 3d pose estimation, video tracking, and motion estimation. real-world applications include face recognition for logging into smartphones, pedestrian and vehicle avoidance in self-driving vehicles, and tumor detection in medical mris. software tools such as matlab® and simulink® are used to develop computer vision techniques.
how it works
most computer vision techniques are developed using an extensive set of real-world data and a workflow of data exploration, model training, and algorithm development. computer vision engineers often modify an existing set of techniques to fit the specific problem of interest. the main types of approaches used in computer vision systems are described below.
deep learning-based techniques
deep learning approaches to computer vision are useful for object detection, object recognition, image deblurring, and scene segmentation. deep learning approaches involve training convoluted neural networks (cnns), which learn directly from data using patterns at different scales. cnn training requires a large set of labeled training images or point clouds. uses pretrained networks to accelerate this process with less training data.
feature-based techniques
feature detection and extraction techniques are computer vision algorithms that identify patterns or structures in images and point clouds for image alignment, video stabilization, object detection, and more. in images, useful feature types include edges, corners, or regions with uniform density, and you can identify these features with detectors such as brisk, surf, or orb. in point clouds, you can use eigenvalue-based extractors.
image processing
image processing techniques are often applied as a preprocessing step in the computer vision workflow. the type of preprocessing depends on the task. relevant image processing techniques include:
point cloud processing
point clouds are a set of data points in 3d space that together represent a 3d shape or object. point cloud processing is typically done to preprocess the data in preparation for the computer vision algorithms that analyze them. point cloud processing typically involves:
3d vision processing
3d vision processing techniques estimate the 3d structure of a scene using multiple images taken with a calibrated camera. these images are typically generated from a monocular camera or stereo camera pair. 3d vision processing techniques include:
why it matters
computer vision is essential in a wide range of real-world applications. some of the most common of these are discussed below.
autonomous systems
aerial or ground autonomous systems use various sensors that collect visual or point cloud data from their environments. the systems use these data with computer vision capabilities, such as simultaneous localization and mapping (slam) and tracking, to map the environment. autonomous systems can use these maps to segment roads, footpaths, or buildings and detect and track humans and vehicles. for example, uses computer vision capabilities in the assisted driving view (adv) to depict surrounding vehicles and identify their types.
industrial applications
computer vision is used in manufacturing applications such as part quality monitoring and infrastructure maintenance. for example, used trained regional convolutional neural networks (r-cnns) to identify tags on machinery. also used the same capabilities for quality control, identifying inferior or defective parts during manufacturing.
construction and agriculture
computer vision is used in construction and agriculture to extract information from aerially captured infrastructure or terrain data. computer vision capabilities such as spectral signature mapping, object detection, and segmentation are applied to analyze images, point clouds, or hyperspectral data from aerial platforms. in japan uses these capabilities to detect damage to dams and bridges with semantic segmentation. farmers also analyze crop health using drones that acquire hyperspectral images of their farms.
photography
the use of computer vision in cameras and smartphones has grown heavily over the last decade. these devices use to focus on faces and to create panoramas. the devices also integrate or to access stored information.
computer vision with matlab
image processing toolbox™, computer vision toolbox™, and lidar toolbox™ in matlab provide apps, algorithms, and trained networks that you can use to build your computer vision capabilities. you can import image or point cloud data, preprocess it, and use built-in algorithms and deep learning networks to analyze the data. the toolboxes provide you with examples to get started.
defect detection using matlab
you can use computer vision toolbox to detect anomalies and defects in objects such as machine parts, electronics circuits, or others. you can increase the chances of detecting the right features by starting with image preprocessing algorithms in image processing toolbox, using capabilities such as , , and .
the defect detection step is often achieved using deep learning. to provide training data for deep learning, you can use the matlab image, video, or lidar labeler apps, which help you label data by creating semantic segmentation or instance segmentation masks. you can then train a deep learning network, either from scratch or by using . you can then use the trained network or one of several pretrained networks to classify the objects based on anomalies or defects.
object detection and tracking using matlab
object detection and tracking is one of the more well-known uses of computer vision for applications such as detecting or , , and . you can use the deep network designer to build deep learning networks in matlab for applications such as . you load labeled training data, , , and evaluate its and against ground truth data. you can then use the network to detect cars and around them.
using computer vision and simulink in autonomous system simulation
you can use object detection and tracking results from computer vision, in a robotic or autonomous system to make decisions. the autonomous emergency braking (aeb) with sensor fusion example demonstrates the ease of building simulink models that integrate computer vision capabilities. the model uses two parts: a computer vision and sensor fusion model to detect obstacles in front of a vehicle, and a forward collision warning (fcw) system to warn the driver and automatically apply the brake. this shows how you can use simulink to integrate computer vision algorithms into a broader system simulation.
localization and mapping with computer vision toolbox
you can use computer vision in matlab to estimate camera positions and map the environment using , to create 3d models of the objects using , and to .
you can estimate the position of a while mapping the environment using built-in capabilities in matlab such as and the . you initialize the map by between the pair of images, and then and feature positions on the map, using to refine the position and orientation of the camera as it moves through the scene.
object counting
you can also use computer vision to count objects in an image or video. in the , you apply to segment the cells, find cell centers using , and count the number of centers found. you then repeat this process for every frame in the video.
apps in matlab, such as the and , provide an interactive user interface to segment objects in an image. helps count objects in an image and calculate their properties, such as area, center of mass, and others.