Seven Computer Vision Algorithms: Quick Overview and Applications
Table of Contents
The number of computer vision algorithms is constantly increasing because of two reasons. The first reason is the specifics of computer vision as an interdisciplinary science. Usually, computer vision scholars have profound expertise in mathematics, statistics, and AI. That is why it is a small wonder that new techniques for identifying imagery are developed constantly by combining the findings in the related fields.
The second reason that contributes to the increasing number of computer vision algorithms is the rapid evolution of computing devices as well as cameras. Unlike in the early stages of development, modern computer vision can rely on unprecedented computing capacity to face fresh challenges.
In other words, new computer vision algorithms are developed every day. So, to help you keep up with the latest tendencies in computer vision, we have decided to come up with a list of both long-established and fresh computer vision algorithms. Keep on reading to learn more about the following seven computer vision algorithms (in alphabetical order) in this post:
Spoiler Alert: If you are running short of time, feel free to skip the main body of the post and go to the conclusions. This is where you find a table introducing a quick overview of all the analyzed algorithms in terms of their founders, objects of analysis, and more.
Computer Vision Algorithms with High Practical Value
Efficient Region Tracking (1998)
This algorithm was first described in Efficient Region Tracking With Parametric Models of Geometry and Illumination. As the efficient region tracking method belongs to computer vision algorithms that apply findings from mathematics, and namely geometry, it is a telling example of most modern works on computer vision being interdisciplinary.
ESM is believed to be better adapted to complex real-time computer vision challenges in comparison to first-order methods. This is one of the computer vision algorithms that can be applied not only to imagery but also to videos. Developed by Selim Benhimane and Ezio Malis in 2004, the ESM method idea is straightforward and yet elegant. According to this algorithm, it is possible to connect the current image with the reference one by extracting the information about several images from one planar surface. First presented in Homography-based 2D Visual Tracking and Servoing in 2004, this algorithm is further used to word and calculate a new image-based control law that relies exclusively on visual data.
Face Recognition Using Eigenfaces by Matthew Turk and Alex Pentland has revolutionized the way faces are recognized by computing devices. Belonging to computer vision algorithms with high practical value, this method is widely applied to identify people based on their head parameters by police and banks. The Eigenface algorithm has also impacted the face recognition procedures on social media. Nor manual interference or visual guidance is required to recognize faces successfully. With more two-dimensional face data fed to the Internet search engines, the Eigenface algorithm has reached a high accuracy level.
First introduced back in the late 1990s, the EigenTracking algorithm was developed by Michael J. Black and Allan D. Jepson. In their work EigenTracking: Robust Matching and Tracking of Articulated Objects Using a View-Based Representation, the scholars presented a fresh approach to identifying hand gestures in motion. Based on the analysis of lengthy image sequences that capture a moving hand, the EigenTracking algorithm allows to track even objects that are shot from different view perspectives.
Introduced in Equivalence and Efficiency of Image Alignment Algorithms, the inverse compositional method has become a logical step in the hierarchy of additive vs. compositional computer vision algorithms. Developed by Simon Baker and Iain Matthews, this method is less computationally costly in comparison to the forward additive and forward compositional algorithms.
For more updated information on how to apply the inverse compositional image alignment algorithm to active appearance models (AAMs), consult Active Appearance Models Revisited by the above-mentioned authors.
Lucas-Kanade Computer Vision Algorithm (1981)
This algorithm was created by Bruce D. Lucas and Takeo Kanade in the early 1980s. It was presented in 1981 in An Iterative Image Registration Technique with an Application to Stereo Vision. The main goal of the Lucas-Kanade algorithm is to estimate the optical flow. In theory, this algorithm argues that the optical flow is originally constant if you look into the immediate neighborhood of pixels. In practice, it means that if you know what nearby pixels look like, you will be able to figure out the rest of the pixels within a defined circle. This algorithm works best either with images or videos where objects change their locations bit by bit.
The authors of this algorithm, Richard Szeliski, and Heung-Yeung Shum, came up with a revolutionary way to create the full view panoramic mosaics already in 2000. In their first work on the subject, namely Creating Full View Panoramic Image Mosaics and Environment Maps, scholars argue the possibility of designing perfect mosaics that do not demand any manual guidance. What is more, Szeliski and Shum claim that their algorithm is not compromised by the quality of cameras used to take photos and possible image distortions? On top of that, the Szeliski-Shum algorithm can be applied to design a detailed environmental map from the processed images.
All the computer vision algorithms discussed above differ in terms of the cost of computing as well as the time required. The accuracy levels can also differ depending on the type of data you need to process.
Table 1
Computer Vision Algorithms: Key Elements
#
Algorithm
Scholars involved
Year of Publication
Objects of Analysis
1
Efficient Region Tracking
Gregory D. Hager, Peter N. Belhumeur
1998
Images and Videos
2
Efficient Second-Order Minimization
Selim Benhimane, Ezio Malis
2004
Images and Videos
3
Eigenface
Matthew Turk, Alex Pentland
1991
Images (Face Recognition)
4
EigenTracking
Michael J. Black, Allan D. Jepson
1998
Video Tracking (Hand Gestures)
5
Inverse Compositional Algorithm
Simon Baker, Iain Matthews
2001
Images
6
Lucas-Kanade
Bruce D. Lucas, Takeo Kanade
1981
Images
7
Szeliski-Shum
Richard Szeliski, Heung-Yeung Shum
2000
Images and Environmental Maps
As you can see from the table below, some of the computer vision algorithms have more practical value if applied to 2D objects, others would work better with 3D models or videos. So, now it is your call to decide what computer vision algorithms to use for your current computer vision project!
This website uses cookies so that we can provide you with the best user experience possible. Cookie information is stored in your browser and performs functions such as recognising you when you return to our website and helping our team to understand which sections of the website you find most interesting and useful.
Strictly Necessary Cookies
Strictly Necessary Cookie should be enabled at all times so that we can save your preferences for cookie settings.
If you disable this cookie, we will not be able to save your preferences. This means that every time you visit this website you will need to enable or disable cookies again.
3rd Party Cookies
This website uses Google Analytics to collect anonymous information such as the number of visitors to the site, and the most popular pages.
Keeping this cookie enabled helps us to improve our website.
Please enable Strictly Necessary Cookies first so that we can save your preferences!
Additional Cookies
This website uses the following additional cookies:
(List the cookies that you are using on the website here.)
Please enable Strictly Necessary Cookies first so that we can save your preferences!