A fancy acronym that stands for the process of COlour Thresholding, Blob Edge Detection, Transformation and CIrcle Detection used for locating 3D objects on a plane. The next series of posts will explain the software algorithms used by ClubElek on 2011 to achieve computer vision. The problems we faced, the solutions we implemented and most importantly what we have learned by doing this project.
These posts are targeted to a wide audience with some background in maths and preferably some background in computer vision. Mostly because there are some maths and magic behind the algorithms used, but I’ll try to keep them as simple and clear as possible. Should you have questions or remarks do not hesitate to comment !
This software was designed to detect “pawns” and “figures” defined by the Eurobot 2011 rules and was demoed during “Industrie Lyon” from 5 april to 8 april 2011. Before continuing reading this post you should read the summary about the rules of the contest so you don’t get lost.
Besides the 5 previously mentioned steps, 2 other steps were necessary before attempting any computer recognition: terrain calibration and colour calibration. These 2 processes will be explained on separate posts as they are far more complex than the COTBLEDTCID itself.
After the pawns' positions had been detected by the means of COTBLEDTCID, they were sent wirelessly to the robot through a XBee connection.
- The software should be used in a real-time environment. The fastest the algorithm, the better.
- The software should be easy to use and fast to configure. (Teams have only 1 minute and 30 seconds before a match to completely set up the robot and its peripherals).
- A Fit-PC-2 disk-less and fan-less computer running a customized ubuntu version controlled through ssh.
- 3 identical Microsoft LifeCam Cinema webcams. Why 3 cameras should you ask. Well, during a match, there are two robots that constantly move around the table and obfuscate large parts of the terrain, with 3 cameras chances are we see most of the objects on the playing table at any time.
- Light intensity remains constant during the match and after calibration.
- The cameras do not move during the match.
Both assumptions resulted to be inaccurate but did not affect the result as the detection and pose estimation algorithm is fairly robust.
What the computer sees (note that the robot’s game field wasn’t entirely finished by the time):
What the computer understands: (compare the pawns' position in both images. You may use the red top corner or the black area at the bottom of the image as a reference)
It’s fairly accurate isn’t it ?
In the next section I will explain how the Colour Thresholding works and why we need it (COT for short).