Traffic Sign Recognition For Computer Vision Project-Based Learning
Traffic Sign Recognition For Computer Vision Project-Based Learning
Traffic Sign Recognition For Computer Vision Project-Based Learning
I. INTRODUCTION
OMPUTER vision (CV) [1] is a subfield of artificial intelligence (AI) aimed at understanding still images and
video sequences; examples of this include recognizing people or
objects, navigating in an environment, reconstructing the threedimensional shape of a scene, or controlling a device (such as
when a robotic arm grasps an object). It is an interdisciplinary
field that draws from computer science, signal processing, and
a number of mathematical fields like geometry, statistics, and
algebra.
During the last two decades, CV has progressively been incorporated into both undergraduate and graduate programs of
computer science studies. In their excellent review on CV education, Bebis et al. [2] review various approaches to teaching
CV. Most of the literature is focused on different approaches
to integrating the topic into undergraduate courses. The most
common approach is to follow the traditional course structure
Manuscript received August 27, 2012; revised November 07, 2012; accepted
December 09, 2012. Date of publication February 01, 2013; date of current
version July 31, 2013. This work was supported by Dr. M. Vanrell, coordinator of the Masters Program of the Department of Computer Science, Universitat Autnoma de Barcelona, and Project Consolider-Ingenio 2010 CSD200700018. The work of D. Gernimo, J. Serrat, and A. M. Lpez was supported by
the Project TRA2011-29454-C03-01. The work of R. Baldrich was supported
by the Project TIN201021771-C02-1.
The authors are with the Computer Vision Center and the Department of Computer Science, Universitat Autnoma de Barcelona, 08193 Bellaterra, Spain
(e-mail: dgeronimo@cvc.uab.es).
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org.
Digital Object Identifier 10.1109/TE.2013.2239997
365
366
Fig. 1. Four project milestones of the traffic sign recognition prototype. Candidate generation of bounding boxes and the learned binary sign/background classifier
are combined in order to create a detection subsystem. A multiclass sign classifier takes the positive bounding boxes and assigns a traffic sign label to each one.
B. Weekly Work
During the course, 1 h per week is dedicated to monitoring
and giving guidance to the otherwise independent student teams.
In the first part of the session (4045 min), the results obtained
in the previous assignment are analyzed by the instructor. He
or she will provide a summary of the common problems, the
most interesting proposals, and a statistical performance of the
teams. At this point, there will also be short presentations by
the teams, submitted via e-mail beforehand, consisting of three
slides summarizing the techniques implemented, the results, and
the problems encountered. Two randomly chosen teams present
their work (it is not mandatory that all the team members speak).
This serves to facilitate communication among the groups and
allows the instructor to spend the majority of this part of the
session studying in greater depth both aspects specific to a particular team and aspects applicable to all. The instructor ensures
a balance between the interest value of the presentation for the
entire class and the frequency with which each team presents
their work. The rest of the groups also participate with comments and questions, not only on the presenting teams work,
but also with respect to their own systems. The submitted work
is not evaluated weekly, but at the milestone checkpoints (see
Section II-C).
In the second part of the session (1520 min), the homework
for the next week is explained. Usually, this is quite straightforward: The main ideas plus the evaluation criteria are explained,
and since the required theoretical concepts are still fresh from
the recent lectures, there are few questions. If any concept has
not been explained (e.g., sliding windows), it is described in detail in this part of the session. In addition, the instructor gives
special instructions on the algorithms that are particularly complicated to implement (e.g., integral image). Finally, it is worth
noting that during the semester, the instructor has one additional hour per week for answering questions from individual
groups or students, during which they can analyze their systems
in greater depth.
367
Fig. 2. Samples from seven of the 14 traffic sign types learned by the multiclass
classifier for sign recognition.
C. Dataset
The students are provided with a dataset of images both
to facilitate the testing of each milestone and to serve as a
common benchmark for all teams. In addition, they are freed
from creating a manually annotated ground-truth set for detection and recognition, which would be very time-consuming.
The dataset used for the project is based on the KUL traffic
sign database [24], [25] created by Dr. Radu Timofte from
the Katholieke Universiteit of Leuven, Leuven, Belgium. The
dataset was adapted to the specific needs of the project, selecting the number of images and those traffic sign types with
more samples. Three subsets are provided to the students:
Detection set: 669 images of 1628 1236 pixels, with annotated ground-truth (bounding boxes and type of traffic
signs present in the image) used in the candidates generation and detection/recognition systems milestones;
Classification set: cropped images to be used for the binary
classifier training and testing. In total, 3000 images in five
groups (circles, inverted triangles, rectangles, octagons and
triangles) plus a background subset;
Recognition set: This is divided into training and testing
and is used to build the multiclass classifier that recognizes
the traffic sign type. The set is limited to 14 types of signs
(Fig. 2), those with more samples. The size of this set is
1225 images.
The final recognition system, in addition to being evaluated on the detection set, is tested on a short sequence
(200 frames) from the KUL dataset and on a challenging sequence (185 frames) from the CVC-02 pedestrian dataset [26].
The latter was recorded with a different camera in a different
country (Spain) and thus contains slightly different signs and
imaging conditions.
D. Evaluation
The project is graded as follows: 20% for each of the four
milestone deliverables, 10% for two oral presentations, and 10%
for peer evaluation. This amounts to 55% of the total course
score, the remaining 45% being determined with a written exam.
Teams are evaluated at the four project milestones based on
their deliverables. These consist of a 10-page report plus the
source code of the corresponding project part. Each weekly
assignment is graded with an E (not submitted), D (failed),
C (meets mandatory requirements), B, or A (exceeds mandatory requirements).1 During the first weeks, it is the instructor
who suggests ways to go beyond the mandatory work. Subsequently, the students propose their own approaches, whose
complexity and originality will determine their grade. As an
1For inclusion in their total marks for the course, students project grades are
,
,
,
, and
.
converted to marks as:
368
Fig. 3. (top) Distribution of the students marks after taking into account deliverables, oral presentations, and intragroup evaluation. (bottom) Distribution
of the grades (see legend) of the deliverables at each milestone.
A. Marks
Fig. 3 (top) shows the distribution of the final student marks
according to the weighting specified in Section II-D. As can be
seen, the marks are high (a mean average of 8.42 and a standard deviation of 1.08 over 10), which means that in general
the teams fulfilled most of the requirements and proposed interesting improvements. Fig. 3 (bottom) illustrates the increase
in most teams grades during the four stages. This increase is
due to two factors. First, the students initially had to adapt to
the weekly milestones scheme, the working framework, and the
language (while almost everybody had medium or high programming skills, 44% of the students had little experience in
MATLAB). Second, they required the first evaluation to understand the type and extent of improvements they had to implement in order to obtain the highest marks.
B. Student Feedback
The feedback from the students was very good. They were
engaged by the weekly assignments and enjoyed the project in
spite of the amount of work it represented. An indication that the
course succeeded in motivating students is that several of them
changed their enrolment from Advanced AI to the Advanced CV
course for the second term of the year.
At the end of the course, the students were asked to fill
in an anonymous assessment form on the project. Based on
this form and on personal communication, some interesting
issues are worth highlighting. On the one hand, the students
valued the continuity of the project, its relevance to the theory
lectures (68%), and the appropriate level of difficulty (87%),
the latter being one of the main concerns during its design. On
the other hand, while the discussion sessions were very highly
valued with respect to their usefulness in developing the project
(93%), some students complained of insufficient time being
dedicated to the explanation of the weekly homework and the
subgoals during these sessions (37%). Keeping an appropriate
369
370
TABLE I
THEORY AND PROJECT CONTENTS OF THE COURSE. PERFEVALPIX AND PERFEVALOBJ ARE INTRODUCED IN WEEKS 2 AND 3 AND THEN USED THROUGHOUT
THE COURSE
ACKNOWLEDGMENT
The authors would like to thank Dr. R. Timofte for making
the KUL Belgium Traffic Signs and Classification Benchmark
datasets publicly available. These datasets were subsequently
adapted for teaching purposes.
REFERENCES
[1] R. Szeliski, Computer Vision: Algorithms and Applications. New
York, NY, USA: Springer, 2010.
[2] G. Bebis, D. Egbert, and M. Shah, Review of computer vision education, IEEE Trans. Educ., vol. 46, no. 1, pp. 221, Feb. 2003.
[3] P. Reimer, A. Albu, and G. Tzanetakis, Raydiance: A tangible interface for teaching computer vision, in Proc. 7th ISVC, Las Vegas, NV,
USA, 2011, vol. 2, pp. 259269.
[4] G. Cielniak, N. Bellotto, and T. Duckett, Integrating mobile robotics
and vision with undergraduate computer science, IEEE Trans. Educ.,
vol. 56, no. 1, pp. 4853, Feb. 2013.
[5] A. Hoover, Computer vision in undergraduate education: Modern embedded computing, IEEE Trans. Educ., vol. 46, no. 2, pp. 235240,
May 2003.
[6] C. Olson, Encouraging the development of undergraduate researchers
in computer vision, ACM SIGCSE Bull., vol. 44, no. 2, pp. 255259,
2006.
[7] R. Capraro and S. Slough, Project-Based Learning. Rotterdam, The
Netherlands: Sense, 2009.
[8] L. Fernandez-Samaca and J. Ramrez, An approach to applying
project-based learning in engineering courses, in Proc. IEEE ANDESCON, Botot, Colombia, 2010, pp. 16.
[9] A. Price, R. Rimington, M. T. Chew, and S. Demidenko, Projectbased learning in robotics and electronics in undergraduate engineering program setting, in Proc. IEEE Int. Symp. Electron. Design,
Test, Appl., Ho Chi Minh City, Vietnam, 2010, pp. 188193.
[10] R. Hsu, Project based learning as a pedagogical tool for embedded
system education, in Proc. 3rd Int. Conf. Inf. Technol., Res. Educ.,
Hsinchu, Taiwan, 2005, pp. 362366.
[11] J. Macias-Guarasa, A project-based learning approach to design
electronic systems curricula, IEEE Trans. Educ., vol. 49, no. 3, pp.
389397, Aug. 2006.
[12] J. Venkataraman, Project based electromagnetics education, in Proc.
Appl. Electron. Conf., 2009, pp. 14.
[13] A. Breiter, G. Fey, and R. Drechsler, Project-based learning in student
team in computer science education, Electron. Energetics, vol. 18, no.
2, pp. 165180, 2005.
[14] F. Martinez, L. Herrero, and S. de Pablo, Project-based learning and
rubrics in the teaching of power supplies and photovoltaic electricity,
IEEE Trans. Educ., vol. 54, no. 1, pp. 8796, Feb. 2011.
[15] J. Nelson, Using project-based assignments in graduate-level digital
signal processing course, in Proc. Digital Signal Process. Workshop
& 4th IEEE Signal Process. Educ. Workshop, 2006, pp. 135140.
[16] Universitat Autnoma de Barcelona, Bellaterra, Spain, Website of the
UAB Master on Computer Vision and Artificial Intelligence, 2012
[Online]. Available: http://mcvai.uab.cat
[17] J. Khan, S. Bhuiyan, and R. Adhami, Image segmentation and shape
analysis for road-sign detection, IEEE Trans. Intell. Transport. Syst.,
vol. 12, no. 1, pp. 8396, Mar. 2011.
David Gernimo received the B.Sc. degree in computer science and M.Sc. degree and Ph.D. degree (with honors) in computer vision from the Universitat
Autnoma de Barcelona (UAB), Cerdanyola, Spain, in 2004, 2006, and 2010,
respectively.
He is currently a Postdoctoral Researcher and Associate Lecturer with the
Computer Science Department, UAB, and the Computer Vision Center. His research interests include object detection, recognition and tracking, feature detection and description, and scene understanding.
371
Joan Serrat received the Ph.D. degree in computer science from the Universitat
Autnoma de Barcelona (UAB), Bellaterra, Spain, in 1990.
He is currently an Associate Professor with the Computer Science Department, UAB, and a member of the Computer Vision Center. At UAB, he is in
charge of two undergraduate courses on Software Engineering and Image Processing and co-coordinates the introductory module of the Master on Computer
Vision. He has coauthored more than 40 papers and three patents. His current
research interest is the application of probabilistic graphical models to computer
vision problems such as feature matching, tracking, and video alignment.
Dr. Serrat has been a Board Member of the Spanish Chapter of the International Association for Pattern Recognition.
Antonio M. Lpez (M10) received the B.Sc. degree in computer science from
the Universitat Politcnica de Catalunya (UPC), Barcelona, Spain, in 1992, and
the M.Sc. and Ph.D. degrees in image processing and artificial intelligence from
the Universitat Autnoma de Barcelona (UAB), Bellaterra, Spain, in 1994 and
2000, respectively.
Since 1992, he has been giving lectures with the Computer Science Department, UAB, where he currently is an Associate Professor. In 1996, he participated in the foundation of the Computer Vision Center at the UAB, where he has
held different institutional responsibilities, currently leading the research group
on advanced driver assistance systems (ADAS) by computer vision. He has led
public and private projects and is a coauthor of more than 100 papers, all in the
field of computer vision.
Ramon Baldrich received the Ph.D. degree in computer science from the Universitat Autnoma de Barcelona (UAB), Bellaterra, Spain, in 2001.
He is an Associate Professor with the Computer Science Department, UAB,
and is attached to the Computer Vision Center as a Researcher. His research
interests are mainly focused in color treatment in computer vision problems,
including color constancy, color induction, and image shadows.