Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3095713.3095746acmotherconferencesArticle/Chapter ViewAbstractPublication PagescbmiConference Proceedingsconference-collections
research-article

Shape Representations for Maya Codical Glyphs: Knowledge-driven or Deep?

Published: 19 June 2017 Publication History

Abstract

This paper investigates two-types of shape representations for individual Maya codical glyphs: traditional bag-of-words built on knowledge-driven local shape descriptors (HOOSC), and Convolutional Neural Networks (CNN) based representations, learned from data. For CNN representations, first, we evaluate the activations of typical CNNs that are pretrained on large-scale image datasets; second, we train a CNN from scratch with all the available individual segments. One of the main challenges while training CNNs is the limited amount of available data (and handling data imbalance issue). Here, we attempt to solve this imbalance issue by introducing class-weights into the loss computation during training. Another possibility is oversampling the minority class samples during batch selection. We show that deep representations outperform the other, but CNN training requires special care for small-scale unbalanced data, that is usually the case in the cultural heritage domain.

References

[1]
1880. Dresden Codex. http://digital.slub-dresden.de/werkansicht/dlf/2967/1/. (1880).
[2]
1883. Madrid Codex. http://www.famsi.org/mayawriting/codices/madrid.html. (1883).
[3]
1887. Paris Codex. http://gallica.bnf.fr/ark:/12148/btv1b8446947j. (1887).
[4]
Serge Belongie, Jitendra Malik, and Jan Puzicha. 2000. Shape context: A new descriptor for shape matching and object recognition. In Conference on Advances in Neural Information Processing Systems, Vol. 2. 3.
[5]
Gulcan Can, Jean-Marc Odobez, and Daniel Gatica-Perez. 2016. Evaluating Shape Representations for Maya Glyph Classification. ACM Journal on Computing and Cultural Heritage (JOCCH) 9, 3 (September 2016).
[6]
Gulcan Can, Jean-Marc Odobez, and Daniel Gatica-Perez. 2017. Maya Codical Glyph Segmentation: A Crowdsourcing Approach. Research Report. Idiap.
[7]
Ken Chatfield, Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. 2014. Return of the Devil in the Details: Delving Deep into Convolutional Nets. In BMVC.
[8]
Navneet Dalal and Bill Triggs. 2005. Histograms of oriented gradients for human detection. In Conference on Computer Vision and Pattern Recognition, Vol. 1. IEEE, 886--893.
[9]
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. In Computer Vision and Pattern Recognition. IEEE, 248--255.
[10]
Jeff Donahue, Yangqing Jia, Oriol Vinyals, Judy Hoffman, Ning Zhang, Eric Tzeng, and Trevor Darrell. DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition.
[11]
Mathias Eitz, James Hays, and Marc Alexa. 2012. How Do Humans Sketch Objects? ACM Trans. Graph. 31, 4, Article 44 (jul 2012), 10 pages.
[12]
Morris Franken and Jan C van Gemert. 2013. Automatic egyptian hieroglyph recognition by retrieving images as texts. In International Conference on Multimedia. ACM, 765--768.
[13]
Rui Hu, Gulcan Can, Carlos Pallan Gayol, Guido Krempel, Jakub Spotak, Gabrielle Vail, Stephane Marchand-Maillet, Jean-Marc Odobez, and Daniel Gatica-Perez. 2015. Multimedia Analysis and Access of Ancient Maya Epigraphy. Signal Processing Magazine 32, 4 (jul 2015), 75--84.
[14]
Sergey Ioffe and Christian Szegedy. 2015. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proc. of International Conference on Machine Learning. 448--456.
[15]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. In Advances in NIPS. 1097--1105.
[16]
Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep learning. Nature 521, 7553 (2015), 436--444.
[17]
Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. 1998. Gradient-based learning applied to document recognition. Proc. IEEE 86, 11 (1998), 2278--2324.
[18]
David G. Lowe. 2004. Distinctive Image Features from Scale-Invariant Keypoints. International Journal of Computer Vision 60, 2 (2004), 91--110.
[19]
Nobuyuki Otsu. 1975. A threshold selection method from gray-level histograms. Automatica 11, 285--296 (1975), 23--27.
[20]
Edgar Roman-Rangel, Gulcan Can, Stephane Marchand-Maillet, Rui Hu, Carlos Pallan Gayol, Guido Krempel, Jakub Spotak, Jean-Marc Odobez, and Daniel Gatica-Perez. 2016. Transferring Neural Representations for Low-dimensional Indexing of Maya Hieroglyphic Art. In ECCV Workshop on Computer Vision for Art Analysis.
[21]
Edgar Roman-Rangel, Jean-Marc Odobez, and Daniel Gatica-Perez. 2013. Evaluating Shape Descriptors for Detection of Maya Hieroglyphs. In Mexican Conference on Pattern Recognition.
[22]
Edgar Roman-Rangel, Carlos Pallan, Jean-Marc Odobez, and Daniel Gatica-Perez. 2011. Analyzing ancient maya glyph collections with contextual shape descriptors. IJCV 94, 1 (2011), 101--117.
[23]
Ali Sharif Razavian, Hossein Azizpour, Josephine Sullivan, and Stefan Carlsson. 2014. CNN Features Off-the-Shelf: An Astounding Baseline for Recognition. In CVPR Workshops.
[24]
Karen Simonyan and Andrew Zisserman. 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. CoRR abs/1409.1556 (2014).
[25]
K. Simonyan and A. Zisserman. 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. CoRR abs/1409.1556 (2014).
[26]
Josef Sivic and Andrew Zisserman. 2003. Video Google: A Text Retrieval Approach to Object Matching in Videos. In ICCV (ICCV '03). 1470--. http://dl.acm.org/citation.cfm?id=946247.946751
[27]
Jason Yosinski, Jeff Clune, Yoshua Bengio, and Hod Lipson. 2014. How transferable are features in deep neural networks?. In Advances in NIPS. 3320--3328.
[28]
Qian Yu, Yongxin Yang, Feng Liu, Yi-Zhe Song, Tao Xiang, and Timothy M. Hospedales. 2016. Sketch-a-Net: A Deep Neural Network that Beats Humans. IJCV (2016), 1--15.
[29]
Qian Yu, Yongxin Yang, Yi-Zhe Song, Tao Xiang, and Timothy Hospedales. 2015. Sketch-a-net that beats humans. In Proc. of BMVC. 7.1--7.12.

Cited By

View all
  • (2019)Improved Hieroglyph Representation for Image RetrievalJournal on Computing and Cultural Heritage 10.1145/328438812:2(1-15)Online publication date: 30-Apr-2019
  • (2018)How to Tell Ancient Signs Apart? Recognizing and Visualizing Maya Glyphs with CNNsJournal on Computing and Cultural Heritage 10.1145/323067011:4(1-25)Online publication date: 5-Dec-2018

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
CBMI '17: Proceedings of the 15th International Workshop on Content-Based Multimedia Indexing
June 2017
237 pages
ISBN:9781450353335
DOI:10.1145/3095713
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 19 June 2017

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Maya glyphs
  2. convolutional neural networks
  3. shape recognition

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

CBMI '17

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)1
  • Downloads (Last 6 weeks)0
Reflects downloads up to 01 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2019)Improved Hieroglyph Representation for Image RetrievalJournal on Computing and Cultural Heritage 10.1145/328438812:2(1-15)Online publication date: 30-Apr-2019
  • (2018)How to Tell Ancient Signs Apart? Recognizing and Visualizing Maya Glyphs with CNNsJournal on Computing and Cultural Heritage 10.1145/323067011:4(1-25)Online publication date: 5-Dec-2018

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media