Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3290420.3290463acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccipConference Proceedingsconference-collections
research-article

A beautiful image or not: a comparative study on classical machine learning and deep learning

Published: 02 November 2018 Publication History

Abstract

With the development of web services and Apps on the Internet, food images are emerging into our life. Consumers from yelp or the dianping service upload a lot of food pictures every day. The images usually express the users' feelings and are shared among the social network. There have been different researches on the images. However, there is few research on how to evaluate the food image is beautiful or not. Therefore, we came up with an idea to classify food pictures by their appearance, which is meaningful in multiple applications, especially picking beautiful pictures to help businesses attract customers. In order to realize this idea, we collected 1067 food images through web crawling and questionnaires. Each image has a unique label: beautiful or not beautiful. Machine learning methods are used in this paper to model the data. CNN models in deep learning: VGGNet, AlexNet, and ResNet can get good results, e.g., ResNet can achieve the accuracy of 95.83%. However, with a good feature engineering job, the classifiers, which are random forest and support vector machine can reach a better accuracy of 99.63%. The experimental results indicate feature engineering is a vital issue in the food image evaluation problem, which lacks of labeled data.

References

[1]
Binh T. Nguyen, Duc-Tien Dang-Nguyen, Dang Xuan Tien, Thai Van Phat, Cathal Gurrin. 2018. A Deep Learning based Food Recognition System for Lifelog Images. ICPRAM : 657--664
[2]
Pan, L., Pouyanfar, S., Chen, H., Qin, J., & Chen, S. C. 2017. DeepFood: Automatic Multi-Class Classification of Food Ingredients Using Deep Learning. IEEE, International Conference on Collaboration and Internet Computing (pp.181--189). IEEE Computer Society.
[3]
Lowe, D. G. 1999. Object Recognition from Local Scale-Invariant Features. IEEE International Conference on Computer Vision (Vol.2, pp. 1150). IEEE.
[4]
Herbert Bay, Tinne Tuytelaars, Luc J. Van Gool.2006. SURF: Speeded Up Robust Features. ECCV (1) : 404--417
[5]
Dalal, N., & Triggs, B. 2005. Histograms of oriented gradients for human detection. IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Vol.1, pp.886--893). IEEE.
[6]
Kao, Y., Huang, K., & Maybank, S. 2016. Hierarchical aesthetic quality assessment using deep convolutional neural networks. Signal Processing Image Communication, 47, 500--510.
[7]
Kong, S., Shen, X., Lin, Z., Mech, R., & Fowlkes, C. 2016. Photo aesthetics ranking network with attributes and content adaptation. 662--679.
[8]
Simonyan, K., & Zisserman, A. 2014. Very deep convolutional networks for large-scale image recognition. Computer Science.
[9]
Krizhevsky, A., Sutskever, I., & Hinton, G. E. 2012. ImageNet classification with deep convolutional neural networks. International Conference on Neural Information Processing Systems (Vol.60, pp.1097--1105). Curran Associates Inc.
[10]
He, K., Zhang, X., Ren, S., & Sun, J. 2016. Deep residual learning for image recognition. CVPR 2016: 770--778.
[11]
Rublee, E., Rabaud, V., Konolige, K., & Bradski, G. 2012. ORB: An efficient alternative to SIFT or SURF. International Conference on Computer Vision (Vol.58, pp.2564--2571). IEEE.
[12]
HARRIS. 1988. A Combined Corner and Edge Detector. Proc. of Fourth Alvey Vision Conference (Vol.53, pp.147--151).
[13]
J. Sivic and A. Zisserman. 2003. Video Google: A text retrieval approach to object matching in videos. in Proc. 9th IEEE Int. Conf Comput. Vis., 2003, pp. 1470--1477.

Index Terms

  1. A beautiful image or not: a comparative study on classical machine learning and deep learning

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICCIP '18: Proceedings of the 4th International Conference on Communication and Information Processing
    November 2018
    326 pages
    ISBN:9781450365345
    DOI:10.1145/3290420
    • Conference Chairs:
    • Jalel Ben-Othman,
    • Hui Yu,
    • Program Chairs:
    • Herwig Unger,
    • Masayuki Arai
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 02 November 2018

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. appearance
    2. classical machine learning
    3. deep learning
    4. food picture

    Qualifiers

    • Research-article

    Conference

    ICCIP 2018

    Acceptance Rates

    Overall Acceptance Rate 61 of 301 submissions, 20%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 105
      Total Downloads
    • Downloads (Last 12 months)4
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 10 Nov 2024

    Other Metrics

    Citations

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media