Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3323873.3325023acmconferencesArticle/Chapter ViewAbstractPublication PagesicmrConference Proceedingsconference-collections
short-paper

Collaborating CNN and SVM for Automatic Image Annotation

Published:05 June 2019Publication History

ABSTRACT

To learn a well-performed image annotation model, a large number of labeled samples are usually required. In this paper, we propose a novel semi-supervised approach based on adaptive weighted fusion for automatic image annotation, which can utilize the labeled data and unlabeled data simultaneously. Firstly, two different classifiers, namely the CNN (convolutional neural network) and the LDA-SVM, are constructed by all the labeled data. These two classifiers are independently represented as different feature views. Then, the most confident data with relevant pseudo-labels are chosen and amalgamated with the whole labeled dataset. After that, the two classifiers are retrained with the new labeled dataset until a stop condition is reached. In each iteration process, the unlabeled samples are labeled by high confidence pseudo-labels that are estimated by an adaptive weighted fusion strategy. Finally, we conduct experiments on two datasets, namely IAPR TC12 and NUS-WIDE, and measure the performance of the model with standard criteria, including precision, recall, F-measure, N+ and mAP. The experimental results show that our approach outperforms many state-of-the-art automatic image annotation approaches.

References

  1. D. M. Blei and M. I. Jordan. 2003. Modeling annotated data. In Proceedings of the 26th annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR'03). ACM, pp.127--134. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. A. Blum and T. Mitchell. 1998. Combining labeled and unlabeled data with co-training. In Proceedings of the 11th annual Conference on Computational Learning Theory (COLT'98). ACM, pp. 92--100. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. G. Carneiro, A. B. Chan, P. J. Moreno and N. Vasconcelos. 2007. Supervised learning of semantic classes for image annotation and retrieval. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(3), 394--410.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. X. Chen, Y. Mu, S. Yan and T.S. Chua. 2010. Efficient large-scale image annotation by probabilistic collaborative multi-label propagation. In Proceedings of the 18th ACM International Conference on Multimedia (MM'10). ACM, pp. 35--44. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. T. S. Chua, J. Tang, R. Hong, H. Li, Z. Luo and Y. Zheng. 2009. NUS-WIDE: A real-world web image database from National University of Singapore. In Proceedings of ACM International Conference on Image and Video Retrieval (CIVR'09). ACM, article 48. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. S. L. Feng, R. Manmatha and V. Lavrenko. 2004. Multiple Bernoulli relevance models for image and video annotation. In Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'04). IEEE, pp. 1002--1009.Google ScholarGoogle Scholar
  7. K. S. Goh, E. Y. Chang and B. Li. 2005. Using one-class and two-class SVMs for multiclass image annotation. IEEE Transactions on Knowledge and Data Engineering, 17(10), 1333--1346. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Y. Gong, Y. Jia, T. K. Leung, A. Toshev and S. Loffe. 2014. Deep convolutional ranking for multilabel image annotation. arXiv preprint arXiv:1312.4894.Google ScholarGoogle Scholar
  9. M. Guillaumin, T. Mensink, J. Verbeek and C. Schmid. 2009. Tagprop: Discriminative metric learning in nearest neighbor models for image auto-annotation. In Proceedings of IEEE International Conference on Computer Vision (ICCV'09). IEEE, pp. 309--316.Google ScholarGoogle Scholar
  10. J. Jeon, V. Lavrenko and R. Manmatha. 2003. Automatic image annotation and retrieval using cross-media relevance models. In: Proceedings of the 26th annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR'03). ACM, pp. 119--126. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. X. Jing, F. Wu, Z. Li, R. Hu and D. Zhang. 2016. Multi-label dictionary learning for image annotation. IEEE Transactions on Image Processing, 25(6), 2712--2725. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. J. Johnson, L. Ballan and Fei-Fei Li. 2015. Love thy neighbors: Image annotation by exploiting image metadata. In Proceedings of IEEE International Conference on Computer Vision (ICCV'15). IEEE, pp. 4624--4632. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. X. Ke, M. Zhou, Y Niu and W. Guo. 2017. Data equilibrium based automatic image annotation by fusing deep model and semantic propagation. Pattern Recognition, 71, 60--77.Google ScholarGoogle ScholarCross RefCross Ref
  14. A. Krizhevsky, I. Sutskever and G. Hinton. 2012. ImageNet classification using deep convolutional neural networks. In Advances in Neural Information Processing Systems (NIPS'12). MIT, pp. 1097--1105. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. V. Lavrenko, R. Manmatha and J. Jeon. 2004. A model for learning the semantics of pictures. In Advances in Neural Information Processing Systems (NIPS'04). MIT, pp.553--560.Google ScholarGoogle Scholar
  16. Z. Li, J. Liu, X. Zhu, T. Liu and H. Lu. 2010. Image annotation using multi-correlation probabilistic matrix factorization. In Proceedings of the 18th ACM International Conference on Multimedia (MM'10). ACM, pp. 1187--1190. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Z. Li, J. Liu, C. Xu and H. Lu. 2013. MLRank: Multi-correlation learning to rank for image annotation. Pattern Recognition, 46(10): 2700--2710. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Z. Li, Z. Shi, X. Liu and Z. Shi. 2011. Modeling continuous visual features for semantic image annotation and retrieval. Pattern Recognition Letters, 32(3): 516--523. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Z. Li, Z. Shi, W. Zhao, Z. Li and Z. Tang. 2013. Learning semantic concepts from image database with hybrid generative/discriminative approach. Engineering Applications of Artificial Intellegence, 26(9): 2143--2152. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. A. Makadia, V. Pavlovic and S. Kumar. 2010. Baselines for image annotation. International Journal of Computer Vision, 90(1): 88--105. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. F. Monay and D. Gatica-Perez. 2007. Modeling semantic aspects for cross-media image indexing. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(10): 1802--1817. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. V. N. Murthy, S. Maji and R. Manmatha. 2015. Automatic image annotation using deep learning representations. In Proceedings of ACM International Conference on Multimedia Retrieval (ICMR'15). ACM, pp. 603--606. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. X. Qi and Y. Han. 2007. Incorporating multiple SVMs for automatic image annotation. Pattern Recognition, 40(2): 728--741. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. K. Simonyan and A. Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv: 1409.1556.Google ScholarGoogle Scholar
  25. A. Tariq and H. Foroosh. 2018. Designing a symmetric classifier for image annotation using multi-layer sparse coding. Image and Vision Computing, 69: 33--43.Google ScholarGoogle ScholarCross RefCross Ref
  26. T. Uricchio, L. Ballan, L. Seidenari and A. Del Bimbo. 2017. Automatic image annotation via label transfer in the semantic space. Pattern Recognition, 71: 144--157.Google ScholarGoogle ScholarCross RefCross Ref
  27. Y. Verma and C. V. Jawahar. 2012. Image annotation using metric learning in semantic neighbourhoods. In Proceedings of European Conference on Computer Vision (ECCV'12). Springer, pp. 836--849.Google ScholarGoogle Scholar
  28. S. Zhang, J. Huang, Y. Huang, Y. Yu, H. Li and D. N. Metaxas. 2010. Automatic image annotation using group sparsity. In Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'10). IEEE, pp.3312--3319.Google ScholarGoogle Scholar
  29. D. Zhang, M. M. Islam and G. Lu. 2012. A review on automatic image annotation techniques. Pattern Recognition, 45(1): 346--362. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. X. Zhu. 2007. Semi-supervised learning literature survey. Technical Report, University of Wisconsin-Madison.Google ScholarGoogle Scholar

Index Terms

  1. Collaborating CNN and SVM for Automatic Image Annotation

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      ICMR '19: Proceedings of the 2019 on International Conference on Multimedia Retrieval
      June 2019
      427 pages
      ISBN:9781450367653
      DOI:10.1145/3323873

      Copyright © 2019 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 5 June 2019

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • short-paper

      Acceptance Rates

      Overall Acceptance Rate254of830submissions,31%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader