• Jan 04, 2024 News!IJFCC will adopt Article-by-Article Work Flow
  • Jun 03, 2024 News!Vol.13, No.2 has been published with online version.   [Click]
  • Dec 05, 2023 News!Vol.12, No.4 has been published with online version.   [Click]
General Information
    • ISSN: 2010-3751 (Print)
    • Frequency: Quarterly
    • DOI: 10.18178/IJFCC
    • Editor-in-Chief: Prof. Pascal Lorenz
    • Executive Editor: Ms. Yoyo Y. Zhou
    • Abstracting/ Indexing: Crossref, Electronic Journals LibraryINSPEC(IET), Google Scholar, EBSCO, etc.
    • E-mail:  ijfcc@ejournal.net 
    • Article Processing Charge: 500 USD
Editor-in-chief

Prof. Pascal Lorenz
University of Haute Alsace, France
 
It is my honor to be the Editor-in-Chief of IJFCC. The journal publishes good papers in the field of future computer and communication. Hopefully, IJFCC will become a recognized journal among the readers in the filed of future computer and communication.

IJFCC 2024 Vol.13(3): 55-59
DOI: 10.18178/ijfcc.2024.13.3.618

Image Caption Generator Using DenseNet201 and ResNet50

Vidhi Khubchandani
Computer Science and Engineering –Artificial Intelligence, Indira Gandhi Delhi Technical University for Women, Delhi, India
Email: vidhi048btcsai20@igdtuw.ac.in (V.K.)

Manuscript received June 12, 2024; revised July 28, 2024; accepted August 12, 2024; published September 14, 2024

Abstract—Image Caption generation is an important research area in computer vision and natural language processing. This paper compares two popular Convolutional Neural Network (CNN) architectures, DenseNet201 and ResNet50, for feature extraction in the title generation task. The study aims to analyze the impact of these architectures on the quality of generated subtitles by measuring their learning curves and Bilingual Evaluation Understudy (BLEU) scores. The study shows that the choice of CNN architecture significantly affects the performance of the captioning model. Densenet201 and Resnet50 have different learning models and BLEU scores, indicating that the former is more effective at capturing high-level features, while the latter is more suitable for capturing local features. This study’s results will help develop more accurate and efficient subtitling models.

Keywords—image caption generator, Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM), DenseNet201, ResNet50

[PDF]

Cite: Vidhi Khubchandani, "Image Caption Generator Using DenseNet201 and ResNet50," International Journal of Future Computer and Communication, vol. 13, no. 3, pp. 55-59, 2024.


Copyright © 2024 by the authors. This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited (CC BY 4.0)

Copyright © 2008-2024. International Journal of Future Computer and Communication. All rights reserved.
E-mail: ijfcc@ejournal.net