Please use this identifier to cite or link to this item: https://dspace.ctu.edu.vn/jspui/handle/123456789/119563
Full metadata record
DC FieldValueLanguage
dc.contributor.authorNguyen, Van Thinh-
dc.contributor.authorTran, Van Lang-
dc.contributor.authorVan, The Thanh-
dc.date.accessioned2025-07-31T02:05:09Z-
dc.date.available2025-07-31T02:05:09Z-
dc.date.issued2024-
dc.identifier.issn1813-9663-
dc.identifier.urihttps://dspace.ctu.edu.vn/jspui/handle/123456789/119563-
dc.description.abstractRecent image captioning works often focus on global features or individual object regions within the image without exploiting the relational information between them, resulting in limited accuracy. In this paper, the proposed image captioning model leverages the relationships between objects in the image to fully understand the content and improve accuracy. The approach goes through the following steps: First, objects in the image are detected using an object detection model combined with a graph convolutional network (GCN). From this, a relationship prediction model based on relational context information and knowledge is proposed to classify relationships between objects to create a relationship graph to represent the image. Subsequently, a dual attention mechanism is built to enable the model to focus on relevant parts of both object regions and vertices in the relationship graph when generating captions. Finally, an LSTM network with dual attention is trained to generate captions relying on the image representation and given captions. Experiments conducted on MS COCO and Visual Genome datasets demonstrate that the proposed model achieves higher accuracy compared to baseline methods and some recently published works. Bộ sưu tập: Journal of Computer Science and Cybernetics.vi_VN
dc.language.isoenvi_VN
dc.relation.ispartofseriesJournal of Computer Science and Cybernetics;Vol.40, No.04 .- P.327-346-
dc.subjectImage captioningvi_VN
dc.subjectObject detectionvi_VN
dc.subjectVisual relationshipvi_VN
dc.subjectAttention mechanismvi_VN
dc.subjectDeep neural networkvi_VN
dc.titleOD-VR-Cap: Image captioning based on detecting and predicting relationships between objectsvi_VN
dc.typeArticlevi_VN
Appears in Collections:Tin học và Điều khiển học (Journal of Computer Science and Cybernetics)

Files in This Item:
File Description SizeFormat 
_file_
  Restricted Access
953.32 kBAdobe PDF
Your IP: 216.73.216.121


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.