Please use this identifier to cite or link to this item: https://dspace.ctu.edu.vn/jspui/handle/123456789/73754
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorLâm, Nhựt Khang-
dc.contributor.authorNguyễn, Duy Lân-
dc.date.accessioned2022-02-21T09:06:00Z-
dc.date.available2022-02-21T09:06:00Z-
dc.date.issued2021-
dc.identifier.otherB1609588-
dc.identifier.urihttps://dspace.ctu.edu.vn/jspui/handle/123456789/73754-
dc.description38 Trvi_VN
dc.description.abstractNowadays, with the strong development of information technology, people are turning to technology to capture information. Therefore, text summarization is becoming more and more essential. As a result of the evolution of machine learning, many Natural Language Processing (NLP) models were introduced. This thesis will apply BERT-Transformer – one of state at the art techniques, and a pre-trained model for implementing a new machine learning model which will receive long sentences as input and take it into shorter ones with the same meaning. To train the model, we used Open-NMT, and the raw material for training the model came from the website vnexpress.net and a few online Vietnamese periodicals. The ROUGE method will be used in this thesis to evaluate the summaries. The results of the model after being evaluated are considered good and acceptable compared to other text summarization models. The result of the thesis introduces an abstractive Vietnamese summarization model. The model enables users to generate a summary that retains the original meaning of the source text at the sentence level.vi_VN
dc.language.isoenvi_VN
dc.publisherTrường Đại Học Cần Thơvi_VN
dc.subjectCÔNG NGHỆ THÔNG TIN-CHẤT LƯỢNG CAOvi_VN
dc.titleVIETNAMESE ABSTRACTIVE SUMMARIZATION USING TRANSFORMER AND BERTvi_VN
dc.typeThesisvi_VN
Appears in Collections:Trường Công nghệ Thông tin & Truyền thông

Files in This Item:
File Description SizeFormat 
_file_
  Restricted Access
716.35 kBAdobe PDF
Your IP: 3.141.166.152


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.