Improvements on a Multi-Task Bert Model

dc.contributor.author Agrali, Mahmut
dc.contributor.author Tekir, Selma
dc.date.accessioned 2024-09-24T15:58:52Z
dc.date.available 2024-09-24T15:58:52Z
dc.date.issued 2024
dc.description.abstract Pre-trained language models have introduced significant performance boosts in natural language processing. Fine-tuning of these models using downstream tasks' supervised data further improves the acquired results. In the fine-tuning process, combining the learning of tasks is an effective approach. This paper proposes a multi-task learning framework based on BERT. To accomplish the tasks of sentiment analysis, paraphrase detection, and semantic text similarity, we include linear layers, a Siamese network with cosine similarity, and convolutional layers to the appropriate places in the architecture. We conducted an ablation study using Stanford Sentiment Treebank (SST), Quora, and SemEval STS datasets for each task to test the framework and its components' effectiveness. The results demonstrate that the proposed multi-task framework improves the performance of BERT. The best results obtained for sentiment analysis, paraphrase detection, and semantic text similarity are accuracies of 0.534 and 0.697 and a Pearson correlation coefficient of 0.345. en_US
dc.identifier.doi 10.1109/SIU61531.2024.10600801
dc.identifier.isbn 9798350388978
dc.identifier.isbn 9798350388961
dc.identifier.issn 2165-0608
dc.identifier.scopus 2-s2.0-85200919611
dc.identifier.uri https://doi.org/10.1109/SIU61531.2024.10600801
dc.identifier.uri https://hdl.handle.net/11147/14815
dc.language.iso tr en_US
dc.publisher Ieee en_US
dc.relation.ispartof 32nd IEEE Signal Processing and Communications Applications Conference (SIU) -- MAY 15-18, 2024 -- Tarsus Univ Campus, Mersin, TURKEY en_US
dc.relation.ispartofseries Signal Processing and Communications Applications Conference
dc.rights info:eu-repo/semantics/closedAccess en_US
dc.subject Multi-Task Learning en_US
dc.subject Sentiment Analysis en_US
dc.subject Paraphrase Detection en_US
dc.subject Semantic Textual Similarity en_US
dc.title Improvements on a Multi-Task Bert Model en_US
dc.title.alternative Çok Görevli Bert Modeli Üzerinde Iyileştirmeler en_US
dc.type Conference Object en_US
dspace.entity.type Publication
gdc.bip.impulseclass C5
gdc.bip.influenceclass C5
gdc.bip.popularityclass C5
gdc.coar.access metadata only access
gdc.coar.type text::conference output
gdc.collaboration.industrial false
gdc.description.department İzmir Institute of Technology en_US
gdc.description.departmenttemp [Agrali, Mahmut; Tekir, Selma] Izmir Yuksek Teknol Enstitusu, Bilgisayar Muhendisligi, Izmir, Turkiye en_US
gdc.description.endpage 4
gdc.description.publicationcategory Konferans Öğesi - Uluslararası - Kurum Öğretim Elemanı en_US
gdc.description.scopusquality N/A
gdc.description.startpage 1
gdc.description.woscitationindex Conference Proceedings Citation Index - Science
gdc.description.wosquality N/A
gdc.identifier.openalex W4400908442
gdc.identifier.wos WOS:001297894700071
gdc.index.type WoS
gdc.index.type Scopus
gdc.oaire.diamondjournal false
gdc.oaire.impulse 0.0
gdc.oaire.influence 2.635068E-9
gdc.oaire.isgreen false
gdc.oaire.popularity 3.0009937E-9
gdc.oaire.publicfunded false
gdc.openalex.fwci 0.0
gdc.openalex.normalizedpercentile 0.2
gdc.opencitations.count 0
gdc.plumx.scopuscites 0
gdc.scopus.citedcount 0
gdc.wos.citedcount 0
relation.isAuthorOfPublication.latestForDiscovery 57639474-3954-4f77-a84c-db8a079648a8
relation.isOrgUnitOfPublication.latestForDiscovery 9af2b05f-28ac-4014-8abe-a4dfe192da5e

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Name:
Improvements_on_a_Multi-task_BERT_Model.pdf
Size:
401.3 KB
Format:
Adobe Portable Document Format
Description:
Article