Improvements on a Multi-Task Bert Model
| dc.contributor.author | Agrali, Mahmut | |
| dc.contributor.author | Tekir, Selma | |
| dc.date.accessioned | 2024-09-24T15:58:52Z | |
| dc.date.available | 2024-09-24T15:58:52Z | |
| dc.date.issued | 2024 | |
| dc.description.abstract | Pre-trained language models have introduced significant performance boosts in natural language processing. Fine-tuning of these models using downstream tasks' supervised data further improves the acquired results. In the fine-tuning process, combining the learning of tasks is an effective approach. This paper proposes a multi-task learning framework based on BERT. To accomplish the tasks of sentiment analysis, paraphrase detection, and semantic text similarity, we include linear layers, a Siamese network with cosine similarity, and convolutional layers to the appropriate places in the architecture. We conducted an ablation study using Stanford Sentiment Treebank (SST), Quora, and SemEval STS datasets for each task to test the framework and its components' effectiveness. The results demonstrate that the proposed multi-task framework improves the performance of BERT. The best results obtained for sentiment analysis, paraphrase detection, and semantic text similarity are accuracies of 0.534 and 0.697 and a Pearson correlation coefficient of 0.345. | en_US |
| dc.identifier.doi | 10.1109/SIU61531.2024.10600801 | |
| dc.identifier.isbn | 9798350388978 | |
| dc.identifier.isbn | 9798350388961 | |
| dc.identifier.issn | 2165-0608 | |
| dc.identifier.scopus | 2-s2.0-85200919611 | |
| dc.identifier.uri | https://doi.org/10.1109/SIU61531.2024.10600801 | |
| dc.identifier.uri | https://hdl.handle.net/11147/14815 | |
| dc.language.iso | tr | en_US |
| dc.publisher | Ieee | en_US |
| dc.relation.ispartof | 32nd IEEE Signal Processing and Communications Applications Conference (SIU) -- MAY 15-18, 2024 -- Tarsus Univ Campus, Mersin, TURKEY | en_US |
| dc.relation.ispartofseries | Signal Processing and Communications Applications Conference | |
| dc.rights | info:eu-repo/semantics/closedAccess | en_US |
| dc.subject | Multi-Task Learning | en_US |
| dc.subject | Sentiment Analysis | en_US |
| dc.subject | Paraphrase Detection | en_US |
| dc.subject | Semantic Textual Similarity | en_US |
| dc.title | Improvements on a Multi-Task Bert Model | en_US |
| dc.title.alternative | Çok Görevli Bert Modeli Üzerinde Iyileştirmeler | en_US |
| dc.type | Conference Object | en_US |
| dspace.entity.type | Publication | |
| gdc.bip.impulseclass | C5 | |
| gdc.bip.influenceclass | C5 | |
| gdc.bip.popularityclass | C5 | |
| gdc.coar.access | metadata only access | |
| gdc.coar.type | text::conference output | |
| gdc.collaboration.industrial | false | |
| gdc.description.department | İzmir Institute of Technology | en_US |
| gdc.description.departmenttemp | [Agrali, Mahmut; Tekir, Selma] Izmir Yuksek Teknol Enstitusu, Bilgisayar Muhendisligi, Izmir, Turkiye | en_US |
| gdc.description.endpage | 4 | |
| gdc.description.publicationcategory | Konferans Öğesi - Uluslararası - Kurum Öğretim Elemanı | en_US |
| gdc.description.scopusquality | N/A | |
| gdc.description.startpage | 1 | |
| gdc.description.woscitationindex | Conference Proceedings Citation Index - Science | |
| gdc.description.wosquality | N/A | |
| gdc.identifier.openalex | W4400908442 | |
| gdc.identifier.wos | WOS:001297894700071 | |
| gdc.index.type | WoS | |
| gdc.index.type | Scopus | |
| gdc.oaire.diamondjournal | false | |
| gdc.oaire.impulse | 0.0 | |
| gdc.oaire.influence | 2.635068E-9 | |
| gdc.oaire.isgreen | false | |
| gdc.oaire.popularity | 3.0009937E-9 | |
| gdc.oaire.publicfunded | false | |
| gdc.openalex.fwci | 0.0 | |
| gdc.openalex.normalizedpercentile | 0.2 | |
| gdc.opencitations.count | 0 | |
| gdc.plumx.scopuscites | 0 | |
| gdc.scopus.citedcount | 0 | |
| gdc.wos.citedcount | 0 | |
| relation.isAuthorOfPublication.latestForDiscovery | 57639474-3954-4f77-a84c-db8a079648a8 | |
| relation.isOrgUnitOfPublication.latestForDiscovery | 9af2b05f-28ac-4014-8abe-a4dfe192da5e |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Improvements_on_a_Multi-task_BERT_Model.pdf
- Size:
- 401.3 KB
- Format:
- Adobe Portable Document Format
- Description:
- Article
