Scopus İndeksli Yayınlar Koleksiyonu / Scopus Indexed Publications Collection

Permanent URI for this collectionhttps://hdl.handle.net/11147/7148

Browse

Search Results

Now showing 1 - 2 of 2
  • Editorial
    Comments on “Relaxed Conditions for the Input-to-State Stability of Switched Nonlinear Time-Varying Systems”
    (Ieee-inst Electrical Electronics Engineers inc, 2025) Sahan, Gokhan; Trenn, Stephan
    This study addresses the deficiencies in the assumptions of the results in (Chen and Yang, 2017) due to the lack of uniformity. We first show the missing hypothesis by presenting a counterexample. Then, we prove why they are wrong in that form and show the errors in the proof of the main result of (Chen and Yang, 2017). Next, we compare the assumptions and related results of (Chen and Yang, 2017) with similar works in the literature. Lastly, we give suggestions to complement the shortcomings of the hypotheses and thus correct them.
  • Conference Object
    Citation - Scopus: 3
    Predicting Software Size and Effort From Code Using Natural Language Processing
    (CEUR-WS, 2024) Tenekeci, S.; Demirörs, Onur; Ünlü, H.; Dikenelli, E.; Selçuk, U.; Kılınç Soylu, G.; Demirörs, O.
    Software Size Measurement (SSM) holds a crucial role in software project management by facilitating the acquisition of software size, which serves as the primary input for development effort and schedule estimation. However, many small and medium-sized companies encounter challenges in conducting objective SSM and Software Effort Estimation (SEE) due to resource constraints and a lack of expert workforce. This often leads to inaccurate estimates and projects exceeding planned time and budget. Hence, organizations need to perform objective SSM and SEE with minimal resources and without relying on an expert workforce. In this research, we introduce two exploratory case studies aimed at predicting the functional size (COSMIC and Event-based size) and effort of software projects from the code using a deep-learning-based NLP model: CodeBERT. For this purpose, we collected and annotated two datasets consisting of 4800 Python and 1100 C# functions. Then, we trained a classification model to predict COSMIC data movements (entry, exit, read, write) and four regression models to predict Event-based size (interaction, communication, process) and effort. Despite utilizing a relatively small dataset for model training, we achieved promising results with an 84.5% accuracy for the COSMIC size, 0.13 normalized mean absolute error (NMAE) for the Event-based size, and 0.18 NMAE for the effort. These findings are particularly insightful as they demonstrate the practical utility of language models in SSM and SEE. © 2024 Copyright for this paper by its authors.