Computer Engineering / Bilgisayar Mühendisliği

Permanent URI for this collectionhttps://hdl.handle.net/11147/10

Browse

Search Results

Now showing 1 - 10 of 32
  • Article
    Citation - WoS: 1
    Citation - Scopus: 3
    Automatic Test Sequence Generation and Functional Coverage Measurement From Uml Sequence Diagrams
    (Igi Global, 2023) Ekici, Nazim Umut; Tuglular, Tugkan
    Sequence diagrams define functional requirements through use cases. However, their visual form limits their usability in the later stages of the development life cycle. This work proposes a method to transform sequence diagrams into graph-based event sequence graphs, allowing the application of graph analysis methods and defining graph-based coverage criteria. This work explores these newfound abilities in two directions. The first is to use coverage criteria along with existing tests to measure their coverage levels, providing a metric of how well they address the scenarios defined in sequence diagrams. The second is to use coverage criteria to automatically generate effective and efficient acceptance test cases based on the scenarios defined in sequence diagrams. The transformation method is validated with over eighty non-trivial projects. The complete method is validated through a non-trivial example. The results show that the test cases generated with the proposed method are more effective at exposing faults and more efficient in test input size than user-generated test cases.
  • Conference Object
    Citation - Scopus: 2
    Repository Landscape in Turkiye and Gcris: the First National Research Information System
    (Elsevier, 2022) Tuğlular, Tuğkan; Gürdal, Gültekin; Kafalı Can, Gönül; Özdemirden, Ahmet Şemsettin
    This paper describes the history and developments of research infrastructures and open science policies in Turkiye. Moreover, it focuses on the GCRIS (Grand Current Research Information Systems), Turkiye's first Research Information System by inter-national standards, emphasizing the need for internationally interoperable research infrastructures in Turkiye. GCRIS Research Information System, implemented on the open-source software DSpace-CRIS 6.3, was developed with data analytics in mind and continues to be improved by Research Ecosystems Inc. As a strategic partner, Izmir Institute of Technology (IZTECH) is the first university to use GCRIS. Other Universities have used GCRIS since then. With the increase in the number of universities using GCRIS, Turkiye's Research Ecosystem will be trackable and measurable much better thanks to GCRIS intelligent reporting sys- tem. Most importantly, not only the research outputs of Turkiye will be more visible, but also research infrastructures' integration will facilitate with the European Open Science Cloud (EOSC) and other initiatives worldwide.
  • Article
    Asking the Right Questions To Solve Algebraic Word Problems
    (TÜBİTAK - Türkiye Bilimsel ve Teknolojik Araştırma Kurumu, 2022) Çelik, Ege Yiğit; Orulluoğlu, Zeynel; Mertoğlu, Rıdvan; Tekir, Selma
    Word algebra problems are among challenging AI tasks as they combine natural language understanding with a formal equation system. Traditional approaches to the problem work with equation templates and frame the task as a template selection and number assignment to the selected template. The recent deep learning-based solutions exploit contextual language models like BERT and encode the natural language text to decode the corresponding equation system. The proposed approach is similar to the template-based methods as it works with a template and fills in the number slots. Nevertheless, it has contextual understanding because it adopts a question generation and answering pipeline to create tuples of numbers, to finally perform the number assignment task by custom sets of rules. The inspiring idea is that by asking the right questions and answering them using a state-of-the-art language model-based system, one can learn the correct values for the number slots in an equation system. The empirical results show that the proposed approach outperforms the other methods significantly on the word algebra benchmark dataset alg514 and performs the second best on the AI2 corpus for arithmetic word problems. It also has superior performance on the challenging SVAMP dataset. Though it is a rule-based system, simple rule sets and relatively slight differences between rules for different templates indicate that it is highly probable to develop a system that can learn the patterns for the collection of all possible templates, and produce the correct equations for an example instance.
  • Article
    Citation - WoS: 1
    Citation - Scopus: 3
    Ignoring Internal Utilities in High-Utility Itemset Mining
    (MDPI, 2022) Oğuz, Damla
    High-utility itemset mining discovers a set of items that are sold together and have utility values higher than a given minimum utility threshold. The utilities of these itemsets are calculated by considering their internal and external utility values, which correspond, respectively, to the quantity sold of each item in each transaction and profit units. Therefore, internal and external utilities have symmetric effects on deciding whether an itemset is high-utility. The symmetric contributions of both utilities cause two major related challenges. First, itemsets with low external utility values can easily exceed the minimum utility threshold if they are sold extensively. In this case, such itemsets can be found more efficiently using frequent itemset mining. Second, a large number of high-utility itemsets are generated, which can result in interesting or important high-utility itemsets that are overlooked. This study presents an asymmetric approach in which the internal utility values are ignored when finding high-utility itemsets with high external utility values. The experimental results of two real datasets reveal that the external utility values have fundamental effects on the high-utility itemsets. The results of this study also show that this effect tends to increase for high values of the minimum utility threshold. Moreover, the proposed approach reduces the execution time.
  • Article
    Citation - WoS: 7
    Citation - Scopus: 12
    A Survey on Organizational Choices for Microservice-Based Software Architectures
    (TÜBİTAK, 2022) Ünlü, Hüseyin; Bilgin, Burak; Demirörs, Onur
    During the last decade, the demand for more flexible, responsive, and reliable software applications increased exponentially. The availability of internet infrastructure and new software technologies to respond to this demand led to a new generation of applications. As a result, cloud-based, distributed, independently deployable web applications working together in a microservice-based software architecture style have gained popularity. The style has been a common practice in the industry and successfully utilized by companies. Adopting this style demands software organizations to transform their culture. However, there is a lack of research studies that explores common practices for microservices. Thus, we performed a survey to explore the organizational choices on software analysis, design, size measurement, and effort estimation when working with microservices. The results provide a snapshot of the software industry that utilizes microservices. We provide insight for software organizations to transform their culture and suggest challenges researchers can focus on in the area.
  • Article
    Citation - WoS: 4
    Citation - Scopus: 4
    Integrative Biological Network Analysis To Identify Shared Genes in Metabolic Disorders
    (Institute of Electrical and Electronics Engineers, 2022) Tenekeci, Samet; Işık, Zerrin
    Identification of common molecular mechanisms in interrelated diseases is essential for better prognoses and targeted therapies. However, complexity of metabolic pathways makes it difficult to discover common disease genes underlying metabolic disorders; and it requires more sophisticated bioinformatics models that combine different types of biological data and computational methods. Accordingly, we built an integrative network analysis model to identify shared disease genes in metabolic syndrome (MS), type 2 diabetes (T2D), and coronary artery disease (CAD). We constructed weighted gene co-expression networks by combining gene expression, protein-protein interaction, and gene ontology data from multiple sources. For 90 different configurations of disease networks, we detected the significant modules by using MCL, SPICi, and Linkcomm graph clustering algorithms. We also performed a comparative evaluation on disease modules to determine the best method providing the highest biological validity. By overlapping the disease modules, we identified 22 shared genes for MS-CAD and T2D-CAD. Moreover, 19 out of these genes were directly or indirectly associated with relevant diseases in the previous medical studies. This study does not only demonstrate the performance of different biological data sources and computational methods in disease-gene discovery, but also offers potential insights into common genetic mechanisms of the metabolic disorders.
  • Article
    Citation - Scopus: 1
    Performance Analysis and Feature Selection for Network-Based Intrusion Detection With Deep Learning
    (Türkiye Klinikleri, 2022) Caner, Serhat; Erdoğmuş, Nesli; Erten, Yusuf Murat
    An intrusion detection system is an automated monitoring tool that analyzes network traffic and detects malicious activities by looking out either for known patterns of attacks or for an anomaly. In this study, intrusion detection and classification performances of different deep learning based systems are examined. For this purpose, 24 deep neural networks with four different architectures are trained and evaluated on CICIDS2017 dataset. Furthermore, the best performing model is utilized to inspect raw network traffic features and rank them with respect to their contributions to success rates. By selecting features with respect to their ranks, sets of varying size from 3 to 77 are assessed in terms of classification accuracy and time efficiency. The results show that recurrent neural networks with a certain level of complexity can achieve comparable success rates with state-of-the-art systems using a small feature set of size 9; while the average time required to classify a test sample is halved compared to the complete set.
  • Article
    Citation - WoS: 1
    Citation - Scopus: 2
    Information Retrieval-Based Bug Localization Approach With Adaptive Attribute Weighting
    (TÜBİTAK - Türkiye Bilimsel ve Teknolojik Araştırma Kurumu, 2021) ErşahIn, Mustafa; Utku, Semih; Kılınç, Deniz; ErşahIn, Buket
    Software quality assurance is one of the crucial factors for the success of software projects. Bug fixing has an essential role in software quality assurance, and bug localization (BL) is the first step of this process. BL is difficult and time-consuming since the developers should understand the flow, coding structure, and the logic of the program. Information retrieval-based bug localization (IRBL) uses the information of bug reports and source code to locate the section of code in which the bug occurs. It is difficult to apply other tools because of the diversity of software development languages, design patterns, and development standards. The aim of this study is to build an adaptive IRBL tool and make it usable by more companies. BugSTAiR solves the aforementioned problem by means of the adaptive attribute weighting (AAW) algorithm and is evaluated on four open-source projects which are well-known benchmark datasets on BL. One of them is BLIA which is the state of the art in bug localization area and another is BLUIR which is a well-known BL tool. According to the promising results of experiments, Top1 rank of BugSTAiR is 2% and MAP is 10% better than BLIA's results on AspectJ and it has localized 4.6% of all bugs in Top1 and its precision is 6.1% better than BLIA on SWT, respectively. On the other side, it is 20% better in the Top1 metric and 30% in precision than BLUIR.
  • Article
    Citation - WoS: 2
    Citation - Scopus: 2
    Incorporating Concreteness in Multi-Modal Language Models With Curriculum Learning
    (MDPI, 2021) Sezerer, Erhan; Tekir, Selma
    Over the last few years, there has been an increase in the studies that consider experiential (visual) information by building multi-modal language models and representations. It is shown by several studies that language acquisition in humans starts with learning concrete concepts through images and then continues with learning abstract ideas through the text. In this work, the curriculum learning method is used to teach the model concrete/abstract concepts through images and their corresponding captions to accomplish multi-modal language modeling/representation. We use the BERT and Resnet-152 models on each modality and combine them using attentive pooling to perform pre-training on the newly constructed dataset, which is collected from the Wikimedia Commons based on concrete/abstract words. To show the performance of the proposed model, downstream tasks and ablation studies are performed. The contribution of this work is two-fold: A new dataset is constructed from Wikimedia Commons based on concrete/abstract words, and a new multi-modal pre-training approach based on curriculum learning is proposed. The results show that the proposed multi-modal pre-training approach contributes to the success of the model.
  • Article
    Citation - WoS: 2
    Citation - Scopus: 1
    Mutant Selection by Using Fourier Expansion
    (Türkiye Klinikleri Journal of Medical Sciences, 2020) Takan, Savaş; Ayav, Tolga
    Mutation analysis is a widely used technique to evaluate the effectiveness of test cases in both hardware and software testing. The original model is mutated systematically under certain fault assumptions and test cases are checked against the mutants created to see whether the test cases can detect the faults or not. Mutation analysis is usually a computationally intensive task, particularly in finite state machine (FSM) testing due to a possibly huge amount of mutants. Random selection could be a practical reduction method under the assumption that each mutant is identical in terms of the probability of occurrence of its associating fault. The present study proposes a mutant selection method based on Fourier analysis of Boolean functions. Fourier helps to identify the most effective transitions on the output so that the mutants related to those transitions can be selected. Such mutants are considered more important since they are more likely to be killed. To evaluate the method, test cases are generated by the well-known W method, which has the capability of detecting every potential fault. The original and reduced sets of mutants are compared with respect to their importance values. Evaluations show that the mutants selected by the proposed technique are more effective, which reduces the cost of mutation analysis without sacrificing the performance of the mutation analysis.