Computer Engineering / Bilgisayar Mühendisliği

Permanent URI for this collectionhttps://hdl.handle.net/11147/10

Browse

Search Results

Now showing 1 - 10 of 217
  • Conference Object
    Enhancing genomic data sharing with blockchain-enabled dynamic consent in beacon V2
    (Springernature, 2024) Binokay, Leman; Celik, Hamit Mervan; Gurdal, Gultekin; Ayav, Tolga; Tuglular, Tugkan; Oktay, Yavuz; Karakulah, Gokhan
  • Article
    Citation - WoS: 1
    Citation - Scopus: 3
    Automatic Test Sequence Generation and Functional Coverage Measurement From Uml Sequence Diagrams
    (Igi Global, 2023) Ekici, Nazim Umut; Tuglular, Tugkan
    Sequence diagrams define functional requirements through use cases. However, their visual form limits their usability in the later stages of the development life cycle. This work proposes a method to transform sequence diagrams into graph-based event sequence graphs, allowing the application of graph analysis methods and defining graph-based coverage criteria. This work explores these newfound abilities in two directions. The first is to use coverage criteria along with existing tests to measure their coverage levels, providing a metric of how well they address the scenarios defined in sequence diagrams. The second is to use coverage criteria to automatically generate effective and efficient acceptance test cases based on the scenarios defined in sequence diagrams. The transformation method is validated with over eighty non-trivial projects. The complete method is validated through a non-trivial example. The results show that the test cases generated with the proposed method are more effective at exposing faults and more efficient in test input size than user-generated test cases.
  • Article
    Link Prediction for Completing Graphical Software Models Using Neural Networks
    (IEEE, 2023) Leblebici, Onur; Tuğlular, Tuğkan; Belli, Fevzi
    Deficiencies and inconsistencies introduced during the modeling of software systems may result in high costs and negatively impact the quality of all developments performed using these models. Therefore, developing more accurate models will aid software architects in developing software systems that match and exceed expectations. This paper proposes a graph neural network (GNN) method for predicting missing connections, or links, in graphical models, which are widely employed in modeling software systems. The proposed method utilizes graphs as allegedly incomplete, primitive graphical models of the system under consideration (SUC) as input and proposes links between its elements through the following steps: (i) transform the models into graph-structured data and extract features from the nodes, (ii) train the GNN model, and (iii) evaluate the performance of the trained model. Two GNN models based on SEAL and DeepLinker are evaluated using three performance metrics, namely cross-entropy loss, area under curve, and accuracy. Event sequence graphs (ESGs) are used as an example of applying the approach to an event-based behavioral modeling technique. Examining the results of experiments conducted on various datasets and variations of GNN reveals that missing connections between events in an ESG can be predicted even with relatively small datasets generated from ESG models. Author
  • Conference Object
    Kurt saldırıları için sentetik irislerde örnek seçilimi
    (IEEE, 2023) Akdeniz, Eyüp Kaan; Erdoğmuş, Nesli
    In this study, samples with higher potential to succeed in wolf attacks are picked among synthetically generated iris images, and the composed subset is shown to pose a more significant threat toward an iris recognition system backed by a Presentation Attack Detection (PAD) module with respect to randomly selected samples. Iris images generated by Deep Convolutional Generative Adversarial Networks (DCGAN) are firstly filtered by rejection sampling on PAD score distribution of real iris image PAD scores. Next, the probability of zero success in all attack attempts is calculated for each synthetic iris image, using real iris images in the training set, and match and non-match score distributions are calculated on those. Synthetic images with the lowest probabilities of zero success are included in the final set. Our hypothesis that this set would be more successful in wolf attacks is tested by comparing its spoofing performances with randomly selected sample sets.
  • Article
    Citation - WoS: 3
    Citation - Scopus: 4
    Application of the Law of Minimum and Dissimilarity Analysis To Regression Test Case Prioritization
    (IEEE, 2023) Ufuktepe, Ekincan; Tuğlular, Tuğkan
    Regression testing is one of the most expensive processes in testing. Prioritizing test cases in regression testing is critical for the goal of detecting the faults sooner within a large set of test cases. We propose a test case prioritization (TCP) technique for regression testing called LoM-Score inspired by the Law of Minimum (LoM) from biology. This technique calculates the impact probabilities of methods calculated by change impact analysis with forward slicing and orders test cases according to LoM. However, this ordering doesn't consider the possibility that consecutive test cases may be covering the same methods repeatedly. Thereby, such ordering can delay the time of revealing faults that exist in other methods. To solve this problem, we enhance the LoM-Score TCP technique with an adaptive approach, namely with a dissimilarity-based coordinate analysis approach. The dissimilarity-based coordinate analysis uses Jaccard Similarity for calculating the similarity coefficients between test cases in terms of covered methods and the enhanced technique called Dissimilarity-LoM-Score (Dis-LoM-Score) applies a penalty with respective on the ordered test cases. We performed our case study on 10 open-source Java projects from Defects4J, which is a dataset of real bugs and an infrastructure for controlled experiments provided for software engineering researchers. Then, we hand-seeded multiple mutants generated by Major, which is a mutation testing tool. Then we compared our TCP techniques LoM-Score and Dis-LoM-Score with the four traditional TCP techniques based on their Average Percentage of Faults Detected (APFD) results.
  • Data Paper
    Citation - WoS: 15
    Citation - Scopus: 20
    Database Covering the Prayer Movements Which Were Not Available Previously
    (Nature Publishing Group, 2023) Mihçin, Şenay; Şahin, Ahmet Mert; Yılmaz, Mehmet; Alpkaya, Alican Tuncay; Tuna, Merve; Akdeniz, Sevinç; Can, Nuray Korkmaz; Tosun, Aliye; Şahin, Serap
    Lower body implants are designed according to the boundary conditions of gait data and tested against. However, due to diversity in cultural backgrounds, religious rituals might cause different ranges of motion and different loading patterns. Especially in the Eastern part of the world, diverse Activities of Daily Living (ADL) consist of salat, yoga rituals, and different style sitting postures. A database covering these diverse activities of the Eastern world is non-existent. This study focuses on data collection protocol and the creation of an online database of previously excluded ADL activities, targeting 200 healthy subjects via Qualisys and IMU motion capture systems, and force plates, from West and Middle East Asian populations with a special focus on the lower body joints. The current version of the database covers 50 volunteers for 13 different activities. The tasks are defined and listed in a table to create a database to search based on age, gender, BMI, type of activity, and motion capture system. The collected data is to be used for designing implants to allow these sorts of activities to be performed.
  • Conference Object
    Citation - WoS: 2
    Citation - Scopus: 2
    Effort Prediction With Limited Data: a Case Study for Data Warehouse Projects
    (IEEE, 2022) Unlu, Huseyin; Yildiz, Ali; Demirors, Onur
    Organizations may create a sustainable competitive advantage against competitors by using data warehouse systems with which they can assess the current status of their operations at any moment. They can analyze trends and connections using up-to-date data. However, data warehouse projects tend to fail more often than other projects as it can be tough to estimate the effort required to build a data warehouse system. Functional size measurement is one of the methods used as an input for estimating the amount of work in a software project. In this study, we formed a measurement basis for DWH projects in an organization based on the COSMIC Functional Size Measurement Method. We mapped COSMIC rules on two different architectures used for DWH projects in the organization and measured the size of the projects. We calculated the productivity of the projects and compared them with the organization's previous projects and DWH projects in the ISBSG repository. We could not create an organization-wide effort estimation model as we had a limited number of projects. As an alternative, we evaluated the success of effort estimation using DWH projects in the ISBSG repository. We also reported the challenges we faced during the size measurement process.
  • Article
    Citation - WoS: 16
    Citation - Scopus: 25
    A Privacy-Preserving Scheme for Smart Grid Using Trusted Execution Environment
    (IEEE, 2023) Akgün, Mete; Üstündağ Soykan, Elif; Soykan, Gürkan
    The increasing transformation from the legacy power grid to the smart grid brings new opportunities and challenges to power system operations. Bidirectional communications between home-area devices and the distribution system empower smart grid functionalities. More granular energy consumption data flows through the grid and enables better smart grid applications. This may also lead to privacy violations since the data can be used to infer the consumer's residential behavior, so-called power signature. Energy utilities mostly aggregate the data, especially if the data is shared with stakeholders for the management of market operations. Although this is a privacy-friendly approach, recent works show that this does not fully protect privacy. On the other hand, some applications, like nonintrusive load monitoring, require disaggregated data. Hence, the challenging problem is to find an efficient way to facilitate smart grid operations without sacrificing privacy. In this paper, we propose a privacy-preserving scheme that leverages consumer privacy without reducing accuracy for smart grid applications like load monitoring. In the proposed scheme, we use a trusted execution environment (TEE) to protect the privacy of the data collected from smart appliances (SAs). The scheme allows customer-oriented smart grid applications as the scheme does not use regular aggregation methods but instead uses customer-oriented aggregation to provide privacy. Hence the accuracy loss stemming from disaggregation is prevented. Our scheme protects the transferred consumption data all the way from SAs to Utility so that possible false data injection attacks on the smart meter that aims to deceive the energy request from the grid are also prevented. We conduct security and game-based privacy analysis under the threat model and provide performance analysis of our implementation. Our results demonstrate that the proposed method overperforms other privacy methods in terms of communication and computation cost. The execution time of aggregation for 10,000 customers, each has 20 SAs is approximately 1 second. The decryption operations performed on the TEE have a linear complexity e.g., 172800 operations take around 1 second while 1728000 operations take around 10 seconds. These results can scale up using cloud or hyper-scalers for real-world applications as our scheme performs offline aggregation.
  • Article
    Citation - WoS: 5
    Citation - Scopus: 9
    P/Key: Puf Based Second Factor Authentication
    (Public Library of Science, 2023) Uysal, Ertan; Akgün, Mete
    One-time password (OTP) mechanisms are widely used to strengthen authentication processes. In time-based one-time password (TOTP) mechanisms, the client and server store common secrets. However, once the server is compromised, the client’s secrets are easy to obtain. To solve this issue, hash-chain-based second-factor authentication protocols have been proposed. However, these protocols suffer from latency in the generation of OTPs on the client side because of the hash-chain traversal. Secondly, they can generate only a limited number of OTPs as it depends on the length of the hash-chain. In this paper, we propose a second-factor authentication protocol that utilizes Physically Unclonable Functions (PUFs) to overcome these problems. In the proposed protocol, PUFs are used to store the secrets of the clients securely on the server. In case of server compromise, the attacker cannot obtain the seeds of clients’ secrets and can not generate valid OTPs to impersonate the clients. In the case of physical attacks, including side-channel attacks on the server side, our protocol has a mechanism that prevents attackers from learning the secrets of a client interacting with the server. Furthermore, our protocol does not incur any client-side delay in OTP generation.
  • Article
    Citation - WoS: 41
    Citation - Scopus: 50
    Bim-Carem: Assessing the Bim Capabilities of Design, Construction and Facilities Management Processes in the Construction Industry
    (Elsevier, 2023) Gökçen, Yılmaz; Akçamete, Aslı; Demirörs, Onur
    BIM adoption has accelerated worldwide since it is an important enabling technology for digitalisation in the construction industry. Adopting BIM requires transforming the traditional building life cycle stages (planning, design, construction and facilities management) into BIM-integrated project deliveries. Assessing the BIM ca- pabilities of these stages helps organisations to identify gaps in their BIM uses and improve them. There is a lack of a comprehensive model in the literature for assessing the BIM capabilities of individual building life cycle stages and their processes. Existing assessment models focus on assessing the BIM maturity of construction projects and organisations which do not inform the required BIM improvements for individual stages and their processes. Hence, we iteratively developed the Building Information Modelling (BIM) Capability Assessment REference Model (BIM-CAREM) and demonstrated its usability through multiple explanatory case studies per- formed with two international design and engineering companies and two general contractors in Turkey. We assessed the BIM capabilities of design, construction and facility management processes of various buildings i.e. residential, stadiums, hospitals and airports. The results showed that the BIM capability levels of design, con- struction and facility management processes vary within and across the companies.