Middleware Security - Online advertisements are a significant element of the Internet ecosystem. Businesses monitor their customers via pushing advertising (Ads). Within minutes, cybercriminals try to defraud and steal data through advertisements. Therefore, the issue of ads must be solved. Ads are obtrusive, a security risk, and they hinder performance and efficiency. Hence, the goal is to create an ad-blocker that would operate across the entire network and prevent advertisement on any website s web pages. To put it another way, it s a little computer with such a SoC (System - On - chip) also referred to as a Raspberry Pi that is merged with a networking system, for which we need to retrain the advertisements. On the home network, software named Pi Hole is used to block websites with advertisements. Any network traffic that passes via devices connected to the home network now passes through the Pi. As a result, the adverts are finally checked out during the Raspberry Pi before they reach the users machine and they will be blocked.
Authored by Harshal Sonawane, Manasi Patil, Shashank Patil, Uma Thakur, Bhavin Patil, Abha Marathe
Middleware Security - Securing IoT networks has been one of recent most active research topics. However, unlike traditional network security, where the emphasis is given on the core network, IoT networks are mostly investigated from the data standpoint. Lightweight data transmission protocols, such as Message Queue Telemetry Transport (MQTT), are often deployed for data-sharing and device authentication due to limited onboard resources. This paper presents the MQTT protocol’s security vulnerabilities by incorporating Elliptic Curve Cryptographybased (ECC-based) security to improve confidentiality issues. We used commercially off-the-shelf (COTS) devices such as Raspberry Pi to build a simplified network topology that connects IoT devices in our smart home laboratory. The results illustrate an ECC-based security application in confidentiality increase of 70.65\% from 29.35\% in time parameter during publish/subscribe communication protocol for the smart home.
Authored by Zainatul Yusoff, Mohamad Ishak, Lukman Rahim, Omer Ali
Middleware Security - Cybersecurity of power hardware is becoming increasingly critical with the emergence of smart and connected devices such as Grid-connected inverters, EVs and their chargers, microgrid controllers, energy storage / energy management controllers, and smart appliances. Cyber-attacks on power hardware have had far-reaching and widespread impacts. For such cyber-physical systems, security must be ensured at all levels in the design - hardware, firmware, software and interfaces. Although previous approaches to cybersecurity have focused mainly on vulnerabilities in the firmware middleware, or software, vulnerabilities in the hardware itself are hard to identify and harder to mitigate, especially when most hardware components are proprietary and not examinable. This paper presents one approach to mitigate this conundrum - a completely open-source implementation of a microcontroller core along with the associated peripherals based on the well-known RISC-V instruction set architecture (ISA). The proof-of-concept architecture presented here uses the “Shakti” E-Class microcontroller core integrated with a fully custom PWM controller implemented in Verilog, and validated on a Xilinx Artix FPGA. For critical applications such designs may be replicated as a custom ASIC thereby guaranteeing total security of the computing hardware.
Authored by S Swakath, Abhijit Kshirsagar, Koteswararao Kondepu, Satish Banavath, Andrii Chub, Dmitri Vinnikov
Middleware Security - Connected devices are being deployed at a steady rate, providing services like data collection. Pervasive applications rely on those edge devices to seamlessly provide services to users. To connect applications and edge devices, using a middleware has been a popular approach. The research is active on the subject as there are many open challenges. The secure management of the edge devices and the security of the middleware are two of them. As security is a crucial requirement for pervasive environment, we propose a middleware architecture easing the secure use of edge devices for pervasive applications, while supporting the heterogeneity of communication protocols and the dynamism of devices. Because of the heterogeneity in protocols and security features, not all edge devices are equally secure. To allow the pervasive applications to gain control over this heterogeneous security, we propose a model to describe edge devices security. This model is accessible by the applications through our middleware. To validate our work, we developed a demonstrator of our middleware and we tested it in a concrete scenario.
Authored by Arthur Desuert, Stéphanie Chollet, Laurent Pion, David Hely
Middleware Security - Web application security is the most important area when it comes to developing a web application. Many web applications having vulnerabilities due to poor implementation of security measures. These web applications will be deployed without fixing the vulnerabilities thus becomes vulnerable to many cyber-attacks. Simple attacks like brute-force and NoSQL injection could give unauthorized access to the user accounts. This leads to user privacy issues which could create huge loss to the organizations. These vulnerabilities can be fixed by implementing the necessary security measures while developing the web application. OWASP (Open Web Application Security Project) is a non-profit organization which gives the severity, impact and prevention methods about Top 10 vulnerabilities in web applications. This research deals with the implementation of bestsecurity practices for Node.js web applications in detail. This research paper proposes the security mechanisms for attacks related to front-end, middleware and backend web development using OWASP suggestions. The main focus of this research paper is on prevention of Denial-of-service attack, Brute force attack, NoSQL injection attack and Unrestricted file upload vulnerability.The proposed prevention methods are implemented in a web application to test the defensive mechanisms against the mentionedvulnerabilities.
Authored by Akshay Kumar, Usha Rani
Middleware Security - Robot Operating System 2 (ROS2) is the latest release of a framework for enabling robot applications. Data Distribution Service (DDS) middleware is used for communication between nodes in a ROS2 cluster. The DDS middleware provides a distributed discovery system, message definitions and serialization, and security. In ROS2, the DDS middleware is accessed through an abstraction layer, making it easy to switch from one implementation to another. The existing middleware implementations differ in a number of ways, e.g., in how they are supported in ROS2, in their support for the security features, their ease of use, their performance, and their interoperability. In this work, the focus is on the ease of use, interoperability, and security features aspects of ROS2 DDS middleware. We compare the ease of installation and ease of use of three different DDS middleware, and test the interoperability of different middleware combinations in simple deployment scenarios. We highlight the difference that enabling the security option makes to interoperability, and conduct performance experiments that show the effect that turning on security has on the communication performance. Our results provide guidelines for choosing and deploying DDS middleware on a ROS2 cluster.
Authored by Max Aartsen, Kanta Banga, Konrad Talko, Dustin Touw, Bertus Wisman, Daniel Meïnsma, Mathias Björkqvist
Metadata Discovery Problem - In order to enable a collaborative Model-based Systems Engineering (MBSE) environment through computer systems, it is completely necessary to enable the possibility of communicating tools (interoperability) and reusing previous engineering designs saving costs and time. In this context, the understanding of the underlying concepts and relationships embedded in the system artifacts becomes a cornerstone to properly exploit engineering artifacts. MBSE tool-chains and suites, such as Matlab Simulink, can be applied to different engineering activities: architecture design (descriptive modeling), simulation (analytical modeling) or verification. Reuse capabilities in specific engineering tools are a kind of non-functional aspect that is usually covered providing a type of search capability based on artifact metadata. In this work, we aim to ease the reuse of the knowledge embedded in Simulink models through a solution called PhysicalModel2Simulink. The proposed approach makes use of an ontology for representing, indexing and retrieving information following a meta-model (mainly to semantically represent concepts and relationships). Under this schema, both meta-data and contents are represented using a common domain vocabulary and taxonomy creating a property graph that can be exploited for system artifact discovery. To do so, a mapping between the Matlab Simulink meta-model and the RSHP (RelationShHiP) meta-model is defined to represent and serialize physical models in a repository. Then, a retrieval process is implemented on top of this repository to allow users to perform text-based queries and look up similar artifacts. To validate the proposed solution, 38 Simulink models have been used and 20 real user queries have been designed to study the effectiveness, in terms or precision and recall, of the proposed solution against the Matlab Simulink searching capabilities.
Authored by Eduardo Cibrian, Roy Mendieta, Jose Alvarez-Rodriguez, Juan Llorens
Metadata Discovery Problem - Collaborative software development platforms like GitHub have gained tremendous popularity. Unfortunately, many users have reportedly leaked authentication secrets (e.g., textual passwords and API keys) in public Git repositories and caused security incidents and finical loss. Recently, several tools were built to investigate the secret leakage in GitHub. However, these tools could only discover and scan a limited portion of files in GitHub due to platform API restrictions and bandwidth limitations. In this paper, we present SecretHunter, a real-time large-scale comprehensive secret scanner for GitHub. SecretHunter resolves the file discovery and retrieval difficulty via two major improvements to the Git cloning process. Firstly, our system will retrieve file metadata from repositories before cloning file contents. The early metadata access can help identify newly committed files and enable many bandwidth optimizations such as filename filtering and object deduplication. Secondly, SecretHunter adopts a reinforcement learning model to analyze file contents being downloaded and infer whether the file is sensitive. If not, the download process can be aborted to conserve bandwidth. We conduct a one-month empirical study to evaluate SecretHunter. Our results show that SecretHunter discovers 57\% more leaked secrets than state-of-the-art tools. SecretHunter also reduces 85\% bandwidth consumption in the object retrieval process and can be used in low-bandwidth settings (e.g., 4G connections).
Authored by Elliott Wen, Jia Wang, Jens Dietrich
Metadata Discovery Problem - Millions of connected devices like connected cameras and streaming videos are introduced to smart cities every year, which are valuable source of information. However, such rich source of information is mostly left untapped. Thus, in this paper, we propose distributed deep neural networks (DNNs) over edge visual Internet of Things (VIoT) devices for parallel, real-time video scene parsing and indexing in conjunction with BigQuery retrieval on stored data in the cloud. The IoT video streams parsed into adaptive meta-data of person, attributes, actions, object, and relations using pre-trained DNNs. The meta-data cached at the edge-cloud for real-time analytics and also continuously transferred to the cloud for data fusion and BigQuery batch processing. The proposed distributed deep learning search platform bridges the gap between edge-to-cloud continuum computation by utilizing state-of-the-art distributed deep learning and BigQuery search algorithms for the geo-distributed Visual Internet of Things (VIoT). We show that our proposed system supports real-time event-driven computing at 122 milliseconds on virtual IoT devices in parallel, and as low as 2.4 seconds batch query response time on multi-table JOIN and GROUP-BY aggregation.
Authored by Arun Das, Mehdi Roopaei, Mo Jamshidi, Peyman Najafirad
Metadata Discovery Problem - To conduct a well-designed and reproducible study, researchers must define and adhere to clear inclusion and exclusion criteria for subjects. Similarly, a well-run journal or conference should publish easily understood inclusion and exclusion criteria that determine which submissions will receive more detailed peer review. This will empower authors to identify the conferences and journals that are the best fit for their manuscripts while allowing organizers and peer reviewers to spend more time on the submissions that are of greatest interest. To provide a more systematic way of representing these criteria, we extend the syntax for concept-validating constraints of the Nexus-PORTAL-DOORS-Scribe cyberinfrastructure, which already serve as criteria for inclusion of records in a repository, to allow description of exclusion criteria.
Authored by Adam Craig, Carl Taswell
Metadata Discovery Problem - We present a methodology for constructing a spatial ontology-based datasets navigation model to allow cross-reference navigation between datasets. We defined the structure of the dataset as metadata, the field names, and the actual values. We defined the relationship between datasets as 3-layer such as metadata layer, field names layer, and data value layer. The relationships in the metadata layer was defined as the correspondence between metadata values. We standardized the field names in dataset to discover the relationships between field names. We designed a method to discover the relationship between data values based on common knowledge datasets for each domain. To confirm the validity of the presented methodology, we applied our methodology to implement an ontology-based knowledge navigation model for actual disasterrelated processes in operation. We built a knowledge navigation model based on spatial common knowledge.
Authored by Yun-Young Hwang, Sumi Shin
Metadata Discovery Problem - We defined and expressed graph-based relationships of pieces of knowledge to allow cross-reference navigation of the knowledge as an ontology. We present a methodology for constructing an ontology-based knowledge navigation model to allow cross-reference navigation between pieces of knowledge, related concepts and datasets. We defined the structure of the dataset as metadata, the field names of the actual values, and the actual values. We defined the relationship between datasets as 3-layer such as metadata layer, field names layer, and data value layer. The relationships in the metadata layer was defined as the correspondence between metadata values. We standardized the field names in dataset to discover the relationships between field names. We designed a method to discover the relationship between data values based on common knowledge for each domain. To confirm the validity of the presented methodology, we applied our methodology to implement an ontology-based knowledge navigation model for actual disaster-related processes in operation. We built a knowledge navigation model based on spatial common knowledge to confirm that the configuration of the knowledge navigation model was correct.
Authored by Yun-Young Hwang, Jiseong Son, Sumi Shin
Metadata Discovery Problem - The OPC UA (Open Platform Communications Unified Architecture) technology is found in many industrial applications as it addresses many of Industry 4.0’s requirements. One of its appeals is its service-oriented architecture. Nonetheless, it requires engineering efforts during deployment and maintenance to bind or associate the correct services to a client or consumer system. We propose the integration of OPC UA with the Eclipse Arrowhead Framework (EAF) to enable automatic service discovery and binding at runtime, reducing delays, costs, and errors. The integration also enables the client system to get the service endpoints by querying the service attributes or metadata. Moreover, this forms a bridge to other industrial communication technologies such as Modbus TCP (Transmission Control Protocol) as the framework is not limited to a specific protocol. To demonstrate the idea, an indexed line with an industrial PLC (programmable logic controller) with an OPC UA server is used to show that the desired services endpoints are revealed at runtime when querying their descriptive attributes or metadata through the EAF’s Orchestrator system.
Authored by Aparajita Tripathy, Jan Van Deventer, Cristina Paniagua, Jerker Delsing
Metadata Discovery Problem - Researchers seeking to apply computational methods are increasingly turning to scientific digital archives containing images of specimens. Unfortunately, metadata errors can inhibit the discovery and use of scientific archival images. One such case is the NSF-sponsored Biology Guided Neural Network (BGNN) project, where an abundance of metadata errors has significantly delayed development of a proposed, new class of neural networks. This paper reports on research addressing this challenge. We present a prototype workflow for specimen scientific name metadata verification that is grounded in Computational Archival Science (CAS), report on a taxonomy of specimen name metadata error types with preliminary solutions. Our 3-phased workflow includes tag extraction, text processing, and interactive assessment. A baseline test with the prototype workflow identified at least 15 scientific name metadata errors out of 857 manually reviewed, potentially erroneous specimen images, corresponding to a ∼ 0.2\% error rate for the full image dataset. The prototype workflow minimizes the amount of time domain experts need to spend reviewing archive metadata for correctness and AI-readiness before these archival images can be utilized in downstream analysis.
Authored by Joel Pepper, Andrew Senin, Dom Jebbia, David Breen, Jane Greenberg
Metadata Discovery Problem - Semantic segmentation is one of the key research areas in computer vision, which has very important applications in areas such as autonomous driving and medical image diagnosis. In recent years, the technology has advanced rapidly, where current models have been able to achieve high accuracy and efficient speed on some widely used datasets. However, the semantic segmentation task still suffers from the inability to generate accurate boundaries in the case of insufficient feature information. Especially in the field of medical image segmentation, most of the medical image datasets usually have class imbalance issues and there are always variations in factors such as shape and color between different datasets and cell types. Therefore, it is difficult to establish general algorithms across different classes and robust algorithms that differ across different datasets. In this paper, we propose a conditional data preprocessing strategy, i.e., Conditional Metadata Embedding (CME) data preprocessing strategy. The CME data preprocessing method will embed conditional information to the training data, which can assist the model to better overcome the differences in the datasets and extract useful feature information in the images. The experimental results show that the CME data preprocessing method can help different models achieve higher segmentation performance on different datasets, which shows the high practicality and robustness of this method.
Authored by Juntuo Wang, Qiaochu Zhao, Dongheng Lin, Erick Purwanto, Ka Man
Metadata Discovery Problem - Open Educational Resources (OER) are educational materials that are available in different repositories such as Merlot, SkillsCommons, MIT OpenCourseWare, etc. The quality of metadata facilitates the search and discovery tasks of educational resources. This work evaluates the metadata quality of 4142 OER from SkillsCommons. We applied supervised machine learning algorithms (Support Vector Machine and Random Forest Classifier) for automatic classification of two metadata: description and material type. Based on our data and model, performances of a first classification effort is reported with the accuracy of 70\%.
Authored by Veronica Segarra-Faggioni, Audrey Romero-Pelaez
Measurement and Metrics Testing - In software regression testing, newly added test cases are more likely to fail, and therefore, should be prioritized for execution. In software regression testing for continuous integration, reinforcement learning-based approaches are promising and the RETECS (Reinforced Test Case Prioritization and Selection) framework is a successful application case. RETECS uses an agent composed of a neural network to predict the priority of test cases, and the agent needs to learn from historical information to make improvements. However, the newly added test cases have no historical execution information, thus using RETECS to predict their priority is more like ‘random’. In this paper, we focus on new test cases for continuous integration testing, and on the basis of the RETECS framework, we first propose a priority assignment method for new test cases to ensure that they can be executed first. Secondly, continuous integration is a fast iterative integration method where new test cases have strong fault detection capability within the latest periods. Therefore, we further propose an additional reward method for new test cases. Finally, based on the full lifecycle management, the ‘new’ additional rewards need to be terminated within a certain period, and this paper implements an empirical study. We conducted 30 iterations of the experiment on 12 datasets and our best results were 19.24\%, 10.67\%, and 34.05 positions better compared to the best parameter combination in RETECS for the NAPFD (Normalized Average Percentage of Faults Detected), RECALL and TTF (Test to Fail) metrics, respectively.
Authored by Fanliang Chen, Zheng Li, Ying Shang, Yang Yang
Measurement and Metrics Testing - The increase of smartphone users in Indonesia is the reason for various sectors to improve their services through mobile applications, including the healthcare sector. The healthcare sector is considered a critical sector as it stores various health data of its users classified as confidential. This is the basis for the need to conduct a security analysis for mobile health applications, which are widely used in Indonesia. MobSF (Mobile Security Framework) and MARA (Mobile Application Reverse Engineering and Analysis) Framework are mobile application security analysis methods capable of assessing security levels based on OWASP (Open Web Application Security Project) Mobile Top 10 2016 classification, CVSS (Common Vulnerability Scoring System) and CWE (Common Weakness Enumeration). It is expected that the test results with MobSF and MARA can provide a safety metric for mobile health applications as a means of safety information for users and application developers.
Authored by Dimas Priambodo, Guntur Ajie, Hendy Rahman, Aldi Nugraha, Aulia Rachmawati, Marcella Avianti
Measurement and Metrics Testing - FIPS 140-3 is the main standard defining security requirements for cryptographic modules in U.S. and Canada; commercially viable hardware modules generally need to be compliant with it. The scope of FIPS 140-3 will also expand to the new NIST Post-Quantum Cryptography (PQC) standards when migration from older RSA and Elliptic Curve cryptography begins. FIPS 140-3 mandates the testing of the effectiveness of “non-invasive attack mitigations”, or side-channel attack countermeasures. At higher security levels 3 and 4, the FIPS 140-3 side-channel testing methods and metrics are expected to be those of ISO 17825, which is based on the older Test Vector Leakage Assessment (TVLA) methodology. We discuss how to apply ISO 17825 to hardware modules that implement lattice-based PQC standards for public-key cryptography – Key Encapsulation Mechanisms (KEMs) and Digital Signatures. We find that simple “random key” vs. “fixed key” tests are unsatisfactory due to the close linkage between public and private components of PQC keypairs. While the general statistical testing approach and requirements can remain consistent with older public-key algorithms, a non-trivial challenge in creating ISO 17825 testing procedures for PQC is the careful design of test vector inputs so that only relevant Critical Security Parameter (CSP) leakage is captured in power, electromagnetic, and timing measurements.
Authored by Markku-Juhani Saarinen
Measurement and Metrics Testing - This paper belongs to a sequence of manuscripts that discuss generic and easy-to-apply security metrics for Strong PUFs. These metrics cannot and shall not fully replace in-depth machine learning (ML) studies in the security assessment of Strong PUF candidates. But they can complement the latter, serve in initial PUF complexity analyses, and are much easier and more efficient to apply: They do not require detailed knowledge of various ML methods, substantial computation times, or the availability of an internal parametric model of the studied PUF. Our metrics also can be standardized particularly easily. This avoids the sometimes inconclusive or contradictory findings of existing ML-based security test, which may result from the usage of different or non-optimized ML algorithms and hyperparameters, differing hardware resources, or varying numbers of challenge-response pairs in the training phase.
Authored by Fynn Kappelhoff, Rasmus Rasche, Debdeep Mukhopadhyay, Ulrich Rührmair
Measurement and Metrics Testing - Fuzz testing is an indispensable test-generation tool in software security. Fuzz testing uses automated directed randomness to explore a variety of execution paths in software, trying to expose defects such as buffer overflows. Since cyber-physical systems (CPS) are often safety-critical, testing models of CPS can also expose faults. However, while existing coverage-guided fuzz testing methods are effective for software, results can be disappointing when applied to CPS, where systems have continuous states and inputs are applied at different points in time.
Authored by Sanaz Sheikhi, Edward Kim, Parasara Duggirala, Stanley Bak
Measurement and Metrics Testing - Nowadays, attackers are increasingly using UseAfter-Free(UAF) vulnerabilities to create threats against software security. Existing static approaches for UAF detection are capable of finding potential bugs in the large code base. In most cases, analysts perform manual inspections to verify whether the warnings detected by static analysis are real vulnerabilities. However, due to the complex constraints of constructing UAF vulnerability, it is very time and cost-intensive to screen all warnings. In fact, many warnings should be discarded before the manual inspection phase because they are almost impossible to get triggered in real-world, and it is often overlooked by current static analysis techniques.
Authored by Haolai Wei, Liwei Chen, Xiaofan Nie, Zhijie Zhang, Yuantong Zhang, Gang Shi
Measurement and Metrics Testing - Software testing is one of the most critical and essential processes in the software development life cycle. It is the most significant aspect that affects product quality. Quality and service are critical success factors, particularly in the software business development market. As a result, enterprises must execute software testing and invest resources in it to ensure that their generated software products meet the needs and expectations of end-users. Test prioritization and evaluation are the key factors in determining the success of software testing. Test suit coverage metrics are commonly used to evaluate the testing process. Soft Computing techniques like Genetic Algorithms and Particle Swarm Optimization have gained prominence in various aspects of testing. This paper proposes an automated Genetic Algorithm approach to prioritizing the test cases and the evaluation through code coverage metrics with the Coverlet tool. Coverlet is a.NET code coverage tool that works across platforms and supports line, branch, and method coverage. Coverlet gathers data from Cobertura coverage test runs, which are then utilized to generate reports. Resultant test suits generated were validated and analyzed and have had significant improvement over the generations.
Authored by Baswaraju Swathi
Measurement and Metrics Testing - Due to the increasing complexity of modern heterogeneous System-on-Chips (SoC) and the growing vulnerabilities, security risk assessment and quantification is required to measure the trustworthiness of a SoC. This paper describes a systematic approach to model the security risk of a system for malicious hardware attacks. The proposed method uses graph analysis to assess the impact of an attack and the Common Vulnerability Scoring System (CVSS) is used to quantify the security level of the system. To demonstrate the applicability of the proposed metric, we consider two open source SoC benchmarks with different architectures. The overall risk is calculated using the proposed metric by computing the exploitability and impact of attack on critical components of a SoC.
Authored by Sujan Saha, Joel Mbongue, Christophe Bobda
Measurement and Metrics Testing - We continue to tackle the problem of poorly defined security metrics by building on and improving our previous work on designing sound security metrics. We reformulate the previous method into a set of conditions that are clearer and more widely applicable for deriving sound security metrics. We also modify and enhance some concepts that led to an unforeseen weakness in the previous method that was subsequently found by users, thereby eliminating this weakness from the conditions. We present examples showing how the conditions can be used to obtain sound security metrics. To demonstrate the conditions’ versatility, we apply them to show that an aggregate security metric made up of sound security metrics is also sound. This is useful where the use of an aggregate measure may be preferred, to more easily understand the security of a system.
Authored by George Yee