Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- ArticleOctober 2024
Emergence in Multi-agent Systems: A Safety Perspective
- Philipp Altmann,
- Julian Schönberger,
- Steffen Illium,
- Maximilian Zorn,
- Fabian Ritz,
- Tom Haider,
- Simon Burton,
- Thomas Gabor
Leveraging Applications of Formal Methods, Verification and Validation. Rigorous Engineering of Collective Adaptive SystemsPages 104–120https://doi.org/10.1007/978-3-031-75107-3_7AbstractEmergent effects can arise in multi-agent systems (MAS) where execution is decentralized and reliant on local information. These effects may range from minor deviations in behavior to catastrophic system failures. To formally define these effects, ...
- ArticleOctober 2024
Establishing the Foundation for Out-of-Distribution Detection in Monument Classification Through Nested Dichotomies
AbstractThis paper introduces a hierarchical approach utilizing nested dichotomies to enhance the MonuMAI framework designed for architectural image classification. The study focuses on developing a foundational layer dedicated to distinguishing between ...
- ArticleOctober 2024
An In-depth Analysis of Jailbreaking Through Domain Characterization of LLM Training Sets
AbstractResearch on large language models (LLMs) is a prominent field in open-world machine learning. Despite their significant capabilities in natural language processing, LLMs face several challenges that must be overcome, namely, consistency, ...
- ArticleSeptember 2024
Bridging the Reality Gap: Assurable Simulations for an ML-Based Inspection Drone Flight Controller
Computer Safety, Reliability, and Security. SAFECOMP 2024 WorkshopsPages 412–424https://doi.org/10.1007/978-3-031-68738-9_33AbstractAutonomous drones have been proposed for many industrial inspection roles including wind farms, railway lines and solar farms. They have many potential benefits, including accessing difficult to reach locations, reduced physical risk to operators ...
- research-articleSeptember 2024
CoSec: On-the-Fly Security Hardening of Code LLMs via Supervised Co-decoding
ISSTA 2024: Proceedings of the 33rd ACM SIGSOFT International Symposium on Software Testing and AnalysisPages 1428–1439https://doi.org/10.1145/3650212.3680371Large Language Models (LLMs) specialized in code have shown exceptional proficiency across various programming-related tasks, particularly code generation. Nonetheless, due to its nature of pretraining on massive uncritically filtered data, prior studies ...
- research-articleJuly 2024
An AI System Evaluation Framework for Advancing AI Safety: Terminology, Taxonomy, Lifecycle Mapping
AIware 2024: Proceedings of the 1st ACM International Conference on AI-Powered SoftwarePages 74–78https://doi.org/10.1145/3664646.3664766The advent of advanced AI underscores the urgent need for comprehensive safety evaluations, necessitating collaboration across communities (i.e., AI, software engineering, and governance). However, divergent practices and terminologies across these ...
- ArticleOctober 2023
Towards a Certified Proof Checker for Deep Neural Network Verification
Logic-Based Program Synthesis and TransformationPages 198–209https://doi.org/10.1007/978-3-031-45784-5_13AbstractRecent developments in deep neural networks (DNNs) have led to their adoption in safety-critical systems, which in turn has heightened the need for guaranteeing their safety. These safety properties of DNNs can be proven using tools developed by ...
- research-articleAugust 2023
User Tampering in Reinforcement Learning Recommender Systems
AIES '23: Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and SocietyPages 58–69https://doi.org/10.1145/3600211.3604669In this paper, we introduce new formal methods and provide empirical evidence to highlight a unique safety concern prevalent in reinforcement learning (RL)-based recommendation algorithms – ’user tampering.’ User tampering is a situation where an RL-...
- ArticleNovember 2021
Mutation Testing of Reinforcement Learning Systems
Dependable Software Engineering. Theories, Tools, and ApplicationsPages 143–160https://doi.org/10.1007/978-3-030-91265-9_8AbstractReinforcement Learning (RL), one of the most active research areas in artificial intelligence, focuses on goal-directed learning from interaction with an uncertain environment. RL systems play an increasingly important role in many aspects of ...
- posterJuly 2020
Safer reinforcement learning through evolved instincts
GECCO '20: Proceedings of the 2020 Genetic and Evolutionary Computation Conference CompanionPages 77–78https://doi.org/10.1145/3377929.3389946An important goal in reinforcement learning is to create agents that can quickly adapt to new goals but at the same time avoid situations that might cause damage to themselves or their environments. One way agents learn is through exploration mechanisms,...
- ArticleSeptember 2019
Risk-averse Distributional Reinforcement Learning: A CVaR Optimization Approach
IJCCI 2019: Proceedings of the 11th International Joint Conference on Computational IntelligencePages 412–423https://doi.org/10.5220/0008175604120423Conditional Value-at-Risk (CVaR) is a well-known measure of risk that has been directly equated to robustness, an important component of Artificial Intelligence (AI) safety. In this paper we focus on optimizing CVaR in the context of Reinforcement ...