default search action
SafeAI@AAAI 2022: Virtual Event
- Gabriel Pedroza, José Hernández-Orallo, Xin Cynthia Chen, Xiaowei Huang, Huáscar Espinoza, Mauricio Castillo-Effen, John A. McDermid, Richard Mallah, Seán Ó hÉigeartaigh:
Proceedings of the Workshop on Artificial Intelligence Safety 2022 (SafeAI 2022) co-located with the Thirty-Sixth AAAI Conference on Artificial Intelligence (AAAI2022), Virtual, February, 2022. CEUR Workshop Proceedings 3087, CEUR-WS.org 2022
Session 1: Bias, Fairness and Value Alignment
- Hal Ashton, Matija Franklin:
The Problem of Behaviour and Preference Manipulation in AI Systems. - Ignacio Serna, Daniel DeAlcala, Aythami Morales Moreno, Julian Fiérrez, Javier Ortega-Garcia:
IFBiD: Inference-Free Bias Detection. - Preston Putzel, Scott Lee:
Blackbox Post-Processing for Multiclass Fairness.
Session 2: Interpretability, Accountability
- Soyeon Jung, Ransalu Senanayake, Mykel J. Kochenderfer:
A Gray Box Model for Characterizing Driver Behavior. - Hal Ashton:
Defining and Identifying the Legal Culpability of Side Effects using Causal Graphs.
Session 3: Robustness and Uncertainty
- Mingjun Ma, Dehui Du, Yuanhao Liu, Yanyun Wang, Yiyang Li:
Efficient Adversarial Sequence Generation for RNN with Symbolic Weighted Finite Automata. - Pascal Gerber, Lisa Jöckel, Michael Kläs:
A Study on Mitigating Hard Boundaries of Decision-Tree-based Uncertainty Estimates for AI Models. - Deebul S. Nair, Nico Hochgeschwender, Miguel A. Olivares-Méndez:
Maximum Likelihood Uncertainty Estimation: Robustness to Outliers.
Session 4: Safe Reinforcement Learning
- Jin Woo Ro, Gerald Lüttgen, Diedrich Wolter:
Reinforcement Learning With Imperfect Safety Constraints. - Chloe He, Borja G. León, Francesco Belardinelli:
Do Androids Dream of Electric Fences? Safety-Aware Reinforcement Learning with Latent Shielding. - Zikang Xiong, Ishika Agarwal, Suresh Jagannathan:
HiSaRL: A Hierarchical Framework for Safe Reinforcement Learning. - Mathieu Godbout, Maxime Heuillet, Sharath Chandra Raparthy, Rupali Bhati, Audrey Durand:
A Game-Theoretic Perspective on Risk-Sensitive Reinforcement Learning.
Session 5: A. I. Testing, Assessment
- Adrian Schwaiger, Kristian Schwienbacher, Karsten Roscher:
Beyond Test Accuracy: The Effects of Model Compression on CNNs. - Ann-Katrin Reuel, Mark Koren, Anthony Corso, Mykel J. Kochenderfer:
Using Adaptive Stress Testing to Identify Paths to Ethical Dilemmas in Autonomous Systems.
Poster Papers
- Sheila Alemany, Niki Pissinou:
The Dilemma Between Data Transformations and Adversarial Robustness for Time Series Application Systems. - John Mern, Sidhart Krishnan, Anil Yildiz, Kyle Hatch, Mykel J. Kochenderfer:
Interpretable Local Tree Surrogate Policies. - Peter Barnett, John Burden:
Oases of Cooperation: An Empirical Evaluation of Reinforcement Learning in the Iterated Prisoner's Dilemma. - Fangqi Li, Lei Yang, Shilin Wang, Alan Wee-Chung Liew:
Leveraging Multi-task Learning for Umambiguous and Flexible Deep Neural Network Watermarking. - Prajit T. Rajendran, Huáscar Espinoza, Agnès Delaborde, Chokri Mraidha:
Human-in-the-loop Learning for Safe Exploration through Anomaly Prediction and Intervention. - Leopold Müller, Lars Böcking, Michael Färber:
Safety Aware Reinforcement Learning by Identifying Comprehensible Constraints in Expert Demonstrations. - Juliette Mattioli, Gabriel Pedroza, Souhaiel Khalfaoui, Bertrand Leroy:
Combining Data-Driven and Knowledge-Based AI Paradigms for Engineering AI-Based Safety-Critical Systems. - Poulami Sinhamahapatra, Rajat Koner, Karsten Roscher, Stephan Günnemann:
Is it all a cluster game? - Exploring Out-of-Distribution Detection based on Clustering in the Embedding Space. - Saeed Bakhshi Germi, Esa Rahtu:
A Practical Overview of Safety Concerns and Mitigation Methods for Visual Deep Learning Algorithms. - Michal Filipiuk, Vasu Singh:
Comparing Vision Transformers and Convolutional Nets for Safety Critical Systems.
Special Sessions
- Edoardo Manino, Danilo S. Carvalho, Yi Dong, Julia Rozanova, Xidan Song, Mustafa A. Mustafa, André Freitas, Gavin Brown, Mikel Luján, Xiaowei Huang, Lucas C. Cordeiro:
EnnCore: End-to-End Conceptual Guarding of Neural Architectures. - Bertrand Braunschweig, Rodolphe Gelin, François Terrier:
The wall of safety for AI: approaches in the Confiance.ai program.
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.