ML Models for Safety-Critical Systems with Lucas García - #705

ML Models for Safety-Critical Systems with Lucas García - #705

Update: 2024-10-14
Share

Digest

This episode of the podcast delves into the complex world of integrating AI models into safety-critical systems, a topic of growing importance in various industries. The guest, Lucas Garcia, a Principal Product Manager for Deep Learning at MathWorks, shares his insights based on his presentation at the NURPS conference. The discussion begins by clarifying the distinction between verification and validation in software development, emphasizing their crucial role in safety-critical systems. Lucas introduces the traditional V model and the W cycle, a specialized development cycle tailored for AI-based systems. He then provides a concrete example of a battery state of charge estimator, illustrating how verification and validation steps are applied in practice. The importance of data management and learning process verification is highlighted. The episode addresses the question of why AI models are preferred over traditional methods in safety-critical systems. Lucas discusses the limitations of traditional models and the advantages of AI in terms of generalization and data-driven approaches. The W cycle workflow is broken down in detail, outlining the specific steps involved in data management, learning process management, model training, learning process verification, implementation, and integration. Lucas further discusses the role of external specifications and regulations in safety-critical systems, highlighting the need for certification and the evolving standards for AI-based systems. The use of formal methods in AI is explored, emphasizing their importance in proving the robustness of AI models. The Deep Learning Toolbox Verification Library, which uses formal verification methods to test the stability of deep learning networks, is introduced. The episode clarifies the role of abstract transformers in the verification process, explaining that they are modeling tools used for verification, not production models. The limitations of formal methods and the need for alternative approaches like scenario-based testing and runtime monitoring are discussed. Lucas discusses neuron coverage, a technique for assessing the coverage of neurons in a neural network. He acknowledges the challenges of connecting neuron coverage to specific objectives and requirements. The use of benchmarks in industry, highlighting the ML Leap project and the importance of real-world use cases, is explored. The use of transformers in safety-critical systems is discussed, highlighting their potential for time series modeling and embedded AI. The challenges of applying large transformers and LLMs in safety-critical environments are acknowledged. The challenges of benchmarking in safety-critical AI are discussed, highlighting the role of certification bodies like ASA and the ML Leap project. A real-world example of a runway sign classifier is provided, illustrating the incremental approach to certification. The concept of architectural mitigation, a technique for achieving higher criticality levels by combining dissimilar components, is explained. This is illustrated with the example of two independent runway sign classifiers, achieving a DALC certification. Lucas introduces constrained deep learning, an approach to training neural networks by incorporating domain-specific constraints. The benefits of this approach in terms of guaranteeing desirable properties like monotonicity and convexity are discussed. The trade-offs and challenges of using convex neural networks are discussed. The potential benefits in terms of reliability and robustness are emphasized, but the challenges of convergence speed and the need for more complex network architectures are acknowledged. The question of why convex neural networks are not more widely used is addressed. The challenges of convergence speed and the lack of a comprehensive framework for constrained deep learning are discussed. The need for more research and development in this area is highlighted. The potential for developing a library of convex models, similar to existing libraries for non-convex models, is discussed. The need for further research and development to make this a reality is emphasized. Future advancements could simplify the verification process by baking desired properties into the network architecture.

Outlines

00:00:00
Introduction and Safety-Critical AI

The episode introduces the topic of incorporating AI models into safety-critical systems, highlighting the importance of verification, validation, and the W cycle workflow for ensuring reliability.

00:01:33
Introducing Lucas Garcia, Expert in Deep Learning

The host introduces Lucas Garcia, Principal Product Manager for Deep Learning at MathWorks, as the guest for the episode.

00:01:58
Verification and Validation in Safety-Critical Systems

The episode focuses on the topic of incorporating AI models into safety critical systems, drawing from Lucas's presentation at the NURPS conference. Lucas explains the differences between verification and validation in software development, emphasizing their importance in safety critical systems. He introduces the traditional V model and the W cycle for AI-based systems.

00:07:51
Concrete Example: Battery State of Charge Estimator

Lucas provides a concrete example of a battery state of charge estimator, illustrating how verification and validation steps are applied in practice. He highlights the importance of data management and learning process verification.

00:10:32
Why Use AI Models in Safety Critical Systems?

Lucas addresses the question of why AI models are preferred over traditional methods in safety critical systems. He discusses the limitations of traditional models and the advantages of AI in terms of generalization and data-driven approaches.

00:16:30
The W Cycle Workflow in Detail

Lucas breaks down the W cycle workflow, outlining the specific steps involved in data management, learning process management, model training, learning process verification, implementation, and integration.

00:23:37
Certification and Regulations for AI-Based Systems

Lucas discusses the role of external specifications and regulations in safety critical systems, highlighting the need for certification and the evolving standards for AI-based systems.

00:25:19
Formal Methods and Robustness in AI

Lucas explores the use of formal methods in AI, emphasizing their importance in proving the robustness of AI models. He introduces the Deep Learning Toolbox Verification Library, which uses formal verification methods to test the stability of deep learning networks.

Keywords

Safety Critical Systems


Systems where failure could lead to significant harm, such as aircraft, medical devices, and nuclear power plants. These systems require rigorous verification and validation processes.

Verification and Validation (V&V)


Processes used to ensure that software meets its intended design and functionality. Verification checks internal consistency, while validation evaluates the system's performance in real-world conditions.

W Cycle


A development cycle specifically designed for AI-based systems, emphasizing data management, learning process management, model training, learning process verification, implementation, and integration.

Formal Methods


Mathematical techniques used to prove the correctness and robustness of software systems. In AI, formal methods can be used to verify the stability and reliability of neural networks.

Deep Learning Toolbox Verification Library


A library developed by MathWorks that uses formal verification methods to test the robustness of deep learning networks. It allows users to mathematically prove the stability of their models.

Neuron Coverage


A technique for assessing the coverage of neurons in a neural network during testing. It aims to ensure that a sufficient number of neurons are activated, providing some guarantees about the network's performance.

Constrained Deep Learning


An approach to training neural networks by incorporating domain-specific constraints into the learning process. This can guarantee desirable properties like monotonicity, convexity, and boundedness.

Convex Neural Networks


Neural networks designed to have convex output functions. This can simplify optimization problems and improve the reliability of the model.

Architectural Mitigation


A technique for achieving higher criticality levels in safety critical systems by combining dissimilar components. This involves using independent systems with different data sets, architectures, and training frameworks.

ML Leap Project


A project aimed at providing a comprehensive guide to the machine learning approval process for safety critical systems. It outlines the steps involved in verification, validation, and certification.

Q&A

  • What are the key differences between verification and validation in software development, especially in the context of safety critical systems?

    Verification ensures that the software correctly implements the intended design, while validation evaluates the system's performance in real-world conditions. Both are crucial for safety critical systems, where failure can have severe consequences.

  • Why are AI models becoming increasingly popular in safety critical systems, despite the challenges they present?

    AI models offer advantages in terms of generalization and data-driven approaches, which can overcome the limitations of traditional models. However, their use requires rigorous verification and validation processes to ensure safety and reliability.

  • What are the main steps involved in the W cycle workflow for AI-based systems?

    The W cycle workflow includes data management, learning process management, model training, learning process verification, implementation, and integration. Each step is crucial for ensuring the safety and reliability of the AI system.

  • How can formal methods be used to improve the robustness of AI models?

    Formal methods provide mathematical guarantees about the behavior of AI models, particularly in terms of their stability and reliability. They can be used to test the robustness of neural networks against adversarial attacks and other perturbations.

  • What are the benefits and challenges of using constrained deep learning for safety critical systems?

    Constrained deep learning can guarantee desirable properties like monotonicity and convexity, improving the reliability and robustness of AI models. However, it can lead to slower convergence and may require more complex network architectures.

  • What are the potential future directions for research and development in the area of convex neural networks?

    Future research could focus on developing a library of convex models, similar to existing libraries for non-convex models. This would simplify the process of building and deploying convex networks. Additionally, advancements in network architecture could further simplify the verification process by baking desired properties into the design.

Show Notes

Today, we're joined by Lucas García, principal product manager for deep learning at MathWorks to discuss incorporating ML models into safety-critical systems. We begin by exploring the critical role of verification and validation (V&V) in these applications. We review the popular V-model for engineering critical systems and then dig into the “W” adaptation that’s been proposed for incorporating ML models. Next, we discuss the complexities of applying deep learning neural networks in safety-critical applications using the aviation industry as an example, and talk through the importance of factors such as data quality, model stability, robustness, interpretability, and accuracy. We also explore formal verification methods, abstract transformer layers, transformer-based architectures, and the application of various software testing techniques. Lucas also introduces the field of constrained deep learning and convex neural networks and its benefits and trade-offs.


The complete show notes for this episode can be found at https://twimlai.com/go/705.

Comments 
In Channel
loading

Table of contents

00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

ML Models for Safety-Critical Systems with Lucas García - #705

ML Models for Safety-Critical Systems with Lucas García - #705

Sam Charrington