This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| en:safeav:maps:validation [2026/03/30 11:14] – [Validation Approaches] airi | en:safeav:maps:validation [2026/03/31 10:14] (current) – [Assessment] airi | ||
|---|---|---|---|
| Line 14: | Line 14: | ||
| | Deterministic | Non-Deterministic | Conventional algorithms are deterministic in nature, and ML algorithms are fundamentally probabilistic in nature. | | | Deterministic | Non-Deterministic | Conventional algorithms are deterministic in nature, and ML algorithms are fundamentally probabilistic in nature. | | ||
| | Known Computational Complexity | Unknown Computational Complexity | Given the analyzable nature of conventional algorithms, one can build a model for computational complexity. This is not always possible for ML techniques, which may require testing to evaluate computational complexity. | | | Known Computational Complexity | Unknown Computational Complexity | Given the analyzable nature of conventional algorithms, one can build a model for computational complexity. This is not always possible for ML techniques, which may require testing to evaluate computational complexity. | | ||
| + | |||
| + | //**Table 1: Contrast of Conventional and Machine Learning Algorithms**// | ||
| + | |||
| + | The introduction of AI as a replacement for traditional software introduces significant validation issues (table 1). Significantly, | ||
| + | |||
| + | Safety standards across automotive, marine, airborne, and space domains are now evolving to address the introduction of AI/ | ||
| + | |||
| + | This remainder of this section presents a practical, simulation-driven illustration to validating the perception, mapping (HD maps/ | ||
| Line 21: | Line 29: | ||
| We decompose the stack into Perception (object detection/ | We decompose the stack into Perception (object detection/ | ||
| - | ====== Perception Validation ====== | + | ====== Perception Validation |
| Line 36: | Line 44: | ||
| Figure 1 explains object comparison. Green boxes are shown for objects captured by ground truth, while Red boxes are shown for objects detected by the AV stack. Threshold-based rules are designed to compare the objects. It is expected to provide specific indicators of detectable vehicles in different ranges for safety and danger areas. | Figure 1 explains object comparison. Green boxes are shown for objects captured by ground truth, while Red boxes are shown for objects detected by the AV stack. Threshold-based rules are designed to compare the objects. It is expected to provide specific indicators of detectable vehicles in different ranges for safety and danger areas. | ||
| - | ====== Mapping / Digital-Twin Validation ====== | + | ====== Mapping / Digital-Twin Validation |
| Line 43: | Line 51: | ||
| Key checks include lane topology fidelity versus survey, geo-consistency in centimeters, | Key checks include lane topology fidelity versus survey, geo-consistency in centimeters, | ||
| - | ====== Localization Validation ====== | + | ====== Localization Validation |
| Line 69: | Line 77: | ||
| A two-stage workflow balances coverage and realism. First, use LF tools (e.g., planner-in-the-loop with simplified sensors and traffic) to sweep large grids of logical scenarios and identify risky regions in parameter space (relative speed, initial gap, occlusion level). Then, promote the most informative concrete scenarios to HF simulation with photorealistic sensors for end-to-end validation of perception and localization interactions. Where appropriate, | A two-stage workflow balances coverage and realism. First, use LF tools (e.g., planner-in-the-loop with simplified sensors and traffic) to sweep large grids of logical scenarios and identify risky regions in parameter space (relative speed, initial gap, occlusion level). Then, promote the most informative concrete scenarios to HF simulation with photorealistic sensors for end-to-end validation of perception and localization interactions. Where appropriate, | ||
| + | |||
| + | ====== Summary ====== | ||
| + | |||
| + | The chapter develops a comprehensive view of perception, mapping, and localization as the foundation of autonomous systems, emphasizing how modern autonomy builds on both historical automation (e.g., autopilots across domains) and recent advances in AI. It explains how perception converts raw sensor data—across cameras, LiDAR, radar, and acoustic systems—into structured understanding through object detection, sensor fusion, and scene interpretation. A key theme is that no single sensor is sufficient; instead, robust autonomy depends on multi-modal sensor fusion, probabilistic estimation, and careful calibration to manage uncertainty. The chapter also highlights the transformative role of AI, particularly deep learning, in enabling scalable perception and scene understanding, | ||
| + | |||
| + | A second major focus is on sources of instability and validation, where the chapter connects environmental effects (weather, electromagnetic interference), | ||
| + | |||
| + | ====== Assessment ====== | ||
| + | |||
| + | ^ # ^ Project Title ^ Description ^ Learning Objectives ^ | ||
| + | | 1 | Multi-Sensor Perception Benchmarking | Build a perception pipeline using at least two sensor modalities (e.g., camera + LiDAR or radar). Evaluate object detection performance under varying conditions (lighting, weather, occlusion) using real or simulated datasets. | Understand strengths/ | ||
| + | | 2 | ODD-Driven Scenario Generation & Validation Study | Define an Operational Design Domain (ODD) for an autonomous system (e.g., urban driving, coastal navigation). Generate a set of test scenarios (including edge cases) and validate system performance using simulation tools. | Define and scope an ODD. Develop scenario-based testing strategies. Understand coverage and edge-case generation. Link scenarios to safety outcomes. | | ||
| + | | 3 | Sensor Failure and Degradation Analysis | Simulate sensor failures (e.g., camera blackout, GNSS loss, radar noise) and analyze system-level impact on perception, localization, | ||
| + | | 4 | AI vs Conventional Algorithm Validation Study | Compare a traditional perception algorithm (e.g., rule-based or classical ML) with a deep learning model on the same dataset. Analyze differences in performance, | ||
| + | | 5 | End-to-End V&V Framework Design (Digital Twin) | Design a validation framework for perception, mapping, and localization using simulation (digital twin). Include KPIs, test conditions (e.g., ISO 26262, SOTIF), simulations, | ||
| + | |||
| + | |||
| + | |||
| + | |||
| + | |||
| + | |||