Table of Contents

Authors

The list of book contributors is presented below.

Tallinn University of Technology (TalTech)
Silesian University of Technology
Riga Technical University
ITT Group
ProDron
Czech Technical University
Technical Editor
External Contributors
Reviewers

 

Project Information

This content was implemented under the project: SafeAV - Harmonizations of Autonomous Vehicle Safety Validation and Verification for Higher Education.

Project number: 2024-1-EE01-KA220-HED-000245441.

Consortium

Erasmus+ Disclaimer
This project has been funded with support from the European Commission.
This publication reflects the views only of the author, and the Commission cannot be held responsible for any use which may be made of the information contained therein.

Copyright Notice
This content was created by the SafeAV Consortium 2024–2027.
The content is copyrighted and distributed under CC BY-NC Creative Commons Licence and is free for non-commercial use.

CC BY-NC

Introduction

Exercises

Autonomous Systems

Productization Lessons and Assessments:

Key lessons for productization include:

  1. Engineers must understand their products operate inside a governance structure consisting of laws, regulations, and standards.
  2. In the case of autonomy, there are many historical standards, but standard development is also under development.
  3. A very key aspect of product design is the expectation function for the product. This expectation function is key to communication from a marketing perspective and also from a legal liability perspective.

Exercises

Section Project Title Objective Technical Scope Deliverables Learning Outcomes
2.0 Autonomous Systems Fundamentals Cross-Domain Autonomy Architecture Design Understand how autonomy architectures differ across ground, airborne, marine, and space domains. Define sensing, compute, control, and communication architecture for one system in each domain; analyze environmental constraints and failure modes. Architecture diagrams (5–10 page report). Understand how environment drives autonomy architecture, safety requirements, and validation strategy.
2.1 Definitions, Classification, and Levels of Autonomy Expectation Function and Autonomy Level Classification Learn how autonomy levels define responsibility and system capability. Select a real-world autonomous system; classify using SAE, UAV, MASS, or ALFUS frameworks; define expectation function and responsibility allocation. Autonomy classification report; expectation function definition; responsibility matrix. Understand autonomy levels as technical, operational, and legal constructs.
2.2 Legal, Ethical, and Regulatory Frameworks Autonomous System Liability Case Study Understand relationship between validation, expectation functions, and legal liability. Analyze a historical accident scenario; determine liability; evaluate compliance with ISO, SAE, FAA, or NASA frameworks. Legal liability analysis report; governance compliance evaluation. Understand how governance frameworks assign responsibility and require validation evidence.
2.3 Introduction to Validation and Verification Operational Design Domain (ODD) and V&V Development Learn how to construct a high-level validation plan for an autonomous system. Define ODD; generate validation scenarios; define correctness criteria; develop validation workflow including simulation and physical tests. Complete high-level V&V plan document; ODD, coverage, and correctness criteria. Understand structure of validation plans and role of ODD, coverage, and correctness criteria.
2.4 Physics-Based vs Decision-Based Validation Comparative Validation of Deterministic vs AI Systems Understand validation complexity differences between physics-based and AI-based systems. Construct a V&V plan for a physics-based function and also for a digital function. Comparative report on testing methodologies. Understand fundamental differences between validating physics-based and AI-based systems.
2.5 Validation Requirements Across Domains Domain-Specific Validation Design Learn how validation requirements differ across ground, airborne, marine, and space domains. Select domain; define hazards, validation methods, certification requirements, and safety argument structure. Domain-specific validation plan; hazard analysis; certification pathway analysis. Understand domain-specific validation constraints and certification requirements.

Hardware and Sensing Technologies

Assessment:

# Assessment Theme Learning Objective Deliverable
1 Evolution of Electronics in Autonomy Understand how semiconductors and electronics transformed ground, airborne, marine, and space systems from isolated functions into integrated autonomous architectures. Paper: comparative essay, or Project: presentation/timeline showing the historical evolution across the four domains.
2 Sensor Fusion Design Explain why autonomous systems require multiple complementary sensors and how sensing choices depend on mission, environment, redundancy, and compute constraints. Paper: analysis of a sensor stack in one domain, or Project: design a sensing architecture with justification for each sensor and compute element.
3 Safety and Governance Analyze how standards and governance frameworks shape hardware design, certification, and risk management in autonomous systems. Paper: standards comparison essay, or Project: briefing/chart mapping ISO 26262, IEC 61508, DO-254, and related frameworks to different domains.
4 Validation and Verification Evaluate how validation, timing, KPIs, scenario-based testing, and simulation contribute to trustworthy autonomy validation beyond simple model-level accuracy. Paper: methodology critique, or Project: create a validation plan with KPIs, scenarios, and simulation/track-test workflow.
5 Supply Chain and Productization Understand how supply chain resilience, certification burden, EMI/EMC compliance, cybersecurity, and obsolescence affect real-world deployment of autonomous systems. Paper: case-based analysis, or Project: risk-mitigation plan for launching and supporting an autonomous product.

Software Systems and Middleware

Assessment:

# Assessment Title Description (Project / Report) Learning Objectives
1 Evolution of Programmable Systems Write a report tracing the evolution from fixed-function hardware to programmable systems (configuration, FPGA, microprocessors) and the abstraction of software as an abstraction. Include historical milestones and examples. Understand the transition from hardware-centric to software-defined systems. Explain key programming paradigms (configuration, assembly, high-level programming). Analyze the role of abstraction architecture (e.g., system stack).
2 Cyber-Physical Software Stack Analysis Develop a structured report analyzing a real-world CPS (e.g., automotive ADAS, UAV, or spacecraft). Map its software stack (HAL, RTOS, middleware, applications) and explain how each layer contributes to overall system functionality. Identify layers in CPS software architectures. Explain the role of RTOS, middleware, and HAL. Analyze real-time and safety constraints in system design.
3 IT vs CPS Supply Chain Comparison Study Produce a comparative analysis of hardware and software supply chains in IT vs CPS, with focus on lifecycle management, dependencies, and update strategies. Include risks and trade-offs. Compare IT and CPS development ecosystems. Evaluate the impact of “innovation cycles” in CPS (cost, obsolescence, certification). Assess risks (safety, cybersecurity) and benefits (flexibility, innovation).
4 Safety Verification and Validation Framework Write a report comparing software validation approaches in IT and CPS, focusing on simulation/emulation (MIL, SIL, HIL) and safety standards (e.g., ISO 26262, DO-178C). Include a case study. Understand verification vs validation in different domains. Explain simulation/emulation methods in CPS. Analyze how safety standards shape software development.
5 Software-Defined System Proposal Develop a conceptual design for a “software-defined” system (e.g., vehicle, drone, or marine system). Describe architecture, update model (OTA), software stack, and lifecycle management approach. Apply concepts of software-defined systems. Design layered, modular architectures. Integrate lifecycle, update, and maintainability considerations.

Perception, Mapping and Localisation

# Project Title Description Learning Objectives
1 Multi-Sensor Perception Benchmarking Build a perception pipeline using at least two sensor modalities (e.g., camera + LiDAR or radar). Evaluate object detection performance under varying conditions (lighting, weather, occlusion) using real or simulated datasets. Understand strengths/limitations of different sensors. Apply sensor fusion concepts. Evaluate detection metrics (precision/recall, distance sensitivity). Analyze environmental impacts on perception.
2 ODD-Driven Scenario Generation & Validation Study Define an Operational Design Domain (ODD) for an autonomous system (e.g., urban driving, coastal navigation). Generate a set of test scenarios (including edge cases) and validate system performance using simulation tools. Define and scope an ODD. Develop scenario-based testing strategies. Understand coverage and edge-case generation. Link scenarios to safety outcomes.
3 Sensor Failure and Degradation Analysis Simulate sensor failures (e.g., camera blackout, GNSS loss, radar noise) and analyze system-level impact on perception, localization, and safety metrics (e.g., time-to-collision). Understand failure modes across sensor types. Evaluate system robustness and redundancy. Apply fault injection techniques. Connect sensor degradation to safety risks.
4 AI vs Conventional Algorithm Validation Study Compare a traditional perception algorithm (e.g., rule-based or classical ML) with a deep learning model on the same dataset. Analyze differences in performance, interpretability, and validation challenges. Distinguish deterministic vs probabilistic systems. Understand validation challenges of AI/ML. Evaluate explainability and traceability. Assess implications for safety certification.
5 End-to-End V&V Framework Design (Digital Twin) Design a validation framework for perception, mapping, and localization using simulation (digital twin). Include KPIs, test conditions (e.g., ISO 26262, SOTIF), simulations, and linkage to safety standards. Design system-level V&V strategies. Define measurable KPIs for autonomy. Understand simulation and digital twin roles. Connect numerical validation to safety standards.

Control, Planning, and Decision-Making

Assessments:

# Project Title Description Learning Objectives
1 Classical vs AI Control Benchmark Study Implement and compare a classical controller (e.g., PID or LQR) with an AI-based controller (e.g., reinforcement learning) for a simplified vehicle model in simulation. Evaluate performance under nominal and disturbed conditions. - Understand differences between model-based and data-driven control
- Analyze stability, robustness, and interpretability trade-offs
- Evaluate controller performance under uncertainty and disturbances
2 Behavioral & Motion Planning Stack Design Design a hierarchical autonomy stack that includes a behavioral layer (FSM or behavior tree) and a motion planner (A*, RRT*, or MPC). Apply it to a scenario such as lane change or obstacle avoidance. * Distinguish between behavioral decision-making and motion planning
* Implement planning algorithms under constraints
* Understand integration between perception, planning, and control
3 Scenario-Based Validation Framework Develop a scenario-based testing framework using parameterized scenarios (e.g., varying speeds, distances, agent behaviors). Use a simulator to evaluate planning/control performance across these scenarios. * Apply design-of-experiments (DOE) to autonomy validation
* Define and measure safety metrics (e.g., TTC, collision rate)
* Understand coverage and edge-case testing challenges
4 Digital Twin & Multi-Fidelity Simulation Study Build a simplified digital twin of a vehicle and environment. Perform validation using both low-fidelity and high-fidelity simulation setups, comparing results and identifying discrepancies. * Understand role of digital twins in V&V
* Analyze trade-offs between simulation fidelity and scalability
* Quantify sim-to-real gaps and their implications
5 Formal Methods for Safety Validation Define safety requirements using a formal specification approach (e.g., temporal logic or rule-based constraints). Apply these to simulation traces and identify violations or edge cases. * Translate safety requirements into formal, testable properties
* Use formal methods for falsification and validation
* Understand limitations of simulation without formal rigor

Human-Machine Communication

Autonomy Validation Tools

Use cases

Use-case requirements

This chapter defines the minimum requirements for the use cases that will be developed, validated, and comparatively assessed within the SafeAV framework (AV shuttle, F1TENTH, mobile robot, UAV). The goal is to align learning outcomes with technical, safety, and regulatory constraints, and to ensure smooth integration with the selected toolchain (e.g., Autoware/ROS2, simulators, and V&V tools). Requirements cover system boundaries and assumptions, environment and scenario descriptions, data flows, performance and safety targets, acceptance criteria, and end-to-end traceability to course outcomes and WP-level KPIs. We explicitly emphasize compliance with relevant standards and regulations (e.g., UNECE; EASA, where applicable), educational reusability (SITL/HITL), and reproducibility: each use case must ship with standardized scenarios, test scripts, and evaluation report templates. Use cases are selected to cover a wide range of AV, both in the ground and air domains.

Conclusions and Decisions

In conclusion, the consortium has evaluated the available verification and validation frameworks based on current research, technical feasibility, and educational applicability. The resulting decisions reflect a balanced consideration of open-source maturity, interoperability, and relevance to the specific SafeAV use cases. The selected frameworks demonstrate strong community support, active ongoing development, and proven suitability for academic integration. Their open-source nature ensures transparency, adaptability, and long-term sustainability, while their functionality aligns closely with the technical and pedagogical goals defined for each use case. Use case No Name Description Selected Frameworks and Software Set Responsible partner #1 AV shuttle Simulation-based V&V use case for an autonomous shuttle used in education, prototyping, and pre-deployment testing. Requirements include multi-level simulation, scenario testing with OpenSCENARIO, Autoware.Universe integration, automated safety metrics, containerized deployment, and educational accessibility. Both planning and perception modules are targeted for validation. Autoware.Universe (software under test), AWSIM (digital twin, sensor simulation, Autoware integration), CARLA (high-fidelity perception & planning validation, synthetic data), Rosbag Replay (real-world perception regression), Scenario Simulator v2 (scenario-based planning validation), CommonRoad (planning benchmarking), SUMO (traffic co-simulation) Scenic (Scenario generator for CARLA) TalTech #2 F1Tenth The F1TENTH use case demonstrates the application of open-source V&V frameworks for validating autonomous driving functions on a 1/10-scale research platform. Verification focuses on trajectory tracking, ensuring that the vehicle follows safe and stable paths without collisions or oscillations when encountering static or dynamic obstacles. In addition, the setup supports perception-level validation, including object and traffic sign detection, persistence of observations, and accuracy of estimated object velocities. Sensor data are also subject to pre-processing validation, guaranteeing their reliability for mapping, localization, and higher-level decision-making. This use case highlights how compact, reproducible, and ROS2-based V&V environments can provide students and researchers with an accessible platform for hands-on testing of autonomous vehicle algorithms and safety assurance methods. Rosbag Replay (real-world perception), F1TENTH Gym (simple 2D simulation) ROSMonitoring (verifying at runtime) RoboFuzz (testing with noisy inputs) CTU #3 Mobile robot The V&V framework for the cooperative indoor logistics robot system ensure that V&V are systematically applied across design, simulation, and real-world operation. Verification focuses on requirement traceability, component and interface testing (ROS2 nodes, MQTT messaging, sensors, DL segmentation, and planners), and formal checks for safety and deadlock freedom. Validation uses simulation, hardware-in-the-loop, and on-site trials to confirm that the system performs safely and efficiently under realistic conditions, including communication losses and operator overrides. Runtime monitoring, safety shields, and continuous testing in the CI/CD pipeline maintain ongoing assurance, ensuring reliable autonomous operation and operator control in dynamic indoor logistics environments. Nav2 and MoveIt 2 (navigation and motion planning) Gazebo/Ignition (simulation and testing) MQTT broker Eclipse Mosquitto for message exchange OpenCV and PCL vision and LIDAR processing ROS2 testing tools RTU #4 Drone Simulation-based verification and validation (V&V) for unmanned aerial vehicles (UAVs) is essential before executing real-world missions to ensure software correctness, safety, and performance. The framework must simulate the drone’s plant model, environmental perception, and control system with sufficient fidelity while maintaining real-time execution capability. Both software-in-the-loop (SITL) and hardware-in-the-loop (HIL) configurations are required to validate autonomous behaviors such as navigation in complex environments, object detection, and precise landing. Modern 3D simulation engines, such as Unreal Engine used in CARLA, now enable highly realistic testing conditions—some even experimentally extended to aerial vehicles—providing a robust basis for safe, repeatable, and educational UAV validation. AirSim (real-world perception) Gazeboo (open source simulator) ArduPilot, QGroundControl (primary framework) UAV Matlab/Simulink Toolbox (real-time simulation and deployment for UAV) CARLA (Unreal Engine based simulator)

Key Findings and Recommendations

ROS-based frameworks thus form a critical part of the SafeAV educational toolchain, ensuring scalability from lightweight student projects to advanced V&V experiments in research and industrial contexts.

The selected frameworks fulfill higher-education requirements:

- Accessibility: open-source and licence-free use. - Ease of setup: Docker-based deployment for classroom or remote use. - Pedagogical link: supports blended learning (MOOCs + hands-on labs). - Interdisciplinary use: applicable in robotics, AI, safety engineering, and mechatronics courses.

These outcomes directly contribute to WP3 educational digital content and WP2 modular curriculum design.

1. Autoware.Universe should be the baseline platform for SafeAV framework adaptation. 2. AWSIM/CARLA + SUMO/Scenic provide complementary environments for digital twin, perception, and traffic-level validation. 3. ROS-based verification tools enable node-level and formal validation, aligning with ISO 26262 / SOTIF principles. 4. Partner use cases ensure coverage of ground, aerial, and hybrid autonomous systems for educational demonstration.

Next Steps → T4.2 Adaptation

- Containerize selected frameworks for student deployment. - Develop a hands-on educational guide linking WP3 digital content to WP4 V&V examples. - Integrate simulation exercises into the SafeAV MOOC platform. - Define data interfaces (ROS bag, OpenSCENARIO) for cross-use of materials. - Establish KPIs for student learning outcomes (practical validation success, reproducibility, safety comprehension).

Use Case #1 AV Shuttle

Autonomous Vehicle Validation through Scenario-Based Simulation and Domain Analysis

This use case is the final validation task of the autonomous vehicle practical course. It builds on the previous course sections, where students first learn Linux, Git, programming, and ROS 2; then perform TurtleBot mapping and autonomous navigation; then use Autoware for mission planning; and finally validate autonomous driving behavior using Autoware Software-in-the-Loop simulation, Scenario Simulator V2, OpenSCENARIO, Scenario Editor, and ScenarioModifiers.

The main purpose is to move students from simply running one autonomous driving scenario to performing a structured validation experiment. Students design an interactive traffic scenario, parameterize it, execute approximately 100 simulation variants, and analyze which parameter combinations lead to unsafe behavior. This follows the validation logic introduced in the SafeAV material: define an Operational Design Domain, generate test scenarios, execute them, and evaluate the results using clear correctness criteria. The SafeAV book describes this V&V process as a sequence of test generation, execution, and correctness evaluation, linked to ODD, coverage, and field-response reasoning.

I. Time and effort

This task is designed as an individual assignment and should be treated as a major practical validation exercise.

Item Recommendation
Estimated student effort 12–18 hours
Recommended workload contribution 4–5 CP equivalent, depending on the local credit model
Suggested calendar duration 2 weeks after the Autoware and scenario simulation lectures
In-class preparation 3–4 hours
Independent work 9–14 hours
Assessment weight 25 points total

Suggested effort distribution:

Activity Estimated effort
Reviewing Autoware, Scenario Simulator V2, and Scenario Editor material 2–3 h
Creating the base interactive scenario 3–4 h
Editing YAML and adding ScenarioModifiers 3–4 h
Running batch simulations 3–5 h
Inspecting result files and failed .xosc cases 3–4 h
Writing report and preparing GitHub repository 4–5 h

II. Resource requirements: hardware and software

Students use a Docker-based Autoware and Scenario Simulator V2 environment. The course material explicitly recommends Docker to avoid complex source installation issues and to simplify the setup of Autoware and the scenario simulator environment.

Minimum Hardware

Software and Infrastructure

Required:

The ROS 2 material assumes Ubuntu 22.04 and ROS 2 Humble as the course baseline, and introduces workspaces, sourcing, nodes, topics, launch files, and `ros2_ws` structure as prerequisites. The Autoware material also uses a Docker workflow with mounted `ros2_ws` and `autoware_map` directories and a `setup.bash` file that sources ROS, Autoware, and the student workspace.

Students may work on their own laptops if the laptop satisfies the requirements. Otherwise, the recommended mode is to use lab workstations or remote servers prepared with the Docker image and maps. The Docker image is distributed through the course platform, so access to Moodle or the course repository is required. No commercial license is required for the core tools used in this task.

1. Prerequisite

Before starting this use case, students should already be able to:

The course prepares these skills progressively. The Linux and Git section introduces terminal commands and GitHub-based version control; the ROS 2 section introduces nodes, topics, workspaces, and launch files; the Autoware section introduces the planning simulator and Ego vehicle mission planning; and the validation section introduces Scenario Simulator V2, Scenario Editor, and OpenSCENARIO-based simulation.

2. Required Reading

- SafeAV book: Chapter 2.3, Introduction to Validation and Verification in Autonomy

Students should focus on the relationship between validation, Operational Design Domain, coverage, test generation, execution, and correctness criteria. This chapter explains that validation must demonstrate reasonable testing across the intended ODD, using generated scenarios, execution, and pass/fail criteria.

- SafeAV book: Chapter 2.4, Validation Requirements across Domains

Students should focus on why ground autonomous vehicles require scenario-based testing and how ODD boundaries, weather, infrastructure, regulation, and human-dense environments affect validation.

- SafeAV book: Chapter 6.3, Validation of Control and Planning

Students should focus on mission-level planning validation, trajectory safety outcomes, collisions, distance-to-collision violations, timeouts, and parameterized scenarios that stress planner thresholds.

- Course material: Autoware – Autonomous Vehicle Software

Students should review how to launch Autoware in Docker, start the planning simulator, set the initial pose, set the goal pose, and start autonomous navigation.

- Course material: Validation – Scenario Simulation

Students should review how Scenario Simulator V2 uses OpenSCENARIO, how the Scenario Editor is used, how predefined and custom scenarios are launched, and how post-simulation results are inspected.

- Task 4 PDF: AV Validation – Create and Simulate an Interactive Scenario

Students should read the full task description, especially the requirements for at least two NPCs, ScenarioModifiers, approximately 100 simulations, failure-boundary analysis, report sections, YAML submission, and GitHub structure.

3. Instruction on the Use Case

The instructor introduces the use case as the final step in the course's autonomy pipeline. The previous tasks show how to make a robot or Ego vehicle move autonomously. This task asks a different question: under which conditions does the autonomous driving stack fail, and how can we provide evidence for that conclusion?

The context is in three parts:

First, the instructor explains the validation motivation. In real autonomous vehicle development, it is not enough to show that the vehicle succeeds once. The system must be tested across a range of traffic situations. The students therefore create a base scenario and mutate key parameters such as Ego starting position, NPC trigger point, NPC speed, pedestrian delay, or timeout.

Second, the instructor explains the technical workflow: create scenario, export YAML, modify YAML paths and Ego configuration, add ScenarioModifiers, run batch simulation through scenario_test_runner, inspect results, identify failed cases, open failed .xosc files, extract parameter values, and define a failure boundary.

Third, the instructor explains the expected analysis. Students must not only report that some simulations failed. They must investigate why they failed and identify critical combinations, such as: Autoware fails when the NPC starts too close to the conflict point while Ego starts beyond a certain distance or speed condition.

Recommended teaching format:

4. Tools Introduction and Description

Autoware

Autoware is the autonomous driving stack under test. In this use case, students evaluate how Autoware behaves when the Ego vehicle is exposed to parameterized traffic interactions.

Scenario Simulator V2

Scenario Simulator V2 is the simulation framework used to execute OpenSCENARIO-based traffic scenarios with Autoware. It can simulate predefined traffic scenarios, validate planning and control modules, and generate machine-readable scenarios for Autoware.

Scenario Editor

Scenario Editor [https://scenario.ci.tier4.jp/] is an online tool for creating and modifying OpenSCENARIO files for Autoware simulation. Students use it to import a Lanelet2 map, place Ego and NPC actors, define interactions, and export the scenario as a YAML file. Google Chrome is recommended for compatibility.

OpenSCENARIO / YAML Scenario Files

The exported YAML file represents the traffic scenario. Students modify this file manually to fix map paths, configure the Ego vehicle, remove unnecessary parts if required, and add ScenarioModifiers.

ScenarioModifiers

ScenarioModifiers are used to replace fixed scenario values with variables. These variables can be defined as ranges or lists to automatically generate many scenario variations. Task 4 requires students to use both lists and ranges and target approximately 100 simulation variants.

Docker

Docker provides the preconfigured environment containing Autoware and Scenario Simulator V2. Students use the provided Docker image and shell script, such as autoware_terminal.sh, to enter the environment.

GitHub

GitHub is used for assignment delivery. Each student must submit a public repository containing the final parameterized YAML file, relevant configuration files, and documentation. The YAML file must be placed in a folder named scenario inside the ros2_ws repository.

Online Assessor

An online assessor is used to score the submitted report according to the defined rubric. The assessor should be treated as a formative and summative support tool: students may use it to check whether their draft addresses the required sections, but the final responsibility for correctness, technical explanation, evidence, and originality remains with the student.

5. Demonstration of Example

The instructor demonstrates a small end-to-end example before students begin their own work. The example should be intentionally simple but complete.

  1. Open Scenario Editor in Google Chrome.
Figure 2: Scenario Editor, online tool
  1. Import the Lanelet2 map used in the course.
  2. Place the Ego vehicle at an initial position before a junction.
  3. Add two NPCs:
    • NPC1: a vehicle approaching the same junction from a crossing lane.
    • NPC2: a pedestrian or second vehicle crossing or entering the conflict zone after a delay.
  4. Define a realistic interaction:
    • Ego drives toward an intersection.
    • NPC1 enters from the side road.
    • NPC2 crosses after a delay.
    • A collision risk appears if timing and distance are unfavorable.
  5. Export the scenario as YAML.
  6. Modify the YAML:
    • update RoadNetwork / LogicFile map path;
    • ensure Ego is configured as isEgo: true for Autoware-in-the-loop execution;
    • remove unsupported or unnecessary elements if required;
    • add a small ScenarioModifiers block.
ScenarioModifiers: 
  ScenarioModifier: 
   # Range Sweep: (35-15)/2 = 11 variations 
   - { name: EGO_START_S, start: 15, step: 2, stop: 35 } 
   # Range Sweep: (45-38)/1 = 8 variations 
   - { name: NPC1_TRIGGER_S, start: 38, step: 1, stop: 45 } 
   # List: 2 variations 
   - name: NPC2_DELAY 
   list: [1, 5]
  1. Run a short batch test, for example 12–20 variants, before showing the full 100-variant requirement.
Figure 3: Scenario Simulator V2, running a test scenario
  1. Open result.junit.xml and identify failed scenario indices.
  2. Open one failed .xosc file and extract the parameter values that caused the failure.
  3. Create a small risk table with columns such as scenario index, Ego start position, NPC trigger point, NPC delay, result, and observed failure.

The instructor should explicitly connect this example to the Task 4 expectation: students must design their own scenario, use at least two NPCs, run approximately 100 variants, and identify a failure boundary.

6. Task Description

Each student must design, execute, and analyze an autonomous vehicle validation experiment using Autoware Scenario Simulator V2.

The student must create an interactive traffic scenario in Scenario Editor. The scenario must include:

The student must export the scenario as YAML and modify it to support batch validation. Instead of hardcoding all values, the student must define ScenarioModifiers for approximately 100 simulations. The variables should represent meaningful domain parameters, such as:

A successful minimum submission includes:

Main simulation command:

ros2 launch scenario_test_runner scenario_test_runner.launch.py \
architecture_type:=awf/universe record:=false \
scenario:='/autoware_map/scenarios/your_scenario.yaml' \
sensor_model:=sample_sensor_kit vehicle_model:=sample_vehicle \
output_directory:='/autoware_map/results' \
global_real_time_factor:=5.0 launch_rviz:=false use_sim_time:=true

If Autoware behaves erratically, students may reduce global_real_time_factor to 2.0 or 3.0.

7. Development of the Task

Mandatory Development

Students must complete the following mandatory parts:

  1. Create a base interactive scenario using Scenario Editor.
  2. Include at least two NPCs.
  3. Export and modify the YAML file.
  4. Add ScenarioModifiers using both ranges and lists.
  5. Generate approximately 100 scenario variations.
  6. Run the simulations with Autoware in the loop.
  7. Analyze result.junit.xml.
  8. Inspect failed .xosc files.
  9. Identify at least one dangerous failed case in detail.
  10. Define a failure boundary or critical threshold.
  11. Submit a structured report and GitHub repository.

Optional Extensions for Higher Quality Work

Students can deepen the task by:

Optional extensions improve the grade only after the mandatory requirements are satisfied.

8. Decision on Group Work

This use case is an individual task.

Each student must submit their own report, their own scenario design, their own parameterized YAML file, their own simulation results, and their own GitHub repository link. Students may discuss general tool usage and troubleshooting with classmates, but the scenario design, YAML parameterization, analysis, and written report must be individually produced.

The scope is intentionally limited to approximately 100 simulation variants so that the task is feasible for one student while still requiring real validation reasoning.

9. Reporting and Evaluation

Students submit:

  1. Detailed report.
  2. Public GitHub repository link.
  3. Final parameterized YAML scenario file.
  4. Relevant configuration files or helper scripts.
  5. Evidence of simulation execution.
  6. Results analysis and failed scenario investigation.

The report must include the following sections:

1. Introduction

Explain the task, validation objective, and why scenario-based simulation is used.

2. Scenario Creation

Describe the selected map, traffic situation, Ego route, NPC actors, and interactions. Include screenshots from Scenario Editor.

3. YAML Modifications

Explain path corrections, Ego configuration, removed or edited blocks, and all ScenarioModifiers. Explain why each variable was chosen.

4. Simulation Process

Show the exact launch command, real-time factor, output directory, Docker environment, and any troubleshooting steps.

5. Results and Analysis

Summarize the number of total simulations, number of passed and failed cases, percentage of failures, failed indices, and key observations.

6. Domain Analysis and Failure Boundary

Identify the critical parameter combinations that lead to unsafe behavior. This must be based on result files and .xosc inspection.

7. Failed Scenario Investigation

Analyze at least one failed scenario in detail. Include the parameter values and explain why the situation became dangerous.

8. Conclusion

Summarize what was learned, limitations of the experiment, and possible improvements.

9. GitHub Link

Provide the public repository link. The final YAML file must be placed inside a folder named scenario in the ros2_ws repository.

The report should also include a short time and cost reflection, including:

10. Evaluation Criteria

Total: 25 points

1. Report Structure and Completeness -- 4 points

The report must be clearly structured and include all required sections: introduction, scenario creation, YAML modifications, simulation process, results and analysis, conclusion, and GitHub link. The content must be logically organized and directly correspond to the assignment objectives.

Score Description
4 points All required sections are present, clearly structured, technically meaningful, and logically connected.
2–3 points Most sections are present, but structure is partially incomplete, unclear, or weakly developed.
0–1 points Major sections are missing, the document lacks logical organization, or the report does not correspond to the assignment.

2. Scenario Design and Interactive Behavior -- 4 points

The student must create an interactive traffic scenario using Scenario Editor with at least two NPCs, such as vehicles or pedestrians. The interaction must be realistic and relevant to AV validation.

Score Description
4 points Scenario is realistic, includes at least two meaningful NPC interactions, and is clearly documented with visual evidence.
2–3 points Scenario is functional but simplistic, weakly explained, or only partially realistic.
0–1 points Scenario is incomplete, unrealistic, missing NPC interaction, or not documented.

3. Parameterized Scenario and YAML Configuration -- 5 points

The YAML file must include correctly implemented ScenarioModifiers using ranges and lists to generate approximately 100 variations. Students must show understanding of parameterized testing rather than hardcoded values.

Score Description
5 points YAML is technically correct, uses ranges and lists properly, generates approximately 100 meaningful variations, includes valid variable definitions, valid simulation parameters, and correct paths. Explanation is specific and technically grounded.
3–4 points Parameterization is mostly correct but incomplete, weakly explained, or produces a less convincing set of variations.
1–2 points Mutation logic is limited, incorrect, mostly hardcoded, or poorly explained.
0 points YAML configuration is missing or does not implement parameterized scenarios.

If AI tools are used to help generate or debug YAML, the usage must be clearly indicated and accompanied by student explanation and comments. Generic AI-generated explanations are not sufficient for full credit.

4. Simulation Execution and Autoware Integration -- 4 points

The student must execute batch simulations using Autoware Scenario Simulator V2 and demonstrate understanding of the workflow.

Score Description
4 points Multiple mutated simulations are successfully executed; launch command, real-time factor, output directory, and configuration are clearly explained; the process is reproducible.
2–3 points Simulation is partially successful or insufficiently documented.
0–1 points Simulation process is missing, incorrect, only a single scenario is run, or results are not reproducible.

5. Domain Analysis and Risk Investigation -- 5 points

The report must include meaningful analysis of the simulation results. Students must analyze failed scenarios using generated .xosc and result files, identify failure patterns, and define a failure boundary or critical threshold.

Score Description
5 points Detailed technical analysis supported by evidence; failure boundary is clearly defined; at least one failed case is investigated in depth; dangerous parameter combinations are explained.
3–4 points Analysis is mostly correct but lacks depth, evidence, or clear threshold definition.
1–2 points Analysis is superficial, generic, or weakly connected to simulation results.
0 points Analysis is missing.

6. GitHub Repository and Project Completeness -- 3 points

The submission must include a public GitHub repository containing the final parameterized YAML file, relevant configuration files, and any additional code used.

Score Description
3 points Repository is complete, well organized, reproducible, and includes the YAML file inside ros2_ws/scenario.
1–2 points Repository is partially complete, missing documentation, or not fully reproducible.
0 points Repository, YAML file, or required scenario folder is missing.

11. AI

Students may use AI tools as assistive tools, not as replacements for engineering work.

Permitted AI use:

Not permitted:

Students must include a short AI usage statement in the report, for example:

AI tools were used to improve the wording of the report and to help debug YAML syntax. All simulation results, screenshots, parameter values, and analysis were produced and verified by the student.

The online assessor may be used as a formative feedback tool before final submission. Its role is to check alignment with the rubric and identify missing sections or weak explanations. The final grade is still based on the submitted technical evidence, scenario files, simulation results, and report quality.

12. External Material or Dataset

Lanelet2 Map

Students use a Lanelet2 map supported by the Scenario Editor and Scenario Simulator V2. The map is imported into the Scenario Editor and referenced by the scenario YAML file.

Autoware Map Directory

The course Autoware workflow uses an autoware_map directory mounted into Docker. The scenario, map files, generated results, and output data should be organized in this directory according to the course instructions.

Scenario Editor Online Platform

Used to create and export the base scenario.

Autoware + Scenario Simulator Docker Image

The Docker image is provided through the course platform. It contains the required Autoware and scenario simulation stack.

GitHub Repository

Each student uses GitHub as the submission and reproducibility platform.

Generated Simulation Outputs

The student-generated output dataset includes:

These generated outputs are not external datasets but are essential evidence for the validation report.

Final Use Case Summary

This use case teaches students how autonomous vehicle validation differs from simple demonstration. A single successful run is not enough. Students must define a traffic interaction, parameterize the scenario, execute many simulation variants, inspect failures, and identify the boundary between safe and unsafe behavior. The task connects practical Autoware simulation skills with the SafeAV validation concepts of Operational Design Domain, scenario generation, execution, correctness criteria, coverage, and failure analysis.

Use Case #2 F1TENTH

The F1TENTH platform is an open-source, small-scale autonomous racing car designed for research and education in autonomous systems. Built on a 1/10-scale RC chassis, it integrates sensors such as LiDAR, camera, and IMU, all running on a ROS2-based control stack. Its modular architecture allows experiments in perception, planning, and control, while remaining low-cost and portable for classroom and laboratory use. The platform is supported by an active international community, offering simulation environments, datasets, and open course materials that make it ideal for hands-on learning and benchmarking in robotics and self-driving research. In academia, it serves as a standardized benchmark for teaching autonomous driving algorithms, allowing students to bridge theory and practice through competitions, lab assignments, and project-based learning. Universities use F1TENTH to demonstrate safety validation, sensor fusion, and real-time decision-making concepts within a manageable and reproducible framework, making it an ideal entry point for higher education in robotics and autonomous vehicle research.

![image.png](attachment:d4ab1d39-89e9-454d-a3f1-571bfdd5a3f1:image.png)

The interface requirements for the F1TENTH use case define how the simulation and control systems communicate within the ROS2 environment (ROS2 Humble, with possible but nontrivial porting to Jazzy). The platform’s sensor outputs include LiDAR scans, RGB or RGB-D camera feeds, and IMU data, which can be noisy and unreliable due to the small size of the vehicle. Additional outputs, such as pose and velocity estimates, may be derived using external localization methods. Control inputs consist of speed commands—either normalized or expressed in meters per second—and steering angle values, typically represented as PWM signals that can be translated into degrees. These interfaces ensure smooth integration between the physical and simulated components, supporting real-time testing, algorithm verification, and reproducibility in educational and research contexts.

The F1TENTH use case defines simulation requirements that ensure accessible and reproducible environments for validating autonomous driving algorithms within academic and research settings.

- The simulation environment must support lightweight, scalable, and educationally accessible setups suitable for university use. - It should utilize kinematic single-track or Ackermann steering models to balance computational efficiency with sufficient accuracy for low-speed vehicles. - The framework must enable reproducible validation of perception, planning, and control algorithms in both 2D and 3D environments. - ROS2 compatibility is required, with optional integration to simplified Autoware vehicle models for advanced testing. - Simulations should run smoothly on standard laptops without the need for high-end GPUs, ensuring broad accessibility for students. - The overall design must support iterative experimentation, open-source deployment, and hands-on learning in line with SafeAV’s educational objectives.

The F1TENTH use case focuses on providing an open, reproducible, and educational platform for validating perception, planning, and control algorithms in small-scale autonomous vehicles. Its simulation and interface requirements emphasize lightweight, ROS2-compatible environments that run efficiently on standard hardware, enabling hands-on learning, iterative testing, and scalable experimentation in autonomous systems education.

Use Case #3 Mobile Robot

The Mobile Robot (RTU) use case focuses on cooperative indoor logistics systems designed to demonstrate autonomous navigation, coordination, and task management in controlled environments. The setup consists of two mobile robot platforms, a central server for planning and task distribution, and MQTT-based communication for asynchronous message exchange. Each robot operates under a ROS2-based control architecture integrating LiDAR, camera, and deep learning–based segmentation for enhanced mapping and path planning. Within the academic context, this use case provides a practical environment for teaching multi-robot coordination, communication reliability, and safety validation, bridging theoretical concepts in robotics and AI with real-world industrial applications.

![rtu_agv.png](attachment:3c93f42d-e4fc-4856-970f-97ca7146f0a4:e6754dc7-26f5-4a35-b070-c49f7f0f754e.png)

The V&V requirements for the Mobile Robot (RTU) use case focus on ensuring safe, reliable, and verifiable operation of cooperative indoor logistics robots within both simulated and physical environments. The framework is designed to validate system behavior across all levels—communication, perception, navigation, and task management—while supporting educational objectives through accessible, open-source tools. These requirements align with SafeAV’s broader goal of integrating real-world industrial practices into higher-education robotics training.

- The verification and validation framework must support systematic testing across design, simulation, and real-world operation, ensuring traceability from requirements to test results. - It should enable component-level and interface testing of ROS2 nodes, MQTT communication, sensor data pipelines, and deep learning–based segmentation modules. - The setup must allow hardware-in-the-loop (HIL) and simulation-based validation, confirming that control, navigation, and planning functions behave safely under realistic indoor conditions. - Runtime monitoring and safety shields should be implemented to detect failures, prevent unsafe actions, and support continuous verification in CI/CD workflows. - The framework must assess system performance under communication losses or operator intervention, ensuring resilience and fault tolerance in distributed multi-robot environments. - All tools and procedures should remain open-source, modular, and reproducible, allowing students to perform iterative experiments, analyze safety properties, and understand industrial V&V practices within an educational context.

The defined V&V requirements establish a comprehensive validation chain that connects design, simulation, and real-world testing. They emphasize fault tolerance, runtime monitoring, and reproducibility using open-source ROS2 and MQTT-based architectures. This ensures that students can study and experiment with advanced verification techniques while developing safe and resilient autonomous robotic systems.

Use Case #4 Drone

The Drone (SUT, PRO) use case focuses on unmanned aerial vehicle (UAV) systems that bridge aviation safety principles with autonomous mobility education. Developed through Prodron’s extensive experience in UAV training and system design, this use case explores real-world challenges such as emergency response, navigation under uncertainty, sensor fusion, and communication reliability. The drones operate using open-source frameworks like ArduPilot and QGroundControl, supporting both software-in-the-loop (SIL) and hardware-in-the-loop (HIL) validation. In the academic context, this use case provides a versatile platform for teaching autonomous flight control, safety verification, and resilience testing, allowing students to apply V&V methodologies from aerial robotics to broader autonomous vehicle domains.

![image.png](attachment:026b44b9-df3c-4124-9206-c8a0cc43b705:image.png)

The UAV use-case requirements focus on defining the essential verification and validation conditions for safe and reliable autonomous flight operations. They emphasize emergency response handling, navigation under uncertainty, sensor integration, and communication robustness, including lessons from MAVLink telemetry that can directly apply to V2X communication challenges in automotive systems. These requirements ensure realistic, reproducible, and educationally relevant validation of UAV systems.

- The framework must enable emergency response validation, including fault injection, communication loss recovery, and system redundancy testing. - It should support navigation under uncertainty, allowing for dynamic obstacle avoidance, no-fly zone compliance, and route optimization under changing conditions. - The simulation environment must facilitate sensor integration testing, including GNSS, IMU, and vision-based systems, with modeling of environmental factors such as wind or weather effects. - Communication reliability and latency simulation should be included to evaluate telemetry, interference resilience, and cybersecurity aspects of UAV operations. - The setup must support SIL and HIL configurations, enabling both pure software and mixed real-hardware validation of control and perception modules. - Tools and models should support, where reasonable, a dual-tier approach, combining solutions with open-source frameworks for advanced experimentation.

The UAV V&V requirements ensure comprehensive testing of autonomous flight systems under realistic and variable conditions, supporting both educational and research-oriented objectives. By integrating open-source simulation, communication reliability testing, and hardware-in-the-loop validation, they provide a robust foundation for safety assurance and hands-on learning. The SafeAV study recommends a dual-tier approach for UAV simulation and validation—combining commercial off-the-shelf (COTS) systems for rapid onboarding with open-source ecosystems for advanced, research-driven experimentation. This structure enables both immediate applicability in training contexts and long-term scalability for integration into academic courses, laboratories, and testbeds.