Skip to the content.

Abstract

This work studies the design of safe control policies for large-scale non-linear systems operating in uncertain environments. In such a case, the robust control framework is a principled approach to safety that aims to maximize the worst-case performance of a system. However, the resulting optimization problem is generally intractable for non-linear systems with continuous states. To overcome this issue, we introduce two tractable methods that are based either on sampling or on a conservative approximation of the robust objective. The proposed approaches are applied to the problem of autonomous driving.

Video

Reproduce the experiments

Install requirements

Run the benchmark

cd <path-to-rl-agents>/scripts/
python experiments.py benchmark configs/RoundaboutEnv/benchmark_robust_control.json \
                      --test --episodes=100 --processes=4

The following agents will be evaluated:

{
    "environments": [
        "configs/RoundaboutEnv/env.json"
    ],
    "agents": [
        "configs/RoundaboutEnv/agents/DeterministicPlannerAgent/baseline.json",
        "configs/RoundaboutEnv/agents/DeterministicPlannerAgent/assume_random_route.json",
        "configs/RoundaboutEnv/agents/DiscreteRobustPlannerAgent/routes.json",
        "configs/RoundaboutEnv/agents/DeterministicPlannerAgent/assume_random_behaviour.json",
        "configs/RoundaboutEnv/agents/IntervalRobustPlannerAgent/behaviours.json"
    ]
}