Drone Autonomy - The AI Grand Prix

The AI Grand Prix is hosting a competition for fully autonomous drone software, so I have decided to register and work on my own tools + autonomy model.

Architecture

Two pipelines: world generation (synthetic training data) and vision-to-control (onboard inference).

World Generator

Diffusion model outputs 3D Gaussian Splatting (3DGS) scenes conditioned on track layout graphs. 3DGS enables photorealistic rendering at arbitrary viewpoints, ~2ms rasterization, and easy domain randomization via splat perturbation.

Vision Model

Mamba SSM backbone instead of transformer: linear complexity (critical at 200 FPS), implicit memory via recurrent state, constant inference time.

Latency Budget

At 150 km/h: 4.2 cm/ms. Target 5ms loop:

Positional error stays under 21 cm—workable for ~1m gate gaps.