Replaces traditional radar signal processing with end-to-end neural networks for self-driving cars.
Using end-to-end neural radar that learns directly from raw signals, adaptive waveform generation, and synthetic radar simulation for scalable training data.

|
Autonomous Vehicles
|
YC W26

Last Updated:
March 19, 2026

Builds AI-native radar systems for self-driving cars, using end-to-end deep learning to replace traditional radar signal processing pipelines with neural network architectures for superior perception.
Research-first culture focused on challenging traditional radar perception and ML reproducibility. No commercial product launch or automotive OEM partnership announced.
Very small PhD-heavy team. Stealth posture likely protecting IP pre-patent. Proprietary breakthroughs in learned radar waveforms or neural radar processing chains. Very early stage.
<p>Replaces the entire traditional radar digital signal processing (DSP) chain with a single end-to-end deep learning model that directly converts raw radar returns into high-fidelity object detections, classifications, and tracks for autonomous driving.</p>
Instead of radar doing math step-by-step like a textbook, a single AI brain looks at the raw radar echoes and instantly tells the car "that's a pedestrian, that's a truck, that's a guardrail."
Traditional automotive radar relies on a rigid, hand-engineered DSP pipeline: matched filtering, CFAR detection, DOA estimation, clustering, and tracking—each stage lossy and tuned for average-case scenarios. Congruent's approach replaces this entire chain with a neural network trained end-to-end on raw ADC (analog-to-digital converter) radar data, learning optimal representations for detection, classification, and tracking simultaneously. This allows the system to extract subtle features—micro-Doppler signatures of pedestrians, occluded vehicle reflections, multi-path ghost suppression—that traditional pipelines discard. The model is trained using simulation-generated synthetic radar data from digital twins and validated against real-world corner cases, enabling rapid iteration without expensive physical test campaigns. The result is a radar perception system that adapts to novel scenarios, degrades gracefully in adverse weather, and improves continuously with new data.
It's like replacing a factory assembly line of 12 specialized workers with one genius savant who sees the raw materials and instantly builds the finished product better than all 12 combined.
<p>Uses deep reinforcement learning to dynamically optimize radar transmit waveforms in real time, adapting to driving context, interference, and environmental conditions to maximize perception quality.</p>
The radar teaches itself to shout in exactly the right "voice" for each driving situation—whispering in clear weather, belting through rain, and dodging other radars' noise—all in real time.
Conventional automotive radars transmit fixed or simply modulated waveforms (FMCW chirps) regardless of the environment, leading to suboptimal performance in congested spectral environments, heavy precipitation, or multi-radar interference scenarios. Congruent's approach uses a deep reinforcement learning agent that observes the current radar environment—clutter profile, interference signatures, target density, weather conditions—and selects or synthesizes optimal transmit waveforms on a pulse-by-pulse or frame-by-frame basis. The RL agent is trained in a high-fidelity electromagnetic simulation environment (digital twin) that models realistic road scenarios, atmospheric effects, and multi-vehicle radar interference. The learned policy generalizes to unseen conditions and continuously improves via online adaptation. This results in dramatically improved detection range, angular resolution, and interference immunity compared to static waveform designs, giving Congruent a fundamental hardware-agnostic advantage that can be deployed on commodity radar front-ends.
It's like a DJ who reads the room in real time and adjusts the music perfectly for every moment, instead of just hitting play on the same playlist every night.
<p>Builds a physics-based digital twin simulation platform that generates synthetic radar data at scale to train, validate, and stress-test AI radar perception models against millions of rare and dangerous driving scenarios without physical road testing.</p>
Instead of driving a million miles to find every weird thing that could go wrong, they build a virtual world where the radar practices against millions of scary scenarios from the safety of a computer.
Validating autonomous vehicle radar perception in the real world is prohibitively expensive and statistically insufficient—critical edge cases (child darting into traffic, black ice reflections, multi-vehicle pileups) occur too rarely to capture at scale. Congruent's digital twin platform models the full radar signal chain: RF propagation, multi-path reflections, material-specific radar cross sections, atmospheric attenuation, and sensor noise characteristics. It procedurally generates diverse driving scenarios—varying weather, lighting, traffic density, road geometry, and adversarial actor behavior—and produces physically accurate synthetic radar returns that are indistinguishable from real sensor data for training purposes. An automated edge case discovery engine uses coverage-guided fuzzing and adversarial scenario generation to systematically find perception model failure modes. Each discovered failure feeds back into the training loop, creating a closed-loop continuous improvement system. This dramatically accelerates development velocity, reduces validation costs, and provides auditable safety evidence for regulatory approval.
It's like a flight simulator for radar—pilots don't learn to handle engine failures by crashing real planes, and Congruent's radar AI doesn't learn to handle black ice by skidding on real highways.
Deep academic radar engineering expertise combined with AI-native design philosophy, potentially bypassing legacy DSP constraints that limit incumbents. Operating independently without pressure to ship prematurely.