What you'll do
Satellite rendezvous requires determining the relative position, velocity, and orientation of two spacecraft moving at several kilometers per second - all in real time. In this role, you will own the development of our relative navigation system: from defining requirements and designing architectures, to building algorithms and validating them with synthetic datasets and hardware-in-the-loop testing. You’ll have the chance to shape this technology from scratch, leveraging computer vision to enable autonomous satellite rendezvous and docking.
Responsibilities
- Design and implement advanced computer vision algorithms for spacecraft navigation
- Design and implement the sensor processing and fusion stack for real-time state estimation
- Generate and validate large-scale synthetic datasets to train and test navigation algorithms
- Optimize vision models for real-time processing on space-grade edge hardware
- Contribute to fault detection, isolation, and recovery systems through robust sensor and state monitoring
- Work closely with GNC and flight software teams to integrate the algorithms into spacecraft software
What we look for
Minimum Requirements
- Bachelor’s degree in Physics, Computer Science, Electrical or any related engineering discipline
- 3+ years of experience in computer vision, robotics, or autonomous systems; graduating PhDs with relevant focus
- Proficiency in Python, C++ and OpenCV and hands-on experience with PyTorch, or TensorFlow
- Proficiency in 3D Computer Vision (SLAM, depth estimation, sensor fusion)
- Proven track record of developing and deploying vision models in real-world applications
- You thrive in fast-paced, dynamic environments, adapt quickly, and enjoy solving hard problems
Preferred qualifications
- Hands-on experience with embedded AI and real-time operating systems