At Serve Robotics, we’re reimagining how things move in cities. Our personable sidewalk robot is our vision for the future. It’s designed to take deliveries away from congested streets, make deliveries available to more people, and benefit local businesses. The Serve fleet has been delighting merchants, customers, and pedestrians along the way in Los Angeles while doing commercial deliveries.
The Serve fleet has been delighting merchants, customers, and pedestrians along the way in Los Angeles while doing commercial deliveries. We’re looking for talented individuals who will grow robotic deliveries from surprising novelty to efficient ubiquity.
We are tech industry veterans in software, hardware, and design who are pooling our skills to build the future we want to live in. We are solving real-world problems leveraging robotics, machine learning and computer vision, among other disciplines, with a mindful eye towards the end-to-end user experience. Our team is agile, diverse, and driven. We believe that the best way to solve complicated dynamic problems is collaboratively and respectfully.
Serve Robotics aims to develop dependable and proficient sidewalk autonomy software. Our Perception & Prediction team is looking for a talent in identified research areas such as artificial intelligence, robotics, and embodied AI, particularly including areas such as representation learning for Embodied AI tasks, end-to-end learning, vision-language-action models, contextually aware AI.
Pioneer the development of embodied intelligence for robotics, enabling unprecedented generalization across diverse environments and embodiments.
Invent/improve novel data-driven paradigms for embodied intelligence, leveraging a variety of modalities (images, video, point clouds (Lidar), text, audio etc).
Create business impactful solutions to improve human robot interaction through embodied intelligence.
Work with ML infrastructure engineers to assess and monitor model performance, ideate on active learning framework from long tail problems and pave innovative solutions to handle generalization capabilities from being data aware.
Collaborate with various teams to understand real-world problems and define tasks, incorporating insights into ML products.
MS with 5 years of industry experience or PhD with minimum 3 years of industry experience with focus in embodied intelligence, AR/VR , Robotics use-cases in a end-to-end learning framework.
Demonstrated leadership in scientific efforts, with contributions to top rated conferences and journals.
Experience with deep learning frameworks (such as pytorch, tensorflow) and computational data science libraries.
Experience with using large-scale datasets and sensor data, with use of various transformer architectures and diffusion models.
Adept with latest technology trends in AI / Robotics and Foundational models. Deep understanding of computer vision, machine learning and deep learning basic concepts.
Comfortable working with SQL queries and ETL logic for data ingress.
Strong programming skills for efficient and robust code.
Experience with multiple sensors such as Lidar, Mono/Stereo cameras, IMU, etc.
Strong communication skills.
Hands-on experience with fine-tuning LLMs or vision-language models and their performance challenges and opportunities.
Experience with MLOps or ML Infrastructure, ability to streamline machine learning workflows.
Familiarity with edge-device perception stack deployment, experience with NVIDIA software libraries such as CUDA or TensorRT.
Open source project contributor.
Experience with GCP or AWS, Kubernetes and Docker.