Serve Robotics

Senior Data Scientist, Machine Learning

Remote US
USD 160k - 200k
Kubernetes Docker Machine Learning Python PyTorch GCP SQL TensorFlow AWS Azure
Description

At Serve Robotics, we’re reimagining how things move in cities. Our personable sidewalk robot is our vision for the future. It’s designed to take deliveries away from congested streets, make deliveries available to more people, and benefit local businesses. The Serve fleet has been delighting merchants, customers, and pedestrians along the way in Los Angeles while doing commercial deliveries.

The Serve fleet has been delighting merchants, customers, and pedestrians along the way in Los Angeles while doing commercial deliveries. We’re looking for talented individuals who will grow robotic deliveries from surprising novelty to efficient ubiquity.

Who We Are

We are tech industry veterans in software, hardware, and design who are pooling our skills to build the future we want to live in. We are solving real-world problems leveraging robotics, machine learning and computer vision, among other disciplines, with a mindful eye towards the end-to-end user experience. Our team is agile, diverse, and driven. We believe that the best way to solve complicated dynamic problems is collaboratively and respectfully.

What you'll be doing

Serve Robotics aims to develop dependable and proficient sidewalk autonomy software. We are looking for a talented Senior Data Scientist who bridges the gap between ML infrastructure and ML engineers. The ideal candidate possesses strong fundamentals in machine learning, with the ability to prototype and train learning-based models using data-centric techniques. This individual should also have expertise in data ETL processes, SQL queries, and building scalable data pipelines to make data accessible for model training.

Responsibilities

  • Prototype and train learning-based models using a data-centric approach, applying techniques such as automated feature engineering, active learning, and fine-tuning on curated datasets.

  • Design, develop, and maintain efficient data and feature extraction pipelines to support ML engineers in accessing high-quality data for model training.

  • Design auto labeling system using ensemble of models that can reason from multi-modal data for different use-cases, For example: image semantic labeling using vision grounded models, intent and path prediction ground truth.

  • Perform complex data extraction, transformation, and loading (ETL) processes, ensuring data is clean, accessible, and well-documented. Write and optimize high-quality SQL queries for data analysis and ingestion from various sources.

  • Partner with data infrastructure and ML engineers to ensure seamless integration of data and machine learning workflows.

  • Produce high-quality, maintainable code and participate in peer code reviews to share knowledge and uphold team standards.

Qualifications

  • Master’s in Computer Science, Data Science, or a related technical field and 5+ years of industry experience in data engineering, machine learning, or a similar domain.

  • Strong proficiency in Python and SQL, with demonstrated experience building data pipelines at scale and ETL workflows that cater to multi-modal data (e.g., images, point clouds, time-series data).

  • Proven ability to work with PB’s of datasets, including structured, semi-structured, and unstructured data.

  • Hands-on experience working with ML frameworks such as TensorFlow, PyTorch, or similar.

  • Solid understanding of ML fundamentals and data-centric techniques for model training.

  • Experience with cloud platforms (GCP, AWS, or Azure) and tools like Kubernetes, Docker, and Airflow.

  • Excellent communication skills and the ability to collaborate with cross-functional teams.

What makes you standout

  • Experience optimizing ML workflows using MLOps tools such as MLflow, TFX, Kubeflow, or similar platforms.

  • Strong understanding of transformer-based models and their application in data-centric AI workflows.

  • Knowledgable in advanced SQL query optimization and ETL pipeline performance tuning.

  • Familiarity with tools for scalable data engineering, such as Apache Beam, Dask, or BigQuery.

There are more than 50,000 engineering jobs:

Subscribe to membership and unlock all jobs

Engineering Jobs

60,000+ jobs from 4,500+ well-funded companies

Updated Daily

New jobs are added every day as companies post them

Refined Search

Use filters like skill, location, etc to narrow results

Become a member

🥳🥳🥳 452 happy customers and counting...

Overall, over 80% of customers chose to renew their subscriptions after the initial sign-up.

To try it out

For active job seekers

For those who are passive looking

Cancel anytime

Frequently Asked Questions

  • We prioritize job seekers as our customers, unlike bigger job sites, by charging a small fee to provide them with curated access to the best companies and up-to-date jobs. This focus allows us to deliver a more personalized and effective job search experience.
  • We've got about 70,000 jobs from 5,000 vetted companies. No fake or sleazy jobs here!
  • We aggregate jobs from 5,000+ companies' career pages, so you can be sure that you're getting the most up-to-date and relevant jobs.
  • We're the only job board *for* software engineers, *by* software engineers… in case you needed a reminder! We add thousands of new jobs daily and offer powerful search filters just for you. 🛠️
  • Every single hour! We add 2,000-3,000 new jobs daily, so you'll always have fresh opportunities. 🚀
  • Typically, job searches take 3-6 months. EchoJobs helps you spend more time applying and less time hunting. 🎯
  • Check daily! We're always updating with new jobs. Set up job alerts for even quicker access. 📅

What Fellow Engineers Say