Python Job: Data Engineer

Job added on


Ocado Group


Toronto - Canada

Job type


Python Job Details

Founded in 2014, Kindred is a robotics and artificial intelligence (AI) company that develops robots to solve real-world problems. Its mission is to enhance the lives of human workers with the power of AI robotics. Kindred’s cutting-edge technology is the foundation of a number of proprietary platforms, including CORE with AutoGrasp™, developed to operate robots autonomously in dynamic environments. Its team of scientists, engineers and business operators have set a new standard for reinforcement learning for robots. The company is co-located in San Francisco and Toronto and is part of the UK-based Ocado Group plc.

As a Data Engineer you’ll be building the foundation for all our Machine Learning services, processes and projects. Real time telemetry, metrics and training events come from a variety of sources, including human annotations. channelled through GCP and AWS. You’ll make sure it’s validated, organised and accessible to developers, enabling them to iterate quickly on Machine Learning experiments.

What you’ll do:

  • Design and deploy APIs to organise internal robotics data to be used in cloud training pipelines
  • Define feature stores to manage evolving data schemas
  • Create ETL pipelines to transform data streams or batches
  • Ensure data validity, consistency, and cleanliness
  • Proactively design data unit tests to ensure quality and gain visibility into potential anomalies
  • Monitor data objects and provide operational support
  • Promoting best practices in data engineering (i.e naming conventions, pipeline design standards, DBT model best practices)
  • Stay up-to-date with industry developments including conferences, launch of new technologies and new patterns in data engineering.

What you bring:

  • Experienced with data engineering and data modelling, including solid SQL knowledge
  • Experienced with Python development applied to data engineering
  • Have worked with cloud data technologies (AWS, GCP) and have built production ETL/ELT pipelines
  • Experienced with data architecture patterns such as event sourcing, change data capture (CDC), incremental loading, data contracts
  • Excited to own development from the ground up
  • Favour long-term thinking and is focused on stability
  • You value collaborative over individual ways of working and are comfortable being challenged by others

Bonus qualifications:

  • Have worked with robotics and/or machine learning
  • Experienced working with DBT, Dataflow, Event Queues