Job Description

What You’ll Do

  • Design, build and maintain optimal data pipelines to transfer and manipulate data for ingestion or analytical purposes, and to operationally own them.
  • Designing, building and maintaining systems to prevent quality issues from manifesting in production.
  • Designing, building and maintaining systems to enable the monitoring of data quality & data processing performance through the data value chain.
  • Build analytics tools that utilise the data pipeline to provide actionable insights into data quality, operational efficiency and other key business performance metrics.
  • Work with stakeholders to assist with data-related technical issues and support their data infrastructure needs.
  • Create data tools for analytics and data scientist team members that assist them in building and optimising our product into an innovative industry leader.
  • Provide Feedback into Platform to optimise queries and propose optimisations

What Experience You Need

  • 2+ years of experience as a data engineer or related role
  • Bachelor’s degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field.
  • Experience with GCP (Cloud certification strongly preferred)
  • Intermediate skills using programming languages such as Python, SQL or scripting languages
  • Basic understanding and experience with Google Cloud Platforms and an overall understanding of cloud computing concepts
  • Experience building and maintaining simple data pipelines, following guidelines,
  • transforming and entering data into a data pipeline in order for the content to be digested and usable for future projects
  • Experience supporting the design and implementation of basic data models
  • Demonstrates proficient Git usage and contributes to team repositories

What Could Set You Apart

  • Master’s Degree
  • Experience with big data tools: Hadoop, Spark, Kafka, etc.
  • Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
  • Experience with data pipeline and workflow management tools: Airflow, GCP dataflow etc.
  • Experience with AI or Machine Learning
  • Experience with Data Visualisation Tools such as Tableau or Looker
  • Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.