Job Description
What You’ll Do
- Design, build and maintain optimal data pipelines to transfer and manipulate data for ingestion or analytical purposes, and to operationally own them.
- Designing, building and maintaining systems to prevent quality issues from manifesting in production.
- Designing, building and maintaining systems to enable the monitoring of data quality & data processing performance through the data value chain.
- Build analytics tools that utilise the data pipeline to provide actionable insights into data quality, operational efficiency and other key business performance metrics.
- Work with stakeholders to assist with data-related technical issues and support their data infrastructure needs.
- Create data tools for analytics and data scientist team members that assist them in building and optimising our product into an innovative industry leader.
- Provide Feedback into Platform to optimise queries and propose optimisations
What Experience You Need
- 2+ years of experience as a data engineer or related role
- Bachelor’s degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field.
- Experience with GCP (Cloud certification strongly preferred)
- Intermediate skills using programming languages such as Python, SQL or scripting languages
- Basic understanding and experience with Google Cloud Platforms and an overall understanding of cloud computing concepts
- Experience building and maintaining simple data pipelines, following guidelines,
- transforming and entering data into a data pipeline in order for the content to be digested and usable for future projects
- Experience supporting the design and implementation of basic data models
- Demonstrates proficient Git usage and contributes to team repositories
What Could Set You Apart
- Master’s Degree
- Experience with big data tools: Hadoop, Spark, Kafka, etc.
- Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
- Experience with data pipeline and workflow management tools: Airflow, GCP dataflow etc.
- Experience with AI or Machine Learning
- Experience with Data Visualisation Tools such as Tableau or Looker
- Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.