We are looking for a data engineer to develop data pipeline solutions to extract, load and transform various forms of event-level data. The data engineer will work with structured, semi structured and unstructured data which will be loaded into a data warehouse / data lake via a centralised data pipeline.
This crucial process is the foundation of deriving knowledge from data which helps us build better products and prepares for AI. In this role, you should have a background in building data engineering solutions using scripting languages such as Python, R and Scala.
Exposure to AWS / Azure / GCP data architecture tools and processes
PostgreSQL (1-2 years)
MySQL (1-2 years)
Linux Server Management and Security (2-4 years)
Ruby (1-2 years)
Technical understanding of Hadoop, MapReduce, HIVE, PIG, Kafka
Worked on a Hadoop project via EMR / HDInsight / Databricks / Dataproc
Demonstrable data visualization skills (Power BI / Tableau / Qlikview)
BSc or BTech majoring in Computer Science will be advantageous, however your ability to demonstrate your in-depth understanding of the web technologies and sound software engineering practices will trump a formal qualification.