Role and Responsibilities
What You'll Do
- Be responsible for system architecture design and data engineering activities including data ingestion, modelling, processing and data governance
- Work with team to improve the data pipeline to process large scale data efficiently
- Be a data steward to educate and promote the data importance and data-driven culture.
- Be constantly challenged to learn and grow with new technologies, identify and solve complex problems via data
- Collaborate with cross-functional teams, including data scientists and business stakeholders.
- Design and implement data architecture that supports analytical needs and scalability
- Implement and maintain cloud-based data solutions for various use cases ranging from batch data processing, near-real time data
Skills and Qualifications:
- Good experience with open-source technologies such as Airflow, Superset, Datahub, Great Expectation
- Good Hadoop ecosystem knowledge (e.g. Hadoop, Spark, Kafka, Hive, Presto)
- Proficient in at least one programming language and scripting languages (e.g. Python, Scala, Java, Bash)
- Good experience with CICD tools such as GitHub Actions
- Familiar with Agile/Scrum software development methodologies
Job Types: Full-time, Permanent