JUMIA is an e-commerce startup with an aim to mimic Amazon’s success by delivering a wide range of items, from toys to generators across the African continent. Shortly after the startup of the business in Nigeria, JUMIA launched warehouses in four other growing countries which include: Egypt, Morocco, Kenya and Cote d'Ivoire.
This AWS Data Engineer role involves designing, developing, and maintaining robust data pipelines using modern cloud technologies, primarily within the AWS ecosystem. This position is crucial for managing large volumes of raw and processed data, orchestrating complex workflows, and ensuring data quality to support Jumia's analytics and reporting requirements.
What you will be doing (Key Responsibilities)
- Design and develop pipelines using Python, PySpark, and SQL.
- Use GitLab as the versioning control system.
- Utilize S3 buckets for storing large volumes of raw and processed data.
- Implement and manage complex data workflows using Apache Airflow (MWAA) to orchestrate tasks.
- Utilize Apache Iceberg (or similar) for managing and organizing data in the data lake.
- Create and maintain data catalogs using AWS Glue Catalog to organize metadata.
- AWS Athena for interactive querying.
- Familiarize with data modeling techniques to support analytics and reporting requirements, as well as knowledge of the data journey stages within a data lake (Medallion Architecture).
Qualifications and Requirements
We are looking for candidates who meet the following criteria:
- Ideally, a degree in Information Technology, Computer Science, or a related field (BA/BSc/HND qualification).
- Ideally, +5 years of experience within the Data Engineering landscape.
- Strong expertise in Python, PySpark, SQL, and the overall AWS data ecosystem.
- Strong problem-solving and analytical skills.
- Ability to explain technical concepts to non-technical users.
- Proficiency to work with Github.
- Terraform and CICD pipelines are a great 'nice-to-have'.