Information TechnologyFull-TimeSenior-level(6+ yrs)
Job Description
Role Overview
In this role, you will design, build, and optimize the data engines that power DTB’s intelligence. You will develop robust data pipelines, feature stores, model‑serving systems, and scalable big‑data platforms that enable advanced credit scoring, fraud detection, customer intelligence, and a wide range of machine‑learning applications.
You will be at the heart of transforming DTB into a data‑driven organization—ensuring that teams across the bank can rely on high‑quality, trusted, and scalable data to drive smarter decisions, stronger governance, and innovative digital solutions. This is a high‑impact role for a builder, a problem‑solver, and a visionary ready to shape the future of data and AI at DTB.
Key Responsibilities
Science & ML Engineering
Build and maintain ETL/ELT pipelines that feed modelling datasets from multiple banking systems (CBS, LMS, CRM, Cards, Mobile Banking, Bureau, Collections systems).
Develop automated data preparation workflows for credit scoring, fraud models, behavioral models, and IFRS9 modelling.
Create end-to-end ML pipelines integrating feature engineering, data validation, model deployment, and monitoring.
Manage and Build other Enterprise ETL using tools like ODI, Informatica, etc.
Big Data Platform Engineering
Develop scalable data-processing workflows using Spark, Hadoop, Kafka, Airflow, Flink, or similar.
Optimize large datasets (transactional, bureau, behavioural, logs) for modelling in batch and real-time environments.
Manage distributed computation and ensure reliability and fault tolerance.
Feature Store & Data Assets Management
Design and maintain a centralized feature store for credit, fraud, marketing, and customer analytics models.
Ensure feature consistency between training and serving environments.
Implement versioning, lineage, documentation, and metadata management for data features.
Model Deployment & MLOps
Collaborate with data scientists to deploy models using MLflow, Docker, Kubernetes, API gateways, and CI/CD pipelines.
Develop automated monitoring pipelines for model performance, drift detection, data quality, and explainability.
Ensure models operate efficiently in real-time decision engines and batch scoring environments.
Data Quality & Governance
Implement robust data validation, profiling, anomaly detection, and reconciliation checks.
Work with Data Governance teams to ensure compliance with IFRS9, Basel, CBK, GDPR, and internal data standards.
Manage data lineage, cataloguing, and documentation to support audits and regulatory reviews.
Collaboration & Stakeholder Support
Partner with Data Scientists, Risk, Credit, Fraud, Marketing, and BI teams to align data pipelines with business use cases.
Work with IT and Infrastructure teams on cluster performance, security, access controls, and SLA adherence.
Participate in sprint planning, architecture reviews, and model implementation committee sessions.
Performance Optimization
Improve the efficiency, scalability, and cost of ML workloads.
Optimize database queries, Spark jobs, Kafka streams, and storage systems.
Qualifications & Experience
Strong academic foundation with a Bachelor’s or Master’s in Computer Science, Data Engineering, Data Science, Information Technology, or a related quantitative field.
3–7+ years of impactful, hands‑on experience in data engineering, big‑data processing, or building scalable ML infrastructure—ideally within fast‑paced, data‑driven environments.
Advanced programming capability, with strong proficiency in Python, SQL, and PySpark; experience with Scala is an added advantage.
Demonstrated expertise in modern data and ML platforms, including: Spark, Hadoop, Kafka, Airflow, MLflow, Docker, Kubernetes, and CI/CD pipelines (GitLab, Jenkins, GitHub Actions).
Experience working with banking systems, risk data, or credit‑modelling datasets is a significant advantage.
Key Competencies
Strong understanding of data structures, distributed systems, and ML workflows.
Excellent problem-solving, debugging, and optimization skills.
Fast learner with ability to adapt to new technologies.
High attention to detail, documentation discipline, and data governance awareness.
Strong collaboration and communication skills.
How to Apply
Interested and qualified candidates should apply online via the Diamond Trust Bank recruitment portal at dtbk.dtbafrica.com.
How to Apply
Interested and qualified candidates should apply directly through the Diamond Trust Bank (DTB) recruitment portal at https://dtbk.dtbafrica.com.