Jobs Career Advice Post Job
X

Send this job to a friend

X

Did you notice an error or suspect this job is scam? Tell us.

  • Posted: Mar 28, 2026
    Deadline: Apr 10, 2026
    • @gmail.com
    • @yahoo.com
    • @outlook.com
  • Never pay for any notarisation, certificate or assessment as part of any recruitment process. When in doubt, contact us

    With over one hundred and thirty branches in Kenya, Tanzania, Uganda, and Burundi, some of which are 24/7 digital branches, DTB is committed to enabling people to advance with confidence and success. The Bank’s heritage and values are articulated in its brand promise, Achieve More, and brought to life through an engaged diverse workforce.
    Read more about this company

     

    Data Science Engineer, Assistant Senior Manager

    Job Purpose:

    In this role, you will design, build, and optimize the data engines that power DTB’s intelligence. You will develop robust data pipelines, feature stores, model‑serving systems, and scalable big‑data platforms that enable advanced credit scoring, fraud detection, customer intelligence, and a wide range of machine‑learning applications.

    You will be at the heart of transforming DTB into a data‑driven organization—ensuring that teams across the bank can rely on high‑quality, trusted, and scalable data to drive smarter decisions, stronger governance, and innovative digital solutions. This is a high‑impact role for a builder, a problem‑solver, and a visionary ready to shape the future of data and AI at DTB

    Key Responsibilities:

     Science & ML

    • Build and maintain ETL/ELT pipelines that feed modelling datasets from multiple banking systems (CBS, LMS, CRM, Cards, Mobile Banking, Bureau, Collections systems).
    • Develop automated data preparation workflows for credit scoring, fraud models, behavioral models, and IFRS9 modelling.
    • Create end-to-end ML pipelines integrating feature engineering, data validation, model deployment, and monitoring.
    • Manage and Build other Enterprise ETL using tools like ODI , informatica etc.

    Big Data Platform Engineering

    • Develop scalable data-processing workflows using Spark, Hadoop, Kafka, Airflow, Flink or similar.
    • Optimize large datasets (transactional, bureau, behavioural, logs) for modelling in batch and real-time environments.
    • Manage distributed computation and ensure reliability and fault tolerance.

    Feature Store & Data Assets Management

    • Design and maintain a centralized feature store for credit, fraud, marketing, and customer analytics models.
    • Ensure feature consistency between training and serving environments.
    • Implement versioning, lineage, documentation, and metadata management for data features.

    Model Deployment & MLOps

    • Collaborate with data scientists to deploy models using MLflow, Docker, Kubernetes, API gateways, CI/CD pipelines.
    • Develop automated monitoring pipelines for model performance, drift detection, data quality, and explainability.
    • Ensure models operate efficiently in real-time decision engines and batch scoring environments.

    Data Quality & Governance

    • Implement robust data validation, profiling, anomaly detection, and reconciliation checks.
    • Work with Data Governance teams to ensure compliance with IFRS9, Basel, CBK, GDPR, and internal data standards.
    • Manage data lineage, cataloguing, and documentation to support audits and regulatory reviews.

    Collaboration & Stakeholder Support

    • Partner with Data Scientists, Risk, Credit, Fraud, Marketing, and BI teams to align data pipelines with business use cases.
    • Work with IT and Infrastructure teams on cluster performance, security, access controls, and SLA adherence.
    • Participate in sprint planning, architecture reviews, and model implementation committee sessions.

    Performance Optimization

    • Improve the efficiency, scalability, and cost of ML workloads.
    • Optimize database queries, Spark jobs, Kafka streams, and storage systems.

    Qualifications & Experience:

    • Strong academic foundation with a Bachelor’s or Master’s in Computer Science, Data Engineering, Data Science, Information Technology, or a related quantitative field.
    • 3–7+ years of impactful, hands‑on experience in data engineering, big‑data processing, or building scalable ML infrastructure—ideally within fast‑paced, data‑driven environments.
    • Advanced programming capability, with strong proficiency in Python, SQL, and PySpark; experience with Scala is an added advantage.
    • Demonstrated expertise in modern data and ML platforms, including:
      • Big‑data technologies: Spark, Hadoop, Kafka, Airflow
      • MLOps & containerization: MLflow, Docker, Kubernetes
      • CI/CD pipelines: GitLab, Jenkins, GitHub Actions
      • Cloud platforms: AWS, GCP, or Azure (highly preferred)
    • Experience working with banking systems, risk data, or credit‑modelling datasets—a significant advantage that accelerates success in this role.

    Key Competencies

    • Strong understanding of data structures, distributed systems, and ML workflows.
    • Excellent problem-solving, debugging, and optimization skills.
    • Fast learner with ability to adapt to new technologies.
    • High attention to detail, documentation discipline, and data governance awareness.
    • Strong collaboration and communication skills.

    Check how your CV aligns with this job

    Method of Application

    Interested and qualified? Go to Diamond Trust Bank (DTB) on dtbk.dtbafrica.com to apply

    Build your CV for free. Download in different templates.

  • Send your application

    View All Vacancies at Diamond Trust Bank (DTB) Back To Home
Average Salary at Diamond Trust Bank (DTB)
KSh 87K from 17 employees
Mysalaryscale.com

Career Advice

View All Career Advice
 

Subscribe to Job Alert

 

Join our happy subscribers

 
 
 
Send your application through

GmailGmail YahoomailYahoomail