Amine T.

Data Engineer

Amine is a Lead Data Engineer and Data Architect with eight years of commercial experience in designing and implementing large-scale data platforms. He specializes in cloud-based data architectures using Databricks, Spark, and AWS, and has strong expertise in Terraform, Airflow, and CI/CD automation.

He delivers enterprise-grade data lakehouses and streaming solutions for organizations such as Vallourec, RATP, and Société Générale, optimizing data pipelines, enabling predictive maintenance, and driving analytics modernization across AWS and Azure.

A Databricks Certified Data Engineer Professional, Amine is passionate about building scalable, cost-efficient, and secure data ecosystems that bridge business needs with engineering excellence.

Main expertise

  • AWS
    AWS 4 years
  • Apache Spark
    Apache Spark 8 years
  • Terraform
    Terraform 4 years

Other skills

    Amine

    Amine T.

    France

    Get started

    Selected experience

    Employment

    • Data Architect | Lead Data Engineer

      Vallourec - 10 months

      • Designed and developed scalable ETL pipelines using AWS Glue and Spark Scala for large-scale data processing.
      • Architected and maintained AWS infrastructure (S3, Glue, Lambda, IAM, Step Functions), ensuring reliability and cost efficiency.
      • Built CI/CD pipelines and enforced engineering standards through code reviews and governance policies.
      • Implemented data quality checks, lineage tracking, and access controls to uphold data integrity and compliance.
      • Led a team of five data engineers, overseeing Terraform provisioning, Azure DevOps pipelines, and Spark performance optimization.
      • Collaborated with data scientists to deploy ML models improving asset efficiency and reducing downtime.
      • Delivered QuickSight dashboards enabling secure, real-time business insights.

      Technologies:

      • Technologies:
      • Databricks Databricks
      • Python Python
      • SQL SQL
      • Terraform Terraform
      • DevOps DevOps
      • ETL ETL
      • Team Leading
      • PySpark PySpark
    • Data Architect | Lead Data Engineer

      RATP Group - 2 years

      • Designed and implemented a Data Mesh architecture on Databricks (AWS) and built governed data repositories in Glue and Collibra.
      • Developed data ingestion and sharing pipelines in Spark/Scala and PySpark with full CI/CD automation.
      • Created Databricks job automation tools using Terraform and orchestrated workflows via AWS MWAA.
      • Designed data pipelines for predictive maintenance and passenger flow analytics using Kafka, Spark, and AWS Glue.
      • Implemented data quality monitoring with Airflow and Great Expectations, ensuring high data reliability.
      • Collaborated with data scientists to operationalize forecasting models for service optimization.

      Technologies:

      • Technologies:
      • AWS AWS
      • Databricks Databricks
      • Apache Spark Apache Spark
      • Python Python
      • Apache Kafka Apache Kafka
      • SQL SQL
      • Terraform Terraform
      • DevOps DevOps
      • Data Engineering
      • Apache Airflow Apache Airflow
      • Data Modeling
      • Redshift Redshift
    • Data Engineer

      Société Générale - 3 years 3 months

      • Migrated production applications from HDP to Cloudera, creating and configuring multiple environments to ensure smooth transition.
      • Developed CI/CD pipelines and Terraform jobs to provision and scale VMs across environments.
      • Supported engineering teams throughout the migration phase, ensuring minimal downtime.
      • Developed and deployed Spark Scala libraries, and orchestrated production jobs for high availability.
      • Designed and implemented NiFi pipelines for ingesting data from external APIs.
      • Built data ingestion and transformation frameworks for market and risk data pipelines using Spark and Hadoop.
      • Automated data quality checks and implemented lineage tracking using Apache Atlas.
      • Collaborated with quant teams to improve risk model data accuracy and reduce latency in downstream analytics.
      • Optimized HDFS and Hive-based data lakes, improving performance and storage efficiency.
      • Contributed to regulatory reporting automation, ensuring compliance with Basel III standards.

      Technologies:

      • Technologies:
      • Apache Spark Apache Spark
      • Python Python
      • SQL SQL
      • Scala Scala
      • Terraform Terraform
      • Data Engineering
      • Apache Hive Apache Hive
      • Hadoop Hadoop
      • Data Quality
    • Big Data Developer

      BNP Paribas - 1 year

      • Built Spark-based data pipelines and ETL workflows in Talend and Kafka for real-time regulatory and anti-fraud reporting.
      • Developed an AWS-based architecture, transforming CSV data to Parquet and optimizing Hive tables for performance.
      • Automated deployment workflows with Jenkins and Ansible, improving development efficiency.
      • Created Oozie bundles and Spark Scala jobs to implement business rules and manage data in Cassandra.
      • Indexed data with Solr to enable fast search capabilities and supported production deployment and monitoring.

      Technologies:

      • Technologies:
      • Apache Spark Apache Spark
      • Python Python
      • Apache Kafka Apache Kafka
      • SQL SQL
      • Apache Hive Apache Hive
      • ETL ETL
      • Hadoop Hadoop

    Education

    • BSc.Diplôme d'ingénieur, Ingénierie

      Tunisia Polytechnic School · 2014 - 2017

    • Standalone courseMathématiques-Physique

      IPEIN - Institut Préparatoire aux Études d'Ingénieur de Nabeul · 2012 - 2014

    Find your next developer within days, not months

    In a short 25-minute call, we would like to:

    • Understand your development needs
    • Explain our process to match you with qualified, vetted developers from our network
    • You are presented the right candidates 2 days in average after we talk

    Not sure where to start? Let’s have a chat