NEW
Proxify is bringing transparency to tech team performance based on research conducted at Stanford. An industry first, built for engineering leaders.
Learn more
Amine T.
Data Engineer
Amine is a Lead Data Engineer and Data Architect with eight years of commercial experience in designing and implementing large-scale data platforms. He specializes in cloud-based data architectures using Databricks, Spark, and AWS, and has strong expertise in Terraform, Airflow, and CI/CD automation.
He delivers enterprise-grade data lakehouses and streaming solutions for organizations such as Vallourec, RATP, and Société Générale, optimizing data pipelines, enabling predictive maintenance, and driving analytics modernization across AWS and Azure.
A Databricks Certified Data Engineer Professional, Amine is passionate about building scalable, cost-efficient, and secure data ecosystems that bridge business needs with engineering excellence.
Main expertise
- AWS 4 years
- Apache Spark 8 years
- Terraform 4 years

Other skills
Selected experience
Employment
Data Architect | Lead Data Engineer
Vallourec - 10 months
- Designed and developed scalable ETL pipelines using AWS Glue and Spark Scala for large-scale data processing.
- Architected and maintained AWS infrastructure (S3, Glue, Lambda, IAM, Step Functions), ensuring reliability and cost efficiency.
- Built CI/CD pipelines and enforced engineering standards through code reviews and governance policies.
- Implemented data quality checks, lineage tracking, and access controls to uphold data integrity and compliance.
- Led a team of five data engineers, overseeing Terraform provisioning, Azure DevOps pipelines, and Spark performance optimization.
- Collaborated with data scientists to deploy ML models improving asset efficiency and reducing downtime.
- Delivered QuickSight dashboards enabling secure, real-time business insights.
Technologies:
- Technologies:
Databricks
Python
SQL
Terraform
DevOps
ETL
- Team Leading
PySpark
Data Architect | Lead Data Engineer
RATP Group - 2 years
- Designed and implemented a Data Mesh architecture on Databricks (AWS) and built governed data repositories in Glue and Collibra.
- Developed data ingestion and sharing pipelines in Spark/Scala and PySpark with full CI/CD automation.
- Created Databricks job automation tools using Terraform and orchestrated workflows via AWS MWAA.
- Designed data pipelines for predictive maintenance and passenger flow analytics using Kafka, Spark, and AWS Glue.
- Implemented data quality monitoring with Airflow and Great Expectations, ensuring high data reliability.
- Collaborated with data scientists to operationalize forecasting models for service optimization.
Technologies:
- Technologies:
AWS
Databricks
Apache Spark
Python
Apache Kafka
SQL
Terraform
DevOps
- Data Engineering
Apache Airflow
- Data Modeling
Redshift
Data Engineer
Société Générale - 3 years 3 months
- Migrated production applications from HDP to Cloudera, creating and configuring multiple environments to ensure smooth transition.
- Developed CI/CD pipelines and Terraform jobs to provision and scale VMs across environments.
- Supported engineering teams throughout the migration phase, ensuring minimal downtime.
- Developed and deployed Spark Scala libraries, and orchestrated production jobs for high availability.
- Designed and implemented NiFi pipelines for ingesting data from external APIs.
- Built data ingestion and transformation frameworks for market and risk data pipelines using Spark and Hadoop.
- Automated data quality checks and implemented lineage tracking using Apache Atlas.
- Collaborated with quant teams to improve risk model data accuracy and reduce latency in downstream analytics.
- Optimized HDFS and Hive-based data lakes, improving performance and storage efficiency.
- Contributed to regulatory reporting automation, ensuring compliance with Basel III standards.
Technologies:
- Technologies:
Apache Spark
Python
SQL
Scala
Terraform
- Data Engineering
Apache Hive
Hadoop
- Data Quality
Big Data Developer
BNP Paribas - 1 year
- Built Spark-based data pipelines and ETL workflows in Talend and Kafka for real-time regulatory and anti-fraud reporting.
- Developed an AWS-based architecture, transforming CSV data to Parquet and optimizing Hive tables for performance.
- Automated deployment workflows with Jenkins and Ansible, improving development efficiency.
- Created Oozie bundles and Spark Scala jobs to implement business rules and manage data in Cassandra.
- Indexed data with Solr to enable fast search capabilities and supported production deployment and monitoring.
Technologies:
- Technologies:
Apache Spark
Python
Apache Kafka
SQL
Apache Hive
ETL
Hadoop
Education
BSc.Diplôme d'ingénieur, Ingénierie
Tunisia Polytechnic School · 2014 - 2017
Standalone courseMathématiques-Physique
IPEIN - Institut Préparatoire aux Études d'Ingénieur de Nabeul · 2012 - 2014
Find your next developer within days, not months
In a short 25-minute call, we would like to:
- Understand your development needs
- Explain our process to match you with qualified, vetted developers from our network
- You are presented the right candidates 2 days in average after we talk
