Zakaria M.

Data Engineer

Zakaria is a skilled Data Engineer with six years of experience in IT, railways, and healthcare industries.

He has expertise in performing ETL/ELT operations using Azure Databricks. In addition to that, he is highly proficient in designing and implementing QlikSense solutions, enabling effective data visualization and analysis.

Zakaria's passion for Artificial Intelligence and Machine Learning drives his continuous exploration of innovative solutions. He actively stays updated with the latest advancements in these fields, seeking opportunities to apply cutting-edge techniques in data analysis and predictive modelling.

Main expertise

  • Apache Spark
    Apache Spark 5 years
  • CSV 6 years
  • Data Engineering 6 years

Other skills

  • Apache Airflow
    Apache Airflow 3 years
  • Qlik Sense
    Qlik Sense 1 years
  • Git
    Git 1 years
Zakaria

Zakaria M.

Portugal

Get started

Selected experience

Employment

  • Sr. Data Engineer

    KWAN - 1 year 4 months

    -Design and deliver scalable ELT pipelines on Databricks using PySpark, Spark SQL, and medallion architecture (Bronze/Silver/Gold). -Implement data governance practices including Unity Catalog for centralized metadata management and data lineage tracking. -Apply partitioning and caching strategies to optimize query performance and reduce processing times. -Manage CI/CD orchestration with Azure DevOps for automated testing and deployment of data assets. -Integrate Databricks jobs with Airflow DAGs for end-to-end workflow orchestration and dependency management.

    Technologies:

    • Technologies:
    • Databricks Databricks
    • Apache Spark Apache Spark
    • Azure Blob storage Azure Blob storage
    • Apache Airflow Apache Airflow
    • Azure Synapse Azure Synapse
    • PySpark PySpark
    • Azure DevOps Azure DevOps
  • Sr. Data Engineer

    loomtech - 6 months

    -Built scalable ELT pipelines with Databricks using PySpark/Spark SQL for high-volume energy data processing. -Leveraged Delta Lake with time travel and CDC logic to ensure data consistency, auditability, and rollback capabilities. -Applied partitioning and caching strategies for performance optimization across large-scale transformations. -Managed CI/CD pipelines using GitLab for version-controlled deployments across environments. -Conducted code reviews and cross-team collaboration to maintain engineering standards and code quality.

    Technologies:

    • Technologies:
    • Databricks Databricks
    • Apache Spark Apache Spark
    • SQL SQL
    • PySpark PySpark
    • GitLab GitLab
    • Azure DevOps Azure DevOps
  • Data Engineer

    SDG Group (Client : Johnson & Johnson) - 4 years 4 months

    • Responsible for performing ETL/ELT operations with Azure Databricks using PySpark and Spark SQL. This involved extracting data from various sources, transforming it into a usable format, and loading it into a target database or data warehouse;
    • Designed, developed, tested, debugged, and implemented QlikSense solutions based on specified requirements from business users related to dashboard goals and objectives. This included creating interactive data visualizations and reports to help stakeholders make informed decisions. The tools used in this role included Azure Databricks, Apache Spark (PySpark, Spark SQL), Qlik Sense, and Qlik NPrinting.

    Technologies:

    • Technologies:
    • Databricks Databricks
    • Apache Spark Apache Spark
    • Qlik Sense Qlik Sense
    • Azure Data Factory Azure Data Factory
    • PySpark PySpark
  • Data Engineer

    ONCF Rabat - 3 years 3 months

    • Responsible for analyzing, designing, and implementing ETLs/ELT workflows for various projects. This included the CRM project used for the management of prospects, customers, and marketing campaigns, as well as the BI project for selling passenger train tickets;
    • Designed and implemented automated ETL/ELT jobs with scheduling to feed the centralized supplier and purchasing management database from different data sources. The tools used in this role included Talend (DI, Big Data, ESB) (Java), Azure Databricks, Hadoop Ecosystem, Apache Spark (Spark SQL, PySpark), Databricks, PostgreSQL, and Power BI.

    Technologies:

    • Technologies:
    • PostgreSQL PostgreSQL
    • Databricks Databricks
    • Apache Spark Apache Spark
    • T-SQL T-SQL
    • SQL SQL
    • Azure Azure
    • Talend Talend
    • PySpark PySpark
  • ETL Engineer

    Akka Technologies - 1 year 8 months

    • Performance tuning: Optimizing the ETL process to ensure it runs efficiently and effectively, including tuning database queries and optimising data flows;
    • Responsible for analyzing detailed functional specifications and elaborating technical specifications for ETL projects. This included designing and executing ETL projects to generate reports and destination files according to the client's needs;
    • Ensured rigour in the validation and delivery phases of these projects. The tool used in this role was Talend ETL, along with PostgreSQL.

    Technologies:

    • Technologies:
    • MSSQL MSSQL
    • T-SQL T-SQL
    • SQL SQL
    • Talend Talend

Education

  • MSc.State Engineer in Computer Science Engineering

    National School of Computer Science and Systems Analysis - ENSIAS Rabat · 2012 - 2016

  • MSc.State Engineer in Computer Science Engineering

    Ecole Nationale Supérieure d'Informatique et d'Analyse des Systèmes - ENSIAS Rabat · 2012 - 2016

Find your next developer within days, not months

In a short 25-minute call, we would like to:

  • Understand your development needs
  • Explain our process to match you with qualified, vetted developers from our network
  • You are presented the right candidates 2 days in average after we talk

Not sure where to start? Let’s have a chat