We are looking for a skilled Azure Data Engineer who enjoys working with large datasets and building efficient, scalable data systems. In this role, you will design and maintain modern data pipelines using Azure technologies and help transform raw data into meaningful insights for business teams.
Key Responsibilities
- Develop and maintain robust data pipelines using PySpark and Azure Databricks
- Build and orchestrate workflows with Azure Data Factory
- Manage and optimize data storage using Azure Data Lake
- Work with structured and unstructured data from multiple sources
- Design efficient data models to support analytics and reporting
- Ensure data quality, performance, and reliability across systems
- Collaborate with analysts, developers, and stakeholders to understand data needs
- Continuously improve pipeline performance and cost efficiency
Required Skills & Experience
- Hands-on experience in Python and PySpark
- Strong working knowledge of:
- Azure Databricks
- Azure Data Factory
- Azure Data Lake
- Azure Synapse / SQL Data Warehouse
- Solid understanding of SQL and relational databases
- Experience with big data ecosystems such as Apache Spark, Kafka, or Hive
- Familiarity with data modeling and ETL processes
Preferred Skills
- Experience working with Delta Lake
- Exposure to CI/CD pipelines and Azure DevOps
- Knowledge of Docker or Kubernetes
- Experience with visualization tools like Power BI or Tableau
- Basic understanding of machine learning workflows