Job Summary
We are looking for a skilled Data Engineer with strong Python and Databricks experience to design, build, and maintain scalable data pipelines and platforms that enable analytics, reporting, and advanced data use cases. You will work closely with Data Scientists, BI Developers, and business stakeholders to deliver high-quality, production-ready data solutions across cloud-based environments.
This role requires strong engineering fundamentals, hands-on Databricks development, and experience working with large, complex datasets.
What you’ll do:
- Design, develop, and maintain scalable data pipelines using Python and Databricks
- Build robust ETL/ELT processes to ingest structured and unstructured data
- Develop and optimize Spark jobs for performance and reliability
- Implement data models to support analytics and downstream consumption
- Integrate multiple data sources (APIs, databases, files, streaming sources)
- Ensure data quality, validation, and governance standards are met
- Collaborate with Data Scientists and BI teams to operationalize analytics solutions
- Implement monitoring, logging, and error-handling mechanisms
- Support production deployments and troubleshoot pipeline issues
- Participate in code reviews and contribute to engineering best practices
- Document solutions and processes clearly
Your Expertise:
- Experience with Azure Databricks or AWS Databricks
- Knowledge of data orchestration tools (Airflow, ADF, etc.)
- Experience with streaming technologies (Kafka, Event Hubs, etc.)
- Exposure to data governance and security frameworks
- Experience supporting machine learning pipelines
- Understanding of DevOps practices
Qualification:
- 5+ years’ experience as a Data Engineer or similar role
- Strong proficiency in Python for data engineering
- Hands-on experience with Databricks (notebooks, jobs, workflows)
- Solid understanding of Apache Spark concepts
- Experience building ETL/ELT pipelines
- Strong SQL skills
- Experience working in cloud environments (AWS / Azure / GCP)
- Familiarity with Delta Lake or similar lakehouse architectures
- Version control (Git) and CI/CD exposure
- Strong problem-solving and analytical skills