We are seeking a skilled Data Engineer with expertise in Databricks, Python, and PySpark to design, build, and optimize data pipelines and workflow orchestration across our infrastructure.
Key Responsibilities:
Develop and maintain ETL/ELT pipelines using Databricks.
Write efficient PySpark & Python scripts for large-scale data processing.
Implement data models, transformations, and CI/CD pipelines for automation.
Requirements:
3+ years in Data Engineering with Databricks.
Strong skills in Python, PySpark orchestration.
Experience with CI/CD, Git, Docker, and Kubernetes.