Job Summary
We are looking for a skilled ETL Developer (Python) with strong hands-on experience in Databricks, PySpark, and Azure Data Factory (ADF). The ideal candidate will be responsible for translating existing Informatica mappings into PySpark-based implementations, ensuring high-quality and timely data pipeline delivery.
Roles & Responsibilities
- Design, develop, and maintain ETL pipelines using Python and PySpark
- Understand and analyze existing Informatica mapping logic
- Rewrite and migrate Informatica logic into PySpark workflows
- Develop and optimize data processing solutions on Databricks
- Work with Azure Data Factory (ADF) for orchestration and scheduling
- Perform SQL development for data extraction, transformation, and validation
- Deliver development tasks independently and within timelines
- Collaborate closely with technical teams to ensure smooth delivery
- Ensure code quality, performance, and scalability of data pipelines
Job Requirements
- 3+ years of experience in the IT industry
- 3+ years of experience as an ETL Developer
- Strong hands-on experience with Databricks and Python (Mandatory)
- Experience with Azure Data Factory (ADF)
- Proficiency in SQL and database development
- Hands-on experience with PySpark
- Exposure to Apache Spark, Hadoop is a plus
- Ability to work independently and efficiently to meet deadlines
- Strong problem-solving skills and attention to detail
- Good communication and teamwork skills
Nice to Have
- Experience in ETL migration projects (Informatica to PySpark)
- Familiarity with cloud-based data platforms
- Performance tuning and optimization experience