Job Description:
We are seeking a talented Data Engineer proficient in Python, Spark, PySpark, and Snowflake to join our dynamic team. In this role, you will be responsible for designing, developing, and maintaining data pipelines and ETL processes to support our data infrastructure.
Note:- notice period 30 days
Responsibilities:
- Develop and maintain scalable data pipelines using Python, Spark, and PySpark.
- Design and optimize ETL processes to move data between various systems, with a focus on Snowflake.
- Collaborate with data scientists and analysts to understand data requirements and deliver actionable insights.
- Ensure data quality and reliability through testing, monitoring, and validation.
- Troubleshoot performance bottlenecks and implement optimizations.
- Work closely with cross-functional teams to integrate new data sources and improve data accessibility.
- Stay current with industry trends and best practices in data engineering, distributed computing, and cloud technologies.
Requirements:
- Bachelor's degree in Computer Science, Engineering, or a related field (or equivalent work experience).
- Proven experience as a Data Engineer or similar role with expertise in Python, Spark, PySpark, and Snowflake.
- Strong understanding of distributed computing principles.
- Experience with data modeling, ETL development, and data warehousing concepts.
- Experience with cloud platforms such as AWS, Azure, or Google Cloud.
- Excellent communication skills and ability to work effectively in a collaborative team environment.
- Problem-solving aptitude and a willingness to learn new technologies.