Job Summary
We are seeking a skilled and motivated Data Engineer with a strong background in data integration, ETL processes, and data warehousing. The ideal candidate will have over 3 years of experience in data engineering, with a good understanding of data architecture, pipeline creation, and big data technologies. This role requires a proactive individual who can design, build, and maintain data systems, collaborate with cross-functional teams, and ensure the highest standards of data quality and performance.
Job Responsibilities
- Design, develop, and maintain scalable data pipelines for data ingestion, processing, and storage.
- Build and optimize data architectures and data models for efficient data storage and retrieval.
- Develop ETL processes to transform and load data from various sources into data warehouses and data lakes.
- Ensure data integrity, quality, and security across all data systems.
- Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions that meet business needs.
- Monitor and troubleshoot data pipelines and workflows to ensure high availability and performance.
- Document data processes, architectures, and data flow diagrams.
Requirements
Candidate Required Skills
- Bachelor's degree in Computer Science, Engineering, or a related field.
- 3+ years of experience in data engineering and data architecture.
- Proficiency in SQL and at least one programming language (e.g., Python, Java, Scala).
- Experience with cloud data platforms (e.g., AWS, Azure, GCP) and their data services.
- Knowledge of ETL tools and frameworks (e.g., Apache NiFi, Talend, Informatica).
- Understanding of data modeling, data structures, and database design.
- Strong analytical and problem-solving skills.
- Strong communication and collaboration skills.
- Ability to work independently and as part of a team.
- Experience with data warehousing solutions (e.g., Redshift, BigQuery, Snowflake).
- Implement and maintain data integration solutions using industry-standard tools and technologies (e.g., Apache Spark, Kafka, Airflow, etc.).
- Hands-on experience with big data technologies (e.g., Hadoop, Spark, Kafka).
- Knowledge of data governance and best practices in data management.
- Familiarity with containerization and orchestration tools (e.g., Docker, Kubernetes).
- Experience with data visualization tools (e.g., Tableau, Power BI).
Benefits
- This role offers the flexibility of working remotely in India.
Sahiba
8296043355