Job Title: Snowflake Data Engineer
Job Summary:
We are seeking a skilled Snowflake Data Engineer to design, implement, and optimize large-scale data intelligence solutions on the Snowflake Cloud Data Warehouse. The ideal candidate will have a strong background in data architecture and analytics, with hands-on experience in building robust data pipelines and managing data workflows.
Key Responsibilities:
- Architect and implement production-grade data solutions utilizing Snowflake Data Warehouse, ensuring high performance and scalability.
- Develop and manage ETL workflows, leveraging tools such as Python, Snowpipe, and other ETL technologies for seamless data integration.
- Design and implement real-time and batch data processing pipelines, ensuring data accuracy and availability.
- Collaborate with cross-functional teams using Agile methodologies (Scrum) and tools like Azure DevOps or JIRA to deliver high-quality solutions.
- Write and optimize stored procedures and functions using Snowflake Scripting or JavaScript for efficient data manipulation.
- Utilize Snowflake features such as data sharing, event-driven architecture, and lake-house patterns to enhance data accessibility.
- Migrate data from RDBMS to Snowflake, employing best practices for data transformation and modeling (star/snowflake schema).
- Understand and implement Role-Based Access Control (RBAC) for secure user and data management.
- Create and manage file formats and external/internal stages for cloud storage integration (AWS S3, Azure Blob).
- Apply knowledge of Snowflake architecture, including storage and compute resources, to optimize data solutions.
- Maintain awareness of Snowflake features like Time Travel and Failsafe to ensure data integrity and availability.
- Translate BI and reporting requirements into effective database and reporting designs.
Basic Qualifications:
- Familiarity with Agile processes and project management tools (e.g., JIRA, Azure DevOps).
- Basic understanding of modern data pipeline automation in cloud environments.
Key Skills:
- Expertise in Snowflake architecture and internals, including data modeling and integration with BI/reporting tools.
- Proficient in programming languages such as Python, Java, or Scala for data processing.
- Experience with structured and semi-structured data formats (e.g., XML, JSON, Parquet).
- Knowledge of Snowflake utilities (SnowSQL, SnowPipe) and best practices for data ingestion and processing.
- Understanding of cloud data warehousing solutions like AWS Redshift or Google BigQuery is a plus.
Keywords: Snowflake, Snowpipe, ETL, Data Warehouse, SQL, Data Engineering, Cloud Solutions, Agile, Pyspark, Data Migration, RBAC.