Looking for an experienced, self-driven, analytical, and strategic Senior Data Engineer. In this role, you will work across a large and complex data lake/warehouse environment. You are passionate about working with disparate datasets in order to bring data together to answer business questions. You should have deep expertise in the creation and management of datasets and the proven ability to translate the data into meaningful insights through collaboration with product managers, Data engineers, business intelligence developer, operation managers and leaders. In this role, you will have ownership of end-to-end development of data engineering solutions to complex questions and youll play an integral role in strategic decision-making.
Knowledge & Skills
- In this role, you will have the opportunity to display and develop your skills in the following areas:
- Interface with PMs, business customers, and Data Architects/Modelers to understand requirements and implement solutions
- Design, develop, and operate highly-scalable, high-performance, low-cost, and accurate data pipelines in distributed data processing platforms with AWS technologies providing ad hoc access to large datasets and computing power
- Explore and learn the latest AWS big data technologies, evaluate and make decisions around the use of new or existing software products to design the data architecture
- Recognize and adopt best practices in data processing, reporting, and analysis: data integrity, test design, analysis, validation, and documentation Basic Qualifications
- Bachelors degree in Computer Science or related technical field, or equivalent work experience.
- 2+ years of work experience with ETL, Data Modeling, and Data Architecture.
- 2+ years of experience with SQL and large data sets, data modeling, ETL development, and data warehousing, or similar skills.
- 2+ year experience with AWS technologies stack including Redshift, RDS, S3, EMR or similar solutions build around Hive/Spark etc. OR 2+ year experience with Azure technologies stack including ADF, Azure BLOB storage, Azure synapse.
- Preferred Qualifications
- Excellent in ETL optimization, designing, coding, and tuning big data processes using Apache Spark or similar technologies.
- Experience operating very large data warehouses or data lakes.
- Experience with building data pipelines and applications to stream and process datasets at low latencies.
- Demonstrate efficiency in handling data tracking data lineage, ensuring data quality, and improving discoverability of data.
- Knowledge of distributed systems and data architecture design and implement batch and stream data processing pipelines, knows how to optimize the distribution, partitioning, and MPP.