E xtract, T ransform, and L oad (ETL) data sets in various file formats such as flat files, excel, Access, PDF, XML , JSON, EDI etc.
Maintain, optimize , and refactor existing ETL processes for improved performance and manageability
Build data pipelines to assemble large, complex sets of data that meet non-functional and functional business requirements
Create and maintain T-SQL scripts, stored procedures, functions, views, triggers, CTEs, Recursive table expressions etc. to handle large and complex datasets
Create and maintain SSIS packages for data updates
Ensure data quality and integrity in databases by developing robust validation scripts, and continuously make improvements to mitigate bugs, redundancies, and inconsistencies
Analyze and query multiple, complex databases and systems to produce various data analysis reports
Create and maintain documentation for data updates process using requirement specifications, schema and record layouts, programs, business requirements and other artifacts
Participate in daily scrum meetings, planning, retrospectives and other agile processes to support two weeks release cycles and assure consistent and accurate delivery of Healthcare data and software
What You Bring:
2+ years of data engineering experience working in partnership with large data sets
Solid experience, developing and implementing DW Architectures, OLAP & OLTP technologies and data modelling
Experienced with T- SQL and ETL automation
Proficient in using SSDT tools for building SQL server relational databases, Integration Services packages and Reporting Services reports
Experience with Python and/or Powershell
Experience with the version control system such as GIT , ADO etc.
Familiarity with BI & dashboarding tools and multi-dimensional modelling
Great problem-solving capabilities, troubleshooting data issues and experience in stabilizing big data systems.
Bonus points:
Experience with US Healthcare data domain is a plus
Certification in one of the cloud platforms (AWS/GCP/Azure)