To be successful in this role, you should meet the following requirements:
- Active Google Cloud Data Engineer Certification or Active Google Professional Cloud Architect Certification
- Experienced in data warehouse ETL build and design
- Hands-on GCP experience with a minimum of an end to end solution designed and implemented at production scale
- Hands on experience of Java, Python and Py-spark programming
- Experienced in designing, building and operational large-scale enterprise data solutions and applications using one or more of GCP data and analytics services in combination with 3rd parties - Cloud DataProc, Cloud Dataflow, Big Table, Cloud BigQuery, Cloud PubSub, Cloud storage, Cloud Functions, Datastream, application Integration, Data Fusion.
- Experienced in designing and building production data pipelines from data ingestion to consumption within a hybrid big data architecture, using Cloud Native GCP, Java, Python etc.
- Experienced in performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement
- Experience on SQL, Hadoop, Spark, Kafka, Airflow, Composer, Teraform, GitHub, Pyspark, Jenkins, Ansible,Attunity, Kafka, Kafka Connect etc.