Job Description
Want to join KANINI
We are looking for a Data Engineer who can build a robust database and its architecture. In this role, you will assess a wide range of requirements and apply relevant database techniques to create a sustainable data architecture before you begin the implementation process and develop the database from scratch.
You Are All Set To
Develop, maintain, evaluate, and test big data solutions. You will be involved in data engineering activities like creating pipelines/workflows for Source-to-Target Data Mapping among others.
You Are Someone Who Can
- You will be involved in the design of data solutions using Hadoop based technologies along with Hadoop, Azure, HDInsight for Cloudera based Data Late using Scala Programming.
- Liaise and be part of our extensive GCP community, contributing in the knowledge exchange learning programme of the platform.
- Be required to showcase your GCP Data engineering experience when communicating with business team on their requirements, turning these into technical data solutions.
- Be required to build and deliver Data solutions using GCP products and offerings.
- Hands on and deep experience working with Google Data Products (e.g. Big Query, Dataflow, Dataproc, AI Building Blocks, Looker, Cloud Data Fusion, Dataprep, etc.).
- Experience in Spark /Scala / Python/Java / Kafka.
- Responsible to Ingest data from files, streams and databases. Process the data with Hadoop, Scala, SQL Database, Spark, ML, IoT
- Develop programs in Scala and Python as part of data cleaning and processing
- Responsible to design and develop distributed, high volume, high velocity multi-threaded event processing systems
- Develop efficient software code for multiple use cases leveraging Python and Big Data technologies for various use cases built on the platform
- Provide high operational excellence guaranteeing high availability and platform stability
- Implement scalable solutions to meet the ever-increasing data volumes, using big data/cloud technologies Pyspark, Kafka, any Cloud computing etc
You Bring In
- Minimum 4+ years of experience in Big Data technologies
- Good to have experience in Cloud, GCP, AWS, Azure Data Engineering with background in Spark/Python/Scala / Java.
- Proficient in any of the programming languages Python, Scala or Java
- Mandatory experience in Mid to Expert Level programming capabilities in a large-scale enterprise
- In-depth experience in modern data platform components such as the Hadoop, Hive, Pig, Spark, Python, Scala, etc
- Experience with Distributed Versioning Control environments such as GIT
- Familiarity with development tools experience on either IntelliJ / Eclipse / VSCode IDE, Build Tool Maven
- Demonstrated experience in modern API platform design including how modern UI's are built consuming services / APIs.
- Experience on Azure cloud including Data Factory, Databricks, Data Lake Storage is highly preferred.
- Solid experience in all phases of Software Development Lifecycle plan, design, develop, test, release, maintain and support, decommission.
Your Qualification Is
B.E/B.Tech/M.C.A/MSc (preferably in Computer Science/IT)