Data Storage and Computation Frameworks, Performance Optimizations
Analytics & Visualizations
Infrastructure & Cloud Computing
Data Management Platforms
Implement scalable architectural models for data processing and storage
Build functionality for data ingestion from multiple heterogeneous sources in batch & realtime mode
Build functionality for data analytics, search and aggregation
Required Skills
Overall 5+ years of IT experience with 3+ years in Data related technologies
Minimum 2.5 years of experience in Big Data technologies and working exposure in at least one cloud platform on related data services (AWS / Azure / GCP)
Hands-on experience with the Hadoop stack HDFS, sqoop, kafka, Pulsar, NiFi, Spark, Spark Streaming, Flink, Storm, hive, oozie, airflow and other components required in building end to end data pipeline
Strong experience in at least of the programming language Java, Scala, Python. Java preferable
Hands-on working knowledge of NoSQL and MPP data platforms like Hbase, MongoDb, Cassandra, AWS Redshift, Azure SQLDW, GCP BigQuery etc
Well-versed and working knowledge with data platform related services on at least 1 cloud platform, IAM and data security