Eviden, part of the Atos Group, with an annual revenue of circa 5 billion is a global leader in data-driven, trusted and sustainable digital transformation. As a next generation digital business with worldwide leading positions in digital, cloud, data, advanced computing and security, it brings deep expertise for all industries in more than 47 countries. By uniting unique high-end technologies across the full digital continuum with 47,000 world-class talents, Eviden expands the possibilities of data and technology, now and for generations to come.
- Experience in Big Data technologies.
- Mandatory - Hands on experience in Python and PySpark.
- Build PySpark applications using Spark Data frames in Python using Jupyter notebook and PyCharm(IDE).
- Worked on optimizing spark jobs that processes huge volumes of data.
- Hands on experience in version control tools like Git.
- Worked on Amazon s Analytics services like Amazon EMR, Amazon Athena, AWS Glue.
- Worked on Amazon s Compute services like Amazon Lambda, Amazon EC2 and Amazon s Storage service like S3 and few other services like SNS.
- Experience/knowledge of bash/shell scripting will be a plus.
- Has built ETL processes to take data, copy it, structurally transform it etc. involving a wide variety of formats like CSV, TSV, XML and JSON.
- Experience in working with fixed width, delimited , multi record file formats etc.
- Good to have knowledge of data warehousing concepts - dimensions, facts, schemas- snowflake, star etc.
- Have worked with columnar storage formats- Parquet,Avro,ORC etc. Well versed with compression techniques - Snappy, Gzip.
- Good to have knowledge of AWS databases (at least one) Aurora, RDS, Redshift, Elastic ache, DynamoDB.
- Hands on experience in tools like Jenkins to build, test and deploy the applications.
- Awareness of Devops concepts and be able to work in an automated release pipeline environment.
- Excellent debugging skills.
- Ready to Work in shift timings of 2.00 -11.00pm