Job Summary
- Familiar with working with data in a distributed computing and parallel processing environments using programming models (MapReduce programming).
- Eagerness to learn new emerging technologies, like Spark on Apache Hadoop : understanding of Yarn framework
- Real time computations - Storm, Spark streaming
- Message broking / inbound data architecture - Flume, Kafka brokers, RabbitMQ (conceptual understanding should be ok)
- SQL with experience on MPP like Greenplum or HAWQ, Asterdata
- Cloud deployment experience (AWS) - familiarity
- Experience performing file management on Hadoop (HDFS)
- Experience with PIG
Didn’t find the job appropriate? Report this Job
Download the iimjobs app to
apply for jobs anywhere, anytime
Download on
App Store
Get it on
Google Play
Scan to Download