Looking for Developer experienced with Big Data ecosystem.
Responsibilities:
* Design and implement map reduce jobs to support distributed processing using java, cascading, python, hive, Impala and pig; Ability to design and implement end to end solution.
* Build libraries, user defined functions, and frameworks around Hadoop
* Research, evaluate and utilize new technologies/tools/frameworks around Hadoop eco system
* Develop user defined functions to provide custom hive and pig capabilities
* Define and build data acquisitions and consumption strategies
* Define & develop best practices
- Experience with Amazon AWS is a plus
* Work with support teams in resolving operational & performance issues
* Work with architecture/engineering leads and other teams on capacity planning
Qualification:
* Strong Java programming, shell scripting, Python, and SQL
* Strong development skills around Hadoop, Map Reduce, Hive, Pig, Impala
* Strong understanding of Hadoop internals
* Good understanding of AVRO and Json and other compression
* Experience with build tools such as Maven
* Experience with performance/scalability tuning, algorithms and computational complexity
* Experience (at least familiarity) with data warehousing, dimensional modeling and ETL development
* Ability to understand and ERDs and relational database schemas
* Proven ability to work cross functional teams to deliver appropriate resolution.
Education: Bachelor's or Higher in Computer Science or related or foreign equivalent