Some background on Hadoop:
MapReduce
Module used to develop Cloud platforms in Hadoop
Paas (Platform as a Service)
Typically have Pig in their background
Storm/Kafka
The Real time processing module in Hadoop
MUST HAVE Hbase/Cassandra
MUST HAVE Kafka experience
Falcon
Module in Hadoop for Data Management
ETL & Pig experience for this module
Positions Available
**For these Technical Staff roles; Hadoop developers won't be a fit for any of my positions. Those guys are users of Hadoop not the people who actually build it. That's why we usually look for people who have built a distributed systems from the ground up and have solid java or C++ development.***
Member of Technical Staff (MapReduce)
Apache Hadoop MapReduce is among the most popular, open-source, data processing systems in the world. We are looking for senior folks with experience in large-scale, distributed systems to help drive Hadoop MapReduce even further. Your primary focus will be scale, performance and scheduling in Apache Hadoop MapReduce.
Requirements:
Member of Technical Staff (Storm/Kafka)
Key responsibilities:
Drive architecture, design, and implementation of Apache Storm core components in collaboration with Apache Storm open source community Work on complex architecture related to real-time processing on Hadoop clusters running on thousands of nodes across data center Understand partner/customer requirements on integration with their existing event stream technologies and frameworks Work with product management and quality assurance teams to ensure deliver high quality products
Requirements:
Big pluses:
Member of Technical Staff – Oozie
Hortonworks is looking for passionate software engineers for the Data Management
development team within the Hortonworks Data Platform. This team is responsible for the components within the Hadoop ecosystem for managing data and moving data into and out of Hadoop, specifically Oozie, Sqoop and Flume. This position will focus initially on Oozie, but will eventually expand to include the other components.
Candidates should be experienced engineers who want to be part of taking Apache Oozie, Flume, Sqoop, and other ETL frameworks to the next level of functionality, stability, and enterprise readiness. To be successful in this position you will need to be able to work well with others in an open source community, share ideas, review designs, and collaborate to achieve optimal results. You must also be passionate about building quality into software from earliest stages of the development lifecycle until final delivery of a commercial quality product.
REQUIREMENTS:
• A MS degree in computer science or equivalent experience in industry
• Advanced Java programmer skills with a good grasp of key computer science
fundamentals including algorithms, data structures, multi-threading
• Advanced C++ can be a substitute for Java
• 3-8 years of relevant hands-on software engineering experience doing system software
design and development including distributed & large scale systems
• Experience with development of data management software, including experience
in distributed systems, workflow and scheduling systems, and/or ETL/ELT
• Highly desirable to have experience with enterprise schedulers such as Oozie,
Quartz, Azkaban or other similar solutions
• Experience with the Hadoop ecosystem is a plus
• Ability to coordinate across teams, including with QA, Doc writers, support, sales,
etc.
• Ability to interact with customers in pre-sales, planning, joint development, and
support situations.
• Strong oral and written communication skills