Our client is looking for candidates who have a broad set of technology skills and who can demonstrate an ability to apply Hadoop solutions to big data problems and learn quickly as the platform develops.

Minimum of 3 years of programming on Hadoop and 5 years of systems development as well as:

o Implementation of ETL applications
o Application/implementation of custom analytics
o Administration of relational databases
o Data migration from existing data stores
o Infrastructure and storage design
o Developing capacity plans for new and existing systems


• Hands-on experience with the Hadoop stack (e.g. MapReduce, Sqoop, Pig, Hive, Hbase, Flume) • Hands-on experience with related/complementary open source software platforms and languages (e.g. Java, Linux, Apache, Perl/Python/PHP, Chef) • Hands-on experience with ETL (Extract-Transform-Load) tools (e.g Informatica, Talend, Pentaho) • Hands-on experience with BI tools and reporting software (e.g. Microstrategy, Cognos, Pentaho) • Hands-on experience with analytical tools, languages, or libraries (e.g. SAS, SPSS, R, Mahout) • Hands-on experience with "productionalizing" Hadoop applications (e.g. administration, configuration management, monitoring, debugging, and performance tuning) • Well versed in installing & managing Cloudera distribution of Hadoop (CDH3, CDH4, Cloudera manager, MapR, Hortonworks etc.) • Hands-on experience with Hadoop based monitoring tools (NagiOS, Ganglia etc.) • Previous experience with high-scale or distributed RDBMS (Teradata, Netezza, Greenplum, Aster Data, Vertica) • Knowledge of cloud computing infrastructure (e.g. Amazon Web Services EC2, Elastic MapReduce) and considerations for scalable, distributed systems • Knowledge of NoSQL platforms (e.g. key-value stores, graph databases, RDF triple stores)

Reference Number



Several Locations



Apply For This Job

Bookmark the permalink.