Hadoop Developer | Data Engineer | Software Engineer | Data Scientist


  • Hive

  • Oracle Database Development

  • SQL Server

  • Java

  • Hadoop

  • Oozie

  • Sqoop

  • Hbase

  • Cloudera Manager

  • C++

  • Zookeeper

  • Pig

  • Flume

  • Mysql

  • Hdfs

  • Cassandra

  • Mapreduce

  • Oracle PL/SQL

Read More

  • Job Posted: Tuesday, March 17, 2015

  • Expires On: Wednesday, March 25, 2015

  • Job Type: Permanent

  • Desired Salary: $130000.00/yr

  • Work Authorization: H1-B

  • 3 years of experience in application development, Database management, Apache Hadoop administration and development.
  • 2+ years of experience in big data technologies : Hadoop HDFS, Map-Reduce, Pig, Hive, Oozie, Flume, Hcatalog, Sqoop, zookeeper, NoSql : Cassandra and Hbase.
  • Experience in solving complex analytical problems and performing joins with Map-Reduce. 
  • Strong knowledge of apache hive administration and development. 
  • Experience in writing UDFS, UDAFS, and UDTFS in java for hive. 
  • Solid experience in Pig administration and development.
  • Experience in writing pig udfs (Eval, Filter, Load and Store) and macros.
  • Familiar with embedding hive and pig in java.
  • Experience in using Hcatalog for Hive, Pig and Hbase.
  • Worked on developing ETL processes to load data from multiple data sources to HDFS using FLUME and SQOOP, perform structural modifications using Map-Reduce, HIVE and analyse data using visualization/reporting tools.
  • Familiar with writing Oozie workflows and Job Controllers for job automation.
  • Familiar with importing and exporting data using Sqoop.
  • Experience in using Flume to stream data into HDFS.
  • Familiar with administering and developing in Cassandra and Hbase.
  • Experience in working with java Hbase APIs.
  • Experience in writing mrunit test cases and pig unit test cases.
  • Solid background in Object-Oriented analysis and design. 
  • In-depth understanding of Data Structure and Algorithms.
  • Good Understanding of Distributed Systems and Parallel Processing architecture.
  • Experience in handling multiple relational databases: Oracle, MySQL, SQL Server.
  • Familiar with Agile Methodology (SCRUM) and Software Testing.
  • Experience in deploying the multinode Hadoop cluster with different Hadoop components (HIVE, PIG, SQOOP, OOZIE, FLUME, HCATALOG, HBASE, ZOOKEEPER) using Cloudera Manager.
  • Experience in benchmarking, performing backup and recovery of NameNode meta data and data residing on cluster 
  • Experience in performing minor and major upgrades, commissioning and decommissioning of nodes on Hadoop cluster.
  • Strong knowledge in configuring NameNode high availability and NameNode federation. 
  • Experience in deploying Hadoop 2.0(YARN). 
  • Google certified professional in Analytics.
  • SAS Certified Advanced Programmer. 
  • Raleigh, NC

  • Menlo Park, CA

  • Mountain View, CA

  • Palo Alto, CA

  • Sacramento, CA

  • San Francisco, CA

  • San Jose, CA

  • Santa Clara, CA

  • Sunnyvale, CA

  • Jacksonville, FL

  • Miami Beach, FL

  • Tampa, FL

  • West Palm Beach, FL

  • Charlotte, NC

  • Austin, TX

  • Dallas, TX

  • San Antonio, TX

Assistant Vice President, Quantitative and Data Analyst | Hadoop ETL Developer @ (HIDDEN)

2/1/2014 - Current

Role Description

Bank of the West is a diversified financial services holding company, headquartered in San Francisco, California. It is a subsidiary of BNP Paribas. The bank has more than 700 branches in the Midwest and Western United States.


Learn More