Hadoop Developer | Data Engineer | Software Engineer | Data Scientist
Hadoop Developer | Data Engineer | Software Engineer | Data Scientist
• 3 years of experience in application development, Database management, Apache Hadoop administration and development.
• 2+ years of experience in big data technologies : Hadoop HDFS, Map-Reduce, Pig, Hive, Oozie, Flume, Hcatalog, Sqoop, zookeeper, NoSql : Cassandra and Hbase.
View my 3D Resume
Hive
Oracle Database Development
SQL Server
Java
Hadoop
Oozie
Sqoop
Hbase
Cloudera Manager
C++
Zookeeper
Pig
Flume
Mysql
Hdfs
Cassandra
Mapreduce
Oracle PL/SQL
Read More
Job Posted: Tuesday, March 17, 2015
Expires On: Wednesday, March 25, 2015
Job Type: Permanent
- 3 years of experience in application development, Database management,
Apache Hadoop administration and development.
- 2+ years of
experience in big data technologies : Hadoop HDFS, Map-Reduce, Pig, Hive, Oozie,
Flume, Hcatalog, Sqoop, zookeeper, NoSql : Cassandra and Hbase.
- Experience in solving complex analytical problems and
performing joins with Map-Reduce.
- Strong knowledge of apache hive administration and
development.
- Experience in writing UDFS, UDAFS, and UDTFS in java for hive.
- Solid experience in Pig administration and development.
- Experience in writing pig udfs (Eval, Filter, Load and Store) and macros.
- Familiar with embedding hive and pig in java.
- Experience in using Hcatalog for Hive, Pig and Hbase.
- Worked on developing ETL
processes to load data from multiple data sources to HDFS using FLUME and
SQOOP, perform structural modifications using Map-Reduce, HIVE and analyse data
using visualization/reporting tools.
- Familiar with writing Oozie workflows and Job
Controllers for job automation.
- Familiar with importing and exporting data using Sqoop.
- Experience in using Flume to stream data into HDFS.
- Familiar with administering and developing in Cassandra and Hbase.
- Experience in working with java Hbase APIs.
- Experience in writing mrunit test cases and pig unit
test cases.
- Solid background in Object-Oriented analysis and design.
- In-depth understanding of Data
Structure and Algorithms.
- Good Understanding of Distributed Systems and Parallel Processing architecture.
- Experience in handling multiple relational databases: Oracle, MySQL, SQL Server.
- Familiar with Agile Methodology
(SCRUM) and Software Testing.
- Experience in
deploying the multinode Hadoop cluster with different Hadoop components (HIVE, PIG, SQOOP, OOZIE, FLUME, HCATALOG, HBASE,
ZOOKEEPER) using Cloudera Manager.
- Experience in
benchmarking, performing backup and recovery of NameNode meta data and data
residing on cluster
- Experience in
performing minor and major upgrades, commissioning and decommissioning of nodes
on Hadoop cluster.
- Strong
knowledge in configuring NameNode high availability and NameNode federation.
- Experience in
deploying Hadoop 2.0(YARN).
- Google certified professional in Analytics.
- SAS Certified Advanced Programmer.
Raleigh, NC
Menlo Park, CA
Mountain View, CA
Palo Alto, CA
Sacramento, CA
San Francisco, CA
San Jose, CA
Santa Clara, CA
Sunnyvale, CA
Jacksonville, FL
Miami Beach, FL
Tampa, FL
West Palm Beach, FL
Charlotte, NC
Austin, TX
Dallas, TX
San Antonio, TX
Assistant Vice President, Quantitative and Data Analyst | Hadoop ETL Developer @ (HIDDEN)
Role Description
Bank of the West is a diversified financial services holding company, headquartered in San Francisco, California. It is a subsidiary of BNP Paribas. The bank has more than 700 branches in the Midwest and Western United States.
Learn More