Hadoop Developer


Responsible for designing and implementing big data analytics solutions for clients. Capable of implementing efficient data workflows, and handling administration tasks for big data clusters. Able to work independently as well as with team members to drive customer satisfaction and successful consulting engagements.


  • Provide a solution design that meets both the business requirements and the best Hadoop technical practices.
  • Hands-on delivery working on popular Hadoop distribution platforms like Cloudera, HortonWorks, and / or MapR.
  • Perform collection, cleansing, processing, and analysis of new and existing data sources, defining and reporting data quality and consistency metrics.
  • Perform admin tasks on Hadoop clusters, including installation, configuration, tuning, maintenance, and monitoring.
  • Acquire Big Data Certifications if required.
  • Participate in working sessions with technical executives and experts.
  • Learn & stay current on Big Data techniques developments/improvements.

Location and Travel

  • Remote. Occasional travel to client worksites.


Qualifications and Skills


  • 2+ years experience in Hadoop development, including HDFS, Hive, Impala, Sqoop, Flume, Kafka.
  • Demonstrated knowledge and proven hands-on experience with HDFS, Hive, Impala, Sqoop Flume, Kafka (knowledge of Cloudera stack is a plus)
  • 2+ years experience designing, configuring and managing full stack Hadoop distribution (preferably Cloudera).
  • 2+ years experience administering Linux production environments.
  • Demonstrated knowledge and hands-on experience with AWS and/or Azure cloud infrastructure.
  • Hands-on experience working with large complex datasets.
  • Experience with implementing and managing Hadoop related security in Linux environments (Kerberos, SSL, etc.).
  • Strong knowledge of Yarn configuration in a multi-tenancy environment.
  • Strong working knowledge of disaster recovery related to Hadoop platforms.
  • Experience with Spark a plus.
  • Experience with Configuration Management tools a plus (i.e. CHEF, Puppet and Ansible).
  • DevOps experience a great plus.


  • Bachelor’s Degree in Computer Science or a relevant technical field. Advanced degree preferred.


  • Strong SQL and HiveQL skills (Java/MapReduce, Python, Scala are a plus).
  • Working knowledge of Hadoop eco-system: YARN, HDFS, Sqoop, Hive/Impala, Oozie, Flume, Kafka, Solr.
  • Proficient in Linux OS and bash scripting (AWK and/or SED is a plus).
  • Understanding of major RDBMS systems like Oracle, MySQL, PostgreSQL, SQL Server, DB2, & Sybase.
  • Working knowledge of data compression techniques, Avro/Parquet schemas, table partitioning, and optimization tuning.
  • Ability to debug and understand Hadoop/YARN log files.
  • Working knowledge of automating/scheduling data flows with Oozie (both utilizing GUI and scripting).
  • Significant experience working with cloud infrastructure in AWS or Azure.
  • Proven track record of implementing effective data solutions.
  • Strong analytical and problem-solving skills with proven communication and consensus building abilities.
  • Proven skills to work effectively across internal functional areas in ambiguous situations.
  • Excellent organization and planning skills.
  • High degree of professionalism.
  • Ability to thrive in a fast-paced environment.


Employment Type

  • Contractor or Full-time Employee



  • Competitive salary
  • Medical