Lead Hadoop Developer (b)

City
Broomfield

State
CO

Job Type
Full-Time Regular

 

Big Data (Hadoop) Architect/Developer

Location: Broomfield CO

Duration: 9 Months

Position Overview:

As a Big Data (Hadoop) Architect/Developer, will be responsible for Cloudera Hadoop development, high-speed querying, managing and deploying Flume, Kafka, HIVE and Spark, and oversee handover to operational teams and propose best practices / standards. Expertise with Designing, building, installing, configuring and developing Hadoop echo system. Familiarity with Pentaho and Nifi a bonus skillset.

Principal Duties and Responsibilities (Essential Functions**):

  • Work with development teams within the data and analytics team to design, develop, and execute solutions to derive business insights and solve clients' operational and strategic problems.
  • Support the development of data and analytics solutions and product that improve existing processes and decision making.
  • Build internal capabilities to better serve clients and demonstrate thought leadership in latest innovations in big data, and advanced analytics.
  • Contribute to business and market development.

Specific skills and abilities:

  • Defining job flows
  • Managing and Reviewing Hadoop Log Files
  • Manage Hadoop jobs using scheduler
  • Cluster Coordination services through Zookeeper
  • Support MapReduce programs running on the Hadoop cluster
  • Ability to write MapReduce jobs
  • Experience in writing Spark scripts
  • Hands on experience in HiveQL
  • Familiarity with data loading tools like Flume, Sqoop
  • Knowledge of workflow/schedulers like Oozie
  • Knowledge of ETL tools like Pentaho

Qualifications & Skills:

  • Bachelor’s degree or related technical field preferred
  • Expertise with HBase, NOSQL, HDFS, JAVA map reduce for SOLR indexing, data transformation, back-end programming, java, JS, Node.js and OOAD
  • 7 + years’ of experience in IT with minimum 2 years’ of experience in Hadoop.

Job ID CTLJP00005529 / Posted Posted 2 Weeks ago
Apply With