Business Analyst & Big Data Architect - "See JD for Skill set" (Senior, Remote)

TopDevz

We are looking for an experienced, senior, Business Analyst and Big Data Architect, who is excited to work on one of our many client projects - both greenfield (new) projects as well as legacy (support) projects in that technology stack. This is a remote position, 40 hours a week.Skills & Requirements
  • Expert level knowledge with 8-10 years of experience in Cloudera Hadoop components such as HDFS, HBase, Impala, Hue, Spark, Hive, Kafka, YARN and ZooKeeper
  • Expertise in architecture, Build and Troubleshooting Hadoop experience on Cloudera Distribution
  • Hands on experience with ON prem and cloud deployments
  • Validated experience in scripting, automation, deployment, set up and installation, trouble shooting and fixing issues across platforms
  • Architecture, Design and Development of Big Data lake
  • Take end to end responsibility of the Hadoop life Cycle in the organization.
  • Detect, analyze and remediate performance problems.
  • 3-5 years of experience in at least one of the following: Python, Java or Shell Scripting and eager to pick up new programming languages on the go
  • The ability to function within a multidisciplinary, global team. Be a self-starter with a strong curiosity for extracting knowledge from data and the ability to elicit technical requirements from a non-technical audience
  • Data Concepts (ETL, near-/real-time streaming, data structures, metadata and workflow management)
  • You have deep understanding of Dev ops and Agile software development methodologies
  • You have strong communication skills and the ability to present deep technical findings to a business audience
  • Hands-on experience working in an agile environment
  • AWS/Azure/Google Cloud experience is a plus.
  • Experience in architecture and troubleshooting skills on tech stacks HDFS,HIVE, Mapreduce, YARN,IMPALA,SPARK,HBASE, KMS, HUE, OOZIE, SENTRY
  • Architecting Large Scale Big Data/Analytics solutions
  • Architecture/Design Patterns for Big Data and Cloud systems
  • Build analytics tools that utilize the data pipeline to provide actionable insights into key business performance metrics.
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc
  • Responsible for defining the Build and Manage part of Hadoop solutions
Subscribe Now