We are looking for an experienced, senior, Business Analyst and Big Data Architect, who is excited to work on one of our many client projects - both greenfield (new) projects as well as legacy (support) projects in that technology stack. This is a remote position, 40 hours a week.
Skills &Requirements
- Expert level knowledge with 8-10 years of experience in Cloudera Hadoop components such as HDFS, HBase, Impala, Hue, Spark, Hive, Kafka, YARN and ZooKeeper
- Expertise in architecture, Build and Troubleshooting Hadoop experience on Cloudera Distribution
- Hands on experience with ON prem and cloud deployments
- Validated experience in scripting, automation, deployment, set up and installation, trouble shooting and fixing issues across platforms
- Architecture, Design and Development of Big Data lake
- Take end to end responsibility of the Hadoop life Cycle in the organization.
- Detect, analyze and remediate performance problems.
- 3-5 years of experience in at least one of the following: Python, Java or Shell Scripting and eager to pick up new programming languages on the go
- The ability to function within a multidisciplinary, global team. Be a self-starter with a strong curiosity for extracting knowledge from data and the ability to elicit technical requirements from a non-technical audience
- Data Concepts (ETL, near-/real-time streaming, data structures, metadata and workflow management)
- You have deep understanding of Dev ops and Agile software development methodologies
- You have strong communication skills and the ability to present deep technical findings to a business audience
- Hands-on experience working in an agile environment
- AWS/Azure/Google Cloud experience is a plus.
- Experience in architecture and troubleshooting skills on tech stacks HDFS,HIVE, Mapreduce, YARN,IMPALA,SPARK,HBASE, KMS, HUE, OOZIE, SENTRY
- Architecting Large Scale Big Data/Analytics solutions
- Architecture/Design Patterns for Big Data and Cloud systems
- Build analytics tools that utilize the data pipeline to provide actionable insights into key business performance metrics.
- Assemble large, complex data sets that meet functional / non-functional business requirements.
- Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc
- Responsible for defining the Build and Manage part of Hadoop solutions