Role : Big Data Engineer
Location : Denver,CO
Duration : 12+ Months
Rate : $60/hr on C2C(Max)
Interview Type : Phone and F2F
Job Description :
Required Skills-
Hadoop, Flume, Kafka, Storm
Basic Qualifications- Parallel Processing, Scale Optimization, realtime Database (memsql), git, jenkins, splunk, pig,
Job Description-
Work Location: (Downtown Denver)
Job Title: Big Data Engineer
JOB DESCRIPTION:
Manages applications engineering team who research, collect and report information on the ability and capacity of company products and competing equipment. Oversees and contributes to the design and development of equipment and systems or redesign to existing equipment to fulfill the needs of customers. Ensures projects are completed on time and within budget. Acts as advisor to applications engineering team regarding projects, tasks, and operations. May require a bachelor's degree in area of specialty and at least 8 years of experience in the field or in a related area. Familiar with a variety of the field's concepts, practices, and procedures. Relies on extensive experience and judgment to plan and accomplish goals. Performs a variety of complex tasks. Leads and directs the work of others. A wide degree of creativity and latitude is expected. Typically reports to head of a unit/department.
Responsibilities:
• Develop solutions to Big Data problems utilizing common tools found in the Hadoop ecosystem.
• Develop solutions to real-time and off line event collecting from various systems.?
• Develop, maintain, and perform analysis within a real-time architecture supporting large amounts of data from various sources.
• Analyze massive amounts of data and help drive prototype ideas for new tools and products.
• Design, build and support APIs and services that are exposed to other internal teams
• Employ rigorous continuous delivery practices managed under an agile software development approach
• Ensure a quality transition to production and solid production operation of the software
Here are some of the specific technologies we use:
• Hadoop
• Flume
• Kafka
• Storm
• MemSQL
• Java
• Maven
• Git
• Jenkins
• Splunk/Hunk
• Apache Pig
• Unix/Linux
Additional Skills:
• 3+ years programming experience
• Bachelors or Masters in Computer Science or related discipline
• Experience in software development of large-scale distributed systems – including proven track record of delivering backend systems that participate in a complex ecosystem.
• Knowledge in Big Data related technologies and open source frameworks preferred.
• Extensive experience programming in Java as well as experience in code optimization and high performance computing.
• Experience with Java servlet containers or application servers such as JBoss, Tomcast, Glassfish, WebLogic, or Jetty.
• Good current knowledge of Unix/Linux environments
• Test-driven development/test automation, continuous integration, and deployment automation
• Enjoy working with data – data analysis, data quality, reporting, and visualization
• Good communicator, able to analyze and clearly articulate complex issues and technologies understandably and engagingly.
• Great design and problem solving skills, with a strong bias for architecting at scale.
• Adaptable, proactive and willing to take ownership.
• Keen attention to detail and high level of commitment.
• Comfortable working in a fast-paced agile environment. Requirements change quickly and our team needs to constantly adapt to moving targets.
Nice to haves:
• Collection, transformation and enrichment frameworks such as Flume
• Messaging middleware or distributed queuing technologies such as Kafka
• MapReduce experience in Hadoop utilizing Pig, Hive, or other query/scripting technology
• Distributed (HBase or Cassandra or equivalent) or NoSQL (e.g. Mongo) database experience
• Expertise in data warehousing and business intelligence
• Scripting tools such as Python
• Git, Maven, Jenkins, Sonar, Nexus, Puppet
• Understanding and/or experience with serialization frameworks such as Thrift, Avro, Google Protocol Buffers, and Kyro preferred.
• Visualization tools libraries, reporting tools, etc. Splunk (Hunk), Tableau, d3.js
• Good understanding in any: advanced mathematics, statistics, and probability.