Returning Candidate?

GHRS, Big Data Engineer

GHRS, Big Data Engineer

Job ID 
Posted Date 
Amazon Corporate LLC
Position Category 
Business Intelligence
Recruiting Team 

Job Description

Job Description: Big Data Engineer
Are you a Big Data/data lake specialist? Do you have Data Warehousing, data lake build, Hadoop/Hive/Spark/EMR learning experience? Do you want to have an impact in the development and use of new data analytics technologies? Would you like a career that gives you opportunities to help customers and partners use Cloud computing services to do build new solutions, faster, and at lower cost?
Responsibilities Include:
• Design, implement and deliver complete analytic solutions for customers.
• Architect, build and maintain high performing ETL processes, including data quality and testing
• Keep up to date with advances in data persistence and big data technologies and run pilots to design the data architecture to scale with the increased data sets.
• Develop and improve the data architecture for Redshift and Hadoop/EMR Cluster.
• Develop analytics with a mind toward accuracy, scalability and high performance
• Provide technical guidance and thought leadership to other programmer analysts in the team.

Basic Qualifications

• Bachelor’s Degree in Computer Science or equivalent degree.
•5+ years hands-on experience with designing and developing on distributed architecture systems up to Tera/petabyte data handling using Open Source software.
• 2+ Years knowledge in modern distributed architectures and compute / data analytics / storage technologies on AWS or related technologies.
• Knowledge of a programming language such as Java/Python/Scala
• Understanding of architectural principles and design patterns / styles using parallel large-scale distributed frameworks such as Hadoop / Spark
• Deep knowledge of RDBMS (MySQL, PostgreSQL, SQL Server) and NoSQL databases such as HBase, Vertica, MongoDB, Dynamo DB, Cassandra
• Demonstrates broad knowledge of technical solutions, design patterns, and code for medium/complex applications deployed in Hadoop.
• Knowledge of working in UNIX environment with fair amount of shell scripting and python experience. Knowledge in spring, Java, MapReduce is expected.
• Hands on experience designing, developing, and maintaining software solutions in Hadoop Production cluster.
• Experience in architecting and building data warehouse systems and BI systems including ETL.
· Experience in performance troubleshooting, SQL optimization, and benchmarking. Strong architectural experience in context of deploying cloud-based data solutions.
• Thorough understanding of service-oriented architectures and data processing in high-volume applications. Full SDLC experience (requirements gathering through production deployment).

Preferred Qualifications

• Outstanding analytical skills, excellent team player and delivery mindset.
• AWS Redshift experience a plus.
• Alteryx , Datameer