Position Details
About this role
This role involves developing and maintaining big data processing systems on AWS, focusing on scalable, distributed data analytics solutions using Hadoop, Spark, and related technologies.
Key Responsibilities
- Designing big data architectures
- Building scalable data pipelines
- Contributing to open source projects
- Developing real-time processing features
- Collaborating with open source community
Technical Overview
The environment includes Hadoop, Spark, Hive, Impala, HBase, and open source big data tools, with a focus on distributed computing and data processing at scale.
Ideal Candidate
The ideal candidate is a mid-level data engineer with experience in big data technologies such as Hadoop, Spark, and Hive, with a strong understanding of distributed systems and open source contributions. They should have excellent programming skills and a passion for scalable data solutions.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
Lack of experience with Hadoop or Spark, No programming skills in relevant languages, No experience with distributed systems, No open source contributions
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile