Position Details
About this role
Software Engineer role focused on building and optimizing data pipelines and analytics solutions in a cloud environment leveraging Java and PySpark.
Key Responsibilities
- Develop and maintain large-scale data processing pipelines (and infrastructure)
- Lead data modeling and schema updates
- Lead code reviews and mentorship
- Drive data quality and data accessibility
- Ensure governance and business alignment
Technical Overview
Work with AWS, Snowflake, Redshift, and Spark-based data processing; implement data modeling techniques and CI/CD for data products; exposure to batch and streaming data processing.
Ideal Candidate
The ideal candidate is a data engineer with 2+ years of experience in Java/PySpark, cloud data warehouses, and data modeling, comfortable building scalable data pipelines in a cloud environment.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
Less than 2 years of relevant experience, No exposure to AWS or cloud data warehouses
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile