Position Details
About this role
This role involves developing and maintaining scalable data pipelines and platforms using Apache Iceberg and related technologies for government clients.
Key Responsibilities
- Develop data pipelines
- Maintain data lake infrastructure
- Implement schema evolution and data versioning
- Troubleshoot data environment issues
- Collaborate with multidisciplinary teams
Technical Overview
The technical environment includes data lake architectures, distributed file systems like S3, HDFS, GCS, and query engines such as Presto, Trino, Spark, and Hive, with programming in Python, Java, and Scala.
Ideal Candidate
The ideal candidate is a mid-level data engineer with at least 2 years of experience in building and maintaining data pipelines, proficient with Apache Iceberg, distributed file systems, and query engines like Presto or Spark. They should have strong troubleshooting skills and a solid understanding of data lake and warehouse architectures.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Clearance & Visa
Keywords for Your Resume
Deal Breakers
Lack of experience with Apache Iceberg or similar table formats, No experience with distributed file systems, No Bachelor's degree, Inability to obtain security clearance, Less than 2 years of relevant experience
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile