Position Details
About this role
Data Engineer responsible for designing and maintaining scalable data pipelines and ETL processes across cloud platforms. The role emphasizes Python/PySpark development, cloud-native services, and data quality governance.
Key Responsibilities
- Data Pipeline Development
- Cloud & Data Architecture
- Data Modeling & Querying
- Data Quality & Operations
- Collaboration & DevOps
Technical Overview
Backend data engineering role focusing on Python, PySpark, and cloud data platforms (AWS, Azure, GCP). Key activities include building ETL pipelines, data modeling in cloud warehouses, and implementing CI/CD for data workflows.
Ideal Candidate
The ideal candidate is a data engineer with 3+ years of Python and PySpark experience, proficient in building scalable ETL pipelines across cloud platforms (AWS, Azure, or GCP). They should be comfortable working in a remote setup and optimizing data modeling and queries in cloud data warehouses.
Must-Have Skills
Nice-to-Have Skills
Tools & Platforms
Required Skills
Hard Skills
Soft Skills
Industry & Role
Keywords for Your Resume
Deal Breakers
No Python or PySpark experience, No cloud platform experience (AWS, Azure, or GCP), No ETL or data pipeline experience
Get matched to jobs like this
Luna finds roles that fit your skills and career goals — no endless scrolling required.
Create a Free Profile