About the Role
We're hiring a Senior Data Engineer to work on our Data Lake Team. As a key member of the team, you will be responsible for building and operating various data platform components, including data quality, data pipelines, infrastructure, and monitoring.
Responsibilities
- Maintain data pipeline job framework
- Develop Data Quality framework ( internal set of tools for internal and external data sources validation )
- Maintain and develop public facing data ingestion service with 17 000+ RPS.
- Maintain and develop core data pipelines in batch and streaming manners.
- Be a last line of support for our internal platform users.
- Take a part in an on-call rotation for data platform incidents (shared across the team).
Requirements
- Fluent English
- 4+ years building production services and data pipelines (batch and/or streaming)
- Strong experience with Python or the readiness to ramp up quickly.
- Hands-on experience with at least one MPP system (Spark, Trino, Redshift etc.)
- Hands-on experience operating services in a cloud environment (AWS preferred)
Nice to Have
- Terraform/CloudFormation or other IaC tools
- ClickHouse or similar analytical databases
- Experiences with data quality/observability tools
Benefits
- Unlimited vacation time – we strongly encourage all employees to take at least 3 weeks per year
- Fully remote team – choose where you live
- Work from home stipend – we want you to have the resources you need to set up your home office
- Apple laptops provided for new employees
- Training and development budget – refreshed each year for every employee
- Maternity & Paternity leave for qualified employees
- Work with smart people who will help you grow and make a meaningful impact
- Base salary: $80k–$120k USD, depending on knowledge, skills, experience, and interview results
- Stock options – offered in addition to the base salary
- Regular team offsites to connect and collaborate
XML job scraping automation by YubHub