As a Senior Data Engineer, you'll be joining a scale-up environment where adaptability and problem-solving are key. This role is essential in designing, building, and optimizing our data architecture to support evolving business needs. With expertise in data pipeline stacks (open-source & AWS) and ClickHouse, you’ll focus on developing scalable, high-performance solutions. Collaborating closely with engineering teams, you’ll implement robust data pipelines that ensure efficiency, reliability, and alignment with business objectives. If you're looking for a role where you can make a real impact, we’d love to hear from you.
Key Responsibilities:
- Data pipeline Design: Robust design experience in follow and implement scalable, high-performance data architectures using AWS services and ClickHouse (or equivalent OLAP databases).
- Data Pipeline Development: Design, build, and maintain robust ETL/ELT pipelines that efficiently handle large-scale data ingestion, transformation, and storage.
- Cloud Infrastructure: Combine open-source data stack and AWS technologies to build and optimize data workflows.
- Data Governance & Quality: Ensure data accuracy and consistency through best practices in data governance, lineage, and monitoring.
- Performance Optimization: Optimize data storage, retrieval, and processing to support high-performance analytical workloads using partitioning, indexing, and query optimization techniques.
- Collaboration & Leadership: Work closely with data scientists, analysts, and software engineers to understand requirements and deliver data-driven solutions, mentoring junior engineers.
- Automation & CI/CD: Implement automated data pipeline deployment and monitoring strategies.
Requirements:
- 5+ years of experience in data engineering with solid experience on open-source data stack, and cloud native experiences.
- Expertise in ClickHouse, or similar OLAP databases is a plus.
- Strong experience in designing and architecting large-scale data systems.
- Proficiency in programming languages like Python, or productivity libraries scripting languages for data processing and pipeline development.
- Deep understanding of ETL/ELT processes, data modeling (Star Schema, Snowflake Schema), and data warehousing.
- Experience with orchestration tools such as Apache Airflow, Step Functions, or Dragster.
- Hands-on experience with infrastructure-as-code (Terraform, CloudFormation, CDK) is a plus
- Strong problem-solving skills and ability to work in a fast-paced environment.
- Knowledge of SQL query performance tuning, materialized views, and sharding strategies for large datasets.
- Familiarity with containerization and serverless computing (Docker, Kubernetes).
- Experience with monitoring and observability tools such as Prometheus, Grafana, AWS CloudWatch.
Here's a glimpse of what's waiting for you:
- A competitive salary package that rewards your hard work.
- 25 paid vacation days. And if that's not enough, you can purchase up to 10 more.
- A world of growth and development opportunities to enhance your skills. You'll have unlimited access to our treasure trove of GoodHabitz resources and MyAcademy .
- Access to mental coaching through our partner, OpenUp, to keep your mind in top shape.
- An annual do-good-day, fully paid, so you can contribute to a cause you're passionate about.
- Travel and expense reimbursement because we've got your journey covered.
- Pension and disability insurance, securing your financial well-being in the long run.
- A hybrid way of working .
- Working in a company that welcomes artificial intelligence and uses it to improve internal processes and push AI-powered features quickly .
- MacBook Pro.