




Summary: We are seeking a Middle/Senior DWH Engineer to design, build, and maintain scalable data warehouse solutions, working across the full data lifecycle with various stakeholders. Highlights: 1. Design, build, and maintain scalable Data Warehouse architectures 2. Develop and optimize ETL/ELT pipelines using Python/Airflow 3. Collaborate with analysts, engineers, BI teams, and business stakeholders We are looking for a Middle/Senior DWH Engineer to design, build, and maintain scalable data warehouse solutions that power analytics and business decision\-making. You will work across the full data lifecycle — from ingestion and modeling to optimization, reliability, and automation — collaborating closely with analysts, developers, and business stakeholders. **Requirements** * Strong SQL skills: complex queries, CTEs, window functions, analytical queries (5\+ years experience); * Knowledge of Python or other scripting languages for data transformations (3\+ years experience); * Deep understanding of DWH concepts: ETL/ELT, Data Vault, Kimball, Star/Snowflake schemas (4\+ years experience); * Experience with Airflow or other data pipeline orchestrators (3\+ years experience); * Hands\-on experience with modern DWH and query engines: BigQuery, Snowflake, Redshift, ClickHouse, Vertica, AWS Athena, Trino (2\+ years experience); * Confident use of Git; experience with team workflows (pull requests, rebasing, merge conflict resolution) (5\+ years experience); * Understanding of server and cloud infrastructure: basic skills in configuration, maintenance, monitoring, and load control (2\+ years experience). **Nice to Have** * Experience with CDC tools and streaming data sources; * Knowledge of Docker, Kubernetes, and Infrastructure as Code (Terraform); * Experience with cloud platforms: AWS, GCP, or Azure; * Familiarity with data governance, data cataloging, and lineage tools. **Responsibilities:** * Design, build, and maintain scalable Data Warehouse architectures aligned with business needs; * Develop and optimize ETL/ELT pipelines using Python/Airflow and custom solutions; * Work with DWH\\Datalake: PostgreSQL, Trino, BigQuery; * Implement incremental loads, CDC, backfills, and reprocessing strategies; * Optimize query performance, data models, and pipeline execution; * Ensure data quality through validation, automated testing, monitoring, and alerting; * Integrate new data sources (APIs, third\-party systems, raw data) without disrupting existing pipelines; * Collaborate with analysts, engineers, BI teams, and business stakeholders to translate requirements into data solutions; * Mentor engineers, review code, and contribute to data standards and best practices. **Benefits** **Flexible payment options:** choose the method that works best for you; **Tax assistance included:** we handle part of your taxes and provide guidance on the local setup; **Financial perks:** Bonuses for holidays, B\-day, work milestones and more \- just to show we care; **Learn \& grow:** We cover courses and certifications — and offer real opportunities to grow your career with us; **Benefit Сafeteria:** Choose what suits you — sports, language courses, therapy sessions, and more; **Stay connected:** From team\-building events to industry conferences — we bring people together online, offline, and on stage; **Modern Equipment:** We provide new laptops along with essential peripherals like monitors and headphones for a comfortable workflow; **Your schedule, your rules:** Start your day at 9, 10, or even 11 — we care about results, not clock\-ins.


