We are looking for an experienced Data Engineer responsible for planning, developing, and maintaining cloud environments for our clients.
About Devapo
At Devapo, we focus on continuous self-development and acquiring new knowledge. If you are a fast learner, want to participate in international projects, are a team player, and can work independently — join us!
We provide our clients with more than just code — we want to equip them with tools that allow their businesses to flourish. Our clients’ success is our success, which is why we ensure that everyone who creates Devapo has a long-term goal in mind.
Key Responsibilities:
Design, implement, and maintain scalable and efficient data pipelines in one of the cloud environments (Azure, AWS, GCP) using tools such as Databricks, Glue, Dataflow, or Azure Data Factory
- Develop and optimize ETL/ELT processes using cloud-native services (e.g., Azure Data Factory, AWS Glue, GCP Dataflow) and Apache Spark/Databricks
- Build Big Data solutions aligned with business and analytical requirements across cloud platforms
- Collaborate with Data Science, BI, and development teams to deliver high-quality, well-structured, and performant data
- Monitor and improve the performance, reliability, and scalability of data processing systems
- Implement robust data governance, security standards, and best practices across cloud environments
- Research and evaluate new tools and technologies within the cloud and data engineering ecosystem
Requirements:
- Minimum 3 years of experience as a Data Engineer or in a similar role
- Hands-on experience with one or more major cloud platforms (Azure, AWS, GCP); deep knowledge of cloud data services such as:
- Azure Data Factory, Azure Data Lake, Synapse Analytics (Azure)
- AWS Glue, S3, Redshift, Athena (AWS)
- GCP Dataflow, BigQuery, Cloud Storage (GCP)
- Extensive experience with Databricks and Apache Spark
- Proficiency in SQL and experience with relational and columnar databases
- Strong programming skills in Python and PySpark
- Experience designing and optimizing data pipelines in distributed, cloud-based architectures
- Familiarity with Delta Lake or other modern data lake architectures
- Solid understanding of data modeling and schema design
What We Offer:
- Salary: 17 800 - 21 500 PLN (B2B contract)
- Co-financing for training and certifications, as well as guaranteed time for learning during working hours
- Private medical care and a Multisport card
- Language classes (English)
- Flexible working hours and the possibility of hybrid work (Warsaw)
- Team integration meetings and company events
- Employee referral program with a bonus
- An individually tailored career development path