
Data Engineer (Azure Databricks)
📝 Twój zakres obowiązków
Your responsibilities, Develop and maintain data pipelines using PySpark and SQL, Work with Azure Databricks for large-scale data processing, Build and maintain ingestion and data transformation workflows, Perform data wrangling and integrate multiple data sources, Optimize Spark jobs (DataFrames, partitioning, clustering, SparkSQL), Work with data formats such as Delta, Parquet, and CSV, Collaborate with stakeholders across multiple time zones, Create and maintain technical documentation
5+ years of experience in Data Engineering, Strong hands-on experience with: PySpark (DataFrames, optimization techniques); Databricks; SQL, Experience with Azure services: Data Factory; Azure SQL / Storage, Knowledge of cloud environments (preferably Azure), Strong problem-solving skills and ability to work independently, Good communication skills in English (B2+)
Optional, Power BI, Azure certifications (AZ-900 or similar), Databricks certification
What we offer, 100% remote work, International project environment, Modern data stack (Azure + Databricks), Fast recruitment process, Long-term cooperation
Benefits, sharing the costs of sports activities, life insurance
ITEAMLY SPÓŁKA Z OGRANICZONĄ ODPOWIEDZIALNOŚCIĄ, As Venture Building, we create profitable technology companies from scratch. We collaborate with partners who share similar aspirations and are open to leveraging the potential of their networks, cultivated over years. We regularly launch new projects founded on long-term, amicable relationships with our associates.
About the project
We are looking for an experienced Data Engineer with strong PySpark skills to join a large-scale Azure data platform project. You will work in a distributed environment, focusing on building and optimizing data pipelines.
📝 Opis główny / Wstęp
About the project
We are looking for an experienced Data Engineer with strong PySpark skills to join a large-scale Azure data platform project. You will work in a distributed environment, focusing on building and optimizing data pipelines.
Your responsibilities
- Develop and maintain data pipelines using PySpark and SQL
- Work with Azure Databricks for large-scale data processing
- Build and maintain ingestion and data transformation workflows
- Perform data wrangling and integrate multiple data sources
- Optimize Spark jobs (DataFrames, partitioning, clustering, SparkSQL)
- Work with data formats such as Delta, Parquet, and CSV
- Collaborate with stakeholders across multiple time zones
- Create and maintain technical documentation
🎁 Co oferujemy (Dodatkowe detale)
Benefits, sharing the costs of sports activities, life insurance
ITEAMLY SPÓŁKA Z OGRANICZONĄ ODPOWIEDZIALNOŚCIĄ, As Venture Building, we create profitable technology companies from scratch. We collaborate with partners who share similar aspirations and are open to leveraging the potential of their networks, cultivated over years. We regularly launch new projects founded on long-term, amicable relationships with our associates.
About the project
We are looking for an experienced Data Engineer with strong PySpark skills to join a large-scale Azure data platform project. You will work in a distributed environment, focusing on building and optimizing data pipelines.