pracaon.pl

Senior / Lead Data Software Engineer (Python, Spark, Azure)

Wrocław, Lower Silesian Voivodeship, Polska, Kraków, Lesser Poland Voivodeship, Polska
Externe Anzeige
EPAM
Partner
56Tg
Gehalt nach Vereinbarung
IT und Telekommunikation
Vollzeit
Hybrid
Wichtige Merkmale des Angebots
  • Feste Vollzeitstelle

  • Hybrides Modell

Anforderungen
  • Proficiency in Python and Spark with at least 3 years in data engineering roles

  • Strong experience with Azure Databricks and PySpark

  • Proven expertise in designing and implementing ETL/ELT solutions

  • Experience migrating big data platforms to Azure-native services

  • Proficiency with Delta tables for model tuning

  • Knowledge of data governance and regulatory compliance frameworks

  • Familiarity with Docker, Kubernetes (AKS), and Terraform for infrastructure automation

  • Ability to manage large data volumes with high efficiency

  • Excellent problem-solving and analytical skills

  • Strong communication and collaboration abilities

  • English proficiency at B2 level or higher

Zakres obowiązków
  • Migrate and optimize over 500 data jobs using Azure Databricks optimization techniques

  • Manage and process 12 TB of data efficiently across platforms

  • Tune machine learning models for Azure environments using Java Spark and Delta tables

  • Update and maintain libraries to address security vulnerabilities

  • Develop and maintain ETL/ELT pipelines using PySpark and related technologies

  • Collaborate with cross-functional teams to integrate GenAI capabilities into data workflows

  • Monitor data quality and implement improvements to ensure accuracy and reliability

  • Automate deployment and operational tasks using Terraform and GitLab CI/CD

  • Support data governance initiatives to comply with regulatory standards

  • Troubleshoot and resolve performance issues in data processing systems

  • Document system processes and provide technical guidance to junior engineers

  • Implement best practices for code quality and data security

  • Participate in code reviews and knowledge sharing sessions

  • Optimize costs associated with data storage and processing

Seniority
  • Senior

Beschreibung

We are seeking a Senior/Lead Data Software Engineer to join our team working on a scalable, ML-ready platform that enhances portfolio model development and deployment with advanced data governance and AI capabilities. You will play a key role in migrating from an IaaS Big Data platform to Azure-native Databricks, optimizing data workflows and improving data quality. Join us to contribute to innovative solutions that boost client services and regulatory compliance. Responsibilities Migrate and optimize over 500 data jobs using Azure Databricks optimization techniques Manage and process 12 TB of data efficiently across platforms Tune machine learning models for Azure environments using Java Spark and Delta tables Update and maintain libraries to address security vulnerabilities Develop and maintain ETL/ELT pipelines using PySpark and related technologies Collaborate with cross-functional teams to integrate GenAI capabilities into data workflows Monitor data quality and implement improvements to ensure accuracy and reliability Automate deployment and operational tasks using Terraform and GitLab CI/CD Support data governance initiatives to comply with regulatory standards Troubleshoot and resolve performance issues in data processing systems Document system processes and provide technical guidance to junior engineers Implement best practices for code quality and data security Participate in code reviews and knowledge sharing sessions Optimize costs associated with data storage and processing Requirements Proficiency in Python and Spark with at least 3 years in data engineering roles Strong experience with Azure Databricks and PySpark Proven expertise in designing and implementing ETL/ELT solutions Experience migrating big data platforms to Azure-native services Proficiency with Delta tables for model tuning Knowledge of data governance and regulatory compliance frameworks Familiarity with Docker, Kubernetes (AKS), and Terraform for infrastructure automation Ability to manage large data volumes with high efficiency Excellent problem-solving and analytical skills Strong communication and collaboration abilities English proficiency at B2 level or higher

Stichwörter / Fähigkeiten
Data Software Engineering
Azure Databricks
ETL/ELT Solutions
PySpark
Das Angebot wurde von einem externen Portal importiert.Anzeigenquelle