- Type of contract: contract of employment (UoP)
- Salary ranges: up to 24,000 PLN gross, month
- Working model: 100% remote
Join our client, one of the leading logistics and transport solutions providers.
About the role:
We’re looking for a Lead Data Engineer who excels not only at coding but, more importantly, at shaping and guiding data architecture. This role is a mix of technical, architectural, and communication responsibilities, requiring someone who can bridge the gap between business and engineering teams.
Responsibilities:
- Understanding in depth both the business and technical problems that the company aims to solve – building tools, platforms, and pipelines to enable teams to clearly and cleanly analyze data, build models, and drive decisions;
- Scaling up from “laptop-scale” to “cluster-scale” problems, in terms of both infrastructure and problem structure and technique;
- Delivering tangible value very rapidly, collaborating with diverse teams of varying backgrounds and disciplines;
- Codifying best practices for future reuse in the form of accessible, reusable patterns,
templates, and code bases; - Interacting with senior technologists from the broader enterprise and outside of the company
(partner ecosystems and customers) to create synergies
Required minimum experience:
- 5+ years equivalent work experience in an information technology or engineering environment
- Extensive knowledge in data engineering and machine learning frameworks, including design, development, and implementation of highly complex systems and data pipelines.
- Extensive knowledge in Information Systems, including design, development, and implementation of large batch or online transaction-based systems
- Strong understanding of the transportation industry, competitors, and evolving technologies
- Experience providing leadership in a general planning or consulting setting.
- Experience as a leader or a senior member of multi-function project teams.
- Strong oral and written communication skills, both in Polish and English
- A related advanced degree may offset the related experience requirements.
Requirements:
- Technical Background: Strong foundation in computer science, software engineering, database systems, and distributed systems.
- Cloud & Distributed Systems: Deep understanding of cloud-based architectures and trade-offs in distributed environments.
- Azure Expertise: In-depth experience with Microsoft Azure tools for large-scale data engineering and deployment.
- Scalable Systems: Proven success designing and deploying robust, complex production systems using modern data science methods.
- Streaming & Ingestion:
- Hands-on experience with Kafka and Structured Streaming
- Skilled in Azure Event Hubs for real-time data ingestion
- Big Data & Analytics:
- Advanced use of Spark (both PySpark and Scala)
- Strong proficiency with Databricks for distributed data processing
- Programming & Data Tools:
- Proficient in Python data stack, Java
- Experience with SQL and NoSQL systems
- Infrastructure & DevOps:
- Strong experience with Kubernetes for orchestration
- Practical knowledge of Terraform and Infrastructure as Code (IaC)
- Familiarity with CI/CD, agile, and DevOps practices
- Problem Solving & Leadership:
- Able to support distributed teams, unblock complex issues, and guide less senior engineers
- Comfortable with context switching and working on evolving problem definitions
- Delivery-focused under tight timelines, with an engineering mindset to maximize impact
- Collaboration & Code Quality:
- Experienced in working with distributed teams
- Skilled in using version control systems and conducting thorough code reviews
- Data Quality & Lineage:
- Capable of performing data investigations, lineage studies, and improving data accessibility and integrity
The offer:
- Benefits like life insurance, sports card, and private healthcare
- Nice equipment: a powerful notebook and additional devices you need
- Development budget
- Remote work (with the possibility to visit the office)