Job Information
IBM Data Engineer in Vilnius, Lithuania
Introduction
In this role, you will work in one of our IBM Consulting Client Innovation Centers (Delivery Centers), where we bring deep technical and industry expertise to public and private sector clients around the world. You will be part of a team that delivers high‑impact solutions and drives adoption of modern data and cloud technologies.
Your role and responsibilities
The successful candidate will design, build, and maintain scalable and reliable data pipelines and platforms used across analytics, AI, and business systems. You will collaborate with cross-functional teams to ensure data is accessible, high-quality, and aligned with business goals. Responsibilities include:
Designing, developing, and optimizing data processing systems, including ETL/ELT pipelines and data orchestration workflows
Building and maintaining data pipelines for batch and real-time (streaming) use cases
Working with data scientists, software engineers, and business stakeholders to deliver high‑quality, production‑ready data solutions
Implementing and enforcing data quality, validation, and governance practices
Ensuring compliance with data security standards, access controls, and regulatory requirements
Monitoring data platform performance and implementing improvements to ensure reliability, scalability, and cost effectiveness
Contributing to standardization, automation, and best practices across data engineering teams
Required technical and professional expertise
Strong Python skills for data processing, pipeline development, and automation
Hands‑on experience with Apache Spark / PySpark for large‑scale distributed data processing
Experience with Databricks and cloud platforms (AWS or Azure), including Delta Lake and related data management tools
Proven experience designing, developing, and maintaining scalable ETL/ELT pipelines and data platform components
Familiarity with building both batch and real‑time (streaming) data workflows, preferably with technologies like Kafka, Event Hubs, or Kinesis
Experience with DevOps practices and Infrastructure as Code (Terraform preferred)
Understanding of data modeling, data warehousing concepts, and modern data architectures (e.g., Lakehouse)
Preferred technical and professional experience
Experience building or integrating with LLM‑powered or AI‑driven solutions
Familiarity with FastAPI and Pydantic for service development
Certification in AWS, Azure, or Databricks
Knowledge of CI/CD pipelines for data workloads
Monthly salary for this position ranges from 3800 EUR gross to 5800 EUR gross. The final offer will be dependent on qualifications, professional experience and competencies.
IBM is committed to creating a diverse environment and is proud to be an equal-opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender, gender identity or expression, sexual orientation, national origin, caste, genetics, pregnancy, disability, neurodivergence, age, veteran status, or other characteristics. IBM is also committed to compliance with all fair employment practices regarding citizenship and immigration status.