Luxoft company
Our benefits:
๐ฉโโ๏ธ Private Medical Care in Luxmed and Life Insurance
๐๏ธโโ๏ธ Multisport Card
๐จโ๐งโ๐ฆ Paid referrals
๐ Self-learning libraries
๐ซ Relocation package for seniors and assistance during all process...and MORE!
๐ Location: Wroclaw or Krakow - Hybrid work (3 days from office)
Project Description:
The project focuses on building a modern cloud-based data processing platform, enabling efficient management and real-time analysis of large datasets. Key aspects include implementing ETL/ELT pipelines, optimizing data processing, and ensuring compliance with industry regulations. The solution leverages Azure, Apache Spark, and Data Lakehouse architecture to support strategic business decisions.
Responsibilities:
We're looking for a Big Data Lead Engineer to:
โข Engineer reliable data pipelines for sourcing, processing, distributing, and storing data in different ways, using cloud (Azure) data platform infrastructure effectively.
โข Transform data into valuable insights that inform business decisions, making use of our internal data platforms and applying appropriate analytical techniques.
โข Develop, train, and apply data engineering techniques to automate manual processes, and solve challenging business problems.
โข Ensure the quality, security, reliability, and compliance of our solutions by applying our digital principles and implementing both functional and non-functional requirements.
โข Build observability into our solutions, monitor production health, help to resolve incidents, and remediate the root cause of risks and issues.
โข Understand, represent, and advocate for client needs.
โข Codify best practices, methodology and share knowledge with other engineers in UBS
โข Shape the Data and Distribution architecture and technology stack within our new cloud-based datalake-house.be a hands-on contributor, senior lead in the big data and data lake space, capable to collaborate and influence architectural and design principles across batch and real time flows
โข Have a continuous improvement mindset, who is always on the look out for ways to automate and reduce time to market for deliveries
Mandatory Skills Description:
Your Expertise
โข Experience in building Data Processing pipeline using various ETL/ELT design patterns and methodologies to Azure data solution, building solutions using ADLSv2, Azure Data factory, Databricks, Python and PySpark.
โข Experience with at least one of the following technologies: Scala/Java or Python
โข Deep understanding of the software development craft, with focus on cloud based (Azure), event driven solutions and architectures, with key focus on Apache Spark batch and streaming, Datalakehouses using medallion architecture. Knowledge of DataMesh principles is added plus.
โข Ability to debug using tools Ganglia UI, expertise in Optimizing Spark Jobs
โข The ability to work across structured, semi-structured, and unstructured data, extracting information and identifying linkages across disparate datasets.
โข Expert in creating data structures optimized for storage and various query patterns for e.g. Parquet and Delta Lake
โข Experience in traditional data warehousing concepts (Kimball Methodology, Star Schema, SCD) / ETL tools (Azure Data factory, Informatica)
โข Experience in data modelling atleast one database technology such as:
o Traditional RDBMS (MS SQL Server, Oracle, PostgreSQL)
o NoSQL (MongoDB, Cassandra, Neo4J, CosmosDB, Gremlin)
o Understanding of Information Security principles to ensure compliant handling and management of data
Nice-to-Have Skills Description:
โข Ability to clearly communicate complex solutions.
โข Strong problem solving and analytical skills.
โข Working experience in Agile methodologies (SCRUM)