Development of a data platform for precision medicine research for the pharma and biotech industries as part of multi-omics integration.
Responsibilities:
Design and deployment of Azure-based data lakehouse solutions, including Azure Databricks. This included configuring the Unity Catalog and setting up the network.
Implement complex data engineering pipelines for ingestion, transformation, and persistence of data into Delta tables using Python, Scala, and SQL, ensuring readiness for analytics and downstream workflows.
Data lake ingestion using real-time streaming and batch technologies like Kafka, Spark Structured Streaming, and Event Hubs.
Automated deployment of the lakehouse platform on Azure DevOps pipelines using IaC with Terraform (azurem) for version control, deployment, and automated testing.
Integration of Databricks with Azure ADLS Gen2 and identity management (managed identities, access connectors) to optimize governance and access control.
Implement data privacy measures to comply with GDPR, including pseudonymization and anonymization of user identifiers, role-based access control (RBAC) with Azure AD integration, data retention policies with automated deletion workflows, column- and row-level security, and data masking.
Ensure data quality using the Expectation Management framework.
Support internal teams by setting up self-service environments, configuring clusters, documenting setup processes, and troubleshooting deployment issues. I am available as of 03.11.2025.
Responsibilities:
Development of cloud applications and their supporting infrastructure.
Design and implementation of a data quality checking and validation strategy for an Azure Data Lakehouse with batch and streaming processing based on the medallion architecture.
Designing and implementing data pipelines within a lakehouse data platform.
Technologies: Azure Synapse, Data Factory, Azure Functions, Event Hubs, Terraform, C#, Python, GX (Expectation Management).
Clients: ORAYLIS GmbH, Telefónica Deutschland
Technical leadership at the project level, coordinating teams of up to five people.
Responsible for designing, implementing, and successfully deploying new cloud data solutions to production.
Close collaboration with stakeholders and technical coaching within the team.
Projects & Results
Migrated an Oracle Exadata on-premise DWH to the cloud.
End-to-end development of data-driven applications, including support and documentation.
Clients: Volkswagen Financial Services (VWFS)
Technologies: Azure Databricks, Spark (Scala, SQL, Python), Synapse Analytics, Kafka, Power BI, Azure DevOps, ADLS Gen2, Spark Structured Streaming, PowerShell Universal, Microsoft Azure Intune.
Lead Developer – Big Data & Software Development
Clients: BMW, Deutsche Börse
In the role of technical lead at the project level, responsible for coordinating development teams of up to five people.
Driving the technical implementation of new products from conception to go-live.
Close collaboration with functional and technical stakeholders to ensure project success and adherence to quality standards.
Planning and executing multiple successful production deployments for both clients.
Technical leadership responsibilities: architecture decisions, technical reviews, and coaching team members.
Deutsche Börse
Technical leadership and co-design of the C7 SCS service (EUREX SCS Reporting Engine), which was successfully adopted in production.
Ensured operational stability and performance of the big data pipelines in live operation.
BMW
Technologies: Hadoop (HDFS), YARN, Spark (Scala, Python), Kafka, Java, AMQP, OpenShift, Kubernetes, Docker, RESTful APIs, Swagger, Oracle, PostgreSQL, RabbitMQ, Jenkins, Kibana, Grafana, Prometheus.
Development of ETL pipelines and web interfaces to automate ordering and delivery processes, inventory checks, payments, shipping, and customer feedback in the e-commerce sector.
Technologies: PHP, Java, Python, JavaScript
Tasks:
ETL and analysis of data from SCADA systems in wind farms to identify performance weaknesses and technical issues.
Development of tools in the statistical software R to calculate the impact of wind and weather on a wind turbine's performance.
Development of internal software tools.
Skills:
C#, R, Statistics
I am an experienced data engineer and entrepreneur with a proven track record in designing, developing, and maintaining big data pipelines for real-time streaming and batch workflows. I have hands-on expertise in microservices architecture, RESTful APIs, full-stack web development, and test automation, which I have applied in industries such as pharma, biotech, and automotive.
I have led projects from data ingestion to visualization, worked in agile environments, and coordinated multiple teams. I develop scalable solutions using tools like Azure, Python, SQL, Kafka, and Spark, ensuring data quality and efficient performance in cloud and on-premise infrastructures.
Discover other experts with similar qualifications and experience
2025 © FRATCH.IO GmbH. All rights reserved.