Christian R.

Data Engineer

Berlin, Germany

Experience

Oct 2024 - Present
1 year 1 month

Data Engineer

Support and consulting services for an SCM application.

Technologies: Google Cloud, GKE, Kubernetes, SQL, Kafka, Kafka Connect, Kotlin, dbt, GitHub, Pekko.

  • Supported the creation of a master data set for a supply chain management application.
  • Designed and implemented ETL workflows.
  • Analyzed data to improve data quality.
  • Built a dashboard to monitor data quality.
  • Set up CI/CD pipelines, monitoring, and alerting for operational ETL processes.
Sep 2023 - Oct 2024
1 year 2 months

Data Engineer

Covestro AG

Support and consulting services as part of a data warehouse implementation.

Technologies: AWS Cloud, CloudFormation, SAP PLM, OpenSearch, Docker, Spring, Flyway, Java, SQL.

  • Advised on data modeling, ETL workflows, and cloud architecture.
  • Designed and implemented multiple ETL workflows.
  • Connected the SAP PLM system for data extraction and preparation.
  • Connected the MES system for data extraction and preparation.
  • Transferred knowledge to internal staff.
Jul 2021 - Oct 2023
2 years 4 months

Data Engineer

GfK SE

Support with migrating a data warehouse from an on-premise environment to the AWS Cloud.

Technologies: AWS Cloud, Airflow, Terraform, Python, Java, SQL, Cloudera, Hadoop, Glue, EMR, LakeFormation, GitLab.

  • Designed and implemented a data warehouse using LakeFormation and Glue Catalog.
  • Migrated existing Spark jobs to AWS Glue and AWS EMR.
  • Implemented workflow management with Airflow.
  • Designed and implemented multiple GitLab CI/CD pipelines.
  • Trained and transferred knowledge to internal staff.
Jun 2020 - Dec 2022
2 years 7 months

Data Architect

RTL Group

Designed and implemented a cloud-based data warehouse for analyzing user data.

Technologies: AWS Cloud, Kinesis, Kubernetes, Istio, Docker, Kafka, Python, PySpark, Pandas, Avro, Kafka Streams, Terraform, Python, Kustomize, PowerBI.

  • Designed and implemented cloud infrastructure for building a data warehouse.
  • Integrated Airflow as a workflow engine.
  • Integrated the Google Spark K8s operator as the runtime for ETL processes.
  • Built a team for implementing ETL processes.
  • Implemented multiple ETL processes with PySpark.
  • Built an event streaming pipeline for real-time analytics.
  • Prepared data for further analysis in PowerBI.
  • Advised on overall data architecture.
Aug 2019 - Dec 2019
5 months

DevOps Engineer

D.Swarovksi KG

Design and implementation of infrastructure for sensor data processing, extension of an existing data science environment.

Technologies: AWS Cloud, Kubernetes, CloudFormation, Kafka, Kafka Connect, PySpark, Bamboo, Java, Python, Docker, Helm Charts, InfluxDB.

  • Provided advice on design and tools for building a Kubernetes-based infrastructure for sensor data processing.
  • Designed and built infrastructure on Kubernetes (Kafka cluster, Spark framework, Zookeeper, ZK manager).
  • Designed and built an ETL pipeline for data ingestion.
  • Designed and built a CI/CD pipeline with Bamboo and Kubernetes.
Jan 2019 - Mar 2020
1 year 3 months

Data Engineer

Volkswagen AG

Design and implementation of a cloud-based data warehouse for analyzing vehicle data, evaluation of the data science environment.

Technologies: AWS Cloud, Lambda, IAM, Kubernetes, Kubeflow, Docker, Databricks, Terraform, Python, OpenSearch, LogStash, Kibana, Helm.

  • Extended and deployed a prototype for mass data processing.
  • Built and launched a CI/CD pipeline.
  • Designed, implemented, and deployed a backend API including Helm chart.
  • Designed and implemented project structure and release management.
  • Evaluated Kubeflow and Databricks.
  • Built ETL processes for data validation and ingestion.
  • Implemented feature extraction from vehicle data.
Oct 2017 - Nov 2018
1 year 2 months

Data Engineer

D. Swarovski KG

Design and implementation of a cloud-based data warehouse / data science environment.

Technologies: AWS Cloud, Kubernetes, Spark, R, NiFi, CloudFormation, Docker, Python, Jupyter Notebook.

  • Designed a dynamically scalable data warehouse.
  • Implemented infrastructure in CloudFormation (Infrastructure as Code).
  • Implemented infrastructure components in Kubernetes.
  • Modeled data warehouse and data storage.
  • Built ETL processes for data ingestion.
May 2017 - Dec 2018
1 year 8 months

Data Architect

aixigo AG

Supported the design and implementation of a microservice architecture.

Technologies: Microservices, Java, Docker, Kafka, LiquiBase, general IT architecture.

  • Coordinated different teams on technology adoption.
  • Helped design core components (de/serializer, data pipeline design, error handling, message handling, database design).
  • Introduced Kafka as the central message bus for microservices.
  • Introduced LiquiBase for database schema management.
  • Provided business and technical support for a specific microservice.
Mar 2017 - Mar 2017
1 month

Requirements Engineer

Open Grid Europe GmbH

Supported the evaluation of vendors in the Big Data field.

Technologies: Hortonworks, Cloudera, SAP Cloud, Apache NiFi, AWS Cloud, MS Azure.

  • Recorded and documented technical and functional requirements for building and running an Apache Hadoop-based data warehouse.
  • Collected quotes from various vendors and prepared the information for decision-making.
  • Implemented a prototype for data ingestion.
Jan 2017 - Aug 2017
8 months

Data Engineer

GfK SE

Designed and implemented a big data warehouse in the AWS Cloud.

Technologies: AWS Cloud, Spark, SparkR, Cloudera, Hadoop, Hive, Python, Jupyter Notebook, R, Bamboo, Terraform.

  • Led the technical project.
  • Designed AWS Cloud infrastructure.
  • Implemented data pipelines.
  • Built data warehouse and workflow management.
  • Data preparation and process management.
Oct 2016 - Oct 2016
1 month

Data Engineer

University Hospital Basel

Big Data Technologies Workshop – Introduction and how it works.

Technologies: Hadoop, Spark, AWS Cloud, MapReduce, Hive, Pig, R.

  • Conducted a 3-day workshop.
  • Introduced the Big Data/Hadoop ecosystem.
  • Hands-on exercise on using Big Data in AWS Cloud.
Jun 2016 - Dec 2016
7 months

Data Engineer

Helix Leisure Pte Ltd

Architecture review, design, and implementation of a streaming layer.

Technologies: Hadoop, Spark, AWS Cloud, Scala, MapReduce, JCascalog, RedShift, CloudFormation.

  • Reviewed and assessed existing architecture and data model.
  • Conducted workshop on data management / Lambda architecture.
  • Designed and implemented real-time layer with Spark RT.
  • Developed concept and implementation for integrating the real-time layer and the batch layer.
Feb 2016 - Jul 2016
6 months

Data Engineer

Otto GmbH & Co. KG

Supported building ETL pipelines for a Hadoop-based data warehouse.

Technologies: Hadoop, Hive, Spark, Redis, Kafka, Avro, Scala, HCatalog, Schedoscope.

  • Planned and implemented a Hive export module.
  • Implemented Kafka & Redis export module as part of an open source project.
  • Implemented an analysis algorithm to evaluate click streams.
Dec 2015 - Aug 2016
9 months

DevOps Engineer

GfK SE

Design and development of a continuous deployment/delivery pipeline for a data-driven application in a cloud environment.

Technologies: AWS Cloud, Hadoop, Spark, Bamboo, Git, Terraform, Vagrant, InfluxDB.

  • Planned and implemented a Big Data infrastructure in AWS Cloud.
  • Planned and implemented a continuous deployment pipeline.
  • Technical lead for an internal team.
Jul 2015 - Oct 2015
4 months

Data Engineer

RadioOpt GmbH

Design and implementation of a data warehouse based on big data technologies – OLAP workload.

Technologies: Hadoop, Impala, Hive, ETL, AWS Cloud.

  • Planned and implemented the cluster infrastructure.
  • Evaluated different input formats for performance.
  • Prepared and conducted load tests.
Jul 2014 - Jun 2015
1 year

Data Engineer

Technicolor SA

Design and implementation of a Big Data system for batch and real-time data processing.

Technologies: Hadoop, Samza, Spark, Kafka, Java, ETL, AWS, CloudFormation.

  • Planning and setup of the deployment environment.
  • Evaluating different technologies for data collection/data processing.
  • Technical lead of a team.
  • Implementing a distributed, fault-tolerant high-throughput messaging and analytics system for machine data (Lambda Architecture).
Mar 2013 - Sep 2014
1 year 7 months

Data Engineer

Ubisoft / BlueByte GmbH

Design and implementation of a Hadoop-based data warehouse for game analytics.

Technologies: Hadoop, Map/Reduce, Kafka, Hive, ETL, Java, Linux.

  • Planning and setup of a data warehouse.
  • Evaluating different approaches to data collection.
  • Selecting suitable technologies.
  • Technical lead and coordination of a distributed team (GER, CN, CAN).
  • Implementing a distributed, fault-tolerant high-throughput messaging system.
Feb 2013 - Jun 2014
1 year 5 months

DevOps Engineer

Deutsche Telekom AG

Design and implementation of a Big Data infrastructure in virtualized environments.

Technologies: Hadoop, OpenStack, Opscode Chef, Java, Linux.

  • Planning and setup of a Big Data deployment infrastructure.
  • Implementing the deployment process for Hadoop clusters on demand in a virtualized environment.
  • Prototyping various algorithms in the Map/Reduce framework.
Nov 2012 - Aug 2015
2 years 10 months

DevOps Engineer

GfK SE

Design and implementation of a Big Data architecture for analyzing telecommunications data.

Technologies: Cloudera, Hadoop, Hive, Flume, Java, Spring, Puppet, Ubuntu Linux, AWS.

  • Planning and setup of the network (VPC).
  • Planning and setup of a Hadoop cluster (100TB capacity).
  • Setting up deployment processes, including monitoring.
  • Implementing a data ingestion framework to store about 300GB of data per day.
May 2012 - Dec 2012
8 months

Data Engineer

exactag GmbH

Design and implementation of a Hadoop cluster.

Technologies: Cloudera, Hadoop, Hive, Pig, Python, Java, Maven, Puppet, Debian Linux.

  • Consulting and designing a Hadoop cluster.
  • Selecting suitable hardware.
  • Setting up a deployment process and rolling out the cluster.
  • Porting existing statistical routines to the Map/Reduce framework.
Jun 2011 - Mar 2012
10 months

Data Engineer

Etracker GmbH

Reimplementation of an analysis tool as a Map/Reduce application.

Technologies: Cloudera, Hadoop/HBase, Java, Maven, Ganglia, Chef, PHP, Debian Linux.

  • Analysis and integration of an existing implementation into the Map/Reduce framework with the Hadoop Streaming API.
  • Installation and configuration of a Hadoop cluster including monitoring.
  • Setup of a deployment process.
Mar 2011 - Sep 2016
5 years 7 months

Data Engineer, DevOps Engineer

LambdaNow.com / AltusInsight GmbH

Design and development of a web application (LambdaNow).

Technologies: Apache Hadoop, Python, Puppet, AWS, OpenStack, Git, RedHat Linux.

Application design.

  • Website and backend implementation.
  • Setup of the deployment process + hosting environment.
  • Setup of a fully automated Apache Hadoop deployment process in the Amazon and OpenStack cloud.
Sep 2010 - Feb 2011
6 months

Backend Developer

Aupeo GmbH

Integration of a payment provider into an existing backend.

Technologies: Ruby/Rails, OAuth, MySQL, Git, Debian Linux.

  • Data mapping and text matching with existing data.
  • Data preparation, conversion and import into the database.
  • Integration of a payment provider.
May 2010 - Sep 2010
5 months

Backend Developer

OpenLimit SignCubes GmbH

Integration of a signature component into an email program (KMail).

Technologies: C++, Qt, KDE, Ubuntu Linux.

  • Setup of the debug environment.
  • Integration of the signature component into KMail.
  • Testing the implementation.
Mar 2010 - May 2010
3 months

Backend Developer

Etracker GmbH

Implementation and refactoring of an analysis tool in C++.

Technologies: C++, MySQL C/C++ API, Doxygen, Hudson, Ubuntu/Debian Linux.

  • Setup of a build environment for C++ projects.
  • Refactoring of the prototype.
  • Adaptation and extension of the software to the production environment (logging, error handling, unit testing).
  • Setup of a deployment process.
  • Setup of a build server (continuous integration).
Jan 2010 - Present
15 years 10 months
Berlin, Germany

Freelance Data Engineer

Ingenieurbüro Christian Richter – Data, Cloud & Container

  • Freelance Data Engineer with an interest in DevOps
  • Involved in over 20 successful projects
Nov 2009 - Feb 2010
4 months

Backend Developer

Ingenieurbüro Christian Richter

Design and development of a web crawler.

Technologies: C++, Fedora/RedHat Linux, Cassandra.

  • Design of the application as a high-performance multithreaded server.
  • Implementation as a distributed application using asynchronous sockets (non-blocking I/O).
May 2009 - Oct 2009
6 months

Data Engineer

MOG Inc.

Extension of an existing indexing framework.

Technologies: Ruby/Rails, MySQL.

  • Adaptation of an indexing framework for music data to match a changed database model.
  • Conversion of existing data (approx. 100 GB).
May 2008 - Apr 2009
1 year

DevOps Engineer

MOG Inc.

Design, setup and deployment of a hosting environment for a large website.

Technologies: Apache, Nginx, HAProxy, Mongrel, MySQL, MySQLProxy, BIND, DHCP, Cobbler, Puppet, RedHat Linux.

  • Design of the hosting environment.
  • Setup of the infrastructure for server provisioning/configuration.
  • Configuration of MySQL master-master replication + MySQLProxy.
  • Installation of server software, monitoring and logging.
  • Migration of the website from the hosting provider to colocation.
  • Analysis and optimization to improve overall system performance.
Oct 2007 - Apr 2008
7 months

Backend Developer

MOG Inc.

Porting an XML-RPC server from Ruby on Rails to C++.

Technologies: C++, XML-RPC, Ruby/Rails, XML Schema, MySQL C/C++ API, RedHat Linux.

  • Analysis of performance issues.
  • Implementation/porting according to specification (given protocol).
  • Replacement of the component and integration into the existing backend.
Jun 2006 - Sep 2007
1 year 4 months

Data Engineer

MOG Inc.

Design and development of an infrastructure to integrate data from external providers.

Technologies: Ruby/Rails, MySQL, Bash, RedHat Linux.

  • Building a music database with data from providers like Allmusic, Muze (Macromedia), Rhapsody, MediaNet.
  • Data preparation, conversion, and import into the database.
  • Mapping data from different providers.
Jan 2006 - May 2006
5 months

Backend Developer

MOG Inc.

Design and development of a spellchecker as a component of an indexing framework.

Technologies: C++, SOAP, CLucene, MySQL C/C++ API, Doxygen, RedHat Linux.

  • Analysis of existing algorithms.
  • Implementation of an algorithm (Levensthein Distance and Ngram Index) in C++.
  • Integration into the existing system.
May 2005 - Oct 2009
4 years 6 months
Berkeley, United States

Software Developer and System Architect

MOG Inc. – Startup in the media/internet sector

  • Responsible for the design, implementation, and deployment of the hosting environment
  • Design and implementation of several software projects in C++
May 2005 - Dec 2005
8 months

Backend Developer

MOG Inc.

Design and development of a collaborative filtering system as a distributed application.

Technologies: C++, MySQL C/C++ API, XML, XML Schema, Perl, Doxygen, RedHat Linux.

  • Evaluation of suitable algorithms.
  • Implementation as a distributed application with database integration.
  • Integration into the existing backend and frontend.
Sep 2004 - Mar 2005
7 months

Backend Developer

Fraunhofer Institut

Implementation/porting of a method for extracting the main melody line.

Technologies: C++, Matlab, Mandrake Linux.

  • Porting an existing Matlab algorithm to C++.
  • Optimization of the implementation for performance.
Jul 2003 - Mar 2005
1 year 9 months
Ilmenau, Germany

Research Associate

Fraunhofer IDMT – Research Institute in the Field of Audio/Video

  • Conducted scientific research in search algorithms
  • Participated in the development and implementation of algorithms
Jul 2003 - Aug 2004
1 year 2 months

Data Analyst

Fraunhofer Institute

Scientific investigation of data structures to determine the nearest neighbor for a Music Information Retrieval System.

Technologies: Matlab, C++, Perl, Apache, CGI, Mandrake Linux.

  • Evaluated several nearest neighbor algorithms for suitability based on the given specifications
  • Designed and implemented an algorithm in C++ and integrated it into the existing system (as a dynamic C++ library)
  • Conducted and analyzed a test series for evaluation
Jan 2003 - Jun 2003
6 months

Data Analyst

ID Analytics Inc.

Collaborated on the development of algorithms for identity theft detection.

Technologies: Java, Octave, Perl, Tomcat, Oracle, Red Hat Linux, Solaris.

Implemented a tool for graph visualization.

  • Analyzed large datasets (~250 GB) for feature extraction using a Java processing framework
  • Developed and implemented algorithms for identity theft detection (regression analysis)
Jan 2003 - Jun 2003
6 months
San Diego, United States
Lorem ipsum dolor sit amet

ID Analytics Inc. – Startup in the Financial Services Sector

  • Collaborated on software for identity theft detection
  • Data preparation and data analysis
  • Contributed to the development and implementation of algorithms
Nov 2001 - Dec 2002
1 year 2 months
Lorem ipsum dolor sit amet

Fraunhofer Institute

Contributed to the design and implementation of a cross-platform GUI (Win32/Linux) for a Query by Humming system.

Technologies: C++, Qt, Win32, Linux.

Aug 2000 - Dec 2002
2 years 5 months
Ilmenau, Germany
Lorem ipsum dolor sit amet

Fraunhofer IDMT – Research Institute for Audio/Video

  • Development of test environments to evaluate algorithms
  • Design and implementation of GUIs to present algorithms at trade fairs
Aug 2000 - Oct 2001
1 year 3 months
Lorem ipsum dolor sit amet

Fraunhofer Institute

Contributed to the scientific evaluation of similarity search algorithms for a query-by-humming system.

Technologies: Matlab, Bash, Linux.

  • Implementation of various algorithms in Matlab.
  • Automation and parallelization of the test environment using Bash scripts.
Jun 1999 - Jul 1999
2 months
Eckental, Germany
Lorem ipsum dolor sit amet

Jumatech – PCB Manufacturing Company

  • Provision of IT infrastructure

Summary

  • Designing and implementing ETL workflows, data pipelines, and ML pipelines
  • GDPR-compliant data management and processing
  • Designing and building cloud-based data warehouse, data lake, and lakehouse implementations
  • Data modeling and analysis, data conversion, and preparation
  • Designing and building data-driven applications on cloud-native infrastructures
  • Requirements analysis, business process analysis, risk analysis

Languages

German
Native
English
Advanced

Education

Oct 1998 - Sep 2004

Technische Universität Ilmenau

Diploma · Theoretical Electrical Engineering · Ilmenau, Germany

Sep 1992 - Jul 1996

Georg-Cantor-Gymnasium Halle (Saale)

High school diploma · Halle (Saale), Germany

Need a freelancer? Find your match in seconds.
Try FRATCH GPT
More actions