Christian R.
Data Engineer
Experience
Data Engineer
Support and consulting services for an SCM application.
Technologies: Google Cloud, GKE, Kubernetes, SQL, Kafka, Kafka Connect, Kotlin, dbt, GitHub, Pekko.
- Supported the creation of a master data set for a supply chain management application.
- Designed and implemented ETL workflows.
- Analyzed data to improve data quality.
- Built a dashboard to monitor data quality.
- Set up CI/CD pipelines, monitoring, and alerting for operational ETL processes.
Data Engineer
Covestro AG
Support and consulting services as part of a data warehouse implementation.
Technologies: AWS Cloud, CloudFormation, SAP PLM, OpenSearch, Docker, Spring, Flyway, Java, SQL.
- Advised on data modeling, ETL workflows, and cloud architecture.
- Designed and implemented multiple ETL workflows.
- Connected the SAP PLM system for data extraction and preparation.
- Connected the MES system for data extraction and preparation.
- Transferred knowledge to internal staff.
Data Engineer
GfK SE
Support with migrating a data warehouse from an on-premise environment to the AWS Cloud.
Technologies: AWS Cloud, Airflow, Terraform, Python, Java, SQL, Cloudera, Hadoop, Glue, EMR, LakeFormation, GitLab.
- Designed and implemented a data warehouse using LakeFormation and Glue Catalog.
- Migrated existing Spark jobs to AWS Glue and AWS EMR.
- Implemented workflow management with Airflow.
- Designed and implemented multiple GitLab CI/CD pipelines.
- Trained and transferred knowledge to internal staff.
Data Architect
RTL Group
Designed and implemented a cloud-based data warehouse for analyzing user data.
Technologies: AWS Cloud, Kinesis, Kubernetes, Istio, Docker, Kafka, Python, PySpark, Pandas, Avro, Kafka Streams, Terraform, Python, Kustomize, PowerBI.
- Designed and implemented cloud infrastructure for building a data warehouse.
- Integrated Airflow as a workflow engine.
- Integrated the Google Spark K8s operator as the runtime for ETL processes.
- Built a team for implementing ETL processes.
- Implemented multiple ETL processes with PySpark.
- Built an event streaming pipeline for real-time analytics.
- Prepared data for further analysis in PowerBI.
- Advised on overall data architecture.
DevOps Engineer
D.Swarovksi KG
Design and implementation of infrastructure for sensor data processing, extension of an existing data science environment.
Technologies: AWS Cloud, Kubernetes, CloudFormation, Kafka, Kafka Connect, PySpark, Bamboo, Java, Python, Docker, Helm Charts, InfluxDB.
- Provided advice on design and tools for building a Kubernetes-based infrastructure for sensor data processing.
- Designed and built infrastructure on Kubernetes (Kafka cluster, Spark framework, Zookeeper, ZK manager).
- Designed and built an ETL pipeline for data ingestion.
- Designed and built a CI/CD pipeline with Bamboo and Kubernetes.
Data Engineer
Volkswagen AG
Design and implementation of a cloud-based data warehouse for analyzing vehicle data, evaluation of the data science environment.
Technologies: AWS Cloud, Lambda, IAM, Kubernetes, Kubeflow, Docker, Databricks, Terraform, Python, OpenSearch, LogStash, Kibana, Helm.
- Extended and deployed a prototype for mass data processing.
- Built and launched a CI/CD pipeline.
- Designed, implemented, and deployed a backend API including Helm chart.
- Designed and implemented project structure and release management.
- Evaluated Kubeflow and Databricks.
- Built ETL processes for data validation and ingestion.
- Implemented feature extraction from vehicle data.
Data Engineer
D. Swarovski KG
Design and implementation of a cloud-based data warehouse / data science environment.
Technologies: AWS Cloud, Kubernetes, Spark, R, NiFi, CloudFormation, Docker, Python, Jupyter Notebook.
- Designed a dynamically scalable data warehouse.
- Implemented infrastructure in CloudFormation (Infrastructure as Code).
- Implemented infrastructure components in Kubernetes.
- Modeled data warehouse and data storage.
- Built ETL processes for data ingestion.
Data Architect
aixigo AG
Supported the design and implementation of a microservice architecture.
Technologies: Microservices, Java, Docker, Kafka, LiquiBase, general IT architecture.
- Coordinated different teams on technology adoption.
- Helped design core components (de/serializer, data pipeline design, error handling, message handling, database design).
- Introduced Kafka as the central message bus for microservices.
- Introduced LiquiBase for database schema management.
- Provided business and technical support for a specific microservice.
Requirements Engineer
Open Grid Europe GmbH
Supported the evaluation of vendors in the Big Data field.
Technologies: Hortonworks, Cloudera, SAP Cloud, Apache NiFi, AWS Cloud, MS Azure.
- Recorded and documented technical and functional requirements for building and running an Apache Hadoop-based data warehouse.
- Collected quotes from various vendors and prepared the information for decision-making.
- Implemented a prototype for data ingestion.
Data Engineer
GfK SE
Designed and implemented a big data warehouse in the AWS Cloud.
Technologies: AWS Cloud, Spark, SparkR, Cloudera, Hadoop, Hive, Python, Jupyter Notebook, R, Bamboo, Terraform.
- Led the technical project.
- Designed AWS Cloud infrastructure.
- Implemented data pipelines.
- Built data warehouse and workflow management.
- Data preparation and process management.
Data Engineer
University Hospital Basel
Big Data Technologies Workshop – Introduction and how it works.
Technologies: Hadoop, Spark, AWS Cloud, MapReduce, Hive, Pig, R.
- Conducted a 3-day workshop.
- Introduced the Big Data/Hadoop ecosystem.
- Hands-on exercise on using Big Data in AWS Cloud.
Data Engineer
Helix Leisure Pte Ltd
Architecture review, design, and implementation of a streaming layer.
Technologies: Hadoop, Spark, AWS Cloud, Scala, MapReduce, JCascalog, RedShift, CloudFormation.
- Reviewed and assessed existing architecture and data model.
- Conducted workshop on data management / Lambda architecture.
- Designed and implemented real-time layer with Spark RT.
- Developed concept and implementation for integrating the real-time layer and the batch layer.
Data Engineer
Otto GmbH & Co. KG
Supported building ETL pipelines for a Hadoop-based data warehouse.
Technologies: Hadoop, Hive, Spark, Redis, Kafka, Avro, Scala, HCatalog, Schedoscope.
- Planned and implemented a Hive export module.
- Implemented Kafka & Redis export module as part of an open source project.
- Implemented an analysis algorithm to evaluate click streams.
DevOps Engineer
GfK SE
Design and development of a continuous deployment/delivery pipeline for a data-driven application in a cloud environment.
Technologies: AWS Cloud, Hadoop, Spark, Bamboo, Git, Terraform, Vagrant, InfluxDB.
- Planned and implemented a Big Data infrastructure in AWS Cloud.
- Planned and implemented a continuous deployment pipeline.
- Technical lead for an internal team.
Data Engineer
RadioOpt GmbH
Design and implementation of a data warehouse based on big data technologies – OLAP workload.
Technologies: Hadoop, Impala, Hive, ETL, AWS Cloud.
- Planned and implemented the cluster infrastructure.
- Evaluated different input formats for performance.
- Prepared and conducted load tests.
Data Engineer
Technicolor SA
Design and implementation of a Big Data system for batch and real-time data processing.
Technologies: Hadoop, Samza, Spark, Kafka, Java, ETL, AWS, CloudFormation.
- Planning and setup of the deployment environment.
- Evaluating different technologies for data collection/data processing.
- Technical lead of a team.
- Implementing a distributed, fault-tolerant high-throughput messaging and analytics system for machine data (Lambda Architecture).
Data Engineer
Ubisoft / BlueByte GmbH
Design and implementation of a Hadoop-based data warehouse for game analytics.
Technologies: Hadoop, Map/Reduce, Kafka, Hive, ETL, Java, Linux.
- Planning and setup of a data warehouse.
- Evaluating different approaches to data collection.
- Selecting suitable technologies.
- Technical lead and coordination of a distributed team (GER, CN, CAN).
- Implementing a distributed, fault-tolerant high-throughput messaging system.
DevOps Engineer
Deutsche Telekom AG
Design and implementation of a Big Data infrastructure in virtualized environments.
Technologies: Hadoop, OpenStack, Opscode Chef, Java, Linux.
- Planning and setup of a Big Data deployment infrastructure.
- Implementing the deployment process for Hadoop clusters on demand in a virtualized environment.
- Prototyping various algorithms in the Map/Reduce framework.
DevOps Engineer
GfK SE
Design and implementation of a Big Data architecture for analyzing telecommunications data.
Technologies: Cloudera, Hadoop, Hive, Flume, Java, Spring, Puppet, Ubuntu Linux, AWS.
- Planning and setup of the network (VPC).
- Planning and setup of a Hadoop cluster (100TB capacity).
- Setting up deployment processes, including monitoring.
- Implementing a data ingestion framework to store about 300GB of data per day.
Data Engineer
exactag GmbH
Design and implementation of a Hadoop cluster.
Technologies: Cloudera, Hadoop, Hive, Pig, Python, Java, Maven, Puppet, Debian Linux.
- Consulting and designing a Hadoop cluster.
- Selecting suitable hardware.
- Setting up a deployment process and rolling out the cluster.
- Porting existing statistical routines to the Map/Reduce framework.
Data Engineer
Etracker GmbH
Reimplementation of an analysis tool as a Map/Reduce application.
Technologies: Cloudera, Hadoop/HBase, Java, Maven, Ganglia, Chef, PHP, Debian Linux.
- Analysis and integration of an existing implementation into the Map/Reduce framework with the Hadoop Streaming API.
- Installation and configuration of a Hadoop cluster including monitoring.
- Setup of a deployment process.
Data Engineer, DevOps Engineer
LambdaNow.com / AltusInsight GmbH
Design and development of a web application (LambdaNow).
Technologies: Apache Hadoop, Python, Puppet, AWS, OpenStack, Git, RedHat Linux.
Application design.
- Website and backend implementation.
- Setup of the deployment process + hosting environment.
- Setup of a fully automated Apache Hadoop deployment process in the Amazon and OpenStack cloud.
Backend Developer
Aupeo GmbH
Integration of a payment provider into an existing backend.
Technologies: Ruby/Rails, OAuth, MySQL, Git, Debian Linux.
- Data mapping and text matching with existing data.
- Data preparation, conversion and import into the database.
- Integration of a payment provider.
Backend Developer
OpenLimit SignCubes GmbH
Integration of a signature component into an email program (KMail).
Technologies: C++, Qt, KDE, Ubuntu Linux.
- Setup of the debug environment.
- Integration of the signature component into KMail.
- Testing the implementation.
Backend Developer
Etracker GmbH
Implementation and refactoring of an analysis tool in C++.
Technologies: C++, MySQL C/C++ API, Doxygen, Hudson, Ubuntu/Debian Linux.
- Setup of a build environment for C++ projects.
- Refactoring of the prototype.
- Adaptation and extension of the software to the production environment (logging, error handling, unit testing).
- Setup of a deployment process.
- Setup of a build server (continuous integration).
Freelance Data Engineer
Ingenieurbüro Christian Richter – Data, Cloud & Container
- Freelance Data Engineer with an interest in DevOps
- Involved in over 20 successful projects
Backend Developer
Ingenieurbüro Christian Richter
Design and development of a web crawler.
Technologies: C++, Fedora/RedHat Linux, Cassandra.
- Design of the application as a high-performance multithreaded server.
- Implementation as a distributed application using asynchronous sockets (non-blocking I/O).
Data Engineer
MOG Inc.
Extension of an existing indexing framework.
Technologies: Ruby/Rails, MySQL.
- Adaptation of an indexing framework for music data to match a changed database model.
- Conversion of existing data (approx. 100 GB).
DevOps Engineer
MOG Inc.
Design, setup and deployment of a hosting environment for a large website.
Technologies: Apache, Nginx, HAProxy, Mongrel, MySQL, MySQLProxy, BIND, DHCP, Cobbler, Puppet, RedHat Linux.
- Design of the hosting environment.
- Setup of the infrastructure for server provisioning/configuration.
- Configuration of MySQL master-master replication + MySQLProxy.
- Installation of server software, monitoring and logging.
- Migration of the website from the hosting provider to colocation.
- Analysis and optimization to improve overall system performance.
Backend Developer
MOG Inc.
Porting an XML-RPC server from Ruby on Rails to C++.
Technologies: C++, XML-RPC, Ruby/Rails, XML Schema, MySQL C/C++ API, RedHat Linux.
- Analysis of performance issues.
- Implementation/porting according to specification (given protocol).
- Replacement of the component and integration into the existing backend.
Data Engineer
MOG Inc.
Design and development of an infrastructure to integrate data from external providers.
Technologies: Ruby/Rails, MySQL, Bash, RedHat Linux.
- Building a music database with data from providers like Allmusic, Muze (Macromedia), Rhapsody, MediaNet.
- Data preparation, conversion, and import into the database.
- Mapping data from different providers.
Backend Developer
MOG Inc.
Design and development of a spellchecker as a component of an indexing framework.
Technologies: C++, SOAP, CLucene, MySQL C/C++ API, Doxygen, RedHat Linux.
- Analysis of existing algorithms.
- Implementation of an algorithm (Levensthein Distance and Ngram Index) in C++.
- Integration into the existing system.
Software Developer and System Architect
MOG Inc. – Startup in the media/internet sector
- Responsible for the design, implementation, and deployment of the hosting environment
- Design and implementation of several software projects in C++
Backend Developer
MOG Inc.
Design and development of a collaborative filtering system as a distributed application.
Technologies: C++, MySQL C/C++ API, XML, XML Schema, Perl, Doxygen, RedHat Linux.
- Evaluation of suitable algorithms.
- Implementation as a distributed application with database integration.
- Integration into the existing backend and frontend.
Backend Developer
Fraunhofer Institut
Implementation/porting of a method for extracting the main melody line.
Technologies: C++, Matlab, Mandrake Linux.
- Porting an existing Matlab algorithm to C++.
- Optimization of the implementation for performance.
Research Associate
Fraunhofer IDMT – Research Institute in the Field of Audio/Video
- Conducted scientific research in search algorithms
- Participated in the development and implementation of algorithms
Data Analyst
Fraunhofer Institute
Scientific investigation of data structures to determine the nearest neighbor for a Music Information Retrieval System.
Technologies: Matlab, C++, Perl, Apache, CGI, Mandrake Linux.
- Evaluated several nearest neighbor algorithms for suitability based on the given specifications
- Designed and implemented an algorithm in C++ and integrated it into the existing system (as a dynamic C++ library)
- Conducted and analyzed a test series for evaluation
Data Analyst
ID Analytics Inc.
Collaborated on the development of algorithms for identity theft detection.
Technologies: Java, Octave, Perl, Tomcat, Oracle, Red Hat Linux, Solaris.
Implemented a tool for graph visualization.
- Analyzed large datasets (~250 GB) for feature extraction using a Java processing framework
- Developed and implemented algorithms for identity theft detection (regression analysis)
ID Analytics Inc. – Startup in the Financial Services Sector
- Collaborated on software for identity theft detection
- Data preparation and data analysis
- Contributed to the development and implementation of algorithms
Fraunhofer Institute
Contributed to the design and implementation of a cross-platform GUI (Win32/Linux) for a Query by Humming system.
Technologies: C++, Qt, Win32, Linux.
Fraunhofer IDMT – Research Institute for Audio/Video
- Development of test environments to evaluate algorithms
- Design and implementation of GUIs to present algorithms at trade fairs
Fraunhofer Institute
Contributed to the scientific evaluation of similarity search algorithms for a query-by-humming system.
Technologies: Matlab, Bash, Linux.
- Implementation of various algorithms in Matlab.
- Automation and parallelization of the test environment using Bash scripts.
Jumatech – PCB Manufacturing Company
- Provision of IT infrastructure
Summary
- Designing and implementing ETL workflows, data pipelines, and ML pipelines
- GDPR-compliant data management and processing
- Designing and building cloud-based data warehouse, data lake, and lakehouse implementations
- Data modeling and analysis, data conversion, and preparation
- Designing and building data-driven applications on cloud-native infrastructures
- Requirements analysis, business process analysis, risk analysis
Languages
Education
Technische Universität Ilmenau
Diploma · Theoretical Electrical Engineering · Ilmenau, Germany
Georg-Cantor-Gymnasium Halle (Saale)
High school diploma · Halle (Saale), Germany
Similar Freelancers
Discover other experts with similar qualifications and experience