In the past I’ve mainly worked with backend systems, that drive business processes, especially information processing and integration with different systems for companies like:

  • Volkswagen Group
  • Mercedes Benz
  • BMW
  • OTTO GmbH
  • ABB
  • Siemens

Some highlighted projects

Development of a cloud platform for measurement data.

Project objective: Provide a platform to find and analyse measurement data for worldwide usage.

My responsibilities:

  • Introducing Apache Airflow as workflow-scheduler and Kubernetes for operations.
  • Code review of existing software.
  • Backlog grooming and User-Story refinements.

Company: EFS (as permanent employee), Gaimersheim

Role: Senior Developer

Software & Tools: Spark, Drill, Airflow, Jenkins, Terraform, AWS, Docker, Kubernetes, Gitlab

Migrate proof-of-concept to Production

Project objective: Scale a proof-of-concept design from Python and PostgreSQL to handle 2TB+ of data.

My responsibilities:

  • Consulting in regards to data-security and data-privacy.
  • Requirements engineering for design and scaling of cluster.
  • Design reviews.
  • Training and coaching: Scala, Test Driven Development, Big-Data, Spark.
  • Implementation of show-cases.
  • Quality assurance.

Company: Mercedes-Benz (as T-Systems permanent employee), Munich

Role: Senior Developer

Software & Tools: Hadoop (Hortonworks), Spark, Atlas, Ranger, Druid, Drill, Zeppelin Notebooks, Superset Dashboards.

Introducing BigData for BI DataScience Department

Project objective: Enable new and senior data-scientists to work with BigData technology.

My responsibilities:

  • Hadoop Cluster Administration (Cloudera Distribution), including Performance Tuning and upgrading hosts and Hadoop, installation of new tools and libraries.
  • Data-Engineering with HiveQL, Spark (Scala and PySpark) for product recommendation, fraud-detection and ad-targeting.
  • Coaching data-scientists and developers on effective Spark programming and optimized HiveQL queries.
  • Scheduling of Data-Hub loading.

Company: OTTO GmbH, Hamburg

Role: Consulting & Coaching on Big-Data Technology

Software & Tools: Hadoop (Cloudera), Spark, Hive, HBase, Cloudera-Manager, Scala, Bash, Python, Java, Zeppelin-Notebooks, Schedoscope, Ansible, Jenkins, Git, Bitbucket.

Integration of Supplier-Relationship data

Project objective: XML and JSON data from SOAP and RESTful webservices needed to be integrated into the customer’s business warehouse.

My responsibilities: Requirements analysis and design of a solution. I came up with the implementation of a Domain Specific Language for transformation of JSON documents into tabular format. The DSL enables projection and pivoting of nested structures in JSON documents and resolves ragged hierarchies.

Company: ABB AG Group Function Supply Chain Management, Mannheim-Zürich

Role: Software-Architect & Lead-developer.

Software & Tools: Clojure, MongoDB, JavaScript, SQL, MS-SQL, DB2

Data engineering for data-science project

Project objective: Data engineering for data-science project in supply chain management domain.

My responsibilities: Realization of the project-mission. This included the definition of the project-goal, collecting the different data-sources, design and documentation of the data-frame, developing transformation rules with Pig & Hive stored procedures, implementing the data-pipeline with HiveQL, coaching the DataScience team (Fraunhofer IAIS & Open Logic Systems) with domain knowledge.

Company: ABB AG Group Function Supply Chain Management, Mannheim-Zürich

Role: Project Manager & Lead-developer.

Software & Tools: Hadoop (Hortonworks), MongoDB, Python, Pig, T-SQL

Introduction of cluster processing

Project objective: Preparation for the introduction of a cluster infrastructure for data processing.

My responsibilities:

  • Quality review for ETL (IBM Data-Stage), BI (IBM Cognos Framework Model), Data-Warehouse Design.
  • Prototype for a data-model for data-science project for pattern analysis on quantitative and qualitative data from supply chain management processes.

Company: ABB AG Group Function Supply Chain Management, Mannheim-Zürich

Role: Software-Architect.

Software & Tools: Hadoop (Hortonworks), R, Python, MongoDB, T-SQL, D3.js

Re-engineering of business warehouse

Project objective: Migrate the business warehouse from a T-SQL based solution to robust and faster technology that’s easier to maintain.

The warehouse handles worldwide reporting data on transactional-level on top of 50+ ERP-systems (SAP and others, non harmonised structures and procedures).

My responsibilities:

  • Support for roll-out of world-wide platform for savings-reporting.
  • Consulting for re-design for a procurement business-intelligence platform. Analysis and documentation of legacy business warehouse design and loading process. Testing and documentation of software.
  • Establishing DevOps Team with internal employees and external consultants.
  • Integrating supplier-lifecycle-management into BI-stack.
  • Implemented new software for management of reference-data and manual savings-entries.
  • Implemented tool for worldwide target-setting process.
  • Implementation of agile project-management practice with Scrum & Kanban.
  • Prototyping of ETL-workload with Hadoop (HiveQL and custom storedprocedures).

Company: ABB AG Group Function Supply Chain Management, Mannheim (permanent employment)

Role: Lead Solution Architect GF-SCMS\

Software & Tools: IBM DataStage, SQL-Server, SSIS, SSAS, Cognos-Suite, Jedox-Suite, Spring-MVC, MongoDB, Python, Java, T-SQL, Jira, Kunagi, Hadoop (Hortonworks).