Projects
In the past I’ve mainly worked with backend systems, that drive business processes, especially information processing and integration with different systems for companies like:
- Volkswagen Group
- Mercedes Benz
- BMW
- OTTO GmbH
- ABB
- Siemens
Some highlighted projects
Development of a cloud platform for measurement data.
Project objective: Provide a platform to find and analyse measurement data for worldwide usage.
My responsibilities:
- Introducing Apache Airflow as workflow-scheduler and Kubernetes for operations.
- Code review of existing software.
- Backlog grooming and User-Story refinements.
Company: EFS (as permanent employee), Gaimersheim
Role: Senior Developer
Software & Tools: Spark, Drill, Airflow, Jenkins, Terraform, AWS, Docker, Kubernetes, Gitlab
Migrate proof-of-concept to Production
Project objective: Scale a proof-of-concept design from Python and PostgreSQL to handle 2TB+ of data.
My responsibilities:
- Consulting in regards to data-security and data-privacy.
- Requirements engineering for design and scaling of cluster.
- Design reviews.
- Training and coaching: Scala, Test Driven Development, Big-Data, Spark.
- Implementation of show-cases.
- Quality assurance.
Company: Mercedes-Benz (as T-Systems permanent employee), Munich
Role: Senior Developer
Software & Tools: Hadoop (Hortonworks), Spark, Atlas, Ranger, Druid, Drill, Zeppelin Notebooks, Superset Dashboards.
Introducing BigData for BI DataScience Department
Project objective: Enable new and senior data-scientists to work with BigData technology.
My responsibilities:
- Hadoop Cluster Administration (Cloudera Distribution), including Performance Tuning and upgrading hosts and Hadoop, installation of new tools and libraries.
- Data-Engineering with HiveQL, Spark (Scala and PySpark) for product recommendation, fraud-detection and ad-targeting.
- Coaching data-scientists and developers on effective Spark programming and optimized HiveQL queries.
- Scheduling of Data-Hub loading.
Company: OTTO GmbH, Hamburg
Role: Consulting & Coaching on Big-Data Technology
Software & Tools: Hadoop (Cloudera), Spark, Hive, HBase, Cloudera-Manager, Scala, Bash, Python, Java, Zeppelin-Notebooks, Schedoscope, Ansible, Jenkins, Git, Bitbucket.
Integration of Supplier-Relationship data
Project objective: XML and JSON data from SOAP and RESTful webservices needed to be integrated into the customer’s business warehouse.
My responsibilities: Requirements analysis and design of a solution. I came up with the implementation of a Domain Specific Language for transformation of JSON documents into tabular format. The DSL enables projection and pivoting of nested structures in JSON documents and resolves ragged hierarchies.
Company: ABB AG Group Function Supply Chain Management, Mannheim-Zürich
Role: Software-Architect & Lead-developer.
Software & Tools: Clojure, MongoDB, JavaScript, SQL, MS-SQL, DB2
Data engineering for data-science project
Project objective: Data engineering for data-science project in supply chain management domain.
My responsibilities: Realization of the project-mission. This included the definition of the project-goal, collecting the different data-sources, design and documentation of the data-frame, developing transformation rules with Pig & Hive stored procedures, implementing the data-pipeline with HiveQL, coaching the DataScience team (Fraunhofer IAIS & Open Logic Systems) with domain knowledge.
Company: ABB AG Group Function Supply Chain Management, Mannheim-Zürich
Role: Project Manager & Lead-developer.
Software & Tools: Hadoop (Hortonworks), MongoDB, Python, Pig, T-SQL
Introduction of cluster processing
Project objective: Preparation for the introduction of a cluster infrastructure for data processing.
My responsibilities:
- Quality review for ETL (IBM Data-Stage), BI (IBM Cognos Framework Model), Data-Warehouse Design.
- Prototype for a data-model for data-science project for pattern analysis on quantitative and qualitative data from supply chain management processes.
Company: ABB AG Group Function Supply Chain Management, Mannheim-Zürich
Role: Software-Architect.
Software & Tools: Hadoop (Hortonworks), R, Python, MongoDB, T-SQL, D3.js
Re-engineering of business warehouse
Project objective: Migrate the business warehouse from a T-SQL based solution to robust and faster technology that’s easier to maintain.
The warehouse handles worldwide reporting data on transactional-level on top of 50+ ERP-systems (SAP and others, non harmonised structures and procedures).
My responsibilities:
- Support for roll-out of world-wide platform for savings-reporting.
- Consulting for re-design for a procurement business-intelligence platform. Analysis and documentation of legacy business warehouse design and loading process. Testing and documentation of software.
- Establishing DevOps Team with internal employees and external consultants.
- Integrating supplier-lifecycle-management into BI-stack.
- Implemented new software for management of reference-data and manual savings-entries.
- Implemented tool for worldwide target-setting process.
- Implementation of agile project-management practice with Scrum & Kanban.
- Prototyping of ETL-workload with Hadoop (HiveQL and custom storedprocedures).
Company: ABB AG Group Function Supply Chain Management, Mannheim (permanent employment)
Role: Lead Solution Architect GF-SCMS\
Software & Tools: IBM DataStage, SQL-Server, SSIS, SSAS, Cognos-Suite, Jedox-Suite, Spring-MVC, MongoDB, Python, Java, T-SQL, Jira, Kunagi, Hadoop (Hortonworks).