Skip to content

No Cost. Just Performance. Start with Exasol Community Edition.

Download now

Building a Data Warehouse in 2025+ (Setup, Creation & Implementation)

Mathias Golombek
· · 21 mins read

A data warehouse is more than a storage system. It is the foundation for analytics, compliance reporting, and strategic decision-making. Building one requires more than provisioning infrastructure: it means defining business goals, choosing a data warehouse architecture that matches data volume and velocity, integrating multiple sources with reliable pipelines, and implementing governance that meets regulatory standards.

This guide explains how to build a data warehouse from the ground up. It covers each stage of the process — from defining requirements to integration, testing, and monitoring — with practical considerations such as cost, disaster recovery metrics (RTO/RPO), and industry-specific compliance.

The goal: a data warehouse that is resilient, efficient, and aligned with business priorities.

What Is a Data Warehouse?

A data warehouse is a centralized system that stores data from multiple sources in a structured form optimized for analysis. Unlike an operational database (OLTP), which manages real-time transactions, a data warehouse supports OLAP workloads: large-scale queries, reporting, and historical analysis.

An operational database answers “What is happening right now?” A data warehouse answers “What has happened over time, and what does it mean?” This distinction makes it the backbone of long-term analytics, compliance, and planning.

Benefits of Building a Data Warehouse

A data warehouse provides more than consolidated storage. It gives organizations a reliable framework for analytics and compliance. The main data warehouse benefits include:

  • Unified decision-making – A single source of truth prevents conflicting reports across business units.
  • Faster intelligence – Optimized structures let queries and dashboards run without slowing transactional systems.
  • Data quality – Transformation processes standardize, cleanse, and validate records before analysis.
  • Historical analysis – Maintain years of data to enable forecasting, audits, and trend discovery.
  • Scalability – A well-structured data warehouse can expand to handle higher data volumes and more sources without full rework.
  • Compliance and sovereignty – Centralization makes it easier to apply GDPR, HIPAA, and regional residency requirements.
  • Resilience – Disaster recovery planning with defined RTO and RPO values ensures business continuity.

To create a data warehouse, follow these 10 must-have steps.

Try Exasol Free: No Costs, Just Speed

Run Exasol locally and test real workloads at full speed.

10 Steps to Building a Data Warehouse

Setting up a data warehouse involves provisioning infrastructure, configuring security, loading initial datasets, and validating reports before wider rollout.

Before you start building, be clear on where your data comes from and which problems you’re solving; otherwise, you risk creating a warehouse that looks complete but delivers little value.

Dirk Beerbohm, Global Partner Solution Architect, Exasol

Step 1 – Define Business Objectives

A modern data warehouse build that begins without explicit objectives almost always ends in scope creep. Start by writing down the business questions your data warehouse must answer. Typical examples:

  • Finance: revenue forecasting, margin analysis, regulatory reporting.
  • Marketing: customer segmentation, attribution modeling, campaign ROI.
  • Operations: supply chain tracking, predictive maintenance.
  • Compliance: audit trails for GDPR, HIPAA, or local residency rules.

Each objective determines technical requirements. A data warehouse for financial compliance demands auditability and immutability; one for marketing analytics needs low-latency pipelines. Define measurable KPIs for success, such as “Sales dashboards must load in under 3 seconds” or “Regulatory reports must include 7 years of history.”

Without this alignment, the project risks becoming an expensive storage repository with little business value.

Step 2 – Assess Data Sources and Requirements

The next step is mapping every source system that will feed the data warehouse. Categories typically include:

  • Internal systems – CRM, ERP, HR, finance, IoT logs, clickstream.
  • External feeds – partner APIs, market data, demographic datasets.
  • Semi-structured and unstructured – JSON logs, images, sensor data.

For each, capture:

  • Volume (rows per day, storage size).
  • Velocity (batch, micro-batch, or streaming).
  • Variety (structured SQL tables vs. NoSQL vs. files).
  • Quality issues (duplicates, missing values, inconsistent IDs).
  • Regulatory restrictions (e.g., EU customer data must stay in Frankfurt).

This assessment informs architecture (centralized vs. lakehouse), integration method (batch vs. streaming), and cost modeling (storage + compute), one of the first steps in setting up a data warehouse.

For example, a retail chain collecting billions of IoT sensor events daily may require a cloud, partitioned lakehouse with tiered storage, while a SaaS business with fewer sources may succeed with a centralized data warehouse and scheduled batch loads.

A survey of integration methods outlines the complexity of combining diverse data sources.

Step 3 – Choose Architecture and Deployment Model

A data warehouse can be centralized, distributed, or combined with a data lake. The right architecture depends on business goals, compliance needs, and expected data volumes. Each option comes with trade-offs:

  • Centralized data warehouse – Strong governance, consistent data models, but bottlenecks if business units cannot iterate independently.
  • Data marts – Faster delivery for departments, but risk of siloing if not integrated.
  • Distributed or data mesh – Domain teams own their pipelines; requires strict interoperability standards.
  • Lakehouse integration – Merges structured and semi-structured data; supports open formats such as Delta Lake, Apache Iceberg, and Hudi for ACID transactions and time-travel queries.

Deployment options:

  • On-premises – Maximum control over performance, security, and data residency. Ideal for organizations with strict compliance or data sovereignty requirements.
  • Hybrid – Sensitive workloads remain on-premises while less regulated data or elastic compute can be handled in the cloud. This approach balances governance with scalability.
  • Cloud – Offers rapid provisioning and elasticity, but can introduce challenges with residency, cost predictability, and vendor lock-in.

Document how the chosen architecture meets the business objectives from Step 1. For example, a bank under strict residency rules may require an on-premises data warehouse with certified security controls, while a multinational manufacturer might adopt a hybrid strategy to keep regulated data local while analyzing IoT data at scale.

Step 4 – Design the Data Model

Data warehouse models define query performance and maintenance effort, and directly shape how you create a data warehouse that supports business needs. Options include:

  • Star schema – Central fact table (e.g., Sales) with dimension tables (e.g., Product, Customer). Fast for BI queries, widely supported.
  • Snowflake schema – Normalized dimensions reduce redundancy but add joins.
  • Data Vault – Hubs, links, and satellites provide resilience to change; suitable for long-term, audit-heavy environments.
  • Anchor modeling – Focused on temporal flexibility; can capture history without complex SCD logic.
  • Wide tables – Flattened schemas optimized for machine learning feature stores, but less flexible for BI.

Define granularity upfront. Daily aggregates minimize storage but limit analysis; transaction-level granularity allows flexible reporting but requires more capacity.

Handle slowly changing dimensions (SCD) explicitly:

  • Type 1: overwrite values (simple, but history lost).
  • Type 2: add new rows with timestamps (maintains history).
  • Type 3–6: hybrid approaches.

Example: A retailer may choose SCD Type 2 for product categories, ensuring category changes over time are preserved for accurate historical sales reporting.

Try Exasol Free: No Costs, Just Speed

Run Exasol locally and test real workloads at full speed.

Step 5 – Select the Technology Stack

Technology choices define how the data warehouse processes, stores, and exposes data. A complete stack usually includes:

  • Database platform – At the core is the data warehouse engine itself. An on-premises or hybrid deployment provides direct control over performance tuning, data residency, and compliance. Evaluate scalability in terms of concurrent queries, response time under load, and ability to handle both structured and semi-structured data.
  • Integration layer – Pipelines must extract, transform, and load (ETL) or extract, load, and transform (ELT) data from multiple sources. Choose frameworks that align with your data velocity needs — batch processing for historical loads, micro-batch for daily reporting, or streaming for near real-time data analytics.
  • Transformation layer – Define how raw data becomes analysis-ready. This may include deduplication, normalization, enrichment, and business rule application. Automation frameworks can reduce manual effort and enforce data quality consistently.
  • Metadata and governance – A catalog or lineage tracker makes it clear where data originated, how it changed, and who owns it. This supports compliance and auditability.
  • Analytics and BI tools – Dashboards and query tools must connect seamlessly to the data warehouse, with role-based access controls to protect sensitive datasets.

When selecting the stack, align every tool with the objectives from Step 1. For example, if regulatory audits are a primary driver, prioritize audit logging and immutable storage. If query concurrency is critical, benchmark the platform under realistic workloads before adoption.

The biggest trap I see is over-engineering. Think big, start small, and let business needs guide the next steps instead of chasing technical perfection.

Dirk Beerbohm, Global Partner Solution Architect, Exasol

Step 6 – Plan Data Integration (ETL and CDC)

Every data warehouse needs reliable pipelines. At a high level, integration means moving data from source systems, keeping it updated, and applying business rules consistently. The details below explain the technical options and trade-offs.

Integration modes

  • Batch loads – Periodic ingestion (hourly, nightly). Efficient for large volumes, but latency can limit decision-making.
  • Micro-batch – Near real-time updates at intervals of minutes. Balances freshness with resource cost.
  • Streaming – Continuous data flow, required for scenarios such as fraud detection or IoT monitoring.

Change Data Capture (CDC)

CDC ensures the data warehouse reflects updates in source systems without full reloads. Common patterns include:

  • Log-based CDC – Reads database transaction logs; efficient and minimally invasive.
  • Trigger-based CDC – Uses database triggers to capture changes; adds overhead but works when logs are unavailable.
  • Table-diff CDC – Compares source and target tables; simple but resource-intensive.

Data transformations

Standardize formats, reconcile identifiers, and apply business rules as data moves into the data warehouse. Typical transformations:

  • Deduplication – Prevents skewed reports.
  • Type conversions – Aligns inconsistent field types across systems.
  • Enrichment – Adds reference data such as geocodes or currency conversions.

Validation and monitoring

Every integration job should include row counts, data type checks, and referential integrity tests. Establish alerts for late or failed loads, because a broken pipeline can make entire dashboards unreliable.

Integration design should also consider data residency. For example, a pipeline moving customer data across regions must comply with GDPR restrictions on cross-border transfers.

Technical Checklist

  • Define SLA for data freshness (e.g., “transactions must appear in the data warehouse within 10 minutes”).
  • Maintain audit tables for pipeline runs (start time, end time, row counts).
  • Implement alerting for load failures and latency breaches.
  • Encrypt data in transit between source and data warehouse.

Step 7 – Implement Data Governance and Security

A data warehouse that lacks governance becomes a liability rather than an asset. Building governance into the data warehouse ensures data can be trusted, audited, and accessed only by the right people.

Governance framework

  • Define ownership for every dataset: who creates it, who approves changes, who consumes it.
  • Document data lineage so every value in a report can be traced back to its origin.
  • Maintain metadata catalogs and business glossaries to prevent inconsistent terminology.

Access and security

  • Enforce role-based access control (RBAC) so users see only the data required for their role.
  • Apply column- and row-level security for sensitive attributes such as salaries or health information.
  • Encrypt data both in transit and at rest to protect against interception or breaches.

Compliance and sovereignty

  • Implement controls aligned to GDPR, HIPAA, SOC 2, or ISO 27001 depending on industry requirements and data sovereignty trends.
  • For organizations in Europe or regulated industries, ensure data residency by hosting the data warehouse in controlled locations or certified facilities.
  • Log all access events and administrative changes to provide a verifiable audit trail for regulators.
  • Security and privacy controls are cataloged in NIST SP 800-53.

Monitoring and audits

  • Schedule periodic audits of access logs, data quality metrics, and governance policies.
  • Use automated validation checks to flag anomalies such as sudden spikes in null values or unauthorized access attempts.

A well-governed data warehouse is not only a technical asset but also a compliance enabler. It reduces audit risk, improves user trust, and ensures the data warehouse remains a reliable foundation for decision-making.

Step 8 – Build for Resilience (RTO/RPO)

Even the best-designed data warehouse is only valuable if it remains available when needed. Resilience planning defines how the system behaves during failures and how quickly it can recover.

Recovery objectives

No system is immune to failure. Planning resilience means defining how quickly the data warehouse must recover and how much data loss is acceptable if something goes wrong.

  • Recovery Time Objective (RTO): the maximum acceptable downtime after an outage.
  • Recovery Point Objective (RPO): the maximum acceptable amount of data loss measured in time (e.g., 15 minutes of transactions).
    Both must be defined with business stakeholders and documented in service-level agreements.

Principles of continuity planning and backup site selection are discussed in research on evaluating DR strategies for high-performance data systems.

High availability measures

  • Redundant hardware or virtualized clusters to remove single points of failure.
  • Synchronous replication between nodes for zero or near-zero data loss.
  • Automated failover processes tested regularly rather than assumed to work.

Disaster recovery planning

  • Maintain geo-redundant backups in separate physical or logical locations.
  • Use immutable backup storage to prevent accidental or malicious deletion.
  • Define rollback procedures so a corrupted load can be reversed without compromising the entire data warehouse.

Testing and validation

  • Run controlled outage drills to verify RTO and RPO targets can be met.
  • Document every test with measured recovery times and update playbooks accordingly.
  • Monitor not only hardware but also pipelines — a pipeline failure can be as disruptive as a hardware outage.

Resilience is not achieved through infrastructure alone. It requires continuous testing, clear documentation, and alignment with the business impact of downtime.

Step 9 – Test, Benchmark, and Optimize

A data warehouse must be validated before it can be trusted. Testing ensures accuracy, benchmarking proves performance under realistic conditions, and optimization keeps the system reliable as data volumes grow.

Testing layers

  • Schema validation – Confirm that tables, constraints, and data types match specifications.
  • Data quality checks – Validate referential integrity, uniqueness, and required fields.
  • Business rule testing – Ensure transformations produce correct outputs (e.g., revenue = price × quantity).
  • Pipeline testing – Verify ETL/ELT jobs complete within required timeframes and deliver expected row counts.
  • User acceptance testing (UAT) – Involve end users to confirm reports and dashboards return correct values.

Automation frameworks

Introduce automated testing frameworks to reduce manual effort and prevent regressions. Examples include:

  • Great Expectations – Declarative validation of data quality and schema constraints.
  • dbt tests – Embedded checks in SQL transformations.
  • Soda – Continuous monitoring of freshness, volume, and integrity.

Benchmarking methodology

Define a repeatable benchmark process before go-live:

  • Load a representative dataset covering peak volumes.
  • Run query suites that simulate expected workloads (reporting dashboards, ad-hoc analysis, machine learning feature extraction).
  • Measure throughput, concurrency, and latency under stress.
  • Record baseline metrics for future comparison.

Optimization practices

  • Partition large fact tables by time or business key to improve query performance.
  • Create materialized views for high-frequency queries.
  • Archive or purge unused historical data into cheaper storage tiers.
  • Monitor query logs to identify inefficient patterns and tune indexes or joins.

Optimization is a continuous process. As new sources and workloads appear, repeat testing and benchmarking to ensure the data warehouse continues to meet performance and business objectives.

Try Exasol Free: No Costs, Just Speed

Run Exasol locally and test real workloads at full speed.

Step 10 – Launch, Monitor, and Iterate

Deployment is not the end of the project. Data warehouse implementation must be continuously monitored and refined to stay reliable and relevant.

Controlled launch

  • After the initial data warehouse setup, start with a limited rollout to a subset of users.
  • Validate critical reports against source systems before opening access to wider teams.
  • Communicate changes to business users so adoption is planned, not forced.

Monitoring in production

  • Track pipeline health: job duration, error rates, and data latency.
  • Monitor query performance and concurrency to detect bottlenecks.
  • Collect data freshness metrics and enforce service-level objectives (SLOs).

Issue response

  • Configure automated alerts for failed loads, missing partitions, or sudden data anomalies.
  • Maintain documented runbooks for common issues such as failed ingestion or schema drift.
  • Review incidents after resolution and update processes to prevent recurrence.

Iteration and optimization

  • Schedule periodic reviews with business stakeholders to confirm the data warehouse still meets reporting and compliance needs.
  • Add new sources incrementally and re-benchmark workloads after each change.
  • Revisit cost models regularly, especially if storage or compute usage grows faster than forecast.

A data warehouse that is actively monitored and iterated upon remains a strategic asset. Without ongoing attention, it risks degrading into an unreliable system that business leaders stop trusting.

Cost and Timeline for Building a Data Warehouse

The cost of a data warehouse project varies with scope, architecture, and regulatory environment. Typical ranges:

  • Pilot or MVP – A limited build focused on a few sources and basic reporting (the fastest way to create a data warehouse). Industry benchmarks indicate a starting cost of around $70,000, with delivery times of 3–4 months.
  • Mid-size implementation – Broader integration across finance, operations, and customer data. Typical data warehouse development projects run 6–9 months with budgets in the $250,000–$500,000 range.
  • Enterprise build – Large-scale deployments with global reach, strict compliance, and high concurrency. These projects often extend 9–12 months or longer, with budgets from $500,000 into the seven figures.

Factors influencing cost

  • Deployment model – On-premises requires capital expenditure for hardware and licenses, while hybrid adds cloud operating costs.
  • Data volume and velocity – Larger, faster sources increase compute and storage demand.
  • Integration complexity – Real-time pipelines, CDC, and third-party data feeds increase engineering effort.
  • Governance and compliance – Enforcing GDPR, HIPAA, or SOC 2 controls raises design and operational costs.
  • Team composition – Skilled architects, engineers, and stewards represent a significant portion of budget.

Hidden costs

  • Ongoing maintenance and optimization.
  • Data quality remediation during initial loads.
  • Audit preparation and external certifications.

Practical timeline guidance

  • Add 15–20% contingency for unexpected data issues.
  • Treat the data warehouse as an iterative build: deliver an MVP quickly, then expand scope in phases rather than aiming for “big bang” delivery.
  • Plan a dedicated period for user training and adoption, often overlooked but critical for project success.
Project scopeTypical cost (USD)TimelineCommon risks
Pilot / MVPFrom $70,0003–4 monthsUnderestimating data quality issues; limited adoption if users aren’t trained
Mid-size build$250,000–$500,0006–9 monthsIntegration complexity, pipeline failures, rising cloud or infra costs
Enterprise build$500,000+ (often seven figures)9–12 months+Compliance gaps, multi-region latency, project sprawl without phased delivery

Migration and Rollback Strategies

Building a new data warehouse often involves migrating data and workloads from existing systems. Without a clear migration and rollback plan, projects risk downtime, data loss, or loss of stakeholder trust.

Migration approaches

  • Lift-and-shift – Move data and schema structures directly with minimal changes. Fast but often carries over inefficiencies.
  • Re-platform – Redesign schemas or pipelines to take advantage of the new data warehouse’s features. Requires more effort but usually results in better performance and maintainability.
  • Hybrid migration – Run old and new data warehouses in parallel for a period, gradually shifting workloads. Reduces risk but increases short-term costs.

Validation methods

  • Reconcile row counts between source and target.
  • Compare aggregates (e.g., total revenue by month) to confirm accuracy.
  • Perform sampling checks to ensure transformations preserved business logic.
  • Use parallel reporting during the dual-run period to confirm consistent outputs.

Rollback planning

  • Define explicit rollback criteria (e.g., more than 2% mismatch in validation checks).
  • Keep snapshots or backups of the legacy system until the new data warehouse is proven reliable.
  • Maintain dual access for business users until validation is complete.
  • Document manual recovery steps in case automated rollback fails.

A well-documented rollback plan prevents a migration issue from turning into a business outage. By running old and new systems in parallel, validating outputs, and keeping clear thresholds for rollback, organizations minimize risk during the transition.

Building the Right Team

A data warehouse project succeeds only if the right mix of roles is in place. Each role covers a distinct responsibility, from technical implementation to governance and user adoption.

Core roles

  • Project manager – Coordinates timelines, budgets, and stakeholder communication.
  • Data architect – Designs the overall architecture and data models.
  • ETL/ELT engineer – Builds and maintains integration pipelines.
  • Database administrator (DBA) – Tunes performance, manages storage, ensures availability.
  • Data steward – Monitors data quality and enforces governance rules.
  • BI developer / analyst – Builds reports and dashboards for business users.
  • Compliance officer or security lead – Ensures regulatory requirements (GDPR, HIPAA, residency) are enforced.

RACI (Responsible, Accountable, Consulted, Informed)

A structured RACI matrix prevents gaps and overlaps in ownership:

  • Responsible: engineers executing the work.
  • Accountable: architect or project manager signing off on deliverables.
  • Consulted: compliance officer, security lead, or subject matter experts.
  • Informed: executives and stakeholders receiving updates.

Staffing guidance

  • A pilot data warehouse may be delivered by a small team (project manager, architect, 1–2 engineers).
  • Enterprise builds require a broader team with dedicated governance and compliance roles.
  • Avoid over-reliance on a single engineer for pipelines or data modeling; redundancy in skills reduces project risk.

The right team balances technical expertise with governance and business alignment, ensuring the data warehouse is not just delivered, but trusted and adopted.

Industry-Specific Blueprints

While the fundamentals of a data warehouse apply across industries, regulatory requirements and data models differ. Tailoring blueprints to industry needs accelerates adoption and reduces compliance risk.

Finance

  • Focus areas: risk reporting, capital adequacy, banking analytics.
  • Compliance: Basel III, SOX, local financial regulations.
  • Modeling considerations: high-granularity transaction data, immutable audit logs, strict retention policies.

Healthcare

  • Focus areas: patient records, clinical outcomes, operational efficiency.
  • Compliance: HIPAA (US), GDPR (EU), HL7/FHIR standards.
  • Modeling considerations: patient-centric schemas, de-identification of sensitive attributes, controlled access by role.

SaaS and Digital Services

  • Focus areas: customer churn, net revenue retention (NRR), product usage analytics.
  • Compliance: SOC 2, GDPR, regional residency rules.
  • Modeling considerations: event-based schemas, real-time ingestion of user activity, multi-tenant data isolation.

Manufacturing

  • Focus areas: supply chain optimization, predictive maintenance, quality assurance.
  • Compliance: ISO 9001, local safety and trade reporting requirements.
  • Modeling considerations: time-series data from IoT sensors, integration of ERP and MES systems, high-frequency streaming.

Retail

  • Focus areas: inventory forecasting, loyalty analytics, omnichannel performance.
  • Compliance: PCI DSS for payment data, GDPR for customer behavior data.
  • Modeling considerations: SKU-level granularity, seasonal trend analysis, integration with POS and e-commerce platforms.

Adapting the data warehouse to industry requirements reduces rework and improves trust in the data. Predefined models, KPI libraries, and compliance mappings help teams move from proof of concept to production faster.

Try Exasol Free: No Costs, Just Speed

Run Exasol locally and test real workloads at full speed.

Build Your Future-Ready Data Warehouse

A data warehouse is not just a technical system. It is the foundation for analytics, compliance, and strategic decision-making. By defining objectives, selecting the right architecture, implementing resilient pipelines, and enforcing governance, organizations can move beyond fragmented reporting and create a single, trusted source of truth.

The most successful projects treat the data warehouse as a long-term program: starting small with a pilot, validating performance and quality, then scaling across departments and regions. With careful planning for resilience, compliance, and cost control, a data warehouse becomes an asset that grows in value as data volumes and business demands increase.

Contact us to discuss how we can help you build a data warehouse tailored to your performance, sovereignty, and compliance requirements.

FAQs on Building a Data Warehouse

Building a data warehouse involves defining business objectives, assessing data sources, selecting architecture, designing data models, creating integration pipelines, applying governance, testing, and then deploying with monitoring.

To create a data warehouse, you set up a database platform, design schemas (e.g., star or snowflake), implement ETL/ELT pipelines, and configure reporting access. The exact process depends on whether you deploy on-premises, hybrid, or cloud.

An individual or small team can create a data warehouse by starting with a pilot: select a lightweight database, integrate 1–2 key data sources, and validate reports. This establishes a foundation that can later scale.

The main components are:

-Data sources

-ETL/ELT processes

-Storage layer

-Metadata and governance layer

-Reporting and analytics tools

1. Data extraction

2. Data transformation and loading

3. Storage and organization in the data warehouse

4. Access and analysis through BI tools

Typical steps include: define business goals, gather and assess data sources, choose architecture, design the data model, build integration pipelines, enforce governance, test and benchmark, then launch with monitoring.

These terms usually describe layered data processing:

L1: raw ingested data.

L2: cleaned and conformed data.

L3: business-ready, aggregated data used for reporting.

A data warehouse developer designs, builds, and maintains the data warehouse. Responsibilities include schema design, ETL/ELT pipeline development, performance tuning, and supporting BI teams.

No. SQL is a query language used to interact with databases, including data warehouses. A data warehouse is a system for storing and analyzing large volumes of data, while SQL is the language to retrieve and manipulate that data.

No. SQL itself is not a tool but a language. However, many data warehouse platforms use SQL as their standard interface.

You define schemas using SQL DDL commands, load data via ETL/ELT processes, and query data for analysis. Building a complete SQL data warehouse requires additional tooling for ingestion, transformation, and governance.

Mathias Golombek
Mathias Golombek

Mathias Golombek is the Chief Technology Officer (CTO) of Exasol. He joined the company as a software developer in 2004 after studying computer science with a heavy focus on databases, distributed systems, software development processes, and genetic algorithms. By 2005, he was responsible for the Database Optimizer team and in 2007 he became Head of Research & Development. In 2014, Mathias was appointed CTO. In this role, he is responsible for product development, product management, operations, support, and technical consulting.