Deutsche Bank - Finance IT - Oracle DWH · For internal use only Deutsche Bank –Finance IT...

19
Deutsche Bank Finance IT Migration Oracle Exadata Dr. Marcus Praetzas March 2012

Transcript of Deutsche Bank - Finance IT - Oracle DWH · For internal use only Deutsche Bank –Finance IT...

For internal use only

Deutsche Bank – Finance ITMigration Oracle Exadata

Dr. Marcus Praetzas

March 2012

For internal use only

1

Motivation

Migration (Phase 2)

3

2

PoC (Phase 1)

4

Observations

Agenda

5

Business Drivers / Background

For internal use only

page 3

Volker Bettag, Architect

Dr. Michael Dreier, Infrastructure Manager

Randolf Geist, Oracle Specialist

Erwin Heute, Oracle Specialist

Jens Koch, MicroStrategy Infrastructure/Project Manager

Deutsche Bank Data Centre

Contact

Dr. Marcus Prätzas, Program ManagerDeutsche Bank AGWilhelm-Fay-Str. 31-37 D-65936 [email protected]

Deutsche Bank – Finance IT – dbArtos/FDW Team

For internal use only

Topic Area

RWA (Basel I / II)

EC / EL / GVA

Output / Activity

RWA calculations, Monthly driver analysis, Quarterly COREP reporting

Monthly Basel II reporting, EPE, MR-RWA

EC / EL / Average Active Equity calculation and reporting

GVA for not impaired corporate credit exposure

Others Group Derivative Bookings, Global Securities Netting, Banking book

collateral, Country risk, ...

Disclosure 20-F Item 11 (Risk Section, 37 pages), Footnotes, Annual Report

Analyst presentations, Interim Report, Financial Data Supplement

German Regulatory KWG Capital, KWG 13 / 14, Financial Conglomerate disclosure

Influence rule making and interpretation

Daily Derivatives Daily derivatives counterparty risk

Provide EPE calculations

Business Background – Drivers

page 4

For internal use only

page 5

Motivation – Daily Processing

Performance Demand 2010Core Process* Run Times by Quarter

Q1 2010

SAS deployed on AMD CPUs

with internal PCIe SSD storage

Q2 2010

InfiniBand private interconnect,

Enhanced parallel processing

Q3 2010

Datawarehouse Infrastructure

PoC using Oracle Exadata and

SSD based storage severs.

Oracle Infrastructure setup.

Q4 2010 – Exadata

Oracle Infrastructure go-live

Q1-Q3 2011 – Exadata

Migration of full environment

Target was ~10h

i.e. 50% reduction in non-

calculation steps required

*

* MR-RWA functionality added

For internal use only

Technical Process

page 6

Distributed

Engines

Netting

Expected

Loss

EC

Basel II

KWG

RWA

Credit Risk Engines

Data Warehouse

Disclosure

Process Control

B/S Netting

GVA

EL / EC

Country Risk

KWG 13 / 14

Principle I/II

Basel IIRegional QA

Regional QA

Daily QA

Monthly

Source

Monthly

Source

Monthly

Source

Monthly

Source

Daily

Source

Daily

Source

Daily

Source

Ext.

CalcSAS

For internal use only

page 7

Data Warehouse – and more …

For internal use only

PoC – Testpoints

Input Area Master Area Reporting Area

Data

Delivery Input Master Report

Calculation

View,

Extract

1

2 3

4

5

Five key production processes have been chosen

A full set of production data ist used for testing

The tests were executed in DB datacenter

The requirement has been set to 50% performance increase compared to the

monthly production setup at the time.

page 8

For internal use only

PoC – Testpoint Characteristics

Testpoint 1 – Integration Function (TP1)

Data Transformation between two Oracle schemas. CPU power consumption (e.g.

currency conversion) as well as large sequential IO operations. The IO is done in parallel

and includes substantial DML.

Testpoint 2/3 – SAS Engine Interface (TP2/3)

Perform a data down- and upload to the SAS Engine environment. As this is not a core

database functionality rather than a regression test of the InfiniBand connection not

further listed here.

Testpoint 4 – Starbuilder (TP4)

Large single threaded operation, where CPU and IO performance are equally essential.

Compared to TP1 these are far less complex operations.

Testpoint 5 – Microstrategy Reporting (TP5)

Random IO and massive parallel execution. Representative set of 110 and 470 reports

from production.

page 9 DDL - Data Definition Language, e.g. create table, partition vs. DML - Data Manipulation Language e.g. insert record

For internal use only

Testpoint 1 Results – Integration Function

54% performance gain on Exadata (V2)

About 25 test-runs with different Oracle / System configuration settings have been

executed for each environment. Minor application changes.

The maximum parallelism causes internal Oracle contention issues. 5 compute

nodes show best performance.

TP1

different Parameter settings per System

page 10

For internal use only

Oracle Exadata Scalability

With the exception of some parts that are executed across all available nodes the

scalability has been tested using a variable number of compute nodes

The optimum is reached with 5 nodes. Beyond that no improvement has been

observed.

page 11

For internal use only

Oracle Exadata Scalability – Data Volume

When doubling the data volume the runtime increases by 7%, for a factor of three the

runtime increases by 19%, with a factor of 4 the runtime gets 34% longer.

Base run-time is 45min

page 12

For internal use only

page 13

Disaster Recovery – Active Data Guard

1. Disaster recovery solutions utilizing Oracle Data-Guard for replication.

2. Utilise the Standby Database for reporting and backup purposes (Microstrategy)

For internal use only

page 14

Result

Oracle Exadata achieves an overall better performance improvement of ~55%

In particular the better reporting performance of the Oracle Exadata adds significant

more value.

The feature of hybrid column compression (available on Exadata only) enables a data

reduction for historical data down to ~25%.

Lower cost than the previous solution (traditional SAN based)

Observation

The PoC showed contention-issues effecting the achievable performance and scalability

of Oracle RAC on the Exadata V2. This occurs in particular when heavily using DDL like

truncating partitioning and rebuilding indexes on other partitions in parallel.

Conclusion

Migration of full environment using V2-8

PoC Results – Decision

For internal use only

page 15

Architecture Solution (2011)

Oracle Exadata V28

Datacentre #1

Full Rack #1

(45 TB available

for data + FRA)

Monthly Prod.(10 TB)

Daily Prod.

(Data Guard Copy)(6 TB)

Flash Recovery Area(all databases - 22 TB)

Cluster Filesystem(Buffer, etc. 4 TB)

Monthly Production

(Data Guard Copy)(10 TB)

Daily Prod.(6 TB)

Flash Recovery Area(all databases – 22TB)

Cluster Filesystem(Buffer, etc. 4 TB)

Oracle Exadata V28

Datacentre #2

Monthly UAT(10 TB)

Daily UAT(6 TB)

Flash Recovery Area(all databases - 22 TB)

Cluster Filesystem(Buffer, etc. 4 TB)

Oracle Exadata V28

Datacentre #2

Oracle Exadata V2

Datacentre #1

INT (10 TB)

DEV (3 TB)

Flash Recovery Area(all databases - 22 TB)

Cluster Filesystem(Buffer, 4 TB)

DR (Data Guard Copy) Clone (Snapshot Copy)

Full Rack #2

(45 TB available

for data + FRA)

Full Rack #3

(45 TB available

for data + FRA)

Full Rack #4

(45 TB available

for data + FRA)

(existing system)

Regional QA(2 TB)

Regional OA. (Data

Guard Copy) (2 TB)

QA UAT(2 TB)

QA INT (1 TB)

QADEV (0.5 TB)

Contingency(7 TB)

For internal use only

Migration started Q1 – 2011

Core functionality was proven & further performance gains indentified (index

usage on ODM) in the PoC

PoC complete and (daily) system live

Oracle Support for go live, environment review, tuning tips. All 12 findings

during POC had been resolved in < 3 weeks and addressed by patch bundle

sets.

Two additional findings since go live resolved.

page 16

For internal use only

Migration Log Book Q2 – Q3 2011

May

3 ODMs have been delivered and handed over from Oracle to Data Centre

HW & SW install in ~10 days (Oracle)

ODMs have been handed over from data centre to project 2 weeks later

June / July

First full environment (incl. SAS, NFS, etc.) established

Migration rehearsal & testing cycles

Integration testing in Jul

August

Dress rehearsalpage 17

September

DataGuard lines established

Improved performance with 10G line to

be compared with Q1 POC on 1GB

October / November

Last cell patches applied on all 3 ODMs

Final test cycles

Go-Live

For internal use only

Summary

More than one year experience with the software stack on Oracle Exadata

processing data on a daily, weekly and monthly data

Performance, cost and storage objectives have been met

No Hardware failures detected so far, important patches applied

Exadata v2.8 configuration is to be rated above commodity level (using SAS

disk only)

Two powerful database nodes proves higher performance & stability vs. a

smaller node

Since September no critical Service open with Oracle

page 18

For internal use only

Effizienteres Kreditrisikoreporting dank

optimierter Data Warehouse Infrastruktur

FAZIT

Der Einsatz der Oracle Exadata

Database Machine für das Data

Warehouse für das Kreditrisiko-

reporting steht für 50% weniger

Laufzeit sowie 75% geringeres

Datenvolumen – und das bei rund 20%

niedrigeren Kosten.

DAS UNTERNEHMEN

• Die Deutsche Bank ist eine führende globale Investmentbank mit einem bedeutenden Privatkundengeschäft sowie sich gegenseitig verstärkenden Geschäftsfeldern.

• Branche: Finanzdienstleistungen• Mitarbeiter: > 100.000

DIE HERAUSFORDERUNG

• Die Analyse von Kreditrisiken und zeitnahes Reporting gewinnt immer größere Bedeutung.

• Die gestiegenen Datenvolumina sowie die umfangreichen Berechnungen stellen eine Herausforderung für das zeitnahe Reporting dar. Dem zu begegnen erfordert den Aufbau einer zukunftsorienterten, performanteren Infrastruktur.

• Mehr als 500 Benutzer greifen aktiv auf die verschiedensten Aspekte im DWH zu. Tausende Abnehmer werden weltweit mit Informationen in unterschiedlichen Formaten versorgt.

ORACLE PRODUKTE & SERVICES

• Oracle Exadata Database Machine• Oracle Linux• Oracle Customer Support

DIE LÖSUNG

• Quartalsweise, monatliche, wöchentliche

bzw. tägliche Bereitstellung der Berichte

mit massiv verbesserter Performance

• Laufzeit zur Generierung der täglichen

Reports um 50% verkürzt

• Dank der Storage-Kompression wurde

das Datenvolumen um 75% reduziert

• Kosteneinsparungen von etwa 20%,

reduzierte Platzanforderungen und

weniger Stromverbrauch

March 2012