+ All Categories
Home > Documents > Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa,...

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa,...

Date post: 31-Mar-2015
Category:
Upload: paloma-fenno
View: 212 times
Download: 0 times
Share this document with a friend
Popular Tags:
12
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers
Transcript
Page 1: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

Forschungszentrum Karlsruhein der Helmholtz-Gemeinschaft

Jos van WezelDoris Ressmann

GridKa, Karlsruhe

TSMas tape storage backend

fordisk pool managers

Page 2: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

tape backends for dCache

• OSM (DESY)• Enstore (FNAL)• HPSS (BNL, CC-IN2P3) • DMF© (SARA/NIKHEF) • TSM (FZK/GridKa)

Page 3: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

Why TSM

• already in use at FZK• takes the burden out of tape handling

– tape/drive replacement– generation migration– reporting, monitoring– etc.

• separates administrator roles– storage management– tape management

• runs on Linux and i386 HW• clients (pool nodes) can directly talk to tape via Storage Agents• documented proven platform• wide spread use and acceptable price• application programmers interface

Page 4: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

dcache as tape front end

• Fresh data is collected per storage class• Each storage class queue has parameters to direct the flush-to-

tape operation– max time between flushes-to-tape– max number of bytes not written to tape– max number of files not written to tape– max number of concurrent writes to tape

• At flush a user defined backend is called– tsmcp– tss

Page 5: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

Disk pool

Oracle FTS database

To Worker nodes

library8 LTO2 drives4 LTO3 drives

2006 projectedlibrary

15 LTO3 drives

10 Gb DFN

GridFTP pool

dCache R/W poolswith TSM Agents

Cisco Catalyst 6059

TSM server

dCache SRM

dCache head node

1/23/2006

FZK/GridKa

26/08/2005

Ethernet2 Gb FC2 Gb FC

2 x 2 Gb FC

FTS server

dCache R pools

dCache environment

Page 6: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

dCache pools

Consolidated NAS box• No SAN fabric, still flexible• Enables other interconnects

• Iban, 10 GE• Easier maintenance• Lower costs• Can function as GPFS IO node

TSM storage agent

dCache

TSM session server

Page 7: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

tsmcp

• uses the TSM API

• starts and closes a session for each store to or retrieve from tape. (could be handled in a script that calls the TSM cli)

• Problem with this approach– session startup time takes inordinate amount of time

– On stores: TSM volume selection algorithm starts cartridge juggle. Efficiency nears zero.

– On retrieves: no control over tape file order

Page 8: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

TSM Session Server properties

• Interfaces directly with TSM via its API– the API libs come with the TSM software

• Single executable, documentation ‘tss –help’• Fan out for all dpm to tape activities

– single session to the TSM server– multiple tape flush/retrieve/rename/log/queries

• Runs on the TSM clients, storage agent or on the server proper• Almost plug-in replacement for the TSM backend that comes with dCache• Sends different type of data to different tape sets

– if known from dcache ‘tag’– groups data that are likely to be retrieved together

• Queues multiple requests (no state is kept, dpm must re-queue if needed)• Work in progress (in cooperation with dcache developers)• Allows to store an exact image of the global name space on tape

– store the ‘site file name’– decoupling of disk pool manager and tape backend– needs ‘rename’ support of the dpm

Page 9: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

Tape Library

TSM Client

EthernetFibre Channel

SAN (tape only)

Ethernet

DataDataMeta-Data

TSMStorage Agent

TSM server&

library manager

Data flow

TTSSSTSS Storage Agent > TSS > dCache

Page 10: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

TSM and TSS in use

• TSM is a viable tape handling system for GridKa• Promising TSS tests results

– up to 150 TSM unary database ops/s– no cartridge juggling– keeps a drive streaming (SAIT at 27 MB/s)

• deployment for SC4 tape challenge– 8 LTO3 drives, (8 LTO1 drives)– 10 dCache write pools/nodes– combined target for this 06/06:

• to disk: 300 - 500 MB/s• to tape: 100 - 150 MB/s

• no known bottleneck in sight– clearly the meta data handling at the server does not scale indef.

• clear cut between online and offline storage operations

Page 11: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

Future enhancements

Reading• Sort retrieve order on tape file sequence

– needs support of the storage manager– announced for dCache

Writing• Improve throughput (LTO3/LTO4)

– decoupling reads and writes– Include sizing estimates on write– throttle or stop writes based on node IO load

Support for xrootd– can use the same interface

10 Gb networking– may use the Ethernet again for tape operations– TSS to TSS communication needed

Page 12: Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.

Jos van Wezel April 3,2006


Recommended