+ All Categories
Home > Documents > I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April 2014

I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April 2014

Date post: 23-Feb-2016
Category:
Upload: miracle
View: 32 times
Download: 0 times
Share this document with a friend
Description:
Challenges of deploying Wide-Area-Network Distributed Storage System under network and reliability constraints – A case study. I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April 2014. Mohd Bazli Ab Karim Advanced Computing Lab MIMOS Berhad , Malaysia. - PowerPoint PPT Presentation
Popular Tags:
16
Challenges of deploying Wide-Area-Network Distributed Storage System under network and reliability constraints – A case study email [email protected] Mohd Bazli Ab Karim Advanced Computing Lab MIMOS Berhad, Malaysia In PRAGMA Student Workshop PRAGMA26, Tainan, Taiwan 9-11 April 2014
Transcript
Page 1: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

Challenges of deploying Wide-Area-Network Distributed Storage System under network and reliability constraints – A case study

[email protected]

Mohd Bazli Ab KarimAdvanced Computing LabMIMOS Berhad, Malaysia

In PRAGMA Student WorkshopPRAGMA26, Tainan, Taiwan9-11 April 2014

Page 2: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

2

Outline

• Distributed Storage System?• PRAGMA25

– DFS over Local Area Network– Ceph vs. GlusterFS

• PRAGMA26– DFS over Wide Area Network– DFS over WAN vs. DFS over LAN

Page 3: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

3

Data Center

SAN/NAS

Disaster Recovery Site(s)

SAN/NAS

Distributed File System

Data Center 1

Local/DAS

Data Center 2

Local/DAS

Data Center n

Local/DAS. . .

One/Multiple Virtual Volume(s)

R/ W

R e p l i c a ti o n s

D a t a S t r i p i n g a n d P a r a l l e l R / W

Page 4: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

4

PRAGMA 25 – DFS on LAN

Dell PowerEdge T110 IIProc: Intel Xeon E3-1220v2 3.10 GHzMemory: 8 GBHard Drives: Seagate Constellation ES 2TB 7200RPM SATARAID Controller: LSI Logic SAS2008Network: 1GbE Operating System: Ubuntu 12.04Ceph: 0.61.7 (Cuttlefish)GlusterFS: 3.4.0

Brick 1/OSD 1

Brick 2/OSD 2

Brick 3/OSD 3

Brick 4/OSD 4

CLIENT Ceph MDSCeph MON

Experiment Network Setup

Experiment Hardware Specification

Page 5: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

5

PRAGMA 25 – Ceph vs GlusterFS

4 K 16 K 64 K 256 K 4 K 16 K 64 K 256 K 4 K 16 K 64 K 256 K 4 K 16 K 64 K 256 K5 MB 50 MB 500 MB 5000 MB

0

5000

10000

15000

20000

Ceph/GlusterFS Sequential Write Profile

Ceph Fuse Ceph Kernel Gluster Native

Kilo

byte

s/se

cond

s

4 K 16 K 64 K 256 K 4 K 16 K 64 K 256 K 4 K 16 K 64 K 256 K 4 K 16 K 64 K 256 K5 MB 50 MB 500 MB 5000 MB

1000

10000

100000

1000000

10000000

Ceph/GlusterFS Sequential Read Profile

Ceph Fuse Ceph Kernel Gluster Native

Kilo

byte

s/se

cond

s

Page 6: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

6

PRAGMA 26 – DFS over WAN

MIMOS Headquarters in Kuala Lumpur and its branch office in Kulim. From Google Map

350km

Page 7: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

7

OSD OSD

OSDOSD

MON/MDSCLIENT

OSD OSD

OSDOSD

MON/MDSCLIENT

OSD OSD

OSDOSD

MON/MDSCLIENT

WAN

Datacenter 1 in HQ officeSGI ALTIX XE 3102 x Dual Core Intel Xeon CPU X5355 2.66 GHz16 Gb RAM3.5 SATA Drive 250GB

Datacenter 2 in HQ officeDELL2 x Dual Core Intel Xeon CPU 5130 2.00 GHz12 Gb RAM3.5 SAS Drive 73GB

Datacenter 1 in Kulim officeSGI ALTIX XE 3102 x Dual Core Intel Xeon CPU X5355 2.66 GHz16 Gb RAM3.5 SATA Drive 250GB

PRAGMA 26 – DFS over WAN (Setup)

The storage pool was set with 3 replica counts with minimum number of replica counts required is 2.

Page 8: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

8

PRAGMA 26 - DFS over WAN (Networking)

Round-trip time in ms

Bandwidth (Mbps)

2 TCP Iperf

Min Avg Max Mdev

DC1 KL to DC1 Kulim

250 96% 13.149 13.491 16.167 0.684

DC2 KL to DC1 Kulim

250 96% 13.176 14.004 17.665 1.079

DC1 KL to DC2 KL

1000 86% 0.422 0.490 1.203 0.136

Page 9: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

9

root@poc-tpm1-mon1:~/ceph-deploy# ceph osd tree# id weight type name up/down reweight-1 2.12 root default-2 0.23 host poc-tpm1-osd10 0.23 osd.0 up 1-3 0.23 host poc-tpm1-osd21 0.23 osd.1 up 1-4 0.23 host poc-tpm1-osd32 0.23 osd.2 up 1-5 0.23 host poc-tpm1-osd43 0.23 osd.3 up 1-6 0.06999 host poc-tpm2-osd14 0.06999 osd.4 up 1-7 0.06999 host poc-tpm2-osd25 0.06999 osd.5 up 1-8 0.06999 host poc-tpm2-osd36 0.06999 osd.6 up 1-9 0.06999 host poc-tpm2-osd47 0.06999 osd.7 up 1-10 0.23 host poc-khtp-osd18 0.23 osd.8 up 1-11 0.23 host poc-khtp-osd29 0.23 osd.9 up 1-12 0.23 host poc-khtp-osd310 0.23 osd.10 up 1-13 0.23 host poc-khtp-osd411 0.23 osd.11 up 1

CRUSH Map - default

Page 10: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

10

# rulesrule data { ruleset 0 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit}rule metadata { ruleset 1 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit}

CRUSH Map Rules - default

Pick one leaf nodeof type host

Page 11: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

11

root@poc-tpm1-mon1:~/ceph-deploy# ceph osd tree# id weight type name up/down reweight-1 2.12 root default-23 0.92 datacenter tpm1-2 0.23 host poc-tpm1-osd10 0.23 osd.0 up 1-3 0.23 host poc-tpm1-osd21 0.23 osd.1 up 1-4 0.23 host poc-tpm1-osd32 0.23 osd.2 up 1-5 0.23 host poc-tpm1-osd43 0.23 osd.3 up 1-24 0.28 datacenter tpm2-6 0.06999 host poc-tpm2-osd14 0.06999 osd.4 up 1-7 0.06999 host poc-tpm2-osd25 0.06999 osd.5 up 1-8 0.06999 host poc-tpm2-osd36 0.06999 osd.6 up 1-9 0.06999 host poc-tpm2-osd47 0.06999 osd.7 up 1-25 0.92 datacenter khtp1-10 0.23 host poc-khtp-osd18 0.23 osd.8 up 1-11 0.23 host poc-khtp-osd29 0.23 osd.9 up 1-12 0.23 host poc-khtp-osd310 0.23 osd.10 up 1-13 0.23 host poc-khtp-osd411 0.23 osd.11 up 1

CRUSH Map - New

Page 12: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

12

# rulesrule data { ruleset 0 type replicated min_size 2 max_size 10 step take default step chooseleaf firstn 0 type datacenter step emit}rule metadata { ruleset 1 type replicated min_size 2 max_size 10 step take default step chooseleaf firstn 0 type datacenter step emit}

CRUSH Map Rules – New

Pick one leaf nodeof type datacenter

Page 13: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

13

DFS on WAN vs. DFS on LAN4

K8

K16

K32

K64

K12

8 K

256

K4

K8

K16

K32

K64

K12

8 K

256

K4

K8

K16

K32

K64

K12

8 K

256

K4

K8

K16

K32

K64

K12

8 K

256

K4

K8

K16

K32

K64

K12

8 K

256

K4

K8

K16

K32

K64

K12

8 K

256

K

1 MB 10 MB 100 MB 1000 MB 2000 MB 4000 MB

0

1000

2000

3000

4000

5000

6000

Ceph Sequential Write Profile

LAN WAN

Kilo

byte

s/se

cond

s

4 K

8 K

16 K

32 K

64 K

128

K25

6 K

4 K

8 K

16 K

32 K

64 K

128

K25

6 K

4 K

8 K

16 K

32 K

64 K

128

K25

6 K

4 K

8 K

16 K

32 K

64 K

128

K25

6 K

4 K

8 K

16 K

32 K

64 K

128

K25

6 K

4 K

8 K

16 K

32 K

64 K

128

K25

6 K

1 MB 10 MB 100 MB 1000 MB 2000 MB 4000 MB

01000000200000030000004000000500000060000007000000

Ceph Sequential Read Profile

LAN WAN

Kilo

byte

s/se

cond

s

Page 14: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

14

Hypothesis: Write performance is slower than read performance. This is due to

write operation requires a creation of new file and also to store overhead information known as metadata, which typically consists of directory information, and space allocation.

DFS IO performs better in LAN compared to WAN due to limited capacity of WAN bandwidth and its latency, jitter etc.

Results: DFS in LAN provides better overall I/O rates compared to DFS in

WAN due to its better network connectivity and bandwidth size. DFS in WAN scores better in writing 64K and 128K block sizes

compared to DFS in LAN.Analysis: DFS in WAN performances in I/O is still acceptable e.g. smaller files

size with 16K, 32K, 64K, 128K block sizes, where DFS in LAN only performs slightly better than in WAN.

Page 15: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

15

Summary

Distributed file system in wide area network works at acceptable I/O rates and it is ideal for usage of smaller file sizes.

Investigating distributed file system in wide area network, focusing on features like: support cloud deployment architecture, ability to provide parallel read and write

operations on a distributed file system with different geographical locations.

Page 16: I n PRAGMA Student Workshop PRAGMA26, Tainan , Taiwan 9-11 April  2014

16


Recommended