Vox files_openstack_swift_voxel_net
-
Upload
dellcloudedge -
Category
Technology
-
view
314 -
download
2
Transcript of Vox files_openstack_swift_voxel_net
About Voxel
• A profitable company comprised of people who really love this stuff (even if you don't) and that's why we're good at it.
• Founded in 1999; first to commercially support load balancing for Linux (RAILs) in 2000.
• Own and operate an International network connecting nearly 20 Voxel Network POPs (AS29791).
• Over 1,000 clients rely on Voxel: Fortune 5000, Web 2.0 startups, Carriers, Media & Entertainment, Advertising Networks, etc.
• Our core competency is providing a 100% uptime SLA on your entire stack: ProManaged Hosting Infrastructure, Voxel IP Network, VoxCLOUD Hybrid Cloud and VoxCAST Content Delivery.
VoxFiles - why?
Compete in the growing Object Storage marketplace Great position in the Managed Hosting worldOther storage systems not meeting needs of customers
o others are unreliable, expensive, proprietary Complements Voxel.Net's rich offerings:
o ProManaged Hosting Infrastructureo Voxel IP Networko VoxCLOUD Hybrid Cloud o VoxCAST Content Delivery Network
OpenStack - why?
Strong, active, responsive developer community
Open Source
All One Language - Python
Big Vision• many products• shared components
Great vendor involvement! Yay, DELL!
What's Great About VoxFiles?
Multi-Datacenter Some Storage Zones Cross Data-Centers
Fully Automated Expansion of Storage (demo in a few minutes)
We wont run out: Big Storage Purchase pre-Thailand Flooding
Multi-Datacenter, you say?
Yes.
All datacenters on different local power, generator backupMultiple independent routes, above 500 year flood plains
VoxFiles Zones
Swift does not yet have the notion of tiered zones.
So I made my own.
3/4 of zones are only within a single datacenter
1/4 of zones span datacenters
Rationale:
Zones that cross datacenters can be more quickly rebuilt when there's connectivity loss. BEWARE SPLIT BRAIN!
VoxFiles - Production
Multi-Petabyte ready storage nodes of commodity hardware. No special hardware. Just shove drives into standard 2Us and put SSDs and lots of RAM in the proxy/account/container servers. We don't want to be an experimental HW shop - we're growing too fast. Keep ready and waiting storage offline (save power/cooling).
Bring storage online (for now) in ~30TB increments as usage exceeds 50%.
Rebuild rings and apply weighting slowly.
VoxFiles - Environments & Automation
We started with an Alpha environment• VMs as Proxy Servers and 6 Storage Nodes• Alpha environment became my Dev system
Hacked Andy's Chef recipes to fit our custom needs• Setup Proxy, Storage, Account, Ring-Builder, Munin• Drives, Networks, OS users • Authentication, Dispersion Testing, Kernel Settings, Logging
I setup a Production Environment with Chef & RunDeck• RunDeck only in Prod• Orchestrates ring rebuilding and node addition/removal
VoxFiles - Auto-Growth in Prod (now)We are still growing our existing zonesMunin alerts Ops when a Zone > 50% Ops adds a storage node via RunDeck: <enter Zone #>RunDeck finds a storage node <node_id> in Ubersmith(tm) inventory and "VoxServers hybrid cloud" powers it on. RunDeck does a "knife voxel server create node_id=<node_id> hostname=storage07.blah.com" and the post-install runs chef-client. Chef sets up the node for storage. RunDeck runs chef-client on ring-builder with low and increasing weights for new storage devices. RunDeck watches disk IO load on each box and if low for 5 minutes, applies the new ring with higher weights.
VoxFiles - AutoGrowth in Prod (later)
Once our zones are big-enoughtm we'll start adding capacity by adding zones.
Adding a zone has similar challenges as adding storage to zones: preventing replication storms.
In the "adding zone" case, we manage weights on the whole zone. Otherwise, it's weights per device.
VoxFiles - Basic Project Plan
alpha 11/11 beta 12/11 production 1/12