Tag Archives: 35 GBps read throughput.

[repost]Product: GlusterFS

original:

Product: GlusterFS

Adapted from their website:
GlusterFS is a clustered file-system capable of scaling to several peta-bytes. It aggregates various storage bricks over Infiniband RDMA or TCP/IP interconnect into one large parallel network file system. Storage bricks can be made of any commodity hardware such as x86-64 server with SATA-II RAID and Infiniband HBA).

Cluster file systems are still not mature for enterprise market. They are too complex to deploy and maintain though they are extremely scalable and cheap. Can be entirely built out of commodity OS and hardware. GlusterFS hopes to solves this problem.

GlusterFS achieved 35 GBps read throughput. The GlusterFS Aggregated I/O Benchmark was performed on 64 bricks clustered storage system over 10 Gbps Infiniband interconnect. A cluster of 220 clients pounded the storage system with multiple dd (disk-dump) instances, each reading / writing a 1 GB file with 1MB block size. GlusterFS was configured with unify translator and round-robin scheduler.

The advantages of GlusterFS are:


* Designed for O(1) scalability and feature rich.
* Aggregates on top of existing filesystems. User can recover the files and folders even without GlusterFS.
* GlusterFS has no single point of failure. Completely distributed. No centralized meta-data server like Lustre.
* Extensible scheduling interface with modules loaded based on user’s storage I/O access pattern.
* Modular and extensible through powerful translator mechanism.
* Supports Infiniband RDMA and TCP/IP.
* Entirely implemented in user-space. Easy to port, debug and maintain.
* Scales on demand.

Related Articles

  • Technical Presentation on GlusterFS
  • Open Fest 5th Annual Conference
  • Zresearch
  • GlusterFS FAQ