WARNING: This is the _old_ Lustre wiki, and it is in the process of being retired. The information found here is all likely to be out of date. Please search the new wiki for more up to date information.

Lustre Publications: Difference between revisions

From Obsolete Lustre Wiki
Jump to navigationJump to search
Line 57: Line 57:
|-
|-
||[http://www.sun.com/software/products/lustre/docs/Peta-Scale_wp.pdf '''Peta-Scale I/O with the Lustre File System''']||Describes low-level infrastructure in the Lustre file system that addresses scalability in very large clusters. Covered topics include scalable I/O, locking policies and algorithms to cope with scale, implications for recovery, and other scalability issues. / ORNL ||February 2008
||[http://www.sun.com/software/products/lustre/docs/Peta-Scale_wp.pdf '''Peta-Scale I/O with the Lustre File System''']||Describes low-level infrastructure in the Lustre file system that addresses scalability in very large clusters. Covered topics include scalable I/O, locking policies and algorithms to cope with scale, implications for recovery, and other scalability issues. / ORNL ||February 2008
|-
||[http://wiki.lustre.org/images/4/49/WP_BestPractices_Lustre_DDN_032108.pdf '''Best Practices for Architecting a Lustre-based Storage Environment''']|| A series of best practices to consider when deploying a highly-reliable, high-performance Lustre environment. Covered topics include storage infrastructure failover, maximizing computational capability by minimizing I/O overhead, ensuring predictable striped file performance, and protecting large, persistent data stores. / DataDirect Networks||2008
|-
|-
||[http://wiki.lustre.org/images/2/20/Lustre_wan_tg07.pdf '''Wide Area Filesystem Performance using Lustre on the TeraGrid''']|| TeraGrid 2007 conference / Indiana University|| June 2007
||[http://wiki.lustre.org/images/2/20/Lustre_wan_tg07.pdf '''Wide Area Filesystem Performance using Lustre on the TeraGrid''']|| TeraGrid 2007 conference / Indiana University|| June 2007

Revision as of 13:06, 23 September 2009

Informative sources regarding Lustre technology and its application.

You'll find user documentation along with videos and podcasts describing Lustre at a high-level, engineering presentations from Lustre engineers, and detailed descriptions of applications in blue prints and white papers produced by research organizations.


Lustre Documentation

Lustre Documentation
Title Description/Source Date
Lustre Operations Manual The Lustre Operations Manual provides users with installation, configuration, tuning, monitoring and troubleshooting information. 2009


Videos & Podcasts

Videos & Podcasts from Sun
Title Description/Source Date
RCE 14: Lustre Cluster File System Research Computing and Engineering interview with Andreas Dilger, one of the principal file system architects for the Lustre file system. 2009
Linux HPC Software Stack Sun Systems Engineer Larry McIntosh provides an overview of Sun HPC software reference stack for Lustre. 2008
Lustre Overview by Peter Bojanic Learn about the Lustre parallel file system, the newest addition to the Sun HPC portfolio, which is designed to meet the demands of the world's largest high performance clusters. December 7, 2007
Sun Storage Cluster Find out how Sun simplifies the deployment of Lustre-based storage. January 9, 2009
Radio HPC - Episode 11 Tony Warner chats with Voltaire's Brian Forbes about their companies' partnership in the InfiniBand space, and Peter Bojanic clues us in on what's new with Sun's Lustre file system. February 3, 2009

White Papers

White Papers
Title Description/Source Date
Lustre File System Networking: High-Performance Features and Flexible Support for a Wide Array of Networks Information about Lustre networking that can be used to plan cluster file system deployments for optimal performance and scalability. Covered topics include Lustre message passing, Lustre Network Drivers, and routing in Lustre networks, and the paper describes how these features can be used to improve cluster storage management. November 2008
Lustre File System: High-Performance Storage Architecture and Scalable Cluster File System Basic information about the Lustre file system. Covered topics include general characteristics and markets in which Lustre has a strong presence, a typical Lustre file system configuration, an overview of Lustre networking (LNET), an introduction of Lustre capabilities that support high availability and rolling upgrades, discussion of file storage in a Lustre file system, additional features, and information about a how a Lustre file system compares to other shared file systems. October 2008
Pathways to Open Petascale Computing Derived from Sun’s innovative design approach and experience with very large , supercomputing deployments, the Sun Constellation System provides the world's first , open petascale computing environment — one built entirely with open and standard , hardware and software technologies. Cluster architects can use the Sun Constellation , System to design and rapidly deploy tightly-integrated, efficient, and cost-effective , supercomputing grids and clusters that scale predictably from a few teraflops to over a , petaflop. With a totally modular approach, processors, memory, interconnect fabric, and storage can all be scaled independently depending on individual needs. June 2008
Peta-Scale I/O with the Lustre File System Describes low-level infrastructure in the Lustre file system that addresses scalability in very large clusters. Covered topics include scalable I/O, locking policies and algorithms to cope with scale, implications for recovery, and other scalability issues. / ORNL February 2008
Wide Area Filesystem Performance using Lustre on the TeraGrid TeraGrid 2007 conference / Indiana University June 2007
Exploiting Lustre File Joining for Effective Collective IO Proceedings of the CCGrid'07 / ORNL May 2007
Tokyo Tech Tsubame Grid Storage Implementation By Syuuichi Ihara May 2007
Optimizing Storage and I/O For Distributed Processing On Enterprise & High Performance Compute(HPC)Systems For Mask Data Preparation Software (CATS) Glenn Newell, Sr.IT Solutions Mgr, Naji Bekhazi, Director of R&D,Mask Data Prep (CATS), Ray Morgan,Sr.Product Marketing Manager,Mask Data Prep(CATS)/ Synopsys 2007
Wide Area Filesystem Performance Using Lustre on the TeraGrid Teragrid 2007 Conference, Madison,WI / TeraGrid 2007
Guidelines for Efficient Parallel I/O on the Cray XT3/XT4 (PDF) Jeff Larkin, Mark Fahey, proceedings of CUG 2007 / Cray User Group 2007
XT7? Integrating and Operating a Conjoined XT3+XT4 System Presented by ORNL on CUG 2007 (Presentation also available.) / Cray User Group 2007
A Center-Wide File System using Lustre Shane Canon, Sarp Oral, proceedings of CUG 2006 \ Cray User Group 2006
Benefits of High Speed Interconnects to Cluster File Systems: A Case Study with Lustre Parallel and Distributed Processing Symposium, 2006. IPDPS 2006. Lustre performance comparison when using InfiniBand and Quadrics interconnects. You can also download the paper at the OSU site. / Ohio State University 2006
Coordinating Parallel Hierarchical Storage Management in Object-base Cluster File Systems MSST2006, Conference on Mass Storage Systems and Technologies (May 2006) / University of Minnesota 2006
Experiences with HP SFS/Lustre at SSCK SGPFS 5 in Stuttgart / Karlsruhe Lustre Talks 4.4.2006
Filesystems on SSCK's HP XC6000 Einführungsveranstaltung im Rechenzentrum / Karlsruhe Lustre Talks 2005
Experiences & Performance of SFS/Lustre Cluster File System in Production HP-CAST 4 in Krakau / Karlsruhe Lustre Talks 10.5.2005
ISC 2005 in Heidelberg Karlsruhe Lustre Talks 24.6.2005
Experiences with 10 Months HP SFS/Lustre in HPC Production HP-CAST 5 in Seattle / Karlsruhe Lustre Talks 11.11.2005
Performance Monitoring in a HP SFS Environment HP-CCN in Seattle / Karlsruhe Lustre Talks 12.11.2005
Selecting a cluster file system CFS Nov 2005
Shared Parallel Filesystem in Heterogeneous Linux Multi-Cluster Environment Proceedings of the 6th LCI International Conference on Linux Clusters: The HPC Revolution (2005). The management issues mentioned in the last part of this paper have been addressed. The same paper is also available at the CU site. / University of Colorado, Boulder 2005
Lustre: Building a cluster file system for 1,000 node clusters A technical presentation about our successes and mistakes during 2002-2003 / CFS Summer 2003

BluePrints

BluePrints
Title Description/Source Date
Lustre File System - Demo Quick Start Guide A simple cookbook for non-Linux experts on how to set up a Linux-based Lustre file system using small servers, workstations, PCs, or other available hardware for demonstration purposes. 2009
Implementing the Lustre File System with Sun Storage Describes an implementation of the Sun Lustre file system as a scalable storage cluster using Sun Fire servers, and highspeed/lowlatency InfiniBand interconnects. 2009
Tokyo Tech Tsubame Grid Storage Implementation This Sun BluePrints™ article describes the storage architecture of the Tokyo Tech , TSUBAME grid, as well as the steps for installing and configuring the Lustre file system , within the storage architecture. 2009
Sun Storage and Archive Solution for HPC: Sun BluePrints Reference Architecture To help customers address an almost bewildering set of architectural challenges, Sun has developed the Sun Storage and Archive Solution for HPC, a reference architecture that , can be easily customized to meet specific application goals and business requirements. May 2008


Presentations

Presentations
Title Description/Source Date
Lustre cluster in production at GSI Presented by SCHöN, Walter / HEPiX Talks May 2008
Final Report from File Systems Working Group Presented by MASLENNIKOV, Andrei / HEPiX Talks May 2008
Setting up a simple Lustre Filesystem Presented by Stephan Wiesand. Slides on HPPix site: Storage Evaluations at BNL / HEPiX Talks May 2008
Storage Evaluations at BNL Presented by Robert Petkus - BNL. Slides on HEPix site: Lustre Experience at CEA/DIF / HEPiX Talks May 2008
Lustre Experience at CEA/DIF Presented by J-Ch Lafoucriere / HEPiX Talks April 23-27, 2008
XT7? Integrating and Operating a Conjoined XT3+XT4 System This paper describes the processes and tools used to move production work from the pre-existing XT3 to the new system incorporating that same XT3, including novel application of Lustre routing capabilities. / Cray User Group (Paper also available) 2007
Using IOR to Analyze the I/O Performance Presented by Hongzhang Shan,John Shalf (NERSC) on CUG 2007 / Cray User Group 2007
Lustre state and production installations Presentation on gelato.org meeting / CFS May 2004
Lustre File System Presentation on the state of Lustre in mid-2003 and the path towards Lustre1.0 / CFS Summer 2003
Lustre: Building a cluster file system for 1,000 node clusters A technical presentation about our successes and mistakes during 2002-2003 / CFS Summer 2003

Presentations from Lustre Engineers - December 2008 and March 2008

Once a year, the Lustre Engineering team gathers to discuss new features under development and testing efforts. This week-long event is known as the Lustre all-hands meeting. The Development and QE presentations made at these all-hands meetings are available here.

Also available at Lustre All-Hands Meeting 12/08 and Lustre All-Hands Meeting_3/08 .


December 2008: Lustre Launch in Beijing
Title Description Date
Lustre Windows Porting Matt Wu December 2008
VBR-TOI Mikhail Pershin December 2008
Simplified Interoperability Recovery Huang Hua December 2008
Recovery Overview Robert Read December 2008
Quotas-TOI Yong Fan December 2008
Quality Initiative Talk Robert Read December 2008
pNFS Speaker Notes Oleg Drokin December 2008
pNFS Oleg Drokin December 2008
OST Pools Nathan Rutman December 2008
OST Migration RAID1 SNS Andreas Dilger December 2008
ORNL Experience Speaker Notes Oleg Drokin December 2008
ORNL Experience Oleg Drokin December 2008
Lustre NRS Simulation Yingjin Qian, Wang Di December 2008
Lustre Interoperability 1.8 Huang Hua December 2008
Lustre All-Hands Meeting 12/2008 - Keynote Peter Bojanic December 2008
Lustre on DMU Alex Zhuravlev December 2008
LLOG Overview Alex Zhuravlev December 2008
HDFS Wang Di December 2008
GIT Overview Robert Read December 2008
COS-TOI COS-TOI Alexander Zarochentsev December 2008
CLIO-TOI Nikita Danilov December 2008
CLIO Nikita Danilov December 2008
Read-Only Cache - TOI Alex Zhuravlev December 2008


March 2008: Lustre Launch in Beijing
Title Description Date
Development
RMG Processes Andreas Dilger March 17, 2008
Lustre Development Strategy Eric Barton March 17, 2008
CMD Yury Umanets March 17, 2008
HEAD MDS Nikita Danilov March 17, 2008
User Space Servers Alex Zhuravlev March 17, 2008
DMU Ricardo Correia March 17, 2008
Recovery Mike Pershin March 17, 2008
DLM Oleg Drokin March 17, 2008
LNET Issac Huang March 17, 2008
Quality Engineering
Day 1Day 2 JD March 17, 2008
Lustre Release & weekly testing Jian Yu March 17, 2008
Build System Yibin Wang March 17, 2008
HEAD testing Zheng Chen March 17, 2008
b1.6 testing Peng Ye March 17, 2008
Performance testing Jack Chen March 17, 2008
Automation Minh Diep March 17, 2008
Acc-small Elena Gryaznova March 17, 2008