Hadoop Distributed File System
Hadoop Distributed File System (HDFS) is designed to reliably store very large files across machines in a large cluster. It is inspired by the GoogleFileSystem.
DFS_requirements. Summarizes the requirements Hadoop DFS should be targeted for, and outlines further development steps towards achieving this requirements.
FAQ (look for the questions starting with HDFS.)