Hadoop Distributed File System
Hadoop Distributed File System (HDFS) is designed to reliably store very large files across machines in a large cluster. It is inspired by the GoogleFileSystem.
- DFS_requirements. Summarizes the requirements Hadoop DFS should be targeted for, and outlines further development steps towards achieving this requirements.
- FAQ (look for the questions starting with HDFS.)