Tahoe-LAFS
==========
-Tahoe-LAFS is a Free Software/Open Source decentralized data store. It
-distributes your filesystem across multiple servers, and even if some of the
-servers fail or are taken over by an attacker, the entire filesystem continues
-to work correctly and to preserve your privacy and security.
+Tahoe-LAFS is a Free and Open decentralized cloud storage system. It distributes your
+data across multiple servers. Even if some of the servers fail or are taken over by an
+attacker, the entire file store continues to function correctly, preserving your privacy
+and security.
To get started please see `quickstart.rst`_ in the docs directory.
A "storage grid" is made up of a number of storage servers. A storage server
has direct attached storage (typically one or more hard disks). A "gateway"
communicates with storage nodes, and uses them to provide access to the
-filesystem over protocols such as HTTP(S), SFTP or FTP.
+grid over protocols such as HTTP(S), SFTP or FTP.
Note that you can find "client" used to refer to gateway nodes (which act as
a client to storage servers), and also to processes or programs connecting to
the user connects to it over HTTPS or SFTP. This means that the operator of
the gateway can view and modify the user's data (the user *relies on* the
gateway for confidentiality and integrity), but the advantage is that the
-user can access the filesystem with a client that doesn't have the gateway
-software installed, such as an Internet kiosk or cell phone.
+user can access the Tahoe-LAFS grid with a client that doesn't have the
+gateway software installed, such as an Internet kiosk or cell phone.
Access Control
==============
4. `Capabilities`_
5. `Server Selection`_
6. `Swarming Download, Trickling Upload`_
-7. `The Filesystem Layer`_
+7. `The File Store Layer`_
8. `Leases, Refreshing, Garbage Collection`_
9. `File Repairer`_
10. `Security`_
(See the `docs/specifications directory`_ for more details.)
-There are three layers: the key-value store, the filesystem, and the
+There are three layers: the key-value store, the file store, and the
application.
The lowest layer is the key-value store. The keys are "capabilities" -- short
due to the limitation of network bandwidth). In practice, values as small as
a few bytes and as large as tens of gigabytes are in common use.
-The middle layer is the decentralized filesystem: a directed graph in which
+The middle layer is the decentralized file store: a directed graph in which
the intermediate nodes are directories and the leaf nodes are files. The leaf
nodes contain only the data -- they contain no metadata other than the length
in bytes. The edges leading to leaf nodes have metadata attached to them
about the file they point to. Therefore, the same file may be associated with
different metadata if it is referred to through different edges.
-The top layer consists of the applications using the filesystem.
+The top layer consists of the applications using the file store.
Allmydata.com used it for a backup service: the application periodically
-copies files from the local disk onto the decentralized filesystem. We later
+copies files from the local disk onto the decentralized file store. We later
provide read-only access to those files, allowing users to recover them.
There are several other applications built on top of the Tahoe-LAFS
-filesystem (see the RelatedProjects_ page of the wiki for a list).
+file store (see the RelatedProjects_ page of the wiki for a list).
.. _docs/specifications directory: specifications
.. _RelatedProjects: https://tahoe-lafs.org/trac/tahoe-lafs/wiki/RelatedProjects
Capabilities sit on the "global+secure" edge of `Zooko's Triangle`_. They are
self-authenticating, meaning that nobody can trick you into accepting a file
that doesn't match the capability you used to refer to that file. The
-filesystem layer (described below) adds human-meaningful names atop the
+file store layer (described below) adds human-meaningful names atop the
key-value layer.
.. _`Zooko's Triangle`: https://en.wikipedia.org/wiki/Zooko%27s_triangle
.. _helper.rst: helper.rst
-The Filesystem Layer
+The File Store Layer
====================
-The "filesystem" layer is responsible for mapping human-meaningful pathnames
+The "file store" layer is responsible for mapping human-meaningful pathnames
(directories and filenames) to pieces of data. The actual bytes inside these
-files are referenced by capability, but the filesystem layer is where the
+files are referenced by capability, but the file store layer is where the
directory names, file names, and metadata are kept.
-The filesystem layer is a graph of directories. Each directory contains a
+The file store layer is a graph of directories. Each directory contains a
table of named children. These children are either other directories or
files. All children are referenced by their capability.
Leases, Refreshing, Garbage Collection
======================================
-When a file or directory in the virtual filesystem is no longer referenced,
-the space that its shares occupied on each storage server can be freed,
-making room for other shares. Tahoe-LAFS uses a garbage collection ("GC")
-mechanism to implement this space-reclamation process. Each share has one or
-more "leases", which are managed by clients who want the file/directory to be
+When a file or directory in the file store is no longer referenced, the space
+that its shares occupied on each storage server can be freed, making room for
+other shares. Tahoe-LAFS uses a garbage collection ("GC") mechanism to
+implement this space-reclamation process. Each share has one or more
+"leases", which are managed by clients who want the file/directory to be
retained. The storage server accepts each share for a pre-defined period of
time, and is allowed to delete the share if all of the leases are cancelled
or allowed to expire.
failure: either temporary downtime (affecting availability of the file), or a
permanent data loss (affecting the preservation of the file). Hard drives
crash, power supplies explode, coffee spills, and asteroids strike. The goal
-of a robust distributed filesystem is to survive these setbacks.
+of a robust distributed file store is to survive these setbacks.
To work against this slow, continual loss of shares, a File Checker is used
to periodically count the number of shares still available for any given
Frontend Configuration
======================
-The Tahoe client process can run a variety of frontend file-access protocols.
-You will use these to create and retrieve files from the virtual filesystem.
-Configuration details for each are documented in the following
-protocol-specific guides:
+The Tahoe-LAFS client process can run a variety of frontend file access
+protocols. You will use these to create and retrieve files from the
+Tahoe-LAFS file store. Configuration details for each are documented in
+the following protocol-specific guides:
HTTP
CLI
The main "bin/tahoe" executable includes subcommands for manipulating the
- filesystem, uploading/downloading files, and creating/running Tahoe
+ file store, uploading/downloading files, and creating/running Tahoe-LAFS
nodes. See CLI.rst_ for details.
SFTP, FTP
1. `Unicode Support`_
3. `Node Management`_
-4. `Filesystem Manipulation`_
+4. `File Store Manipulation`_
1. `Starting Directories`_
2. `Command Syntax Summary`_
========
Tahoe-LAFS provides a single executable named "``tahoe``", which can be used to
-create and manage client/server nodes, manipulate the filesystem, and perform
+create and manage client/server nodes, manipulate the file store, and perform
several debugging/maintenance tasks.
This executable lives in the source tree at "``bin/tahoe``". Once you've done a
The "``tahoe``" tool provides access to three categories of commands.
* node management: create a client/server node, start/stop/restart it
-* filesystem manipulation: list files, upload, download, unlink, rename
+* file store manipulation: list files, upload, download, unlink, rename
* debugging: unpack cap-strings, examine share files
To get a list of all commands, just run "``tahoe``" with no additional
start using their changes.
-Filesystem Manipulation
+File Store Manipulation
=======================
-These commands let you exmaine a Tahoe-LAFS filesystem, providing basic
+These commands let you exmaine a Tahoe-LAFS file store, providing basic
list/upload/download/unlink/rename/mkdir functionality. They can be used as
primitives by other scripts. Most of these commands are fairly thin wrappers
around web-API calls, which are described in `<webapi.rst>`__.
-By default, all filesystem-manipulation commands look in ``~/.tahoe/`` to
+By default, all file store manipulation commands look in ``~/.tahoe/`` to
figure out which Tahoe-LAFS node they should use. When the CLI command makes
web-API calls, it will use ``~/.tahoe/node.url`` for this purpose: a running
Tahoe-LAFS node that provides a web-API port will write its URL into this
--------------------
As described in `docs/architecture.rst <../architecture.rst>`__, the
-Tahoe-LAFS distributed filesystem consists of a collection of directories
+Tahoe-LAFS distributed file store consists of a collection of directories
and files, each of which has a "read-cap" or a "write-cap" (also known as
a URI). Each directory is simply a table that maps a name to a child file
or directory, and this table is turned into a string and stored in a
``tahoe ls subdir``
- This lists a subdirectory of your filesystem.
+ This lists a subdirectory of your file store.
``tahoe webopen``
All Tahoe-LAFS client nodes can run a frontend SFTP server, allowing regular
SFTP clients (like ``/usr/bin/sftp``, the ``sshfs`` FUSE plugin, and many
-others) to access the virtual filesystem. They can also run an FTP server,
-so FTP clients (like ``/usr/bin/ftp``, ``ncftp``, and others) can too. These
+others) to access the file store. They can also run an FTP server, so FTP
+clients (like ``/usr/bin/ftp``, ``ncftp``, and others) can too. These
frontends sit at the same level as the web-API interface.
Since Tahoe-LAFS does not use user accounts or passwords, the SFTP/FTP
Basic Concepts: GET, PUT, DELETE, POST
======================================
-As described in `docs/architecture.rst`_, each file and directory in a Tahoe
-virtual filesystem is referenced by an identifier that combines the
+As described in `docs/architecture.rst`_, each file and directory in a
+Tahoe-LAFS file store is referenced by an identifier that combines the
designation of the object with the authority to do something with it (such as
read or modify the contents). This identifier is called a "read-cap" or
"write-cap", depending upon whether it enables read-only or read-write
URL) will return information about the object, such as metadata. GET
operations are required to have no side-effects.
-PUT is used to upload new objects into the filesystem, or to replace an
+PUT is used to upload new objects into the file store, or to replace an
existing link or the contents of a mutable file. DELETE is used to unlink
objects from directories. Both PUT and DELETE are required to be idempotent:
performing the same operation multiple times must have the same side-effects
these tasks. In general, everything that can be done with a PUT or DELETE can
also be done with a POST.
-Tahoe's web API is designed for two different kinds of consumer. The first is
-a program that needs to manipulate the virtual file system. Such programs are
+Tahoe-LAFS' web API is designed for two different kinds of consumer. The
+first is a program that needs to manipulate the file store. Such programs are
expected to use the RESTful interface described above. The second is a human
-using a standard web browser to work with the filesystem. This user is given
-a series of HTML pages with links to download files, and forms that use POST
-actions to upload, rename, and unlink files.
+using a standard web browser to work with the file store. This user is
+presented with a series of HTML pages with links to download files, and forms
+that use POST actions to upload, rename, and unlink files.
When an error occurs, the HTTP response code will be set to an appropriate
400-series code (like 404 Not Found for an unknown childname, or 400 Bad Request
=======================
Now that we know how to build URLs that refer to files and directories in a
-Tahoe virtual filesystem, what sorts of operations can we do with those URLs?
+Tahoe-LAFS file store, what sorts of operations can we do with those URLs?
This section contains a catalog of GET, PUT, DELETE, and POST operations that
can be performed on these URLs. This set of operations are aimed at programs
that use HTTP to communicate with a Tahoe node. A later section describes
``PUT /uri``
This uploads a file, and produces a file-cap for the contents, but does not
- attach the file into the filesystem. No directories will be modified by
+ attach the file into the file store. No directories will be modified by
this operation. The file-cap is returned as the body of the HTTP response.
This method accepts format= and mutable=true as query string arguments, and
Create a new empty directory and return its write-cap as the HTTP response
body. This does not make the newly created directory visible from the
- filesystem. The "PUT" operation is provided for backwards compatibility:
+ file store. The "PUT" operation is provided for backwards compatibility:
new code should use POST.
This supports a format= argument in the query string. The format=
a link to a child is created -- i.e. when there was not previously a link
under that name.
-Note however, that if the edge in the Tahoe filesystem points to a mutable
-file and the contents of that mutable file is changed, then the
+Note however, that if the edge in the Tahoe-LAFS file store points to a
+mutable file and the contents of that mutable file is changed, then the
'tahoe':'linkmotime' value on that edge will *not* be updated, since the
edge itself wasn't updated -- only the mutable file was.
values of the 'mtime'/'ctime' pair, and this API is used by the
"tahoe backup" command (in Tahoe v1.3.0 and later) to set the 'mtime' and
'ctime' values when backing up files from a local filesystem into the
-Tahoe filesystem. As of Tahoe v1.4.0, the set_children API cannot be used
-to set anything under the 'tahoe' key of the metadata dict -- if you
+Tahoe-LAFS file store. As of Tahoe v1.4.0, the set_children API cannot be
+used to set anything under the 'tahoe' key of the metadata dict -- if you
include 'tahoe' keys in your 'metadata' arguments then it will silently
ignore those keys.
There are several ways that the 'ctime' field could be confusing:
1. You might be confused about whether it reflects the time of the creation
- of a link in the Tahoe filesystem (by a version of Tahoe < v1.7.0) or a
- timestamp copied in by "tahoe backup" from a local filesystem.
+ of a link in the Tahoe-LAFS file store (by a version of Tahoe < v1.7.0)
+ or a timestamp copied in by "tahoe backup" from a local filesystem.
2. You might be confused about whether it is a copy of the file creation
time (if "tahoe backup" was run on a Windows system) or of the last
``PUT /uri/$DIRCAP/[SUBDIRS../]CHILDNAME?t=uri``
This attaches a child object (either a file or directory) to a specified
- location in the virtual filesystem. The child object is referenced by its
+ location in the Tahoe-LAFS file store. The child object is referenced by its
read- or write- cap, as provided in the HTTP request body. This will create
intermediate directories as necessary.
This section describes the HTTP operations that provide support for humans
running a web browser. Most of these operations use HTML forms that use POST
-to drive the Tahoe node. This section is intended for HTML authors who want
-to write web pages that contain forms and buttons which manipulate the Tahoe
-filesystem.
+to drive the Tahoe-LAFS node. This section is intended for HTML authors who
+want to write web pages containing user interfaces for manipulating the
+Tahoe-LAFS file store.
Note that for all POST operations, the arguments listed can be provided
either as URL query arguments or as form body fields. URL query arguments are
``POST /uri?t=mkdir``
- This creates a new empty directory, but does not attach it to the virtual
- filesystem.
+ This creates a new empty directory, but does not attach it to any other
+ directory in the Tahoe-LAFS file store.
If a "redirect_to_result=true" argument is provided, then the HTTP response
will cause the web browser to be redirected to a /uri/$DIRCAP page that
``POST /uri?t=upload``
This uploads a file, and produces a file-cap for the contents, but does not
- attach the file into the filesystem. No directories will be modified by
- this operation.
+ attach the file to any directory in the Tahoe-LAFS file store. That is, no
+ directories will be modified by this operation.
The file must be provided as the "file" field of an HTML encoded form body,
produced in response to an HTML form like this::
for debugging. This is a table of (path, filecap/dircap), for every object
reachable from the starting directory. The path will be slash-joined, and
the filecap/dircap will contain a link to the object in question. This page
- gives immediate access to every object in the virtual filesystem subtree.
+ gives immediate access to every object in the file store subtree.
This operation uses the same ophandle= mechanism as deep-check. The
corresponding /operations/$HANDLE page has three different forms. The
==================
The portion of the web namespace that begins with "/uri" (and "/named") is
-dedicated to giving users (both humans and programs) access to the Tahoe
-virtual filesystem. The rest of the namespace provides status information
-about the state of the Tahoe node.
+dedicated to giving users (both humans and programs) access to the Tahoe-LAFS
+file store. The rest of the namespace provides status information about the
+state of the Tahoe-LAFS node.
``GET /`` (the root page)
Node information: library versions, local nodeid, services being provided.
- Filesystem Access Forms: create a new directory, view a file/directory by
+ File store access forms: create a new directory, view a file/directory by
URI, upload a file (unlinked), download a file by
URI.
- Grid Status: introducer information, helper information, connected storage
+ Grid status: introducer information, helper information, connected storage
servers.
``GET /status/``
============================================
Summary: use explicit file- and dir- caps whenever possible, to reduce the
-potential for surprises when the filesystem structure is changed.
+potential for surprises when the file store structure is changed.
-Tahoe provides a mutable filesystem, but the ways that the filesystem can
+Tahoe-LAFS provides a mutable file store, but the ways that the store can
change are limited. The only thing that can change is that the mapping from
child names to child objects that each directory contains can be changed by
adding a new child name pointing to an object, removing an existing child name,
Overview
========
-When a file or directory in the virtual filesystem is no longer referenced,
+When a file or directory in a Tahoe-LAFS file store is no longer referenced,
the space that its shares occupied on each storage server can be freed,
making room for other shares. Tahoe currently uses a garbage collection
("GC") mechanism to implement this space-reclamation process. Each share has
.TH TAHOE 1 "July 2011" "Tahoe-LAFS \[em] tahoe command" "User Commands"
.SH NAME
.PP
-tahoe - Secure distributed filesystem.
+tahoe - Secure distributed file store.
.SH SYNOPSIS
.PP
tahoe \f[I]COMMAND\f[] [\f[I]OPTION\f[]]... [\f[I]PARAMETER\f[]]...
Display help and exit
.RS
.RE
-.SS USING THE FILESYSTEM
+.SS USING THE FILE STORE
.TP
.B \f[B]mkdir\f[]
Create a new directory.
The lossmodel.lyx file is the source document for an in-progress paper
-that analyzes the probability of losing files stored in a Tahoe
-Least-acces File System under various scenarios. It describes:
+that analyzes the probability of losing files stored in a Tahoe-LAFS
+file store under various scenarios. It describes:
1. How to estimate peer reliabilities, based on peer MTBF failure
data.
Run Tahoe-LAFS
--------------
-Now you are ready to deploy a decentralized filesystem. The ``tahoe``
+Now you are ready to deploy a decentralized file store. The ``tahoe``
executable in the ``bin`` directory can configure and launch your Tahoe-LAFS
nodes. See `<running.rst>`__ for instructions on how to do that.
Prefer the command-line? Run “``tahoe --help``” (the same command-line tool
that is used to start and stop nodes serves to navigate and use the
-decentralized filesystem). To get started, create a new directory and mark it
+decentralized file store). To get started, create a new directory and mark it
as the 'tahoe:' alias by running “``tahoe create-alias tahoe``”. Once you've
done that, you can do “``tahoe ls tahoe:``” and “``tahoe cp LOCALFILE
-tahoe:foo.txt``” to work with your filesystem. The Tahoe-LAFS CLI uses
+tahoe:foo.txt``” to work with your file store. The Tahoe-LAFS CLI uses
similar syntax to the well-known scp and rsync tools. See CLI.rst_ for more
details.
a collection of three layers. The lowest layer is the key-value store: it
provides operations that accept files and upload them to the grid, creating
a URI in the process which securely references the file's contents.
-The middle layer is the filesystem, creating a structure of directories and
-filenames resembling the traditional unix/windows filesystems. The top layer
-is the application layer, which uses the lower layers to provide useful
+The middle layer is the file store, creating a structure of directories and
+filenames resembling the traditional Unix or Windows filesystems. The top
+layer is the application layer, which uses the lower layers to provide useful
services to users, like a backup application, or a way to share files with
friends.
-This document examines the middle layer, the "filesystem".
+This document examines the middle layer, the "file store".
1. `Key-value Store Primitives`_
-2. `Filesystem goals`_
-3. `Dirnode goals`_
+2. `File Store Goals`_
+3. `Dirnode Goals`_
4. `Dirnode secret values`_
5. `Dirnode storage format`_
6. `Dirnode sizes, mutable-file initial read sizes`_
replace(mutable_uri, new_data)
data = get(mutable_uri)
-Filesystem Goals
+File Store Goals
================
-The main goal for the middle (filesystem) layer is to give users a way to
+The main goal for the middle (file store) layer is to give users a way to
organize the data that they have uploaded into the grid. The traditional way
to do this in computer filesystems is to put this data into files, give those
files names, and collect these names into directories.
ciphertext to guess the number of children of each node, and might be able to
guess the length of the child names (or at least their sum). From this, the
attacker may be able to build up a graph with the same shape as the plaintext
-filesystem, but with unlabeled edges and unknown file contents.
+file store, but with unlabeled edges and unknown file contents.
Integrity failures in the storage servers
of generating two random primes).
When a backup program is run for the first time, it needs to copy a large
-amount of data from a pre-existing filesystem into reliable storage. This
-means that a large and complex directory structure needs to be duplicated in
-the dirnode layer. With the one-object-per-dirnode approach described here,
-this requires as many operations as there are edges in the imported
-filesystem graph.
+amount of data from a pre-existing local filesystem into reliable storage.
+This means that a large and complex directory structure needs to be
+duplicated in the dirnode layer. With the one-object-per-dirnode approach
+described here, this requires as many operations as there are edges in the
+imported filesystem graph.
Another approach would be to aggregate multiple directories into a single
storage object. This object would contain a serialized graph rather than a
for example. The UI will need to give the sending user an opportunity to
indicate whether they want to grant read-write or read-only access to the
recipient. The recipient then needs an interface to drag the new folder into
-their vdrive and give it a home.
+their file store and give it a home.
Revocation
==========
2. `Directory URIs`_
3. `Internal Usage of URIs`_
-Each file and directory in a Tahoe filesystem is described by a "URI". There
-are different kinds of URIs for different kinds of objects, and there are
-different kinds of URIs to provide different kinds of access to those
+Each file and directory in a Tahoe-LAFS file store is described by a "URI".
+There are different kinds of URIs for different kinds of objects, and there
+are different kinds of URIs to provide different kinds of access to those
objects. Each URI is a string representation of a "capability" or "cap", and
there are read-caps, write-caps, verify-caps, and others.
a different file or directory for each person or process that wants to
write.
-If mutable parts of a filesystem are accessed via sshfs, only a single
+If mutable parts of a file store are accessed via sshfs, only a single
sshfs mount should be used. There may be data loss if mutable files or
directories are accessed via two sshfs mounts, or written both via sshfs
and from other clients.
-ANNOUNCING Tahoe, the Least-Authority File System, v1.10
+ANNOUNCING Tahoe, the Least-Authority File Store, v1.10
The Tahoe-LAFS team is pleased to announce the immediate
availability of version 1.10.0 of Tahoe-LAFS, an extremely
WHAT IS IT GOOD FOR?
-With Tahoe-LAFS, you distribute your filesystem across
+With Tahoe-LAFS, you distribute your file store across
multiple servers, and even if some of the servers fail or are
-taken over by an attacker, the entire filesystem continues to
+taken over by an attacker, the entire file store continues to
work correctly, and continues to preserve your privacy and
security. You can easily share specific files and directories
with other people.
setup_args["version"] = version
setup(name=APPNAME,
- description='secure, decentralized, fault-tolerant filesystem',
+ description='secure, decentralized, fault-tolerant file store',
long_description=open('README.rst', 'rU').read(),
author='the Tahoe-LAFS project',
author_email='tahoe-dev@tahoe-lafs.org',
+ startstop_node.subCommands
+ GROUP("Debugging")
+ debug.subCommands
- + GROUP("Using the filesystem")
+ + GROUP("Using the file store")
+ cli.subCommands
)