webapi for metadata in vdrive #117

Closed
opened 2007-08-20 19:39:53 +00:00 by warner · 11 comments

For a backup application (in which the vdrive is used to mirror the
filesystem that the user wants backed up), it would be useful to know quickly
that a given file was already present in the vdrive at a given location. The
DHT layer is responsible for efficiently handling files that are already
present in the grid, but that still requires a good bit of encoding work (and
convergence must be enabled). The goal here is to avoid getting to the
encoding phase at all.

I'm thinking that the dirnode edges could contain some metadata that serves
to roughly identify the file (like a quick hash of the file contents, or
maybe just timestamp+length), and the backup process could refrain from
re-pushing anything that looks like it hasn't changed. The exact criteria
used would depend upon how the user wants to make the tradeoff between
performance and robustness. If they're feeling casual and don't want to spend
a lot of CPU or disk IO, then they set it to compare just timestamp+length.
If they're paranoid that they've replaced one file on their disk with a
different file of the exact same length (and also set the timestamp back to
match), then they can set their node to use a full secure hash for comparison
purposes. An intermediate position is possible too: in some circumstances, we
could use a fast non-cryptographic hash here (perhaps just a CRC).

Note that the vdrive as it stands today has no concept of metadata: it is
purely a graph of named edges connecting non-terminal dirnodes with terminal
filenodes and other dirnodes. All of the usual disk-based filesystem metadata
like ctimes and mtimes are left out, as well as even-less-appropriate
metadata like "owner" and "permissions". A backup application, serving as a
bridge between a disk-based filesystem and the vdrive, will want some of this
metadata around (so that the app can restore the metadata as well as the data
itself). This metadata doesn't necessarily have to go on the vdrive edges,
though.. it depends upon how we want to build the backup app.

Also note that a backup application is not obligated to conform to the vdrive
layer (although it would be handy for browsing if it did). The app could, for
example, upload all files to the mesh, get their URIs, then write out a
single big file with one pathname/metadata/URI tuple per line, and upload
that, and remember its URI. (i.e. bypass the vdrive completely). It could
also store the files in the vdrive normally, but stash just the metadata in a
separately-uploaded file.

For a backup application (in which the vdrive is used to mirror the filesystem that the user wants backed up), it would be useful to know quickly that a given file was already present in the vdrive at a given location. The DHT layer is responsible for efficiently handling files that are already present in the grid, but that still requires a good bit of encoding work (and convergence must be enabled). The goal here is to avoid getting to the encoding phase at all. I'm thinking that the dirnode edges could contain some metadata that serves to roughly identify the file (like a quick hash of the file contents, or maybe just timestamp+length), and the backup process could refrain from re-pushing anything that looks like it hasn't changed. The exact criteria used would depend upon how the user wants to make the tradeoff between performance and robustness. If they're feeling casual and don't want to spend a lot of CPU or disk IO, then they set it to compare just timestamp+length. If they're paranoid that they've replaced one file on their disk with a different file of the exact same length (and also set the timestamp back to match), then they can set their node to use a full secure hash for comparison purposes. An intermediate position is possible too: in some circumstances, we could use a fast non-cryptographic hash here (perhaps just a CRC). Note that the vdrive as it stands today has no concept of metadata: it is purely a graph of named edges connecting non-terminal dirnodes with terminal filenodes and other dirnodes. All of the usual disk-based filesystem metadata like ctimes and mtimes are left out, as well as even-less-appropriate metadata like "owner" and "permissions". A backup application, serving as a bridge between a disk-based filesystem and the vdrive, will want some of this metadata around (so that the app can restore the metadata as well as the data itself). This metadata doesn't necessarily have to go on the vdrive edges, though.. it depends upon how we want to build the backup app. Also note that a backup application is not obligated to conform to the vdrive layer (although it would be handy for browsing if it did). The app could, for example, upload all files to the mesh, get their URIs, then write out a single big file with one pathname/metadata/URI tuple per line, and upload *that*, and remember its URI. (i.e. bypass the vdrive completely). It could also store the files in the vdrive normally, but stash just the metadata in a separately-uploaded file.
warner added the
c/code
p/major
t/enhancement
v/0.5.0
labels 2007-08-20 19:39:53 +00:00

On my MacBook Pro, adler32 (from Python's zlib module) hashes a 120-million-byte file in 0.475 seconds, and sha256 (from Python's hashlib module) hashes it in 1.717 seconds.

adler32 does a 734 million byte file in 2.646 seconds and sha256 does it in 10.166 seconds.

This kind of thing makes me wish we used the Tiger hash. The new implementation of Tiger hash in the crypto++ library is about 2.5 times as fast as sha-256.

http://cryptopp.com/benchmarks.html

That C++/assembly implementation of Tiger hashes 217 MiB/second on a modern desktop, which means that the limiting factors would really be disk speed and Python interpreter. For comparison, the Crypto++ implementation of SHA-256 does 81 MiB/second.

This makes me wonder how much overhead we have just to read in a file in Python. Here, the nullsum test just reads in the data. It processes the 734 million byte file in 2.224 seconds.

On my [MacBook](wiki/MacBook) Pro, adler32 (from Python's zlib module) hashes a 120-million-byte file in 0.475 seconds, and sha256 (from Python's hashlib module) hashes it in 1.717 seconds. adler32 does a 734 million byte file in 2.646 seconds and sha256 does it in 10.166 seconds. This kind of thing makes me wish we used the Tiger hash. The new implementation of Tiger hash in the crypto++ library is about 2.5 times as fast as sha-256. <http://cryptopp.com/benchmarks.html> That C++/assembly implementation of Tiger hashes 217 MiB/second on a modern desktop, which means that the limiting factors would really be disk speed and Python interpreter. For comparison, the Crypto++ implementation of SHA-256 does 81 MiB/second. This makes me wonder how much overhead we have just to read in a file in Python. Here, the nullsum test just reads in the data. It processes the 734 million byte file in 2.224 seconds.

Attachment sha256sum.py (311 bytes) added

sha256sum.py

**Attachment** sha256sum.py (311 bytes) added sha256sum.py

Attachment nullsum.py (256 bytes) added

nullsum.py

**Attachment** nullsum.py (256 bytes) added nullsum.py

Attachment adler32sum.py (306 bytes) added

adler32sum.py

**Attachment** adler32sum.py (306 bytes) added adler32sum.py
Author

I think that leveraging existing backup software will work pretty well if we can just preserve ctime/mtime/length and the usual unix/windows permission fields.

I'm currently thinking that the dirnode structure should be enhanced to record a dictionary of metadata next to the child URI.

I think that leveraging existing backup software will work pretty well if we can just preserve ctime/mtime/length and the usual unix/windows permission fields. I'm currently thinking that the dirnode structure should be enhanced to record a dictionary of metadata next to the child URI.
Author

we have a proposal for the metadata API floating around somewhere.. we should
gather some more feedback from Mike and the others and then actually go and
implement it.

The dirnode storage layout has a place for metadata already, as a
JSON-encoded dictionary. We're just lacking the dirnode methods to modify it
and the HTTP interfaces to drive them.

we have a proposal for the metadata API floating around somewhere.. we should gather some more feedback from Mike and the others and then actually go and implement it. The dirnode storage layout has a place for metadata already, as a JSON-encoded dictionary. We're just lacking the dirnode methods to modify it and the HTTP interfaces to drive them.
zooko modified the milestone from 0.9.0 (Allmydata 3.0 final) to 0.8.0 (Allmydata 3.0 Beta) 2008-01-23 02:23:45 +00:00
Author

Here's the proposals that we threw together back in october.


d. metadata

The set of directories and files in a tahoe virtual drive forms a graph: each
directory is a node, each file is a leaf node, and each name that you use to
look at a child file or directory is an edge.

In Tahoe, all file metadata is kept on those edges. The filenode itself is
just a unadorned sequence of bytes. Everything else that is traditionally
associated with a "file" (filename, type, timestamps, notions like "owners"
and "permissions", etc) are recorded in the parent directory in the record
that points to the file's URI.

Note that this implies that two separate directories which link to the same
filenode will have completely independent metadata for that file.

Application programs can add arbitrary metadata to these edges. We are
establishing conventions for well-known metadata types, but apps are free to
add whatever other data they like. Backup applications which copy data from a
normal disk-based filesystem can use this metadata to record enough
information to recreate the original filesystem (using attributes like owner,
permissions, and timestamps).

This metadata is expressed as a dictionary, which maps ASCII strings to any
data type that can be serialized as JSON, which includes booleans, numbers
(ints and floats), strings (ascii or unicode), lists, and string-keyed
dictionaries. This does not include sets, so the convention is that sets are
expressed as lists with manual duplicate-pruning.

Metadata is retrieved by reading the dirnode using the "GET t=json" method
described below. Metadata is set by providing a dictionary of commands to the
"POST t=change-metadata" method also described below.


e. examining files or directories

  GET $URL?t=json

  This returns machine-parseable information about the indicated file or
  directory in the HTTP response body. The JSON always contains a list of two
  elements, and the first element of the list is always a flag that indicates
  whether the referenced object is a file or a directory. The second element
  is a dictionary of data about the object.

  If it is a file, then the information includes file size and URI, like this:

   [ 'filenode', { 'ro_uri': file_uri,
                   'size': bytes,
 } ]

  If it is a directory, then it includes information about the children of
  this directory, as a mapping from child name to a set of data about the
  child (the same data that would appear in a corresponding GET?t=json of the
  child itself, plus metadata). Like this:

   [ 'dirnode', { 'rw_uri': read_write_uri,
                  'ro_uri': read_only_uri,
                  'children': children } ]

  In the above example, 'children' is a dictionary in which the keys are
  child names and the values depend upon whether the child is a file or a
  directory:

   'foo.txt': [ 'filenode', { 'ro_uri': uri, 'size': bytes 
                              'ctime': 1192728870.6729021,
                              'tags': ['personal', 'pictures'], } ]
   'subdir':  [ 'dirnode', { 'rw_uri': rwuri, 'ro_uri': rouri } ]

  note that the value is the same as the JSON representation of the child
  object, except that directories do not recurse (the "children" entry of the
  child is omitted), and the value's second element contains any metadata for
  the child in addition to the URIs.

  The rw_uri field will be present in the information about a directory if
  and only if you have read-write access to that directory,

  So a complete example of the return value for 'GET $URL?t=json' for a
  directory could be:

[
 "dirnode", 
 {
  "rw_uri": "URI:DIR:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:gqu1fub33exw9cu63718yzx6gr", 
  "ro_uri": "URI:DIR-RO:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:3sjp7fwkrsojm9xsymtgkk5khh", 
  "children": {
   "subdir": [
    "dirnode", 
    {
     "rw_uri": "URI:DIR:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:wibsf933cn1y1zowume4uy1ajr", 
     "ro_uri": "URI:DIR-RO:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:97srnzip6muckorjopfe9gk8ae",
     "tags": ["personal"]
    }
   ], 
   "webapi.txt": [
    "filenode", 
    {
     "ro_uri": "URI:CHK:35p7umig6cc3omtn1go4j5nbgw:9ab9ek8b6nagrrk1g4nzepa4kgta9eh434o6shddxx9hfbrrydky:3:10:19653", 
     "size": 19653,
     "tags": ["work", "tahoe"],
     "ctime": 1192728870.6729021,
     "mtime": 1192728899.8800000
    }
   ], 
   "architecture.txt": [
    "filenode", 
    {
     "ro_uri": "URI:CHK:iah546ukk6eqntehth1s3ndeoh:it9o1k5db7mjwjbj3tdd7qjyghf3qjkq7ntjd6ad5e3ufii3uwwo:3:10:35769", 
     "size": 35769,
     "tags": ["work", "tahoe"],
     "ctime": 1192728870.6729021,
     "mtime": 1192728899.8800000
    }
   ]
  }
 }
]



k. modifying metadata

  POST $URL?t=change-metadata

  This operation is used to modify the metadata attached to a directory. The
  body of the POST method is a specially-formatted JSON data structure with
  commands to set and remove metadata keys. Because the body is not an
  encoded HTML form, this method is for use by programmatic clients, rather
  than a web browser. At present there is no provision for a web browser to
  modify metadata.

 API PROPOSAL ONE:

  The POST body is a JSON-encoded dictionary, with one member for each edge
  (file or subdirectory) that you want to modify. The key of this member is
  the child name (filename or subdir name), and the value associated with
  that key is a "command dictionary", which describes how you want to modify
  that edge's metadata.

  There are currently two commands defined in the command dictionary: "set"
  and "delete". The value associated with the "set" command is a dictionary
  that maps metadata-key to metadata-value. For each metadata key that
  appears in this dictionary, any existing metadata is replaced with the
  given value.

  The value associated with the "delete" key is a list of metadata key names.
  For each such name, the associated metadata is removed.

  For example, if the client wishes to set ctime and mtime, and remove an
  associated icon, it could use the following:

   {
     "webapi.txt": {"set": {"ctime": 1192728870,
                            "mtime": 1192728899, },
                    "delete": ["icon"],
                   },
   }

 API PROPOSAL TWO:

  The POST body is a JSON-encoded list, with one element for each
  modification you wish to make to the metadata. Each element is "command",
  encoded as as a list of length 3 (for "delete" commands) or 4 (for "set"
  commands).

  The "set" command provides the name of the child to be modified, the name
  of the metadata key to be modified, and the new value for that metadata
  key. The new value completely replaces the old one. For example:

    ["set", "webapi.txt", "tags", ["work", "tahoe"]]

  The "delete" command provides the child name and the name of the metadata
  key to be deleted:

    ["delete", "webapi.txt, "icon"]

  If the client wishes to set ctime and mtime, and remove any icon, it could
  use the following as the body of the POST request:

    [
     ["delete", "webapi.txt", "icon"],
     ["set", "webapi.txt", "ctime", 1192728870],
     ["set", "webapi.txt", "mtime", 1192728899],
    ]

the feedback we've received so far is from Mike:

The metadata seems to be very flexible.  I think that working with proposal
two is slightly easier, but I personally prefer "PROPOSAL ONE" because it's
less verbose, and I like how the metadata is grouped together logically by
command, according to the file node or directory node that it describes.
Plus, "PROPOSAL ONE" more closely resembles the format of the JSON returned
when you examine the file or directory with the GET $URL?t=JSON HTTP
request. 

It might be nice to have a way to set metadata for a file or directory node
in the same request that the node is created in too.

-Mike
Here's the proposals that we threw together back in october. ``` d. metadata The set of directories and files in a tahoe virtual drive forms a graph: each directory is a node, each file is a leaf node, and each name that you use to look at a child file or directory is an edge. In Tahoe, all file metadata is kept on those edges. The filenode itself is just a unadorned sequence of bytes. Everything else that is traditionally associated with a "file" (filename, type, timestamps, notions like "owners" and "permissions", etc) are recorded in the parent directory in the record that points to the file's URI. Note that this implies that two separate directories which link to the same filenode will have completely independent metadata for that file. Application programs can add arbitrary metadata to these edges. We are establishing conventions for well-known metadata types, but apps are free to add whatever other data they like. Backup applications which copy data from a normal disk-based filesystem can use this metadata to record enough information to recreate the original filesystem (using attributes like owner, permissions, and timestamps). This metadata is expressed as a dictionary, which maps ASCII strings to any data type that can be serialized as JSON, which includes booleans, numbers (ints and floats), strings (ascii or unicode), lists, and string-keyed dictionaries. This does not include sets, so the convention is that sets are expressed as lists with manual duplicate-pruning. Metadata is retrieved by reading the dirnode using the "GET t=json" method described below. Metadata is set by providing a dictionary of commands to the "POST t=change-metadata" method also described below. e. examining files or directories GET $URL?t=json This returns machine-parseable information about the indicated file or directory in the HTTP response body. The JSON always contains a list of two elements, and the first element of the list is always a flag that indicates whether the referenced object is a file or a directory. The second element is a dictionary of data about the object. If it is a file, then the information includes file size and URI, like this: [ 'filenode', { 'ro_uri': file_uri, 'size': bytes, } ] If it is a directory, then it includes information about the children of this directory, as a mapping from child name to a set of data about the child (the same data that would appear in a corresponding GET?t=json of the child itself, plus metadata). Like this: [ 'dirnode', { 'rw_uri': read_write_uri, 'ro_uri': read_only_uri, 'children': children } ] In the above example, 'children' is a dictionary in which the keys are child names and the values depend upon whether the child is a file or a directory: 'foo.txt': [ 'filenode', { 'ro_uri': uri, 'size': bytes 'ctime': 1192728870.6729021, 'tags': ['personal', 'pictures'], } ] 'subdir': [ 'dirnode', { 'rw_uri': rwuri, 'ro_uri': rouri } ] note that the value is the same as the JSON representation of the child object, except that directories do not recurse (the "children" entry of the child is omitted), and the value's second element contains any metadata for the child in addition to the URIs. The rw_uri field will be present in the information about a directory if and only if you have read-write access to that directory, So a complete example of the return value for 'GET $URL?t=json' for a directory could be: [ "dirnode", { "rw_uri": "URI:DIR:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:gqu1fub33exw9cu63718yzx6gr", "ro_uri": "URI:DIR-RO:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:3sjp7fwkrsojm9xsymtgkk5khh", "children": { "subdir": [ "dirnode", { "rw_uri": "URI:DIR:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:wibsf933cn1y1zowume4uy1ajr", "ro_uri": "URI:DIR-RO:pb:\/\/xextf3eap44o3wi27mf7ehiur6wvhzr6@207.7.153.180:56677,127.0.0.1:56677\/vdrive:97srnzip6muckorjopfe9gk8ae", "tags": ["personal"] } ], "webapi.txt": [ "filenode", { "ro_uri": "URI:CHK:35p7umig6cc3omtn1go4j5nbgw:9ab9ek8b6nagrrk1g4nzepa4kgta9eh434o6shddxx9hfbrrydky:3:10:19653", "size": 19653, "tags": ["work", "tahoe"], "ctime": 1192728870.6729021, "mtime": 1192728899.8800000 } ], "architecture.txt": [ "filenode", { "ro_uri": "URI:CHK:iah546ukk6eqntehth1s3ndeoh:it9o1k5db7mjwjbj3tdd7qjyghf3qjkq7ntjd6ad5e3ufii3uwwo:3:10:35769", "size": 35769, "tags": ["work", "tahoe"], "ctime": 1192728870.6729021, "mtime": 1192728899.8800000 } ] } } ] k. modifying metadata POST $URL?t=change-metadata This operation is used to modify the metadata attached to a directory. The body of the POST method is a specially-formatted JSON data structure with commands to set and remove metadata keys. Because the body is not an encoded HTML form, this method is for use by programmatic clients, rather than a web browser. At present there is no provision for a web browser to modify metadata. API PROPOSAL ONE: The POST body is a JSON-encoded dictionary, with one member for each edge (file or subdirectory) that you want to modify. The key of this member is the child name (filename or subdir name), and the value associated with that key is a "command dictionary", which describes how you want to modify that edge's metadata. There are currently two commands defined in the command dictionary: "set" and "delete". The value associated with the "set" command is a dictionary that maps metadata-key to metadata-value. For each metadata key that appears in this dictionary, any existing metadata is replaced with the given value. The value associated with the "delete" key is a list of metadata key names. For each such name, the associated metadata is removed. For example, if the client wishes to set ctime and mtime, and remove an associated icon, it could use the following: { "webapi.txt": {"set": {"ctime": 1192728870, "mtime": 1192728899, }, "delete": ["icon"], }, } API PROPOSAL TWO: The POST body is a JSON-encoded list, with one element for each modification you wish to make to the metadata. Each element is "command", encoded as as a list of length 3 (for "delete" commands) or 4 (for "set" commands). The "set" command provides the name of the child to be modified, the name of the metadata key to be modified, and the new value for that metadata key. The new value completely replaces the old one. For example: ["set", "webapi.txt", "tags", ["work", "tahoe"]] The "delete" command provides the child name and the name of the metadata key to be deleted: ["delete", "webapi.txt, "icon"] If the client wishes to set ctime and mtime, and remove any icon, it could use the following as the body of the POST request: [ ["delete", "webapi.txt", "icon"], ["set", "webapi.txt", "ctime", 1192728870], ["set", "webapi.txt", "mtime", 1192728899], ] ``` the feedback we've received so far is from Mike: ``` The metadata seems to be very flexible. I think that working with proposal two is slightly easier, but I personally prefer "PROPOSAL ONE" because it's less verbose, and I like how the metadata is grouped together logically by command, according to the file node or directory node that it describes. Plus, "PROPOSAL ONE" more closely resembles the format of the JSON returned when you examine the file or directory with the GET $URL?t=JSON HTTP request. It might be nice to have a way to set metadata for a file or directory node in the same request that the node is created in too. -Mike ```
Author

We have three conceivable uses for this metadata. The first is to record enough information to allow a roundtrip (from a normal unix-ish filesystem, through tahoe, and back again) to restore most of the original file. This means a timestamp, at least, and maybe some mode information (the same sort of information that 'tar' records).

The second is a set of hashes, something application-specific, to allow a custom backup tool to be more efficient.

The third is arbitrary application data, perhaps tags or something.

The most important need right now is for the first use case (timestamps), so I'm retitling this ticket. We'll probably implement the framework for all three at the same time, so I'm not creating a separate ticket for the second two use cases.

We have three conceivable uses for this metadata. The first is to record enough information to allow a roundtrip (from a normal unix-ish filesystem, through tahoe, and back again) to restore most of the original file. This means a timestamp, at least, and maybe some mode information (the same sort of information that 'tar' records). The second is a set of hashes, something application-specific, to allow a custom backup tool to be more efficient. The third is arbitrary application data, perhaps tags or something. The most important need right now is for the first use case (timestamps), so I'm retitling this ticket. We'll probably implement the framework for all three at the same time, so I'm not creating a separate ticket for the second two use cases.
warner modified the milestone from 0.8.0 (Allmydata 3.0 Beta) to 0.9.0 (Allmydata 3.0 final) 2008-02-04 23:11:24 +00:00
warner changed title from metadata in vdrive to improve backup-application performance to metadata in vdrive to record timestamps 2008-02-04 23:11:24 +00:00
Author

It occurred to me today that a good starting point would just be to automatically add timestamps every time we do DirectoryNode.add_child. Rob pointed out that the algorithm could be:

 metadata['mtime'] = now
 if 'ctime' not in metadata:
  metadata['ctime'] = now

that would probably give us enough information for the FUSE plugins to show meaningful timestamps.

To allow 'cp -r' to look right, we'd still need to provide a way to set this metadata, of course. The tasks are:

  • present metadata in the JSON output
  • implement the webapi set-metdata call
It occurred to me today that a good starting point would just be to automatically add timestamps every time we do `DirectoryNode.add_child`. Rob pointed out that the algorithm could be: ``` metadata['mtime'] = now if 'ctime' not in metadata: metadata['ctime'] = now ``` that would probably give us enough information for the FUSE plugins to show meaningful timestamps. To allow 'cp -r' to look right, we'd still need to provide a way to set this metadata, of course. The tasks are: * present metadata in the JSON output * implement the webapi set-metdata call
Author

I've modified the dirnode methods to add metadata setters on
add_node/add_uri, and arranged the defaults to automatically add ctime/mtime
keys. I've also added metadata the the ?t=json output (i.e. zooko's "WAPI"),
and to the human-oriented HTML web page (zooko's "WUI").

This is enough for front-ends to display timestamps, and might be sufficient
for 1.0 . In the longer run, we still need:

  • web API setters for metadata, to allow things like 'cp -p' to preserve
    timestamps from some other filesystem
  • web API setters/modifiers for generalized metadata
I've modified the dirnode methods to add metadata setters on add_node/add_uri, and arranged the defaults to automatically add ctime/mtime keys. I've also added metadata the the ?t=json output (i.e. zooko's "WAPI"), and to the human-oriented HTML web page (zooko's "WUI"). This is enough for front-ends to display timestamps, and might be sufficient for 1.0 . In the longer run, we still need: * web API setters for metadata, to allow things like 'cp -p' to preserve timestamps from some other filesystem * web API setters/modifiers for generalized metadata
warner added
p/minor
and removed
p/major
labels 2008-02-12 02:13:58 +00:00
zooko modified the milestone from 0.9.0 (Allmydata 3.0 final) to undecided 2008-03-08 04:15:46 +00:00
warner changed title from metadata in vdrive to record timestamps to webapi for metadata in vdrive 2008-06-01 20:55:11 +00:00
warner added
c/code-frontend-web
and removed
c/code
labels 2009-03-08 22:09:53 +00:00

This was fixed long ago.

This was fixed long ago.
daira added the
r/fixed
label 2010-04-13 00:59:03 +00:00
daira closed this issue 2010-04-13 00:59:03 +00:00
Sign in to join this conversation.
No labels
c/code
c/code-dirnodes
c/code-encoding
c/code-frontend
c/code-frontend-cli
c/code-frontend-ftp-sftp
c/code-frontend-magic-folder
c/code-frontend-web
c/code-mutable
c/code-network
c/code-nodeadmin
c/code-peerselection
c/code-storage
c/contrib
c/dev-infrastructure
c/docs
c/operational
c/packaging
c/unknown
c/website
kw:2pc
kw:410
kw:9p
kw:ActivePerl
kw:AttributeError
kw:DataUnavailable
kw:DeadReferenceError
kw:DoS
kw:FileZilla
kw:GetLastError
kw:IFinishableConsumer
kw:K
kw:LeastAuthority
kw:Makefile
kw:RIStorageServer
kw:StringIO
kw:UncoordinatedWriteError
kw:about
kw:access
kw:access-control
kw:accessibility
kw:accounting
kw:accounting-crawler
kw:add-only
kw:aes
kw:aesthetics
kw:alias
kw:aliases
kw:aliens
kw:allmydata
kw:amazon
kw:ambient
kw:annotations
kw:anonymity
kw:anonymous
kw:anti-censorship
kw:api_auth_token
kw:appearance
kw:appname
kw:apport
kw:archive
kw:archlinux
kw:argparse
kw:arm
kw:assertion
kw:attachment
kw:auth
kw:authentication
kw:automation
kw:avahi
kw:availability
kw:aws
kw:azure
kw:backend
kw:backoff
kw:backup
kw:backupdb
kw:backward-compatibility
kw:bandwidth
kw:basedir
kw:bayes
kw:bbfreeze
kw:beta
kw:binaries
kw:binutils
kw:bitcoin
kw:bitrot
kw:blacklist
kw:blocker
kw:blocks-cloud-deployment
kw:blocks-cloud-merge
kw:blocks-magic-folder-merge
kw:blocks-merge
kw:blocks-raic
kw:blocks-release
kw:blog
kw:bom
kw:bonjour
kw:branch
kw:branding
kw:breadcrumbs
kw:brians-opinion-needed
kw:browser
kw:bsd
kw:build
kw:build-helpers
kw:buildbot
kw:builders
kw:buildslave
kw:buildslaves
kw:cache
kw:cap
kw:capleak
kw:captcha
kw:cast
kw:centos
kw:cffi
kw:chacha
kw:charset
kw:check
kw:checker
kw:chroot
kw:ci
kw:clean
kw:cleanup
kw:cli
kw:cloud
kw:cloud-backend
kw:cmdline
kw:code
kw:code-checks
kw:coding-standards
kw:coding-tools
kw:coding_tools
kw:collection
kw:compatibility
kw:completion
kw:compression
kw:confidentiality
kw:config
kw:configuration
kw:configuration.txt
kw:conflict
kw:connection
kw:connectivity
kw:consistency
kw:content
kw:control
kw:control.furl
kw:convergence
kw:coordination
kw:copyright
kw:corruption
kw:cors
kw:cost
kw:coverage
kw:coveralls
kw:coveralls.io
kw:cpu-watcher
kw:cpyext
kw:crash
kw:crawler
kw:crawlers
kw:create-container
kw:cruft
kw:crypto
kw:cryptography
kw:cryptography-lib
kw:cryptopp
kw:csp
kw:curl
kw:cutoff-date
kw:cycle
kw:cygwin
kw:d3
kw:daemon
kw:darcs
kw:darcsver
kw:database
kw:dataloss
kw:db
kw:dead-code
kw:deb
kw:debian
kw:debug
kw:deep-check
kw:defaults
kw:deferred
kw:delete
kw:deletion
kw:denial-of-service
kw:dependency
kw:deployment
kw:deprecation
kw:desert-island
kw:desert-island-build
kw:design
kw:design-review-needed
kw:detection
kw:dev-infrastructure
kw:devpay
kw:directory
kw:directory-page
kw:dirnode
kw:dirnodes
kw:disconnect
kw:discovery
kw:disk
kw:disk-backend
kw:distribute
kw:distutils
kw:dns
kw:do_http
kw:doc-needed
kw:docker
kw:docs
kw:docs-needed
kw:dokan
kw:dos
kw:download
kw:downloader
kw:dragonfly
kw:drop-upload
kw:duplicity
kw:dusty
kw:earth-dragon
kw:easy
kw:ec2
kw:ecdsa
kw:ed25519
kw:egg-needed
kw:eggs
kw:eliot
kw:email
kw:empty
kw:encoding
kw:endpoint
kw:enterprise
kw:enum34
kw:environment
kw:erasure
kw:erasure-coding
kw:error
kw:escaping
kw:etag
kw:etch
kw:evangelism
kw:eventual
kw:example
kw:excess-authority
kw:exec
kw:exocet
kw:expiration
kw:extensibility
kw:extension
kw:failure
kw:fedora
kw:ffp
kw:fhs
kw:figleaf
kw:file
kw:file-descriptor
kw:filename
kw:filesystem
kw:fileutil
kw:fips
kw:firewall
kw:first
kw:floatingpoint
kw:flog
kw:foolscap
kw:forward-compatibility
kw:forward-secrecy
kw:forwarding
kw:free
kw:freebsd
kw:frontend
kw:fsevents
kw:ftp
kw:ftpd
kw:full
kw:furl
kw:fuse
kw:garbage
kw:garbage-collection
kw:gateway
kw:gatherer
kw:gc
kw:gcc
kw:gentoo
kw:get
kw:git
kw:git-annex
kw:github
kw:glacier
kw:globalcaps
kw:glossary
kw:google-cloud-storage
kw:google-drive-backend
kw:gossip
kw:governance
kw:grid
kw:grid-manager
kw:gridid
kw:gridsync
kw:grsec
kw:gsoc
kw:gvfs
kw:hackfest
kw:hacktahoe
kw:hang
kw:hardlink
kw:heartbleed
kw:heisenbug
kw:help
kw:helper
kw:hint
kw:hooks
kw:how
kw:how-to
kw:howto
kw:hp
kw:hp-cloud
kw:html
kw:http
kw:https
kw:i18n
kw:i2p
kw:i2p-collab
kw:illustration
kw:image
kw:immutable
kw:impressions
kw:incentives
kw:incident
kw:init
kw:inlineCallbacks
kw:inotify
kw:install
kw:installer
kw:integration
kw:integration-test
kw:integrity
kw:interactive
kw:interface
kw:interfaces
kw:interoperability
kw:interstellar-exploration
kw:introducer
kw:introduction
kw:iphone
kw:ipkg
kw:iputil
kw:ipv6
kw:irc
kw:jail
kw:javascript
kw:joke
kw:jquery
kw:json
kw:jsui
kw:junk
kw:key-value-store
kw:kfreebsd
kw:known-issue
kw:konqueror
kw:kpreid
kw:kvm
kw:l10n
kw:lae
kw:large
kw:latency
kw:leak
kw:leasedb
kw:leases
kw:libgmp
kw:license
kw:licenss
kw:linecount
kw:link
kw:linux
kw:lit
kw:localhost
kw:location
kw:locking
kw:logging
kw:logo
kw:loopback
kw:lucid
kw:mac
kw:macintosh
kw:magic-folder
kw:manhole
kw:manifest
kw:manual-test-needed
kw:map
kw:mapupdate
kw:max_space
kw:mdmf
kw:memcheck
kw:memory
kw:memory-leak
kw:mesh
kw:metadata
kw:meter
kw:migration
kw:mime
kw:mingw
kw:minimal
kw:misc
kw:miscapture
kw:mlp
kw:mock
kw:more-info-needed
kw:mountain-lion
kw:move
kw:multi-users
kw:multiple
kw:multiuser-gateway
kw:munin
kw:music
kw:mutability
kw:mutable
kw:mystery
kw:names
kw:naming
kw:nas
kw:navigation
kw:needs-review
kw:needs-spawn
kw:netbsd
kw:network
kw:nevow
kw:new-user
kw:newcaps
kw:news
kw:news-done
kw:news-needed
kw:newsletter
kw:newurls
kw:nfc
kw:nginx
kw:nixos
kw:no-clobber
kw:node
kw:node-url
kw:notification
kw:notifyOnDisconnect
kw:nsa310
kw:nsa320
kw:nsa325
kw:numpy
kw:objects
kw:old
kw:openbsd
kw:openitp-packaging
kw:openssl
kw:openstack
kw:opensuse
kw:operation-helpers
kw:operational
kw:operations
kw:ophandle
kw:ophandles
kw:ops
kw:optimization
kw:optional
kw:options
kw:organization
kw:os
kw:os.abort
kw:ostrom
kw:osx
kw:osxfuse
kw:otf-magic-folder-objective1
kw:otf-magic-folder-objective2
kw:otf-magic-folder-objective3
kw:otf-magic-folder-objective4
kw:otf-magic-folder-objective5
kw:otf-magic-folder-objective6
kw:p2p
kw:packaging
kw:partial
kw:password
kw:path
kw:paths
kw:pause
kw:peer-selection
kw:performance
kw:permalink
kw:permissions
kw:persistence
kw:phone
kw:pickle
kw:pip
kw:pipermail
kw:pkg_resources
kw:placement
kw:planning
kw:policy
kw:port
kw:portability
kw:portal
kw:posthook
kw:pratchett
kw:preformance
kw:preservation
kw:privacy
kw:process
kw:profile
kw:profiling
kw:progress
kw:proxy
kw:publish
kw:pyOpenSSL
kw:pyasn1
kw:pycparser
kw:pycrypto
kw:pycrypto-lib
kw:pycryptopp
kw:pyfilesystem
kw:pyflakes
kw:pylint
kw:pypi
kw:pypy
kw:pysqlite
kw:python
kw:python3
kw:pythonpath
kw:pyutil
kw:pywin32
kw:quickstart
kw:quiet
kw:quotas
kw:quoting
kw:raic
kw:rainhill
kw:random
kw:random-access
kw:range
kw:raspberry-pi
kw:reactor
kw:readonly
kw:rebalancing
kw:recovery
kw:recursive
kw:redhat
kw:redirect
kw:redressing
kw:refactor
kw:referer
kw:referrer
kw:regression
kw:rekey
kw:relay
kw:release
kw:release-blocker
kw:reliability
kw:relnotes
kw:remote
kw:removable
kw:removable-disk
kw:rename
kw:renew
kw:repair
kw:replace
kw:report
kw:repository
kw:research
kw:reserved_space
kw:response-needed
kw:response-time
kw:restore
kw:retrieve
kw:retry
kw:review
kw:review-needed
kw:reviewed
kw:revocation
kw:roadmap
kw:rollback
kw:rpm
kw:rsa
kw:rss
kw:rst
kw:rsync
kw:rusty
kw:s3
kw:s3-backend
kw:s3-frontend
kw:s4
kw:same-origin
kw:sandbox
kw:scalability
kw:scaling
kw:scheduling
kw:schema
kw:scheme
kw:scp
kw:scripts
kw:sdist
kw:sdmf
kw:security
kw:self-contained
kw:server
kw:servermap
kw:servers-of-happiness
kw:service
kw:setup
kw:setup.py
kw:setup_requires
kw:setuptools
kw:setuptools_darcs
kw:sftp
kw:shared
kw:shareset
kw:shell
kw:signals
kw:simultaneous
kw:six
kw:size
kw:slackware
kw:slashes
kw:smb
kw:sneakernet
kw:snowleopard
kw:socket
kw:solaris
kw:space
kw:space-efficiency
kw:spam
kw:spec
kw:speed
kw:sqlite
kw:ssh
kw:ssh-keygen
kw:sshfs
kw:ssl
kw:stability
kw:standards
kw:start
kw:startup
kw:static
kw:static-analysis
kw:statistics
kw:stats
kw:stats_gatherer
kw:status
kw:stdeb
kw:storage
kw:streaming
kw:strports
kw:style
kw:stylesheet
kw:subprocess
kw:sumo
kw:survey
kw:svg
kw:symlink
kw:synchronous
kw:tac
kw:tahoe-*
kw:tahoe-add-alias
kw:tahoe-admin
kw:tahoe-archive
kw:tahoe-backup
kw:tahoe-check
kw:tahoe-cp
kw:tahoe-create-alias
kw:tahoe-create-introducer
kw:tahoe-debug
kw:tahoe-deep-check
kw:tahoe-deepcheck
kw:tahoe-lafs-trac-stream
kw:tahoe-list-aliases
kw:tahoe-ls
kw:tahoe-magic-folder
kw:tahoe-manifest
kw:tahoe-mkdir
kw:tahoe-mount
kw:tahoe-mv
kw:tahoe-put
kw:tahoe-restart
kw:tahoe-rm
kw:tahoe-run
kw:tahoe-start
kw:tahoe-stats
kw:tahoe-unlink
kw:tahoe-webopen
kw:tahoe.css
kw:tahoe_files
kw:tahoewapi
kw:tarball
kw:tarballs
kw:tempfile
kw:templates
kw:terminology
kw:test
kw:test-and-set
kw:test-from-egg
kw:test-needed
kw:testgrid
kw:testing
kw:tests
kw:throttling
kw:ticket999-s3-backend
kw:tiddly
kw:time
kw:timeout
kw:timing
kw:to
kw:to-be-closed-on-2011-08-01
kw:tor
kw:tor-protocol
kw:torsocks
kw:tox
kw:trac
kw:transparency
kw:travis
kw:travis-ci
kw:trial
kw:trickle
kw:trivial
kw:truckee
kw:tub
kw:tub.location
kw:twine
kw:twistd
kw:twistd.log
kw:twisted
kw:twisted-14
kw:twisted-trial
kw:twitter
kw:twn
kw:txaws
kw:type
kw:typeerror
kw:ubuntu
kw:ucwe
kw:ueb
kw:ui
kw:unclean
kw:uncoordinated-writes
kw:undeletable
kw:unfinished-business
kw:unhandled-error
kw:unhappy
kw:unicode
kw:unit
kw:unix
kw:unlink
kw:update
kw:upgrade
kw:upload
kw:upload-helper
kw:uri
kw:url
kw:usability
kw:use-case
kw:utf-8
kw:util
kw:uwsgi
kw:ux
kw:validation
kw:variables
kw:vdrive
kw:verify
kw:verlib
kw:version
kw:versioning
kw:versions
kw:video
kw:virtualbox
kw:virtualenv
kw:vista
kw:visualization
kw:visualizer
kw:vm
kw:volunteergrid2
kw:volunteers
kw:vpn
kw:wapi
kw:warners-opinion-needed
kw:warning
kw:weapi
kw:web
kw:web.port
kw:webapi
kw:webdav
kw:webdrive
kw:webport
kw:websec
kw:website
kw:websocket
kw:welcome
kw:welcome-page
kw:welcomepage
kw:wiki
kw:win32
kw:win64
kw:windows
kw:windows-related
kw:winscp
kw:workaround
kw:world-domination
kw:wrapper
kw:write-enabler
kw:wui
kw:x86
kw:x86-64
kw:xhtml
kw:xml
kw:xss
kw:zbase32
kw:zetuptoolz
kw:zfec
kw:zookos-opinion-needed
kw:zope
kw:zope.interface
p/blocker
p/critical
p/major
p/minor
p/normal
p/supercritical
p/trivial
r/cannot reproduce
r/duplicate
r/fixed
r/invalid
r/somebody else's problem
r/was already fixed
r/wontfix
r/worksforme
t/defect
t/enhancement
t/task
v/0.2.0
v/0.3.0
v/0.4.0
v/0.5.0
v/0.5.1
v/0.6.0
v/0.6.1
v/0.7.0
v/0.8.0
v/0.9.0
v/1.0.0
v/1.1.0
v/1.10.0
v/1.10.1
v/1.10.2
v/1.10a2
v/1.11.0
v/1.12.0
v/1.12.1
v/1.13.0
v/1.14.0
v/1.15.0
v/1.15.1
v/1.2.0
v/1.3.0
v/1.4.1
v/1.5.0
v/1.6.0
v/1.6.1
v/1.7.0
v/1.7.1
v/1.7β
v/1.8.0
v/1.8.1
v/1.8.2
v/1.8.3
v/1.8β
v/1.9.0
v/1.9.0-s3branch
v/1.9.0a1
v/1.9.0a2
v/1.9.0b1
v/1.9.1
v/1.9.2
v/1.9.2a1
v/cloud-branch
v/unknown
No milestone
No project
No assignees
3 participants
Notifications
Due date
The due date is invalid or out of range. Please use the format "yyyy-mm-dd".

No due date set.

Dependencies

No dependencies set.

Reference: tahoe-lafs/trac#117
No description provided.