Repair used default shares.happy #1212

Closed
opened 2010-09-25 14:22:55 +00:00 by eurekafag · 56 comments
eurekafag commented 2010-09-25 14:22:55 +00:00
Owner

I've tried to repair a file and got:

<class 'allmydata.interfaces.UploadUnhappinessError'>: shares could be
placed on only 5 server(s) such that any 3 of them have enough shares
to recover the file, but we were asked to place shares on at least 7
such servers. (placed all 10 shares, want to place shares on at least
7 servers such that any 3 of them have enough shares to recover the
file, sent 7 queries to 5 peers, 7 queries placed some shares, 0
placed none (of which 0 placed none due to the server being full and 0
placed none due to an error))

Everything worked fine on 1.7.1 and shares.happy = 3 (didn't changed it after upgrade). So I did a little investigation and found the problem. It's immutable/repairer.py, line 60:

happy = upload.BaseUploadable.default_encoding_param_happy

Why do we use default happy here? It definitely should be read from
the config. I didin't dig further but replaced it with ugly hack:

happy = 3 #upload.BaseUploadable.default_encoding_param_happy

...and the problem has gone! Repairing works with just 6 servers
online.

I've tried to repair a file and got: ``` <class 'allmydata.interfaces.UploadUnhappinessError'>: shares could be placed on only 5 server(s) such that any 3 of them have enough shares to recover the file, but we were asked to place shares on at least 7 such servers. (placed all 10 shares, want to place shares on at least 7 servers such that any 3 of them have enough shares to recover the file, sent 7 queries to 5 peers, 7 queries placed some shares, 0 placed none (of which 0 placed none due to the server being full and 0 placed none due to an error)) ``` Everything worked fine on 1.7.1 and shares.happy = 3 (didn't changed it after upgrade). So I did a little investigation and found the problem. It's immutable/repairer.py, line 60: ``` happy = upload.BaseUploadable.default_encoding_param_happy ``` Why do we use default happy here? It definitely should be read from the config. I didin't dig further but replaced it with ugly hack: ``` happy = 3 #upload.BaseUploadable.default_encoding_param_happy ``` ...and the problem has gone! Repairing works with just 6 servers online.
tahoe-lafs added the
c/code-network
p/major
t/defect
v/1.8β
labels 2010-09-25 14:22:55 +00:00
tahoe-lafs added this to the undecided milestone 2010-09-25 14:22:55 +00:00
kevan commented 2010-09-27 03:19:23 +00:00
Author
Owner

We eventually decided that this should be 0 when designing and implementing Servers of Happiness; see #778, around comment 45 for the discussion on that.

If no one has claimed this ticket by Tuesday, I'll fix it.

(I'm also setting the version to 1.8.0; if I understand your mailing list message, that's what your client is running. Feel free to change it back if I'm wrong. :-)

We eventually decided that this should be 0 when designing and implementing Servers of Happiness; see [#778, around comment 45](@@http://tahoe-lafs.org/trac/tahoe-lafs/ticket/778#[comment:-1](/tahoe-lafs/trac/issues/1212#issuecomment--1)@@) for the discussion on that. If no one has claimed this ticket by Tuesday, I'll fix it. (I'm also setting the version to 1.8.0; if I understand your mailing list message, that's what your client is running. Feel free to change it back if I'm wrong. :-)
tahoe-lafs added
v/1.8.0
and removed
v/1.8β
labels 2010-09-27 03:19:23 +00:00
eurekafag commented 2010-09-27 05:29:29 +00:00
Author
Owner

Yes, it's 1.8.0 but there was no such a version in a drop-down list when I created the ticket. This issue is important because it breaks the repair feature on small grids with <7 nodes online. And even more important because repairing is needed more often on such small networks.

Yes, it's 1.8.0 but there was no such a version in a drop-down list when I created the ticket. This issue is important because it breaks the repair feature on small grids with <7 nodes online. And even more important because repairing is needed more often on such small networks.
kevan commented 2010-09-28 04:21:33 +00:00
Author
Owner

Attachment 1212.dpatch (3258 bytes) added

**Attachment** 1212.dpatch (3258 bytes) added
kevan commented 2010-09-28 04:22:36 +00:00
Author
Owner

I think that the patch in attachment:1212.dpatch fixes this issue.

I think that the patch in [attachment:1212.dpatch](/tahoe-lafs/trac/attachments/000078ac-618d-cdfb-0cd9-5eeaa15c03d8) fixes this issue.
eurekafag commented 2010-09-28 07:28:07 +00:00
Author
Owner

Attachment 1212.patch (2810 bytes) added

**Attachment** 1212.patch (2810 bytes) added
2.7 KiB
eurekafag commented 2010-09-28 07:29:29 +00:00
Author
Owner

Why do you provide patches in dpatch format? It's debian only and I didn't find sources in google to compile it for OpenSuSE (which is RPM-based). So I patched by hand and here is unified diff patch: attachment:1212.patch

Why do you provide patches in dpatch format? It's debian only and I didn't find sources in google to compile it for [OpenSu](wiki/OpenSu)SE (which is RPM-based). So I patched by hand and here is unified diff patch: [attachment:1212.patch](/tahoe-lafs/trac/attachments/000078ac-618d-cdfb-0cd9-b3ba1ef420ff)
eurekafag commented 2010-09-28 07:46:25 +00:00
Author
Owner

Looks like this version works, thanks.

Looks like this version works, thanks.

Replying to [eurekafag]comment:6:

Why do you provide patches in dpatch format? It's debian only and I didn't find sources in google to compile it for OpenSuSE (which is RPM-based). So I patched by hand and here is unified diff patch: attachment:1212.patch

That was actually a darcs patch rather than a debian patch. I didn't realize that "dpatch" stood for "debian patch". I've updated wiki/Patches to suggest that people named their darcs patches thing.darcspatch.txt instead of thing.dpatch.

Replying to [eurekafag]comment:6: > Why do you provide patches in dpatch format? It's debian only and I didn't find sources in google to compile it for [OpenSu](wiki/OpenSu)SE (which is RPM-based). So I patched by hand and here is unified diff patch: [attachment:1212.patch](/tahoe-lafs/trac/attachments/000078ac-618d-cdfb-0cd9-b3ba1ef420ff) That was actually a *d*arcs patch rather than a *d*ebian patch. I didn't realize that "dpatch" stood for "debian patch". I've updated [wiki/Patches](wiki/Patches) to suggest that people named their darcs patches `thing.darcspatch.txt` instead of `thing.dpatch`.

Reviewed and applied in changeset:ec4f87a98c034dac, thanks!
By the way, I think we should do more work here. This patch corrects the regression from v1.7.1 to v1.8.0 (introduced in changeset:797828f47fe1aa44), in that v1.7.1 would repair with a servers-of-happiness (H) of 0 and v1.8.0 would repair with an H of 7. While I agree that this was a regression and that we should put it back to 0, I actually think that the old behavior of 0 was wrong and that we should have been using "currently configured H" instead!

That is: if you have configured your servers-of-happiness H to be 3, like eurekafag did, and the number of servers currently reachable on your grid is 2, and you do a repair, then I think the repair should stop with an explicit error message instead of proceeding and then giving you a report at the end that mentions (if you know how to read it) that it actually only put the shares onto 2 servers.

(In other words, I think I was wrong when I suggested letting repairer use H==0 in comment:373318. Or at least, what we did then was to keep the behavior of repairer from v1.6 when we made v1.7, but what I'm suggesting to do now is improve that behavior for the next release.)

Kevan, David-Sarah, Brian, eurekafag: do you agree? If so, let's open a new ticket saying to make the H used by repair be the same as the H that would be used by an upload. (Also in the new code we should make the H value be a parameter passed to the repairer instead of letting the repairer query the node-wide configuration. This is in keeping with CodingStandards regarding configuration and will facilitate some possible future work where people can pass explicit K, M, and H for a given upload or repair, e.g. as options to the tahoe put command line or optional fields in the WUI.

Reviewed and applied in changeset:ec4f87a98c034dac, thanks! By the way, I think we should do more work here. This patch corrects the regression from v1.7.1 to v1.8.0 (introduced in changeset:797828f47fe1aa44), in that v1.7.1 would repair with a servers-of-happiness (`H`) of `0` and v1.8.0 would repair with an `H` of `7`. While I agree that this was a regression and that we should put it back to `0`, I actually think that the old behavior of `0` was wrong and that we should have been using "currently configured `H`" instead! That is: if you have configured your servers-of-happiness `H` to be 3, like eurekafag did, and the number of servers currently reachable on your grid is 2, and you do a repair, then I think the repair should stop with an explicit error message instead of proceeding and then giving you a report at the end that mentions (if you know how to read it) that it actually only put the shares onto 2 servers. (In other words, I think I was wrong when I suggested letting repairer use `H==0` in [comment:373318](/tahoe-lafs/trac/issues/778#issuecomment-373318). Or at least, what we did then was to keep the behavior of repairer from v1.6 when we made v1.7, but what I'm suggesting to do now is improve that behavior for the next release.) Kevan, David-Sarah, Brian, eurekafag: do you agree? If so, let's open a new ticket saying to make the `H` used by repair be the same as the `H` that would be used by an upload. (Also in the new code we should make the `H` value be a parameter passed to the repairer instead of letting the repairer query the node-wide configuration. This is in keeping with [CodingStandards regarding configuration](wiki/CodingStandards#configuration) and will facilitate some possible future work where people can pass explicit `K`, `M`, and `H` for a given upload or repair, e.g. as options to the `tahoe put` command line or optional fields in the WUI.
zooko added the
r/fixed
label 2010-09-29 05:05:22 +00:00
zooko modified the milestone from undecided to soon 2010-09-29 05:05:22 +00:00
zooko closed this issue 2010-09-29 05:05:22 +00:00

Just testing syntax highlighting by uploading an attachment named "thing.darcspatch.txt"...

Just testing syntax highlighting by uploading an attachment named "thing.darcspatch.txt"...

Attachment 1212.darcspatch.txt (3258 bytes) added

**Attachment** 1212.darcspatch.txt (3258 bytes) added

Attachment 1212again.darcspatch.txt (3258 bytes) added

**Attachment** 1212again.darcspatch.txt (3258 bytes) added

Attachment 1212.darcs.patch (3258 bytes) added

**Attachment** 1212.darcs.patch (3258 bytes) added
eurekafag commented 2010-09-29 07:06:15 +00:00
Author
Owner

I do agree that zero happiness should be changed to H. There is no need to create a ticket because I've mentioned that already:
It definitely should be read from the config.
The temporary solution is nice but not complete to close this ticket.

I do agree that zero happiness should be changed to H. There is no need to create a ticket because I've mentioned that already: `It definitely should be read from the config.` The temporary solution is nice but not complete to close this ticket.
kevan commented 2010-09-29 20:12:23 +00:00
Author
Owner

If we do that, we lose the property that the repairer will always try to place whichever shares are missing onto some storage servers, even if the end result isn't optimally distributed.

If I have a cron job that does a deep repair of my rootcap, and the rootcap or some other important dircap or filecap only has k or k+1 shares available, and it is stored on a grid with a lot of churn, I probably care more about the fact that there are more than a few shares of that cap around than I do about where they are, and I certainly wouldn't want the repairer to not even bother generating new ones because it couldn't satisfy my distribution criteria; IOW, I'm better off with more shares that are poorly distributed than I am with no repair action (I'm oversimplifying, and it depends on the specific situation, but having more shares will make things better in some situations and generally won't make things worse, AFAICT without doing the math).

On the other hand, I think that the repairer should definitely tell the user whether the file is distributed correctly or not, and an exception message certainly does that. I can also make my node's repair go for broke with share regeneration by changing the value of happiness in tahoe.cfg to be 0. This is a chore, but it means that people who really want the repairer to try to place new shares regardless of where can still get that behavior.

Maybe the best approach is to fix #614 with this in mind. The repairer could regenerate and try to place all of the missing shares, as it does now, but also tell the caller (in the post repair results) whether the repair was ultimately successful or not based on how the shares are distributed, using the client's configured happiness value for that check.

Edit: I didn't read Zooko's comment closely enough. Is what I describe in the third paragraph what the repairer already does? If so, what don't you like about that?

If we do that, we lose the property that the repairer will always try to place whichever shares are missing onto *some* storage servers, even if the end result isn't optimally distributed. If I have a cron job that does a deep repair of my rootcap, and the rootcap or some other important dircap or filecap only has `k` or `k+1` shares available, and it is stored on a grid with a lot of churn, I probably care more about the fact that there are more than a few shares of that cap around than I do about where they are, and I certainly wouldn't want the repairer to not even bother generating new ones because it couldn't satisfy my distribution criteria; IOW, I'm better off with more shares that are poorly distributed than I am with no repair action (I'm oversimplifying, and it depends on the specific situation, but having more shares will make things better in some situations and generally won't make things worse, AFAICT without doing the math). On the other hand, I think that the repairer should definitely tell the user whether the file is distributed correctly or not, and an exception message certainly does that. I can also make my node's repair go for broke with share regeneration by changing the value of happiness in `tahoe.cfg` to be 0. This is a chore, but it means that people who really want the repairer to try to place new shares regardless of where can still get that behavior. Maybe the best approach is to fix #614 with this in mind. The repairer could regenerate and try to place all of the missing shares, as it does now, but also tell the caller (in the post repair results) whether the repair was ultimately successful or not based on how the shares are distributed, using the client's configured happiness value for that check. Edit: I didn't read Zooko's comment closely enough. Is what I describe in the third paragraph what the repairer already does? If so, what don't you like about that?

Replying to kevan:

If we do that, we lose the property that the repairer will always try to place whichever shares are missing onto some storage servers, even if the end result isn't optimally distributed.

Doesn't this mean that H is effectively 0 for you when you are doing this?

I can also make my node's repair go for broke with share regeneration by changing the value of happiness in tahoe.cfg to be 0. This is a chore, but it means that people who really want the repairer to try to place new shares regardless of where can still get that behavior.

Right. If you want this behavior, set H==0. If you want the other behavior (abort the repair) set H to something else. With the v1.7.1 behavior and the current trunk behavior (since 20100927200102-b8d28-9111a341188a4264e5070f91b52364a2addcb3dc), setting H in your tahoe.cfg has no effect on repairer behavior—repairer always acts as though H==0.

Maybe the best approach is to fix #614 with this in mind. The repairer could regenerate and try to place all of the missing shares, as it does now, but also tell the caller (in the post repair results) whether the repair was ultimately successful or not based on how the shares are distributed, using the client's configured happiness value for that check.

Oh, good catch. Yes, if we fix #614 then repairer would be using H (during the check/verify step) to determine whether or not to trigger a repair. Once it triggered the repairer, then it could also use H to determine whether to abort the repair, or it could instead treat H as effectively 0 for the purpose of the repair.

Now that I've thought about it more and read your comments, Kevan, I think I agree that we should have the latter behavior, as long as we fix #614 so that the output reported by the repairer can be easily understood by the user as indicating "unhealthy" when the servers of happiness is less than H.

Oh, in fact, what I really want is for repairer to proceed and to do its best even if it knows that it can't reach servers of happiness greater than or equal to H (instead of aborting the way uploader does), but then to return a failure result saying that it wasn't able to repair the file back to health.

Does that make sense?

Okay, I'm done changing my mind for the moment. What do you think?

Edit: I didn't read Zooko's comment closely enough. Is what I describe in the third paragraph what the repairer already does? If so, what don't you like about that?

Sorry: I don't understand this question. Hopefully I answered it above.

Replying to [kevan](/tahoe-lafs/trac/issues/1212#issuecomment-381270): > If we do that, we lose the property that the repairer will always try to place whichever shares are missing onto *some* storage servers, even if the end result isn't optimally distributed. Doesn't this mean that `H` is effectively `0` for you when you are doing this? > I can also make my node's repair go for broke with share regeneration by changing the value of happiness in `tahoe.cfg` to be 0. This is a chore, but it means that people who really want the repairer to try to place new shares regardless of where can still get that behavior. Right. If you want this behavior, set `H==0`. If you want the other behavior (abort the repair) set `H` to something else. With the v1.7.1 behavior and the current trunk behavior (since 20100927200102-b8d28-9111a341188a4264e5070f91b52364a2addcb3dc), setting `H` in your `tahoe.cfg` has no effect on repairer behavior—repairer always acts as though `H==0`. > Maybe the best approach is to fix #614 with this in mind. The repairer could regenerate and try to place all of the missing shares, as it does now, but also tell the caller (in the post repair results) whether the repair was ultimately successful or not based on how the shares are distributed, using the client's configured happiness value for that check. Oh, good catch. Yes, if we fix #614 then repairer would be using `H` (during the check/verify step) to determine whether or not to trigger a repair. Once it triggered the repairer, then it could *also* use `H` to determine whether to abort the repair, or it could instead treat `H` as effectively `0` for the purpose of the repair. Now that I've thought about it more and read your comments, Kevan, I think I agree that we should have the latter behavior, as long as we fix #614 so that the output reported by the repairer can be easily understood by the user as indicating "unhealthy" when the servers of happiness is less than `H`. Oh, in fact, what I *really* want is for repairer to *proceed* and to do its best even if it knows that it can't reach servers of happiness greater than or equal to `H` (instead of aborting the way uploader does), but then to return a failure result saying that it wasn't able to repair the file back to health. Does that make sense? Okay, I'm done changing my mind for the moment. What do you think? > Edit: I didn't read Zooko's comment closely enough. Is what I describe in the third paragraph what the repairer already does? If so, what don't you like about that? Sorry: I don't understand this question. Hopefully I answered it above.

This is how I think the repairer should work (I think this is violently agreeing with Zooko's comment:13, but with more detail):

  • let k and N be the shares-needed and total number of shares for this file, and let H be the happiness threshold read from tahoe.cfg.

  • if there are fewer than k connected servers, report that the repair failed completely.

  • construct a server map for this file by asking all connected servers which shares they have. (In the case of a mutable file, construct a server map for the latest retrievable version.)

  • construct a maximum matching M : server -> share, of size |M|, for this file (preferring to include servers that are earlier on the permuted list when there is a choice).

  • while |M| < N, and we have not tried to put shares on all connected servers:

    • pick a share not in M, and the server not in M that is next on the permuted list (wrapping around if necessary). Try to extend M by putting that share onto that server.
  • if |M| < k, report that the repair failed completely. If k <= |M| < H, report that the file is retrievable but unhealthy. In any case report what |M| is.

(The while loop should be done in parallel, with up to N - |M| outstanding requests.)

This is how I think the repairer should work (I *think* this is violently agreeing with Zooko's comment:13, but with more detail): * let k and N be the shares-needed and total number of shares for this file, and let H be the happiness threshold read from tahoe.cfg. * if there are fewer than k connected servers, report that the repair failed completely. * construct a server map for this file by asking all connected servers which shares they have. (In the case of a mutable file, construct a server map for the latest retrievable version.) * construct a maximum matching M : server -> share, of size |M|, for this file (preferring to include servers that are earlier on the permuted list when there is a choice). * while |M| < N, and we have not tried to put shares on all connected servers: * pick a share not in M, and the server not in M that is next on the permuted list (wrapping around if necessary). Try to extend M by putting that share onto that server. * if |M| < k, report that the repair failed completely. If k <= |M| < H, report that the file is retrievable but unhealthy. In any case report what |M| is. (The while loop should be done in parallel, with up to N - |M| outstanding requests.)
zooko removed the
r/fixed
label 2010-10-01 04:20:24 +00:00
zooko reopened this issue 2010-10-01 04:20:24 +00:00
zooko modified the milestone from soon to 1.8.1 2010-10-07 22:04:24 +00:00
kevan commented 2010-10-13 23:12:49 +00:00
Author
Owner

comment:381272 seems sensible to me.

I'm not sure if milestone 1.8.1 is about the little regression that I submitted a patch for, or the broader, likely to be fixed by #614 issue of how the repairer should work. If the latter, then 1.8.1 might be a little optimistic; fixing #614 correctly will require (unless I'm missing something obvious) a decent chunk of engineering, since the immutable file repairer is currently very simple. I would at least be more confident in my ability to get #614 done by 1.9.0 than by 1.8.1.

[comment:381272](/tahoe-lafs/trac/issues/1212#issuecomment-381272) seems sensible to me. I'm not sure if milestone 1.8.1 is about the little regression that I submitted a patch for, or the broader, likely to be fixed by #614 issue of how the repairer should work. If the latter, then 1.8.1 might be a little optimistic; fixing #614 correctly will require (unless I'm missing something obvious) a decent chunk of engineering, since the immutable file repairer is currently very simple. I would at least be more confident in my ability to get #614 done by 1.9.0 than by 1.8.1.

Replying to davidsarah:

This is how I think the repairer should work (I think this is violently agreeing with Zooko's comment:13, but with more detail):
...

  • if there are fewer than k connected servers, report that the repair failed completely.

Why this step?

Replying to [davidsarah](/tahoe-lafs/trac/issues/1212#issuecomment-381272): > This is how I think the repairer should work (I *think* this is violently agreeing with Zooko's comment:13, but with more detail): ... > * if there are fewer than k connected servers, report that the repair failed completely. Why this step?

Replying to [zooko]comment:19:

Replying to davidsarah:

This is how I think the repairer should work (I think this is violently agreeing with Zooko's comment:13, but with more detail):
...

  • if there are fewer than k connected servers, report that the repair failed completely.

Why this step?

Just a shortcut; this case would fail in the last step anyway.

Replying to davidsarah:
[...]

  • while |M| < N, and we have not tried to put shares on all connected servers:
    • pick a share not in M, and the server not in M that is next on the permuted list (wrapping around if necessary). Try to extend M by putting that share onto that server.

A small refinement of this step would be that once |M| >= H, we could allow placing the remaining N-H shares on servers that are already in the matching, if we're unable to place them on servers that are not in the matching.

Replying to [zooko]comment:19: > Replying to [davidsarah](/tahoe-lafs/trac/issues/1212#issuecomment-381272): > > This is how I think the repairer should work (I *think* this is violently agreeing with Zooko's comment:13, but with more detail): > ... > > * if there are fewer than k connected servers, report that the repair failed completely. > > Why this step? Just a shortcut; this case would fail in the last step anyway. Replying to [davidsarah](/tahoe-lafs/trac/issues/1212#issuecomment-381272): [...] > * while |M| < N, and we have not tried to put shares on all connected servers: > * pick a share not in M, and the server not in M that is next on the permuted list (wrapping around if necessary). Try to extend M by putting that share onto that server. A small refinement of this step would be that once |M| >= H, we could allow placing the remaining N-H shares on servers that are already in the matching, if we're unable to place them on servers that are not in the matching.

Replying to davidsarah:

  • if |M| < k, report that the repair failed completely. If k <= |M| < H, report that the file is retrievable but unhealthy. In any case report what |M| is.

Okay, but why this one?

We definitely need to classify health into several types: unrecoverable, 100% (|M|>=N), and servers-of-happiness-satisfying (|M|>=H) (needs a better name! "healthy" ?).

Do we also need another type to show that servers_of_happiness >= K?

Replying to [davidsarah](/tahoe-lafs/trac/issues/1212#issuecomment-381272): > * if |M| < k, report that the repair failed completely. If k <= |M| < H, report that the file is retrievable but unhealthy. In any case report what |M| is. Okay, but why this one? We definitely need to classify health into several types: *unrecoverable*, *100%* (`|M|`>=`N`), and *servers-of-happiness-satisfying* (`|M|`>=`H`) (needs a better name! *"healthy"* ?). Do we also need another type to show that `servers_of_happiness` >= `K`?

I think we should distinguish between levels of happiness at which the uploader or repairer will

  • (a) abort the upload or repair,
  • (b) treat the file as "unhealthy" which means
  • (b.1) a check-and-repair will trigger the repair step and
  • (b.2) a check results or a repair results will show the file (with clear, noticeable presentation) as being in danger ),
  • (c) treat the file as "perfectly healthy" which means that a repair would not make any change to it.

One of the questions in this ticket -- comment:381279 -- is whether (a) should trigger when |M| < K or not. Sometimes people would rather that the uploader/repairer get the file out there, even if all the shares are on a single server! Other times people might prefer that the uploader/repairer avoid wasting bandwidth on that and instead stop and raise the alarm.

#614 is all about whether (b) should trigger when |M| < N (current behavior) or |M| < H (proposed new behavior).

I think we should distinguish between levels of happiness at which the uploader or repairer will * (a) abort the upload or repair, * (b) treat the file as *"unhealthy"* which means * (b.1) a check-and-repair will trigger the repair step and * (b.2) a check results or a repair results will show the file (with clear, noticeable presentation) as being in danger ), * (c) treat the file as *"perfectly healthy"* which means that a repair would not make any change to it. One of the questions in this ticket -- [comment:381279](/tahoe-lafs/trac/issues/1212#issuecomment-381279) -- is whether (a) should trigger when `|M| < K` or not. Sometimes people would rather that the uploader/repairer get the file out there, even if all the shares are on a single server! Other times people might prefer that the uploader/repairer avoid wasting bandwidth on that and instead stop and raise the alarm. #614 is all about whether (b) should trigger when `|M| < N` (current behavior) or `|M| < H` (proposed new behavior).

Oh, what I just proposed in comment:381280 is a significant new behavior if we allow the level of happiness that triggers (a) to be different than the level of happiness that triggers (b)! Currently uploader/repairer aborts the upload or repair if it knows that it cannot achieve "health", i.e. |M| >= H. There are even unit tests to ensure that buildbot will go red if uploader/repairer proceeds to do an upload when it can't reach that level of happiness. :-)

Oh, what I just proposed in [comment:381280](/tahoe-lafs/trac/issues/1212#issuecomment-381280) is a significant new behavior if we allow the level of happiness that triggers (a) to be different than the level of happiness that triggers (b)! Currently uploader/repairer aborts the upload or repair if it knows that it cannot achieve *"health"*, i.e. `|M| >= H`. There are even unit tests to ensure that buildbot will go red if uploader/repairer proceeds to do an upload when it can't reach that level of happiness. :-)

Ah, I confusingly said "|M| < k" when I actually meant to say "the file is not retrievable". (It might be retrievable if there are >= k shares, but on less than k distinct servers.)

I think we should only abort a repair if the file is not retrievable (in which case we can't repair it anyway).

Ah, I confusingly said "|M| < k" when I actually meant to say "the file is not retrievable". (It might be retrievable if there are >= k shares, but on less than k distinct servers.) I think we should only *abort* a repair if the file is not retrievable (in which case we can't repair it anyway).

Hmm, why shouldn't a check-and-repair always try to restore a file to happiness N? The only reason I can think of is that it might result in redundant shares if there are a few servers that are sometimes disconnected, but wouldn't that tend to stablise after a few repair cycles?

Hmm, why shouldn't a check-and-repair always try to restore a file to happiness N? The only reason I can think of is that it might result in redundant shares if there are a few servers that are sometimes disconnected, but wouldn't that tend to stablise after a few repair cycles?

I guess something that I haven't made up my mind about yet is how repair jobs (either tahoe repair command on the cli or clicking on the "check-and-repair" button on the wui) should handle the case that the upload/repair fails, or partially fails on some of the files.

Should it proceed to completion, generate a report saying to what degree each attempt to repair a file succeeded, and exit with a "success" code (i.e. exit code 0 from tahoe repair), or should it abort the attempt to repair this one file, and should it also abort any other file repair attempts from the current deep-repair job?

For example, suppose you ask it to repair a single file with K=3, H=7, N=10, and it finds out that there are only two storage servers currently connected. One storage server has 3 shares and the other has 0. Then should it abort the upload immediately? Or should it upload a few shares (3?) to the second storage server which currently has none, and then report to you that the file is still unhealthy?

Here is one set of principles to answer this question (not sure if this is the best set):

  1. Idempotence if you run an upload-or-repair job, and it does some work (uploads some shares), and then you run it again when nothing has changed among the servers (there are no servers that joined or left and none of them acquired or lost shares), then the second run will not upload any shares.

  2. Forward progress if you run a repair job (not necessarily an upload job!), and it is possible for it to make |M| greater than it was before, then it will do so.

If we use these principles then we give up on an alternate principle:

  1. Network efficiency if you run an upload or repair job, and it is impossible for it to make |M| >= H, then it does not use any bulk network bandwidth. (Also, if it looks like it is possible at first, but after it has started uploading then one of the servers fails and it becomes impossible, then it aborts right then and does not use any more of your network bandwidth.)

I think people (including me) intuitively wanted principle 3 for uploads, but now that we are thinking about repairs instead of uploads we intuitively want principle 2.

I guess something that I haven't made up my mind about yet is how repair jobs (either `tahoe repair` command on the cli or clicking on the "check-and-repair" button on the wui) should handle the case that the upload/repair fails, or partially fails on some of the files. Should it proceed to completion, generate a report saying to what degree each attempt to repair a file succeeded, and exit with a "success" code (i.e. exit code 0 from `tahoe repair`), or should it abort the attempt to repair this one file, and should it also abort any other file repair attempts from the current deep-repair job? For example, suppose you ask it to repair a single file with `K=3, H=7, N=10`, and it finds out that there are only two storage servers currently connected. One storage server has 3 shares and the other has 0. Then should it abort the upload immediately? Or should it upload a few shares (3?) to the second storage server which currently has none, and then report to you that the file is still unhealthy? Here is one set of principles to answer this question (not sure if this is the best set): 1. *Idempotence* if you run an upload-or-repair job, and it does some work (uploads some shares), and then you run it again when nothing has changed among the servers (there are no servers that joined or left and none of them acquired or lost shares), then the second run will not upload any shares. 2. *Forward progress* if you run a repair job (not necessarily an upload job!), and it is possible for it to make `|M|` greater than it was before, then it will do so. If we use these principles then we give up on an alternate principle: 3. *Network efficiency* if you run an upload or repair job, and it is impossible for it to make `|M| >= H`, then it does not use any bulk network bandwidth. (Also, if it looks like it is possible at first, but after it has started uploading then one of the servers fails and it becomes impossible, then it aborts right then and does not use any *more* of your network bandwidth.) I think people (including me) intuitively wanted principle 3 for uploads, but now that we are thinking about repairs instead of uploads we intuitively want principle 2.

One possibility would be to make the behavior of uploader different than of repairer. Perhaps people prefer for their initial uploads to fail quickly and network-efficiently (principle 3) if they won't be able to achieve happiness level of H, but prefer for their repairs to proceed and do their best (principle 2). However, making the two behave differently would make things more complicated in the source code and would also make things more complicated in the usage, because principle 1 -- idempotence -- would not apply to "first upload and then repair" or "first repair and then upload". Sometimes an upload would abort itself and return failure but then a subsequent repair would do a lot of work to make progress, or a repair would do a lot of work to make progress but then an upload would abort itself and return failure.

Unless we are really sure that we need to support two different modes, I would prefer to err on the side of simplicity and find a mode that is good enough for both upload and repair. One good way to estimate "complication in usage" is to think how much documentation we would need to write to explain the different behavior of upload and repair in the different cases. :-)

One possibility would be to make the behavior of uploader different than of repairer. Perhaps people prefer for their initial uploads to fail quickly and network-efficiently (principle 3) if they won't be able to achieve happiness level of `H`, but prefer for their repairs to proceed and do their best (principle 2). However, making the two behave differently would make things more complicated in the source code and would also make things more complicated in the usage, because principle 1 -- *idempotence* -- would not apply to "first upload and then repair" or "first repair and then upload". Sometimes an upload would abort itself and return failure but then a subsequent repair would do a lot of work to make progress, or a repair would do a lot of work to make progress but then an upload would abort itself and return failure. Unless we are really sure that we need to support two different modes, I would prefer to err on the side of simplicity and find a mode that is good enough for both upload and repair. One good way to estimate "complication in usage" is to think how much documentation we would need to write to explain the different behavior of upload and repair in the different cases. :-)

Replying to zooko:

One possibility would be to make the behavior of uploader different than of repairer. Perhaps people prefer for their initial uploads to fail quickly and network-efficiently (principle 3) if they won't be able to achieve happiness level of H, but prefer for their repairs to proceed and do their best (principle 2). However, making the two behave differently would make things more complicated in the source code and would also make things more complicated in the usage, because principle 1 -- idempotence -- would not apply to "first upload and then repair" or "first repair and then upload". Sometimes an upload would abort itself and return failure but then a subsequent repair would do a lot of work to make progress, or a repair would do a lot of work to make progress but then an upload would abort itself and return failure.

Unless we are really sure that we need to support two different modes, I would prefer to err on the side of simplicity and find a mode that is good enough for both upload and repair.

I'm not sure that two different modes would add much complexity. Almost all of the code would be shared, and the upload/repair flag would just enable the fast abort in the upload case.

Replying to [zooko](/tahoe-lafs/trac/issues/1212#issuecomment-381285): > One possibility would be to make the behavior of uploader different than of repairer. Perhaps people prefer for their initial uploads to fail quickly and network-efficiently (principle 3) if they won't be able to achieve happiness level of `H`, but prefer for their repairs to proceed and do their best (principle 2). However, making the two behave differently would make things more complicated in the source code and would also make things more complicated in the usage, because principle 1 -- *idempotence* -- would not apply to "first upload and then repair" or "first repair and then upload". Sometimes an upload would abort itself and return failure but then a subsequent repair would do a lot of work to make progress, or a repair would do a lot of work to make progress but then an upload would abort itself and return failure. > > Unless we are really sure that we need to support two different modes, I would prefer to err on the side of simplicity and find a mode that is good enough for both upload and repair. I'm not sure that two different modes would add much complexity. Almost all of the code would be shared, and the upload/repair flag would just enable the fast abort in the upload case.

At some point, possibly in email to tahoe-dev, davidsarah convinced me that two modes was appropriate because people who are uploading a file are not yet committed to the file being up, so it is better for them to abort in case of unsatisfying distribution, but people who are repairing an existing file are already committed to the file being out there, so it is better for them to do your best to make some improvement even in case of unsatisfying distribution.

At some point, possibly in email to tahoe-dev, davidsarah convinced me that two modes was appropriate because people who are uploading a file are not yet committed to the file being up, so it is better for them to abort in case of unsatisfying distribution, but people who are repairing an existing file are already committed to the file being out there, so it is better for them to do your best to make some improvement even in case of unsatisfying distribution.

Okay, I've now re-read this long, confusing ticket and I now agree that the patch Kevan already applied to make H be 0 during repair is correct. This means that repair processes always try to make progress (principle 2 from comment:381284) instead of trying to conserve network bandwidth (principle 3 from comment:381284), but upload processes (which aren't repairs) choose principle 3 instead of principle 2.

Also, yes, we really ought to fix #614 by defining healthy as "satisfying the servers-of-happiness level that my user has chosen". :-)

I don't think there's anything else to do but add a source:NEWS entry and then we can close this ticket. Does anyone else who is reading this agree?

Okay, I've now re-read this long, confusing ticket and I now agree that the patch Kevan already applied to make `H` be `0` during repair is correct. This means that repair processes always try to make progress (principle 2 from [comment:381284](/tahoe-lafs/trac/issues/1212#issuecomment-381284)) instead of trying to conserve network bandwidth (principle 3 from [comment:381284](/tahoe-lafs/trac/issues/1212#issuecomment-381284)), but upload processes (which aren't repairs) choose principle 3 instead of principle 2. Also, yes, we really ought to fix #614 by defining `healthy` as "satisfying the servers-of-happiness level that my user has chosen". :-) I don't think there's anything else to do but add a source:NEWS entry and then we can close this ticket. Does anyone else who is reading this agree?

Replying to zooko:

Okay, I've now re-read this long, confusing ticket and I now agree that the patch Kevan already applied to make H be 0 during repair is correct. This means that repair processes always try to make progress (principle 2 from comment:381284) instead of trying to conserve network bandwidth (principle 3 from comment:381284), but upload processes (which aren't repairs) choose principle 3 instead of principle 2.

Also, yes, we really ought to fix #614 by defining healthy as "satisfying the servers-of-happiness level that my user has chosen". :-)

I don't think there's anything else to do but add a source:NEWS entry and then we can close this ticket. Does anyone else who is reading this agree?

Yes. There are still things we want to fix about repair (at least #614, #1124, and giving more complete information about the health of a file after repair), but let's address those for v1.9.0.

Replying to [zooko](/tahoe-lafs/trac/issues/1212#issuecomment-381288): > Okay, I've now re-read this long, confusing ticket and I now agree that the patch Kevan already applied to make `H` be `0` during repair is correct. This means that repair processes always try to make progress (principle 2 from [comment:381284](/tahoe-lafs/trac/issues/1212#issuecomment-381284)) instead of trying to conserve network bandwidth (principle 3 from [comment:381284](/tahoe-lafs/trac/issues/1212#issuecomment-381284)), but upload processes (which aren't repairs) choose principle 3 instead of principle 2. > > Also, yes, we really ought to fix #614 by defining `healthy` as "satisfying the servers-of-happiness level that my user has chosen". :-) > > I don't think there's anything else to do but add a source:NEWS entry and then we can close this ticket. Does anyone else who is reading this agree? Yes. There are still things we want to fix about repair (at least #614, #1124, and giving more complete information about the health of a file after repair), but let's address those for v1.9.0.

In changeset:cb764da0edc2b161:

NEWS: entries for #1190 and #1212, and minor cleanups. refs #1190, #1212
In changeset:cb764da0edc2b161: ``` NEWS: entries for #1190 and #1212, and minor cleanups. refs #1190, #1212 ```
daira added the
r/fixed
label 2010-11-02 02:29:42 +00:00
daira closed this issue 2010-11-02 02:29:42 +00:00

Replying to davidsarah:

  • construct a maximum matching M : server -> share, of size |M|, for this file (preferring to include servers that are earlier on the permuted list when there is a choice).

and preferring to include servers that have least available space (especially those that are full), since that will allow uploads to succeed in more cases by placing new shares on servers that have available space.

Replying to [davidsarah](/tahoe-lafs/trac/issues/1212#issuecomment-381272): > * construct a maximum matching M : server -> share, of size |M|, for this file (preferring to include servers that are earlier on the permuted list when there is a choice). and preferring to include servers that have least available space (especially those that are full), since that will allow uploads to succeed in more cases by placing new shares on servers that have available space.

Diego "sickness" Righi is dissatisfied with this solution. He has 10 storage servers, and sets M=10 and H=10. His desire is that he never gets more than one share on one storage server. Current uploader does what he wants -- it never places more than one share on one storage server. But repairer does what he doesn't want -- if fewer than 10 storage servers are available then repairer uploads extra shares to some of the available servers.

To my way of thinking, uploading extra shares is making the file more available. For example if you have 8 servers with 1 share each and 1 server is 2 shares (and K=5), then if you lost the first five of your servers (each of which had 1 share) you could still recover your file from the remaining four servers. If instead you have 9 servers with one share each, K=5, and you lost the first five of your servers then the file would be lost.

So, now I'm going to stop here and ask sickness: does this cause you to change your mind so that now you want repairer to upload a second share to one of the existing servers in the case that there are only 9 servers available? Or do you still prefer that it should not do that?

Diego "sickness" Righi is dissatisfied with this solution. He has 10 storage servers, and sets M=10 and H=10. His desire is that he never gets more than one share on one storage server. Current uploader does what he wants -- it never places more than one share on one storage server. But repairer does what he doesn't want -- if fewer than 10 storage servers are available then repairer uploads extra shares to some of the available servers. To my way of thinking, uploading extra shares is making the file more available. For example if you have 8 servers with 1 share each and 1 server is 2 shares (and `K=5`), then if you lost the first five of your servers (each of which had 1 share) you could still recover your file from the remaining four servers. If instead you have 9 servers with one share each, `K=5`, and you lost the first five of your servers then the file would be lost. So, now I'm going to stop here and ask sickness: does this cause you to change your mind so that now you *want* repairer to upload a second share to one of the existing servers in the case that there are only 9 servers available? Or do you still prefer that it should not do that?
zooko removed the
r/fixed
label 2011-06-08 13:52:15 +00:00
zooko reopened this issue 2011-06-08 13:52:15 +00:00
Author
Owner

I think sickness's desire for not having more than one share on a server blurs two things. One is having adequate redundancy, and a behavior of adding shares s.t. a server has two (in the 9 servers present case) helps. But, when the 10th server is back on line, if it doesn't have a share, then repair should consider the file unhealthy and place a share on the 10th server such that 10 servers have a distinct share. Further the lease on the extra share probably shouldn't be renewed.

If sickness also desires some form of storage efficiency, to avoid placing the 2nd share, then I think it's a misuse of servers-of-happiness and there should be some max-shares-per-server config, defaulting to infinite.

This all becomes difficult in the middle, when you have a 3/10 encoding and 3 or 4 servers. You want to set H to 3 or 4, but a share distribution of 7/1/1/1 isn't really ok - you want it to be more balanced. But I think we should figure out if this is a a reliability concern or an efficiency concern and treat them separately.

I think sickness's desire for not having more than one share on a server blurs two things. One is having adequate redundancy, and a behavior of adding shares s.t. a server has two (in the 9 servers present case) helps. But, when the 10th server is back on line, if it doesn't have a share, then repair should consider the file unhealthy and place a share on the 10th server such that 10 servers have a distinct share. Further the lease on the extra share probably shouldn't be renewed. If sickness also desires some form of storage efficiency, to avoid placing the 2nd share, then I think it's a misuse of servers-of-happiness and there should be some max-shares-per-server config, defaulting to infinite. This all becomes difficult in the middle, when you have a 3/10 encoding and 3 or 4 servers. You want to set H to 3 or 4, but a share distribution of 7/1/1/1 isn't really ok - you want it to be more balanced. But I think we should figure out if this is a a reliability concern or an efficiency concern and treat them separately.
Author
Owner

Sorry, I didn't mean sickness's desire was blurry. I meant that on reading it, there are two issues possibly behind it, and we should be clear on which we are addressing and why.

Sorry, I didn't mean sickness's desire was blurry. I meant that on reading it, there are two issues possibly behind it, and we should be clear on which we are addressing and why.

Replying to zooko:

Diego "sickness" Righi is dissatisfied with this solution.

The original problem in this ticket was that the repairer was using the default value for happiness, which was certainly wrong. Let's not overload the ticket; sickness' complaint is that the current repairer often places shares in a way that doesn't increase happiness, when another different placement of the same number of shares would have done so. That's covered by #1130.

Replying to [zooko](/tahoe-lafs/trac/issues/1212#issuecomment-381294): > Diego "sickness" Righi is dissatisfied with this solution. The original problem in this ticket was that the repairer was using the default value for happiness, which was certainly wrong. Let's not overload the ticket; sickness' complaint is that the current repairer often places shares in a way that doesn't increase happiness, when another different placement of the same number of shares would have done so. That's covered by #1130.
daira added the
r/fixed
label 2011-06-09 00:13:16 +00:00
daira closed this issue 2011-06-09 00:13:16 +00:00

Hm, perhaps we should take this to tahoe-dev. Because I don't think that is sickness's complaint--I think his complaint is that it uploads more than one share to a server. I'll try to write a post for tahoe-dev.

Hm, perhaps we should take this to tahoe-dev. Because I don't think that is sickness's complaint--I think his complaint is that it uploads more than one share to a server. I'll try to write a post for tahoe-dev.
kmarkley86 commented 2011-12-22 02:12:16 +00:00
Author
Owner

Reopening this ticket. I'm affected by the same fundamental problem, but by a different path. The fix identified earlier was to immutable/repairer.py, but I'm getting an error from immutable/upload.py.

Scenario:
I'm using 2-of-4 encoding with shares.happy=4 on tahoe 1.8.1. From the CLI I do a tahoe check --repair on a file with shares {0, 2, 3} already existing on the grid but share 1 not existing, and I get an UploadUnhappinessError complaining that "we were asked to place shares on at least 7" servers. There are only 4 servers on my grid -- hence my choice of shares.happy=4.

I observed that in immutable/upload.py, BaseUploadable has a statement "default_encoding_param_happy = 7". I tried the experiment of changing this value to 4 (the shares.happy value in my tahoe.cfg) and then the repair succeeds without error.

So there must be a path through this code where the default_encoding_param_happy value is actually used instead of being overridden by the value in tahoe.cfg. (I think it smells a little that this object has defaults at all, instead of requiring the parameters to be provided.)

Reopening this ticket. I'm affected by the same fundamental problem, but by a different path. The fix identified earlier was to immutable/repairer.py, but I'm getting an error from immutable/upload.py. Scenario: I'm using 2-of-4 encoding with shares.happy=4 on tahoe 1.8.1. From the CLI I do a tahoe check --repair on a file with shares {0, 2, 3} already existing on the grid but share 1 not existing, and I get an [UploadUnhappinessError](wiki/UploadUnhappinessError) complaining that "we were asked to place shares on at least 7" servers. There are only 4 servers on my grid -- hence my choice of shares.happy=4. I observed that in immutable/upload.py, [BaseUploadable](wiki/BaseUploadable) has a statement "default_encoding_param_happy = 7". I tried the experiment of changing this value to 4 (the shares.happy value in my tahoe.cfg) and then the repair succeeds without error. So there must be a path through this code where the default_encoding_param_happy value is actually used instead of being overridden by the value in tahoe.cfg. (I think it smells a little that this object has defaults at all, instead of requiring the parameters to be provided.)
tahoe-lafs removed the
r/fixed
label 2011-12-22 02:12:16 +00:00
kmarkley86 reopened this issue 2011-12-22 02:12:16 +00:00

Since this is a regression, I think we should consider trying to fix it for Tahoe-LAFS v1.9.1. Advice and help would be appreciated...

Since this is a regression, I think we should consider trying to fix it for Tahoe-LAFS v1.9.1. Advice and help would be appreciated...
zooko modified the milestone from 1.8.1 to 1.9.1 2011-12-29 00:25:37 +00:00

Please note that the scope of this ticket is just the fact that immutable/upload.py is incorrectly using default_encoding_param_happy = 7. As far as I know, we're not trying either to fix #1130 or to apply the refactoring/improvements to share placement in #1382 in Tahoe-LAFS 1.9.1.

Please note that the scope of *this* ticket is just the fact that immutable/upload.py is incorrectly using `default_encoding_param_happy = 7`. As far as I know, we're not trying either to fix #1130 or to apply the refactoring/improvements to share placement in #1382 in Tahoe-LAFS 1.9.1.
kevan commented 2012-01-14 19:41:59 +00:00
Author
Owner

kmarkley86: A stack trace would help me fix this. Could you provide one?

kmarkley86: A stack trace would help me fix this. Could you provide one?

The problem described in comment:381303 is critical to fix for v1.9.2 (or 1.10.0 if we decide to call it that; the next release, anyway).

The problem described in [comment:381303](/tahoe-lafs/trac/issues/1212#issuecomment-381303) is critical to fix for v1.9.2 (or 1.10.0 if we decide to call it that; the next release, anyway).
daira added
p/critical
and removed
p/major
labels 2012-03-31 23:15:25 +00:00

Oh wait, no hold on--this is a php script? No php on tahoe-lafs.org! Sorry.

Oh wait, no hold on--this is a php script? No php on tahoe-lafs.org! Sorry.

Wrong ticket. (should have been #1417)

Wrong ticket. (should have been #1417)
warner was unassigned by daira 2012-06-12 16:50:42 +00:00
daira self-assigned this 2012-06-12 16:50:42 +00:00
daira changed title from Repairing fails if less than 7 servers available to Upload (sometimes?) ignores shares.happy in tahoe.cfg 2012-06-15 18:56:43 +00:00

In changeset:196bd583b6c4959c:

Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212
In changeset:196bd583b6c4959c: ``` Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212 ```
david-sarah <david-sarah@jacaranda.org> commented 2012-06-19 03:25:10 +00:00
Author
Owner

In changeset:196bd583b6c4959c:

Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212
In changeset:196bd583b6c4959c: ``` Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212 ```

kmarkley86: can you try again to reproduce the problem in comment:381303 using trunk?

kmarkley86: can you try again to reproduce the problem in [comment:381303](/tahoe-lafs/trac/issues/1212#issuecomment-381303) using trunk?

In changeset:5521/1.9.2:

Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212
In changeset:5521/1.9.2: ``` Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212 ```

In changeset:5522/1.9.2:

Rollback last patch, which  was pushed unintentionally to the 1.9.2 branch (Add assertions to make sure that set_default_encoding_parameters is always called...) refs #1212
In changeset:5522/1.9.2: ``` Rollback last patch, which was pushed unintentionally to the 1.9.2 branch (Add assertions to make sure that set_default_encoding_parameters is always called...) refs #1212 ```

We decided to defer actually fixing the bug (if it still exists) to 1.10.

We decided to defer actually fixing the bug (if it still exists) to 1.10.
daira modified the milestone from 1.9.2 to 1.10.0 2012-06-21 20:51:13 +00:00

In changeset:5883/cloud-backend:

[rebased for cloud-backend] Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212
In changeset:5883/cloud-backend: ``` [rebased for cloud-backend] Add assertions to make sure that set_default_encoding_parameters is always called, rather than using hardcoded 3/7/10 defaults. Also update affected tests. Note that this by itself cannot fix the bug mentioned in ticket:1212#comment:41, but it might make it easier to reproduce. refs #1212 ```

Kyle: this ticket is blocked on you attempting to reproduce comment:381303 using the new code, which has assertions that will let us learn more about the bug.

Kyle: this ticket is blocked on you attempting to reproduce [comment:381303](/tahoe-lafs/trac/issues/1212#issuecomment-381303) using the new code, which has assertions that will let us learn more about the bug.

Moved to #1830. The original problem was fixed in 1.8.1 I think. See #1130 and #1382 for other improvements to share placement and servers-of-happiness.

Moved to #1830. The original problem was fixed in 1.8.1 I think. See #1130 and #1382 for other improvements to share placement and servers-of-happiness.
daira added the
r/fixed
label 2012-10-26 02:50:29 +00:00
daira modified the milestone from 1.10.0 to 1.8.1 2012-10-26 02:50:29 +00:00
daira closed this issue 2012-10-26 02:50:29 +00:00
daira changed title from Upload (sometimes?) ignores shares.happy in tahoe.cfg to Repair used default shares.happy 2012-10-26 02:54:56 +00:00

There was discussion of this issue on tahoe-dev: [//pipermail/tahoe-dev/2013-March/008091.html]

There was discussion of this issue on tahoe-dev: [//pipermail/tahoe-dev/2013-March/008091.html]

Replying to zooko:

There was discussion of this issue on tahoe-dev: [//pipermail/tahoe-dev/2013-March/008091.html]

I'm sure that's not the same issue (nor is it the same issue as #1830).

Replying to [zooko](/tahoe-lafs/trac/issues/1212#issuecomment-381332): > There was discussion of this issue on tahoe-dev: [//pipermail/tahoe-dev/2013-March/008091.html] I'm sure that's not the same issue (nor is it the same issue as #1830).
Sign in to join this conversation.
No labels
c/code
c/code-dirnodes
c/code-encoding
c/code-frontend
c/code-frontend-cli
c/code-frontend-ftp-sftp
c/code-frontend-magic-folder
c/code-frontend-web
c/code-mutable
c/code-network
c/code-nodeadmin
c/code-peerselection
c/code-storage
c/contrib
c/dev-infrastructure
c/docs
c/operational
c/packaging
c/unknown
c/website
kw:2pc
kw:410
kw:9p
kw:ActivePerl
kw:AttributeError
kw:DataUnavailable
kw:DeadReferenceError
kw:DoS
kw:FileZilla
kw:GetLastError
kw:IFinishableConsumer
kw:K
kw:LeastAuthority
kw:Makefile
kw:RIStorageServer
kw:StringIO
kw:UncoordinatedWriteError
kw:about
kw:access
kw:access-control
kw:accessibility
kw:accounting
kw:accounting-crawler
kw:add-only
kw:aes
kw:aesthetics
kw:alias
kw:aliases
kw:aliens
kw:allmydata
kw:amazon
kw:ambient
kw:annotations
kw:anonymity
kw:anonymous
kw:anti-censorship
kw:api_auth_token
kw:appearance
kw:appname
kw:apport
kw:archive
kw:archlinux
kw:argparse
kw:arm
kw:assertion
kw:attachment
kw:auth
kw:authentication
kw:automation
kw:avahi
kw:availability
kw:aws
kw:azure
kw:backend
kw:backoff
kw:backup
kw:backupdb
kw:backward-compatibility
kw:bandwidth
kw:basedir
kw:bayes
kw:bbfreeze
kw:beta
kw:binaries
kw:binutils
kw:bitcoin
kw:bitrot
kw:blacklist
kw:blocker
kw:blocks-cloud-deployment
kw:blocks-cloud-merge
kw:blocks-magic-folder-merge
kw:blocks-merge
kw:blocks-raic
kw:blocks-release
kw:blog
kw:bom
kw:bonjour
kw:branch
kw:branding
kw:breadcrumbs
kw:brians-opinion-needed
kw:browser
kw:bsd
kw:build
kw:build-helpers
kw:buildbot
kw:builders
kw:buildslave
kw:buildslaves
kw:cache
kw:cap
kw:capleak
kw:captcha
kw:cast
kw:centos
kw:cffi
kw:chacha
kw:charset
kw:check
kw:checker
kw:chroot
kw:ci
kw:clean
kw:cleanup
kw:cli
kw:cloud
kw:cloud-backend
kw:cmdline
kw:code
kw:code-checks
kw:coding-standards
kw:coding-tools
kw:coding_tools
kw:collection
kw:compatibility
kw:completion
kw:compression
kw:confidentiality
kw:config
kw:configuration
kw:configuration.txt
kw:conflict
kw:connection
kw:connectivity
kw:consistency
kw:content
kw:control
kw:control.furl
kw:convergence
kw:coordination
kw:copyright
kw:corruption
kw:cors
kw:cost
kw:coverage
kw:coveralls
kw:coveralls.io
kw:cpu-watcher
kw:cpyext
kw:crash
kw:crawler
kw:crawlers
kw:create-container
kw:cruft
kw:crypto
kw:cryptography
kw:cryptography-lib
kw:cryptopp
kw:csp
kw:curl
kw:cutoff-date
kw:cycle
kw:cygwin
kw:d3
kw:daemon
kw:darcs
kw:darcsver
kw:database
kw:dataloss
kw:db
kw:dead-code
kw:deb
kw:debian
kw:debug
kw:deep-check
kw:defaults
kw:deferred
kw:delete
kw:deletion
kw:denial-of-service
kw:dependency
kw:deployment
kw:deprecation
kw:desert-island
kw:desert-island-build
kw:design
kw:design-review-needed
kw:detection
kw:dev-infrastructure
kw:devpay
kw:directory
kw:directory-page
kw:dirnode
kw:dirnodes
kw:disconnect
kw:discovery
kw:disk
kw:disk-backend
kw:distribute
kw:distutils
kw:dns
kw:do_http
kw:doc-needed
kw:docker
kw:docs
kw:docs-needed
kw:dokan
kw:dos
kw:download
kw:downloader
kw:dragonfly
kw:drop-upload
kw:duplicity
kw:dusty
kw:earth-dragon
kw:easy
kw:ec2
kw:ecdsa
kw:ed25519
kw:egg-needed
kw:eggs
kw:eliot
kw:email
kw:empty
kw:encoding
kw:endpoint
kw:enterprise
kw:enum34
kw:environment
kw:erasure
kw:erasure-coding
kw:error
kw:escaping
kw:etag
kw:etch
kw:evangelism
kw:eventual
kw:example
kw:excess-authority
kw:exec
kw:exocet
kw:expiration
kw:extensibility
kw:extension
kw:failure
kw:fedora
kw:ffp
kw:fhs
kw:figleaf
kw:file
kw:file-descriptor
kw:filename
kw:filesystem
kw:fileutil
kw:fips
kw:firewall
kw:first
kw:floatingpoint
kw:flog
kw:foolscap
kw:forward-compatibility
kw:forward-secrecy
kw:forwarding
kw:free
kw:freebsd
kw:frontend
kw:fsevents
kw:ftp
kw:ftpd
kw:full
kw:furl
kw:fuse
kw:garbage
kw:garbage-collection
kw:gateway
kw:gatherer
kw:gc
kw:gcc
kw:gentoo
kw:get
kw:git
kw:git-annex
kw:github
kw:glacier
kw:globalcaps
kw:glossary
kw:google-cloud-storage
kw:google-drive-backend
kw:gossip
kw:governance
kw:grid
kw:grid-manager
kw:gridid
kw:gridsync
kw:grsec
kw:gsoc
kw:gvfs
kw:hackfest
kw:hacktahoe
kw:hang
kw:hardlink
kw:heartbleed
kw:heisenbug
kw:help
kw:helper
kw:hint
kw:hooks
kw:how
kw:how-to
kw:howto
kw:hp
kw:hp-cloud
kw:html
kw:http
kw:https
kw:i18n
kw:i2p
kw:i2p-collab
kw:illustration
kw:image
kw:immutable
kw:impressions
kw:incentives
kw:incident
kw:init
kw:inlineCallbacks
kw:inotify
kw:install
kw:installer
kw:integration
kw:integration-test
kw:integrity
kw:interactive
kw:interface
kw:interfaces
kw:interoperability
kw:interstellar-exploration
kw:introducer
kw:introduction
kw:iphone
kw:ipkg
kw:iputil
kw:ipv6
kw:irc
kw:jail
kw:javascript
kw:joke
kw:jquery
kw:json
kw:jsui
kw:junk
kw:key-value-store
kw:kfreebsd
kw:known-issue
kw:konqueror
kw:kpreid
kw:kvm
kw:l10n
kw:lae
kw:large
kw:latency
kw:leak
kw:leasedb
kw:leases
kw:libgmp
kw:license
kw:licenss
kw:linecount
kw:link
kw:linux
kw:lit
kw:localhost
kw:location
kw:locking
kw:logging
kw:logo
kw:loopback
kw:lucid
kw:mac
kw:macintosh
kw:magic-folder
kw:manhole
kw:manifest
kw:manual-test-needed
kw:map
kw:mapupdate
kw:max_space
kw:mdmf
kw:memcheck
kw:memory
kw:memory-leak
kw:mesh
kw:metadata
kw:meter
kw:migration
kw:mime
kw:mingw
kw:minimal
kw:misc
kw:miscapture
kw:mlp
kw:mock
kw:more-info-needed
kw:mountain-lion
kw:move
kw:multi-users
kw:multiple
kw:multiuser-gateway
kw:munin
kw:music
kw:mutability
kw:mutable
kw:mystery
kw:names
kw:naming
kw:nas
kw:navigation
kw:needs-review
kw:needs-spawn
kw:netbsd
kw:network
kw:nevow
kw:new-user
kw:newcaps
kw:news
kw:news-done
kw:news-needed
kw:newsletter
kw:newurls
kw:nfc
kw:nginx
kw:nixos
kw:no-clobber
kw:node
kw:node-url
kw:notification
kw:notifyOnDisconnect
kw:nsa310
kw:nsa320
kw:nsa325
kw:numpy
kw:objects
kw:old
kw:openbsd
kw:openitp-packaging
kw:openssl
kw:openstack
kw:opensuse
kw:operation-helpers
kw:operational
kw:operations
kw:ophandle
kw:ophandles
kw:ops
kw:optimization
kw:optional
kw:options
kw:organization
kw:os
kw:os.abort
kw:ostrom
kw:osx
kw:osxfuse
kw:otf-magic-folder-objective1
kw:otf-magic-folder-objective2
kw:otf-magic-folder-objective3
kw:otf-magic-folder-objective4
kw:otf-magic-folder-objective5
kw:otf-magic-folder-objective6
kw:p2p
kw:packaging
kw:partial
kw:password
kw:path
kw:paths
kw:pause
kw:peer-selection
kw:performance
kw:permalink
kw:permissions
kw:persistence
kw:phone
kw:pickle
kw:pip
kw:pipermail
kw:pkg_resources
kw:placement
kw:planning
kw:policy
kw:port
kw:portability
kw:portal
kw:posthook
kw:pratchett
kw:preformance
kw:preservation
kw:privacy
kw:process
kw:profile
kw:profiling
kw:progress
kw:proxy
kw:publish
kw:pyOpenSSL
kw:pyasn1
kw:pycparser
kw:pycrypto
kw:pycrypto-lib
kw:pycryptopp
kw:pyfilesystem
kw:pyflakes
kw:pylint
kw:pypi
kw:pypy
kw:pysqlite
kw:python
kw:python3
kw:pythonpath
kw:pyutil
kw:pywin32
kw:quickstart
kw:quiet
kw:quotas
kw:quoting
kw:raic
kw:rainhill
kw:random
kw:random-access
kw:range
kw:raspberry-pi
kw:reactor
kw:readonly
kw:rebalancing
kw:recovery
kw:recursive
kw:redhat
kw:redirect
kw:redressing
kw:refactor
kw:referer
kw:referrer
kw:regression
kw:rekey
kw:relay
kw:release
kw:release-blocker
kw:reliability
kw:relnotes
kw:remote
kw:removable
kw:removable-disk
kw:rename
kw:renew
kw:repair
kw:replace
kw:report
kw:repository
kw:research
kw:reserved_space
kw:response-needed
kw:response-time
kw:restore
kw:retrieve
kw:retry
kw:review
kw:review-needed
kw:reviewed
kw:revocation
kw:roadmap
kw:rollback
kw:rpm
kw:rsa
kw:rss
kw:rst
kw:rsync
kw:rusty
kw:s3
kw:s3-backend
kw:s3-frontend
kw:s4
kw:same-origin
kw:sandbox
kw:scalability
kw:scaling
kw:scheduling
kw:schema
kw:scheme
kw:scp
kw:scripts
kw:sdist
kw:sdmf
kw:security
kw:self-contained
kw:server
kw:servermap
kw:servers-of-happiness
kw:service
kw:setup
kw:setup.py
kw:setup_requires
kw:setuptools
kw:setuptools_darcs
kw:sftp
kw:shared
kw:shareset
kw:shell
kw:signals
kw:simultaneous
kw:six
kw:size
kw:slackware
kw:slashes
kw:smb
kw:sneakernet
kw:snowleopard
kw:socket
kw:solaris
kw:space
kw:space-efficiency
kw:spam
kw:spec
kw:speed
kw:sqlite
kw:ssh
kw:ssh-keygen
kw:sshfs
kw:ssl
kw:stability
kw:standards
kw:start
kw:startup
kw:static
kw:static-analysis
kw:statistics
kw:stats
kw:stats_gatherer
kw:status
kw:stdeb
kw:storage
kw:streaming
kw:strports
kw:style
kw:stylesheet
kw:subprocess
kw:sumo
kw:survey
kw:svg
kw:symlink
kw:synchronous
kw:tac
kw:tahoe-*
kw:tahoe-add-alias
kw:tahoe-admin
kw:tahoe-archive
kw:tahoe-backup
kw:tahoe-check
kw:tahoe-cp
kw:tahoe-create-alias
kw:tahoe-create-introducer
kw:tahoe-debug
kw:tahoe-deep-check
kw:tahoe-deepcheck
kw:tahoe-lafs-trac-stream
kw:tahoe-list-aliases
kw:tahoe-ls
kw:tahoe-magic-folder
kw:tahoe-manifest
kw:tahoe-mkdir
kw:tahoe-mount
kw:tahoe-mv
kw:tahoe-put
kw:tahoe-restart
kw:tahoe-rm
kw:tahoe-run
kw:tahoe-start
kw:tahoe-stats
kw:tahoe-unlink
kw:tahoe-webopen
kw:tahoe.css
kw:tahoe_files
kw:tahoewapi
kw:tarball
kw:tarballs
kw:tempfile
kw:templates
kw:terminology
kw:test
kw:test-and-set
kw:test-from-egg
kw:test-needed
kw:testgrid
kw:testing
kw:tests
kw:throttling
kw:ticket999-s3-backend
kw:tiddly
kw:time
kw:timeout
kw:timing
kw:to
kw:to-be-closed-on-2011-08-01
kw:tor
kw:tor-protocol
kw:torsocks
kw:tox
kw:trac
kw:transparency
kw:travis
kw:travis-ci
kw:trial
kw:trickle
kw:trivial
kw:truckee
kw:tub
kw:tub.location
kw:twine
kw:twistd
kw:twistd.log
kw:twisted
kw:twisted-14
kw:twisted-trial
kw:twitter
kw:twn
kw:txaws
kw:type
kw:typeerror
kw:ubuntu
kw:ucwe
kw:ueb
kw:ui
kw:unclean
kw:uncoordinated-writes
kw:undeletable
kw:unfinished-business
kw:unhandled-error
kw:unhappy
kw:unicode
kw:unit
kw:unix
kw:unlink
kw:update
kw:upgrade
kw:upload
kw:upload-helper
kw:uri
kw:url
kw:usability
kw:use-case
kw:utf-8
kw:util
kw:uwsgi
kw:ux
kw:validation
kw:variables
kw:vdrive
kw:verify
kw:verlib
kw:version
kw:versioning
kw:versions
kw:video
kw:virtualbox
kw:virtualenv
kw:vista
kw:visualization
kw:visualizer
kw:vm
kw:volunteergrid2
kw:volunteers
kw:vpn
kw:wapi
kw:warners-opinion-needed
kw:warning
kw:weapi
kw:web
kw:web.port
kw:webapi
kw:webdav
kw:webdrive
kw:webport
kw:websec
kw:website
kw:websocket
kw:welcome
kw:welcome-page
kw:welcomepage
kw:wiki
kw:win32
kw:win64
kw:windows
kw:windows-related
kw:winscp
kw:workaround
kw:world-domination
kw:wrapper
kw:write-enabler
kw:wui
kw:x86
kw:x86-64
kw:xhtml
kw:xml
kw:xss
kw:zbase32
kw:zetuptoolz
kw:zfec
kw:zookos-opinion-needed
kw:zope
kw:zope.interface
p/blocker
p/critical
p/major
p/minor
p/normal
p/supercritical
p/trivial
r/cannot reproduce
r/duplicate
r/fixed
r/invalid
r/somebody else's problem
r/was already fixed
r/wontfix
r/worksforme
t/defect
t/enhancement
t/task
v/0.2.0
v/0.3.0
v/0.4.0
v/0.5.0
v/0.5.1
v/0.6.0
v/0.6.1
v/0.7.0
v/0.8.0
v/0.9.0
v/1.0.0
v/1.1.0
v/1.10.0
v/1.10.1
v/1.10.2
v/1.10a2
v/1.11.0
v/1.12.0
v/1.12.1
v/1.13.0
v/1.14.0
v/1.15.0
v/1.15.1
v/1.2.0
v/1.3.0
v/1.4.1
v/1.5.0
v/1.6.0
v/1.6.1
v/1.7.0
v/1.7.1
v/1.7β
v/1.8.0
v/1.8.1
v/1.8.2
v/1.8.3
v/1.8β
v/1.9.0
v/1.9.0-s3branch
v/1.9.0a1
v/1.9.0a2
v/1.9.0b1
v/1.9.1
v/1.9.2
v/1.9.2a1
v/cloud-branch
v/unknown
No milestone
No project
No assignees
3 participants
Notifications
Due date
The due date is invalid or out of range. Please use the format "yyyy-mm-dd".

No due date set.

Dependencies

No dependencies set.

Reference: tahoe-lafs/trac#1212
No description provided.