We received a tor bug bounty report from jaym about a congestion attack variant that can cause bandwidth stats watermark.
The bug uses the fact that Tor increments the read bytes counter before adding the cell to the output buffer: If the circuit gets killed before the cell gets relayed to the next hop, then the write bytes counter will never be updated, making the read bytes counter having a higher value than the write bytes counter. The attacker could exploit this assymetry to find relays using their bandwidth graph.
The attacker can kill the circuit using the OOM killer by saturating its output queue with cells until circuits_handle_oom() gets called and kills the circuit.
We should figure out whether this attack is practical (the paper claims it is) and whether it's worthwhile fixing it. Just fixing this issue won't solve the general issue of congestion attacks, and it might even allow other kinds of attacks.
The most practical fix right now seem to be to hack circuit_handle_oom() to actually decrement the read counters before killing a circuit. However, that's a very specific fix that might solve this very specific bug, but leave the rest of the bug class open.
Another approach would be removing the bandwidth graphs, or aggregating them over a greater period of time, or adding noise. We should consider these approaches carefully since bandwidth graphs see great use by academic papers and also by relay operators (to gauge their contribution).
Edited
To upload designs, you'll need to enable LFS and have an admin enable hashed storage. More information
Child items ...
Show closed items
Linked items 0
Link issues together to show that they're related.
Learn more.
My current thinking is that this ticket should be about the specific fix, and we should include more general fixes in legacy/trac#22729 (moved) or subtickets.
I think we should also look into other counters and correct their values even if they are not published. The control protocol can be used to dump stats about circuits, including counter of relayed cells, etc. They are subjected to the same kind of bug.
During Montreal, Roger mentioned there is a second piece to this. I tried to save it here but trac ate my changes. Roger, can you say what that was again?
Mike: I talked to Jaym about this topic at PETS. I think at the time I was under the impression that we are incrementing the write counter, that is, the outbound connection, even though we never actually push the bytes out to the network because we kill the outbuf in the oom killer.
So I had thought that the bug was "we say that we wrote out the bytes even though we didn't", which allows an attacker to queue up a bunch of bytes for us to outbuf, and generate an artificially large signal.
If that is the issue (which contradicts asn's summary above I think), then there would be a second piece to the fix, since if we only decremented the write count to stop including the bytes we didn't actually write, then we would have an asymmetry between the (still much higher) read count and the (now corrected) write count. So the second part of the fix would be to decrement the read count by the corresponding amount too, so things match up.
Mike: I talked to Jaym about this topic at PETS. I think at the time I was under the impression that we are incrementing the write counter, that is, the outbound connection, even though we never actually push the bytes out to the network because we kill the outbuf in the oom killer.
So I had thought that the bug was "we say that we wrote out the bytes even though we didn't", which allows an attacker to queue up a bunch of bytes for us to outbuf, and generate an artificially large signal.
Yes, because of the first version of the paper you reviewed I considered the wrong counter. This counter actually behaves exactly like you describe above but this was not the counter used to report bandwidth in the extra-info descriptor. With your review, you made me realize that this problem actually happens with the read counter.
The up-to-date version of the paper I sent you in the beginning of September corrects the explanation and gives more details (given to asn too, which explains why you see difference in asn's description with you original thoughts). I also added chutney experiments to observe the difference in read/write that can be reproduced with this code (https://github.com/popets18dropping/code_and_data/tree/master/hs_drop_attack).
If that is the issue (which contradicts asn's summary above I think), then there would be a second piece to the fix, since if we only decremented the write count to stop including the bytes we didn't actually write, then we would have an asymmetry between the (still much higher) read count and the (now corrected) write count. So the second part of the fix would be to decrement the read count by the corresponding amount too, so things match up.
Yep. If needed I can help testing the fix with the code linked above.
This buffering fills the memory until circuit_handle_oom () is called to kill the circuits and recover the memory from its queues. In the mean- time, the cell payload is counted in the read statistics and ignored in the write statistics.
This makes it sound like asn's proposed fix in the description should actually be correct, and that arma's comment may have been based on an earlier paper revision.
Sure would be nice to get that github url back to actually reproduce some science here..
I made that branch off of 0.3.2, because yesterday dgoulet told me that the network is still experiencing continuous OOM attacks, trigging circuit oomkiller. This very well could be (one of) the reasons for such attacks. So I think we should backport. Certainly plenty of relays are experiencing circuit OOMs and reporting asymmetric stats.
Instead of trying to guess when the bytes arrived and subtract them from the appropriate read totals, I just report that we wrote them instead. Much simpler, easier to backport, etc.
Downsides of this fix (and probably any other fix): We don't know how many bytes the TLS headers took up. For this reason, I also didn't bend over backwards to count bytes for var cells, wide circ ids, etc. Do we think that is sufficient? Should we lie and add ~1 TLS header of bytes per cell?
Are there other places where we kill circuits like this?
Dgoulet - what about the DoS/DESTROY queue handling?
Trac: Cc: arma to arma, dgoulet Status: new to needs_review
I would really want us to not forget to document #define TLS_PER_CELL_OVERHEAD 29. I know we are waiting on pastly to do the impossible job of recalling how 29 came about during the KIST development process.
In any case, now KIST and this will at least use the same value.
The unit test sometimes fails (approx 1 time out of 10) due (I think) to our use of approx_time() in the patch, but time() in rephist.c I will redo the branches for this fix, merge forward as before, and then make a master merge commit.
Ok I changed the patch's use of approx_time() to time() for consistency with rephist.c (which fixes the flapping test for me), and I now check the exact byte count in the tests. Otherwise the following branches are identical to v3. Setting back to merge_ready.
Dhalgren - What produced that graph, and for what Tor relay version? That does not look like a metrics graph.
Exits can have asymmetry between bytes read and bytes written due to cell packing. If there is a lot of intermittent small-packet activity (eg SSH) being read from a public internet IP by your exit, your exit will package will package those small packets into their own cells, and end up writing much more than it has been reading as it sends those mostly empty cells to the middle node. So this could be unrelated.
Pastly pointed out that TLS_PER_CELL_OVERHEAD is exactly half of the TLS header size for TLS as used in Tor. He and Rob observed that exactly 2 Tor cells got packed into TLS frames whenever there were 2 or more in queue.
This means that if the number of cells in the circuit queue is odd, we should probably add one extra TLS_PER_CELL_OVERHEAD to the written bytes.
I can do that in another fixup, and comment the TLS_PER_CELL_OVERHEAD define.
Dhalgren - What produced that graph, and for what Tor relay version? That does not look like a metrics graph.
The graph is an obscured snip of the high-res view provided by ISP.
Seeing a bit of this lately, investigated and it appears mostly DIR response traffic--visible in published statistics. Perhaps some new form of dubious behavior from at-large.
Jaym - how did you test this? What type of node did you test with (guard, middle, exit)? And how large was this write discrepancy?
It is surprising that it is reporting more written than read -- we made conservative assumptions about TLS overhead (we report as if there are always 2 cells written per TLS payload, which is the minimum possible written overhead) that should mean that the reported write totals are always <= the read totals.