minor edit, and adding to the end

This commit is contained in:
loeding
2025-07-23 21:03:37 +00:00
parent 38d0a68371
commit 90d8c6cd13

View File

@@ -530,3 +530,11 @@ This experiment makes us think of two learning tasks.
3) Could you get the correct entry even with partial information? 3) Could you get the correct entry even with partial information?
4) Probabilistic attack for guessing the correct sequence? 4) Probabilistic attack for guessing the correct sequence?
First split shuffle evenly splits things, so you don't gain much. Subsequent shuffles will start to bias toward the correct passcode. The trade-off is that if you don't shuffle a position then the intruder can just use the same input as before and can enter the correct key without knowing the correct icon, but if you do shuffle, then the intruder learns more information. First split shuffle evenly splits things, so you don't gain much. Subsequent shuffles will start to bias toward the correct passcode. The trade-off is that if you don't shuffle a position then the intruder can just use the same input as before and can enter the correct key without knowing the correct icon, but if you do shuffle, then the intruder learns more information.
One key Im understanding better now: Theres a trade-off between (1) information being learned by the intruder and (2) the chances that an intruder could key in a correct key-sequence for each subsequent observation / login attempt.
If an attribute is not shuffled, then that increases the chance for (2), and if an attribute is shuffled that increases the chance for (1).
I think that when the selection of the sets is randomized like youre doing, youre getting the optimal trade-off between these two things. However, theres also probably a strategy that might be even more optimal than just randomly choosing a split each time. If each shuffle chooses for its set of half of the attributes: half from the attributes that were previously shuffled and half from the attributes that werent shuffled in the previous shuffle. You could look back essentially log_2(p) steps (where p is the number of attributes) and similarly subdivide. This seems to be a way to balance the tradeoff for consecutive observations.