You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the meeting on 03-08-2021 we went over some example computations the aggregate API could support (slides) that satisfy differential privacy. These included:
Hierarchical domains (possibly with multiple queries), to prune a larger domain smaller in some flexible way. Thresholding can be used to make the computations more efficient, though may not be strictly required by DP.
"Sparse vector" techniques to handle truly massive domains (e.g. 2^64 or 2^128 entries from hashing a string), which requires thresholding to preserve DP, but will never report on a key that wasn't present (see this doc)
Example MPC: something like what we documented in private_histograms_mpc.md, although more work is needed to evaluate these techniques.
These techniques have different pros and cons (and these techniques are obviously not exhaustive). I'm filing this issue to solicit more feedback. Some evaluation criteria:
Developer ergonomics (especially with regard to figuring out a dense encoding of aggregation keys)
Utility of output (e.g. bias introduced by thresholding)
MPC simplicity
MPC security guarantees (zero-knowledge, etc.)
MPC computation / communication costs
Privacy of output (e.g. smaller domain sizes can encode less information about users)
The text was updated successfully, but these errors were encountered:
In the meeting on 03-08-2021 we went over some example computations the aggregate API could support (slides) that satisfy differential privacy. These included:
These techniques have different pros and cons (and these techniques are obviously not exhaustive). I'm filing this issue to solicit more feedback. Some evaluation criteria:
The text was updated successfully, but these errors were encountered: