You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Private model training has been recently mentioned here. One of the privacy considerations is to include DP in the training loop through DP-SGD.
There are cases when DP-SGD would make the training process considerably slower as it destroys the sparsity of the gradients calculated during backprop, rendering impossible to use optimization techniques that rely on such sparsity. This is usually the case when some features are categorical features or working with embedding tables in the case of LLMs. I am aware there is research around this topic to remedy it, although it is not clear from the explainer linked above if this is something that has been considered in the context of Protected Audience API.
Are there any techniques that are being considered to face this or thoughts about this topic?
Thanks
The text was updated successfully, but these errors were encountered:
Hi @Lufe44. Thank you very much for the feedback. It is useful to know that this is a concern.
Yes we are aware of some techniques to preserve sparsity in DP-SGD, e.g. https://arxiv.org/abs/2404.10881, and indeed we are exploring supporting these kinds of techniques in the private model training infrastructure. However, since we are in the research phase we aren't able to make any commitments yet to whether it will be feasible for us to support.
Did you have any specific approaches you wanted us to consider?
Hello,
Private model training has been recently mentioned here. One of the privacy considerations is to include DP in the training loop through DP-SGD.
There are cases when DP-SGD would make the training process considerably slower as it destroys the sparsity of the gradients calculated during backprop, rendering impossible to use optimization techniques that rely on such sparsity. This is usually the case when some features are categorical features or working with embedding tables in the case of LLMs. I am aware there is research around this topic to remedy it, although it is not clear from the explainer linked above if this is something that has been considered in the context of Protected Audience API.
Are there any techniques that are being considered to face this or thoughts about this topic?
Thanks
The text was updated successfully, but these errors were encountered: