diff --git a/SECURITY.md b/SECURITY.md index 79514f2c282..3baa145df79 100644 --- a/SECURITY.md +++ b/SECURITY.md @@ -3,6 +3,7 @@ - [**Reporting a Vulnerability**](#reporting-a-vulnerability) - [**Using Pytorch Securely**](#using-pytorch-securely) - [Untrusted models](#untrusted-models) + - [TorchScript models](#torchscript-models) - [Untrusted inputs](#untrusted-inputs) - [Data privacy](#data-privacy) - [Using distributed features](#using-distributed-features) @@ -38,6 +39,10 @@ Important Note: The trustworthiness of a model is not binary. You must always de https://arxiv.org/abs/2312.04748 https://arxiv.org/abs/2401.05566 +### TorchScript models + +TorchScript models should treated the same way as locally executable code from an unknown source. Only run TorchScript models if you trust the provider. Please note, that tools for introspecting TorchScript models (such as `torch.utils.model_dump`) may also execute partial or full code stored in those models, therefore they should be used only if you trust the provider of the binary you are about to load. + ### Untrusted inputs during training and prediction If you plan to open your model to untrusted inputs, be aware that inputs can also be used as vectors by malicious agents. To minimize risks, make sure to give your model only the permissions strictly required, and keep your libraries updated with the latest security patches.