Response to U.S. AI Safety Institute’s Request for Comment on Managing Misuse Risk For Dual-Use Foundation Models

Stanford HAI aggrees with and supports the U.S. AI Safety Institute’s (US AISI) draft guidelines for improving the safety, security, and trustworthiness of dual-use foundation models.
Introduction
Overall, we agree with and support the U.S. AI Safety Institute’s (US AISI) draft guidelines (hereafter “the guidelines”) for improving the safety, security, and trustworthiness of dual-use foundation models, which were issued in line with obligations under the October 2023 Executive Order on Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence (AI). The guidelines provide useful and actionable recommendations for model developers to manage misuse risk. We encourage the US AISI to develop similar guidance for other actors in the foundation model supply chain as we believe model developers can contribute to, but are not best positioned for, mitigating all types of risk. In addition, while not captured by the Executive Order’s focus on foundation model misuse, we encourage the US AISI to develop guidance for non-misuse risks. While elements of the guidance (e.g., API monitoring) are more appropriate for some foundation model release strategies, we encourage the US AISI to explicitly affirm that these should not dissuade the open release of foundation models absent evidence of marginal risk.