Abuse monitoring

As outlined in Section 4.3 "Generative AI Safety and Abuse" of Google Cloud Platform Terms of Service, Google uses the following process to detect potential abuse and violations of its Acceptable Use Policy and Prohibited Use Policy as part of providing Generative AI Services to customers.

  • Automated detection: Google uses automated safety classifiers to detect potential abuse and violations. For technical details on how safety classifiers work, see Configure safety filters.
  • Prompt logging: If automated safety classifiers detect suspicious activity that requires further investigation into whether a customer has violated our policies, then Google may log customer prompts solely for the purpose of examining whether a violation of the AUP or Prohibited Use Policy has occurred. This data won't be used to train or fine-tune any AI/ML models. This data is stored securely for up to 30 days in the same region or multi-region selected by the customer for their project and adheres to Google Cloud assurances, such as Data Residency, Access Transparency and VPC Service Controls. Customers also have the option to request an opt-out from abuse logging (see below).
  • Action: Authorized Google employees may assess the flagged prompts and may reach out to the customer for clarification. Failure to address the behavior—or recurring or severe abuse—may result in suspension or termination of the customer's access to Vertex AI or Google Cloud services.
  • Services in scope: Vertex AI API, when used with Google's large language models.
  • Customers in scope: Only customers whose use of Google Cloud is governed by the Google Cloud Platform Terms of Service and who don't have an Invoiced Cloud Billing account are subject to prompt logging for abuse monitoring.
  • Customer opt-out: Customers may request for an exception by filling out this form. If approved, Google won't store any prompts associated with the approved Google Cloud account.

What's next