Tech »  Topic »  Introducing BigQuery managed and SQL-native inference for open models

Introducing BigQuery managed and SQL-native inference for open models


BigQuery provides access to a variety of LLMs for text and embedding generation, including Google's Gemini models, Google-managed models from partners like Anthropic and Mistral. Using Gemini models and Google-managed partner models in BigQuery is simple — just create the model with the foundation model name and run inference directly in SQL queries. Today, we are bringing this same simplicity and power to any model you may choose from Hugging Face or Vertex AI Model Garden.

A SQL-native workflow with automated management

With the launch of managed third-party generative AI inference in BigQuery (Preview), you can now run open models using just two SQL statements. 

This new capability delivers four key benefits:

  1. Simplified deployment: Deploy open models using a single CREATE MODEL SQL statement with the model id string (e.g., google/gemma-3-1b-it). BigQuery automatically provisions the compute resources with default configurations.
  2. Automated resource management: BigQuery automatically releases idle compute ...

Copyright of this story solely belongs to google cloudblog . To see the full text click HERE