Connect with us

Hi, what are you looking for?

Top Stories

BigQuery Launches SQL-Native Inference for Open Models, Simplifying AI Deployments

Google’s BigQuery introduces SQL-native inference for open models, enabling users to deploy advanced AI with just two SQL statements, simplifying access to generative AI technologies.

Google’s BigQuery has introduced a new feature that enhances accessibility to various large language models (LLMs) for text and embedding generation, including its own Gemini models and those managed in collaboration with partners like Anthropic and Mistral. This capability, which facilitates the use of these models directly within SQL queries, aims to simplify the deployment and management of generative AI models for users, regardless of their technical expertise. Alongside this, BigQuery is extending its support to models available on platforms such as Hugging Face and Vertex AI Model Garden, marking a significant advancement in database management and AI integration.

With the launch of managed third-party generative AI inference in BigQuery (currently in Preview), users can execute open models with just two SQL statements. This streamlined approach offers four primary advantages: simplified deployment, automated resource management, granular resource control, and a unified SQL interface. The deployment process is designed to be straightforward; users can create an open model by issuing a single CREATE MODEL statement that includes the model ID string, such as google/gemma-3-1b-it. BigQuery handles the provisioning of compute resources automatically, making it accessible even for those less familiar with AI model management.

One of the standout features is automated resource management. BigQuery actively releases idle compute resources, which helps prevent unexpected costs for users. This functionality can be customized through the endpoint_idle_ttl configuration, allowing users to define how long resources should remain active without use. Additionally, users have the option to customize backend computing resources, adjusting parameters like machine types and minimum or maximum replicas within the CREATE MODEL statement. This flexibility ensures that users can tailor the performance and expenses of their models to suit their needs.

To illustrate how the process works, one can create a BigQuery managed open model by first executing a CREATE MODEL statement with the appropriate open model ID. Depending on the size of the model and the chosen machine type, the query typically completes within a few minutes. For models sourced from Hugging Face, users must specify the hugging_face_model_id in a format that includes the provider name and model name—an example being sentence-transformers/all-MiniLM-L6-v2.

This initiative reflects a broader commitment by Google to democratize access to advanced AI capabilities while ensuring that organizations can leverage these technologies without extensive resources or expertise. As generative AI continues to evolve and become more integrated within various sectors, the implications of such developments are significant. The ability to utilize powerful models through a familiar SQL interface could redefine workflows across industries, enabling more users to harness the potential of AI-driven analytics and insights.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

Top Stories

Anthropic accuses DeepSeek and two other Chinese firms of executing 16 million distillation attacks to illegally enhance their AI models, threatening U.S. tech dominance.

AI Technology

Australia mandates major tech firms like Apple and Google to implement age verification for AI services by March 9 or face penalties up to...

AI Research

New research reveals that high-income countries utilize AI tools four times more than middle- and low-income nations, highlighting a critical global adoption gap.

Top Stories

Perplexity Computer introduces a $200/month multi-model AI platform, streamlining workflows by integrating 19 AI models for enhanced productivity in enterprise settings.

AI Government

Trump halts all federal use of Anthropic's AI tools, citing security concerns over unrestricted access to the company's chatbot, Claude, within six months.

Top Stories

SURXRAT expands its malware capabilities by incorporating a 23GB LLM module from Hugging Face, enhancing surveillance and exploitation tactics for cybercriminals.

Top Stories

A 14-year-old's suicide linked to an AI chatbot prompts a lawsuit against Character.AI, highlighting urgent calls for stronger protections for vulnerable users.

AI Tools

Google enhances Google Translate with AI-driven features for contextual translations in the U.S. and India, improving communication clarity across diverse interactions.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.