Connect with us

Hi, what are you looking for?

Top Stories

BigQuery Launches SQL-Native Inference for Open Models, Simplifying AI Deployments

Google’s BigQuery introduces SQL-native inference for open models, enabling users to deploy advanced AI with just two SQL statements, simplifying access to generative AI technologies.

Google’s BigQuery has introduced a new feature that enhances accessibility to various large language models (LLMs) for text and embedding generation, including its own Gemini models and those managed in collaboration with partners like Anthropic and Mistral. This capability, which facilitates the use of these models directly within SQL queries, aims to simplify the deployment and management of generative AI models for users, regardless of their technical expertise. Alongside this, BigQuery is extending its support to models available on platforms such as Hugging Face and Vertex AI Model Garden, marking a significant advancement in database management and AI integration.

With the launch of managed third-party generative AI inference in BigQuery (currently in Preview), users can execute open models with just two SQL statements. This streamlined approach offers four primary advantages: simplified deployment, automated resource management, granular resource control, and a unified SQL interface. The deployment process is designed to be straightforward; users can create an open model by issuing a single CREATE MODEL statement that includes the model ID string, such as google/gemma-3-1b-it. BigQuery handles the provisioning of compute resources automatically, making it accessible even for those less familiar with AI model management.

One of the standout features is automated resource management. BigQuery actively releases idle compute resources, which helps prevent unexpected costs for users. This functionality can be customized through the endpoint_idle_ttl configuration, allowing users to define how long resources should remain active without use. Additionally, users have the option to customize backend computing resources, adjusting parameters like machine types and minimum or maximum replicas within the CREATE MODEL statement. This flexibility ensures that users can tailor the performance and expenses of their models to suit their needs.

To illustrate how the process works, one can create a BigQuery managed open model by first executing a CREATE MODEL statement with the appropriate open model ID. Depending on the size of the model and the chosen machine type, the query typically completes within a few minutes. For models sourced from Hugging Face, users must specify the hugging_face_model_id in a format that includes the provider name and model name—an example being sentence-transformers/all-MiniLM-L6-v2.

This initiative reflects a broader commitment by Google to democratize access to advanced AI capabilities while ensuring that organizations can leverage these technologies without extensive resources or expertise. As generative AI continues to evolve and become more integrated within various sectors, the implications of such developments are significant. The ability to utilize powerful models through a familiar SQL interface could redefine workflows across industries, enabling more users to harness the potential of AI-driven analytics and insights.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

Top Stories

Mistral unveils Connectors in Studio, enabling seamless API integration for enterprise AI applications, streamlining workflows and reducing setup time significantly.

AI Marketing

Reddit captures over 9% of AI citations, compelling brands to overhaul AEO strategies and engage authentically in community-driven discourse.

AI Cybersecurity

New analysis warns that Anthropic's Mythos AI tool could empower cyberattacks on small businesses, making them vulnerable to exploitation by advanced AI threats.

AI Generative

Microsoft launches MAI-Image-2, ranking third on Arena.ai with advanced photorealism and text generation, but faces significant usage limitations.

Top Stories

A new BMJ Open study reveals that five AI chatbots, including ChatGPT and Grok, deliver 49.6% problematic health responses, raising urgent oversight concerns.

AI Regulation

Anthropic appoints Novartis CEO Vas Narasimhan to its board, enhancing AI governance with his extensive healthcare and regulatory expertise.

AI Finance

70% of finance teams in Australia and New Zealand use shadow AI tools like ChatGPT, risking data governance with only 16% confident in data...

AI Generative

InVideo launches an AI video generator powered by over 200 models, enabling complete video creation for just $28 a month, streamlining content production for...

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.