Connect with us

Hi, what are you looking for?

Top Stories

BigQuery Launches SQL-Native Inference for Open Models, Simplifying AI Deployments

Google’s BigQuery introduces SQL-native inference for open models, enabling users to deploy advanced AI with just two SQL statements, simplifying access to generative AI technologies.

Google’s BigQuery has introduced a new feature that enhances accessibility to various large language models (LLMs) for text and embedding generation, including its own Gemini models and those managed in collaboration with partners like Anthropic and Mistral. This capability, which facilitates the use of these models directly within SQL queries, aims to simplify the deployment and management of generative AI models for users, regardless of their technical expertise. Alongside this, BigQuery is extending its support to models available on platforms such as Hugging Face and Vertex AI Model Garden, marking a significant advancement in database management and AI integration.

With the launch of managed third-party generative AI inference in BigQuery (currently in Preview), users can execute open models with just two SQL statements. This streamlined approach offers four primary advantages: simplified deployment, automated resource management, granular resource control, and a unified SQL interface. The deployment process is designed to be straightforward; users can create an open model by issuing a single CREATE MODEL statement that includes the model ID string, such as google/gemma-3-1b-it. BigQuery handles the provisioning of compute resources automatically, making it accessible even for those less familiar with AI model management.

One of the standout features is automated resource management. BigQuery actively releases idle compute resources, which helps prevent unexpected costs for users. This functionality can be customized through the endpoint_idle_ttl configuration, allowing users to define how long resources should remain active without use. Additionally, users have the option to customize backend computing resources, adjusting parameters like machine types and minimum or maximum replicas within the CREATE MODEL statement. This flexibility ensures that users can tailor the performance and expenses of their models to suit their needs.

To illustrate how the process works, one can create a BigQuery managed open model by first executing a CREATE MODEL statement with the appropriate open model ID. Depending on the size of the model and the chosen machine type, the query typically completes within a few minutes. For models sourced from Hugging Face, users must specify the hugging_face_model_id in a format that includes the provider name and model name—an example being sentence-transformers/all-MiniLM-L6-v2.

This initiative reflects a broader commitment by Google to democratize access to advanced AI capabilities while ensuring that organizations can leverage these technologies without extensive resources or expertise. As generative AI continues to evolve and become more integrated within various sectors, the implications of such developments are significant. The ability to utilize powerful models through a familiar SQL interface could redefine workflows across industries, enabling more users to harness the potential of AI-driven analytics and insights.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Marketing

Higgsfield secures $80M in funding, boosting its valuation to $1.3B as demand for AI-driven video content surges, targeting social media marketers.

Top Stories

Critical security flaws in Nvidia, Salesforce, and Apple’s AI libraries expose Hugging Face models to remote code execution risks, threatening open-source integrity.

Top Stories

Walmart partners with Google to integrate shopping into Gemini AI, signaling a pivotal shift in commerce that may marginalize smaller retailers.

AI Generative

Z.ai's GLM-Image surpasses Google's Nano Banana Pro with an impressive 91.16% accuracy, signaling a major shift towards open-source dominance in AI text rendering.

AI Generative

Google's Veo 3.1 update enhances generative AI video production with native vertical support, character consistency, and 4K upscaling for professional use.

Top Stories

Google invites Gemini users to opt into Personal Intelligence, enhancing AI responses with personalized data from Gmail, Photos, and YouTube.

Top Stories

Microsoft plans to invest $500 million annually in Anthropic to integrate AI technology into its cloud offerings, enhancing revenue and competitive positioning.

Top Stories

Apple and Google unite to enhance Siri and Google Assistant, potentially reshaping AI market dynamics and user experiences across smart devices.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.