Docker has introduced the Docker Model Runner in its latest preview release of Docker Desktop 4.40 for macOS, specifically tailored for Apple Silicon. This new feature enables developers to run machine learning models locally and refine application code without interrupting their existing container workflows. By leveraging local large language models (LLMs), developers can benefit from reduced costs, enhanced data privacy, minimized network latency, and greater control over model execution. The Docker Model Runner aims to alleviate common challenges faced by developers integrating LLMs into containerized applications, such as managing external tools and models outside their container environments.
The Docker Model Runner includes a robust inference engine built on llama.cpp and offers accessibility through the OpenAI API. This integration allows developers to bypass the performance overhead associated with virtual machines by utilizing host-based execution, which runs models directly on Apple Silicon while capitalizing on GPU acceleration. Additionally, Docker employs the Open Container Initiative (OCI) standard for model distribution, fostering a unified workflow for both containers and models. Developers can seamlessly pull models from Docker Hub and will soon have the capability to push their own models, integrating fully with any container registry. The introduction of the docker model command streamlines model management, enabling commands like docker model pull ai/smollm2:360M-Q4_K_M and docker model run ai/smollm2:360M-Q4_K_M "Give me a fact about whales." to be performed easily, all without the need to create containers.
Furthermore, the Model Runner can be utilized with any OpenAI-compatible client or framework via its endpoint at http://model-runner.docker.internal/engines/v1. This endpoint is accessible both within containers and from the host machine, provided that TCP host access is enabled. Docker Hub hosts a variety of ready-to-use models for Model Runner, including smollm2, llama3.3, and gemma3. In addition, Docker has released tutorials to assist developers in integrating these models into applications, providing valuable resources for maximizing their utility.
In a related development, Docker has rolled out the Docker Model Context Protocol (MCP) Catalog and Toolkit, offering a centralized platform for developers to discover verified and curated MCP tools. This initiative is designed to simplify the development of AI agents by combining Docker’s inherent ease of use and security with the capabilities of the MCP ecosystem. The MCP Catalog allows developers to explore and manage over 100 MCP servers from a variety of providers directly through Docker Desktop. The MCP Toolkit further enhances usability by enabling developers to run, authenticate, and manage these tools effortlessly, fostering a developer-first ecosystem for MCP tools.
With the rapid rise of AI applications, the emphasis on secure software delivery has become increasingly critical. The Docker MCP Catalog addresses these concerns by ensuring that tools are verified, secure, and straightforward to operate, allowing developers to devote more time to application development rather than grappling with complex integrations. Future enhancements will enable teams to publish and manage their own MCP servers, which will bolster security controls and support compliance efforts.
The recent launch of Docker Desktop 4.40, featuring the beta version of Docker Model Runner, marks a significant step in simplifying GenAI application development while supporting scalability. The integration of local model execution, GPU acceleration tailored for Apple Silicon, and standardized model packaging using OCI Artifacts represents a comprehensive advancement for developers.
Moreover, the Docker AI Agent has been upgraded with MCP integration, facilitating smoother communication with various tools and applications. This enhanced agent now supports a range of developer tasks, including executing shell commands, managing files, and performing Git operations. The AI Tool Catalog extension allows developers to discover different MCP servers and link the Docker AI agent to other tools or LLMs, simplifying the management of diverse server configurations and credential arrangements.
As the AI landscape evolves, the importance of secure authentication methods cannot be overstated. SSOJet offers solutions targeting enterprises, providing secure Single Sign-On (SSO), Multi-Factor Authentication (MFA), and Passkey management. With features like directory synchronization, SAML, and OIDC support, SSOJet aims to streamline user management while enhancing security for its enterprise clientele. For more details on their services, visit ssojet.com.
See also
U.S. Generative AI Market Set to Reach $1,022.41 Billion by 2032 with 47.53% CAGR
LG Unveils AI Cabin Platform at CES 2026, Transforming In-Vehicle Experiences with Generative AI
Study Shows Rude Prompts Yield 84.8% Accuracy in AI Interactions, Penn State Research Reveals
Citrine Informatics Launches New AI Tools to Accelerate Sustainable Materials Development by 80%
OpenAI Licenses 200+ Disney Characters for Sora Videos in $1B Deal with Disney


















































