https://ollama.com/blog/coding-models Ollama Models GitHub Discord Docs Cloud [ ] Sign in Download [ ] Models Download GitHub Discord Docs Cloud Sign in New coding models & integrations October 16, 2025 Illustration of Ollama coding GLM-4.6 and Qwen3-coder-480B are available on Ollama's cloud service with easy integrations to the tools you are familiar with. Qwen3-Coder-30B has been updated for faster, more reliable tool calling in Ollama's new engine. Get started GLM-4.6 ollama run glm-4.6:cloud Qwen3-Coder-480B ollama run qwen3-coder:480b-cloud For users with more than 300GB of VRAM, qwen3-coder:480b is also available locally. Qwen3-Coder-30B ollama run qwen3-coder:30b Example prompts Create a single-page app in a single HTML file with the following requirements: Name: Ollama's Adventure Goal: Jump over obstacles to survive as long as possible. Features: Increasing speed, high score tracking, retry button, and funny sounds for actions and events. The UI should be colorful, with parallax scrolling backgrounds. The characters should look cartoonish, related to alpacas and be fun to watch. The game should be enjoyable for everyone. Example code by GLM-4.6 in a single prompt example image of the HTML site running example image 2 of the HTML site running Usage with VS Code First, pull the coding models so they can be accessed via VS Code: ollama pull glm-4.6:cloud ollama pull qwen3-coder:480b-cloud 1. Open the copilot chat sidebar 2. Select the model dropdown - Manage models 3. Click on Ollama under Provider Dropdown, then select desired models 4. Select the model dropdown - and choose the model (e.g. glm-4.6) Usage with Zed First pull the coding models so they can be accessed via Zed: ollama pull glm-4.6:cloud ollama pull qwen3-coder:480b-cloud Then, open Zed (now available for Windows!) 1. Click on the agent panel button (glittering stars) 2. Click on the model dropdown - Configure 3. Select LLM providers - Ollama 4. Confirm the Host URL is http://localhost:11434, then click Connect 5. Select a model under Ollama Usage with Droid First, install Droid: curl -fsSL https://app.factory.ai/cli | sh Add the following configuration to ~/.factory/config.json: { "custom_models": [ { "model_display_name": "GLM-4.6", "model": "glm-4.6:cloud", "base_url": "http://localhost:11434/v1", "api_key": "not-needed", "provider": "generic-chat-completion-api", "max_tokens": 16384 }, { "model_display_name": "Qwen3-Coder-480B", "model": "qwen3-coder:480b-cloud", "base_url": "http://localhost:11434/v1", "api_key": "not-needed", "provider": "generic-chat-completion-api", "max_tokens": 16384 } ] } Then run Droid and type /model to change to the model: +--------------------------------------------------+ | > GLM-4.6 [current] | | Qwen3-Coder-480B | | | | |/| to navigate, Enter to select, ESC to go back | +--------------------------------------------------+ Integrations Ollama's documentation now includes sections on using Ollama with popular coding tools: * Codex * Cline * VS Code * Zed * Droid * Roo code Cloud API access Cloud models such as glm-4.6 and qwen3-coder:480b can also be accessed directly via ollama.com's cloud API: First, create an API key, and set it in your environment export OLLAMA_API_KEY="your_api_key_here" Then, call ollama.com's API curl https://ollama.com/api/chat \ -H "Authorization: Bearer $OLLAMA_API_KEY" \ -d '{ "model": "glm-4.6", "messages": [{ "role": "user", "content": "Write a snake game in HTML." }] }' For more information see the Ollama's API documentation. (c) 2025 Ollama Download Blog Docs GitHub Discord X (Twitter) Contact Us * Blog * Download * Docs * GitHub * Discord * X (Twitter) * Meetups (c) 2025 Ollama Inc.