612 Downloads Updated 2 months ago
My sincere apologies and gratitude for each of the ones who tried this model ,Thanks a lot for supporting me Hope you have a great use of the model
This repository provides a custom Ollama Modelfile for OpenAI’s gpt-oss-20b, the state-of-the-art open weight model designed for powerful reasoning and agentic tasks.
This version is specifically configured to be an “unfettered” developer edition. It delivers the raw agentic capabilities of the model by removing the built-in, hardcoded tools (browser, python). This gives you, the developer, complete control and transparency over tool implementation.
The official gpt-oss model from OpenAI is fantastic, but it comes with pre-packaged tools. This version is for developers who need more control:
browse tool that uses Selenium instead of a simple GET request? You can build it. Need a Python sandbox with specific libraries? It’s all up to you.gpt-oss)This model retains all the powerful features of the original gpt-oss release:
analysis channel) for easier debugging and increased trust.Modelfile defaults to medium).ollama run mashriram/gpt-oss-Regular
—
This model uses a unique format for tool calling. Your application needs to be able to parse the model’s output and provide responses in the correct format.
Example Flow: Using a custom get_weather tool
Initial Request (Your App -> Ollama)
You provide the tool definitions in the .Tools field of your API request.
Model’s Response (Ollama -> Your App) The model will first output its reasoning, then the tool call.
<|start|>assistant<|channel|>analysis<|message|>The user is asking for the weather in a specific location. I need to use the `get_weather` tool.<|end|>
<|start|>assistant to=functions.get_weather<|channel|>commentary json<|message|>{"location": "San Francisco"}<|call|>
Tool Execution (Your App)
Your code parses the to=functions.get_weather and the arguments {"location": "San Francisco"}. You execute your function and get a result, e.g., {"temperature": "65F", "conditions": "Foggy"}.
Tool Response (Your App -> Ollama) You send the result back to the model, specifying the tool name.
Final Answer (Ollama -> Your App) The model processes the tool result and generates a final answer.
<|start|>assistant<|channel|>final<|message|>The weather in San Francisco is currently 65°F and Foggy.<|end|>
OpenAI uses MXFP4 quantization for the MoE weights in gpt-oss models. This Modelfile uses the raw GGUF, and Ollama’s engine supports this format natively, ensuring the highest possible quality without additional conversions.
The base gpt-oss model is licensed under the Apache 2.0 license, granting permissive use for commercial and private applications.