airgap_qwen
What it is
- A module that defines a preconfigured
ChatModelinstance for the Qwen3 model running in “airgap” mode via a local Docker Model Runner endpoint.
Public API
- Module constants
MODEL_ID:"ai/qwen3"PROVIDER:"qwen"NAME:"qwen3-airgap"DESCRIPTION: Human-readable description of the model configuration.IMAGE: URL for a representative image.CONTEXT_WINDOW:8192(declared, not wired into theChatModelin this file)
- Module variable
model: ChatModel: AChatModelwrapping alangchain_openai.ChatOpenAIclient configured to talk to a local endpoint.
Configuration/Dependencies
- Python dependencies
langchain_openai.ChatOpenAInaas_abi_core.models.Model.ChatModel
- Runtime dependency
- A service reachable at:
http://localhost:12434/engines/v1
- A service reachable at:
- Model configuration
ChatOpenAI(model="ai/qwen3", temperature=0.7, base_url="http://localhost:12434/engines/v1")
Usage
from naas_abi.models.airgap_qwen import model
# Access metadata
print(model.model_id) # "ai/qwen3"
print(model.provider) # "qwen"
print(model.name) # "qwen3-airgap"
# Access the underlying LangChain chat model/client
llm = model.model
print(llm)
Caveats
- The module only defines configuration; successful usage requires a compatible local server running at
http://localhost:12434/engines/v1. CONTEXT_WINDOWis declared but not applied to theChatOpenAI/ChatModelconfiguration in this file.