MiniMax M2.5 is LiveMiniMax M2.5 is Live on Canopy Wave. Try it Now!DeepSeek V3.1
GLM 4.6 API
CODELLM

GLM 4.6 API

All You Need To Know About GLM 4.6 API

Overview

Model Provider:Zai-org
Model Type:CODE/LLM
State:Take offline

Key Specs

Quantization:BF16
Parameters:355B
Context:128k
Take offline
Quick Start
Reserve Dedicated Endpoint

Introduction

As the latest iteration in the GLM series, GLM-4.6 introduces major improvements over GLM-4.5, and the new glm 4.6 api brings those enhancements directly to developers. Key upgrades include a significantly expanded 200K-token context window for more complex workloads, stronger real-world coding performance, and more advanced reasoning supported by improved tool-use integration. With the glm-4.6 api, teams can leverage these capabilities seamlessly in production applications.

In addition, GLM-4.6 provides more advanced agentic behaviors and more natural, human-aligned output—capabilities that become fully accessible through the glm-4.6 api. Benchmark results show that GLM-4.6 achieves notable improvements over GLM-4.5 while matching or surpassing other top-tier models. With the glm 4.6 api, developers can directly leverage these gains for reasoning, coding, and agentic applications.

GLM 4.6 API Usage

Model

Endpoint

zai/glm-4.6


        1
        curl -X POST https://inference.canopywave.io/v1 \
      
        2
          -H "Content-Type: application/json" \
      
        3
          -H "Authorization: Bearer $CANOPYWAVE_API_KEY" \
      
        4
          -d '{
      
        5
            "model": "zai/glm-4.6",
      
        6
            "messages": [
      
        7
              {"role": "system", "content": "You are a helpful assistant."},
      
        8
              {"role": "user", "content": "please tell me a story."}
      
        9
            ],
      
        10
          }'
      
PromotionContact us