Sign up now! New useSign up now! New users get $20 in free creditsDeepSeek V3.1
GLM 4.6 API
CODELLM

GLM 4.6 API

All You Need To Know About GLM 4.6 API

Overview

Model Provider:Zai-org
Model Type:CODE/LLM
State:Take offline

Key Specs

Quantization:BF16
Parameters:355B
Context:128k
Pricing:$0.45 input / $1.50 output
Take offline
Quick Start
Reserve Dedicated Endpoint

Introduction

As the latest iteration in the GLM series, GLM-4.6 introduces major improvements over GLM-4.5, and the new glm 4.6 api brings those enhancements directly to developers. Key upgrades include a significantly expanded 200K-token context window for more complex workloads, stronger real-world coding performance, and more advanced reasoning supported by improved tool-use integration. With the glm-4.6 api, teams can leverage these capabilities seamlessly in production applications.

In addition, GLM-4.6 provides more advanced agentic behaviors and more natural, human-aligned output—capabilities that become fully accessible through the glm-4.6 api. Benchmark results show that GLM-4.6 achieves notable improvements over GLM-4.5 while matching or surpassing other top-tier models. With the glm 4.6 api, developers can directly leverage these gains for reasoning, coding, and agentic applications.

GLM 4.6 API Usage

Model

Endpoint

zai/glm-4.6


        1
        curl -X POST https://inference.canopywave.io/v1 \
      
        2
          -H "Content-Type: application/json" \
      
        3
          -H "Authorization: Bearer $CANOPYWAVE_API_KEY" \
      
        4
          -d '{
      
        5
            "model": "zai/glm-4.6",
      
        6
            "messages": [
      
        7
              {"role": "system", "content": "You are a helpful assistant."},
      
        8
              {"role": "user", "content": "please tell me a story."}
      
        9
            ],
      
        10
          }'
      
Contact us