Qwen3.5-397B-A17B & Qwen3.5-397B-A17B & MiniMax M2.5 is Live on Canopy Wave. Try it Now!DeepSeek V3.1

Open Code

Step 1. Install Open Code

npm install -g opencode-ai

Step 2. Set Canopy Wave API

    1. 1. Configuration file locations: You can place opencode.json in the project root directory based on the project's location, or set it as the global ~/.config/opencode/opencode.json.
    2. 2. Get your model api key from Model API Key Canopy Wave.
    3. 3. Copy the json file below. Replace the Bearer token key with your actual model API key you got from Model API Key in the previous step. opencode.json should look like this
    {
      "$schema": "https://opencode.ai/config.json",
      "provider": {
        "myprovider": {
          "npm": "@ai-sdk/openai-compatible",
          "name": "Canopy Wave",
          "options": {
            "baseURL": "https://inference.canopywave.io/v1",
            "headers": {
              "Authorization": "Bearer your_canopywave_key"
            }
          },
          "models": {
            "zai/glm-4.7": {
              "name": "glm47"
            }
          }
        }
      }
    }

Step 3. Use

    1. 1. Type opencode will open the console to talk to the model.
    2. Use OpenCode
    3. 2. Use the /connect command to add provider:
    4. Add Provider
    5. 3. Enter the provider name you just set in the JSON.
    6. Add Provider
    7. 4. Enter your model API key in 2(3). Press Enter.
    8. Add Provider
    9. 5. Start using our best open-source models on your OpenCode.
    10. Add Provider

If you have questions, please contact our support team at support@canopywave.com, and we will gladly get back to you!

PromotionContact us