New Users: 7-Day FreNew Users: 7-Day Free Trial of Kimi K2.6 - Limited Time OnlyDeepSeek V3.1
DeepSeek-V4-Flash API
ChatLLM

DeepSeek-V4-Flash API

All You Need To Know About DeepSeek-V4-Flash API

Overview

Model Provider:DeepSeek
Model Type:LLM
State:Ready

Key Specs

Quantization:FP8
Parameters:284B
Context:1M
Pricing:$0.14 input / $0.28 output / $0.028 cache
Try Model API
Quick Start
Reserve Dedicated Endpoint
Referral Program Banner

More referrals, More rewards

Invite new users and earn:

$20 free credits, $40 coupons, and 30-Day free plan.

The bigger your network, the bigger the bonus.

Introduction

DeepSeek-V4-Flash is a streamlined, high-efficiency model purpose-built by DeepSeek to achieve the perfect balance between blazing-fast inference and exceptional affordability. With a leaner parameter count and reduced activation overhead, V4-Flash powers an API that is both remarkably swift and budget-friendly. At its heart, the model delivers robust reasoning performance that strongly rivals the V4-Pro edition. Although it features a somewhat condensed repository of world knowledge, it remains fully capable of handling the vast majority of real-world use cases. In agent-based applications, V4-Flash matches the Pro version on routine and foundational tasks. As the go-to choice for developers prioritizing massive concurrency, minimal latency, and cost efficiency, DeepSeek-V4-Flash stands as the ideal solution for deploying large-scale, high-frequency, and lightweight AI workloads.

DeepSeek-V4-Flash API Usage

Model

Endpoint

deepseek/deepseek-v4-flash


        1
        curl -X POST https://inference.canopywave.io/v1/chat/completions \
      
        2
          -H "Content-Type: application/json" \
      
        3
          -H "Authorization: Bearer $CANOPYWAVE_API_KEY" \
      
        4
          -d '{
      
        5
            "model": "deepseek/deepseek-v4-flash",
      
        6
            "messages": [
      
        7
              {"role": "user", "content": "tell me a story"}
      
        8
            ],
      
        9
            "max_tokens": 1000,
      
        10
            "temperature": 0.7
      
        11
          }'