deepseek-v4-pro-vindex
Per-expert gate-vector vindex for deepseek-ai/DeepSeek-V4-Pro, built by the Divinci-AI team for use with LarQL (Chris Hay) and adjacent feature-routing inference research.
Vindex specs
- Source model:
deepseek-ai/DeepSeek-V4-Pro - Architecture:
deepseek_v4(61 layers, 7168 hidden, 3072 moe_intermediate) - Experts: 384 routed + 1 shared, 6 per token
- Layers indexed: 61 MoE layers (L00-L60)
- Features per expert: 64 (top-K right singular vectors of
gate_proj) - Format: float32, mmap-friendly contiguous binary
- Total size: 42.98 GB
What this is
gate_vectors.binโ flat float32 binary, layout[moe_layers, n_experts, num_feats, hidden_size]. Each per-expert chunk is the top-64 right singular vectors (Vt[:K, :]) of that expert'sgate_projweight after fp8/MXFP4 dequantization.gate_vectors_index.jsonโ sidecar with per-layerfile_offset(bytes),shape, and SVD stats (median_var64,q25_var64,q75_var64). Lookup table for mmap.phase1_moe_svd.jsonโ full per-layer Phase 1 stats (routed/shared/router decomposition).phase2_router_svd.jsonโ router weight SVD per layer (top-K variance, effective rank, s0/s1 ratio).
What this is not
- Not a runnable model (no inference path on its own).
- Not raw weights โ only top-K right singular vectors of
gate_proj, with the singular values not retained. Reconstruction is lossy. - Not a fine-tune or quantization of the base model.
Usage
import numpy as np
arr = np.memmap("gate_vectors.bin", dtype=np.float32, mode="r")
import json
idx = json.load(open("gate_vectors_index.json"))
moe = idx["model_config"]["moe"]
n_experts = moe["n_routed_experts"]
n_feats = idx["num_feats"]
hidden = moe["hidden_size"]
def get_layer(L):
meta = idx["layers"][str(L)]
offset = meta["file_offset"] // 4
n = n_experts * n_feats * hidden
return arr[offset:offset+n].reshape(n_experts, n_feats, hidden)
V_L1 = get_layer(1)
print("L1 expert 0 top vector L2 norm:", np.linalg.norm(V_L1[0, 0]))
Citation
@misc{divinci_deepseek_v4_pro_vindex_2026,
title = {deepseek-v4-pro-vindex: per-expert gate-vector vindex for deepseek-ai/DeepSeek-V4-Pro},
author = {Divinci-AI},
year = {2026},
url = {https://huggingface.co/Divinci-AI/deepseek-v4-pro-vindex},
}
Built using moe_vindex_builder.py.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for Divinci-AI/deepseek-v4-pro-vindex
Base model
deepseek-ai/DeepSeek-V4-Pro