havenoammo/Qwen3.6-35B-A3B-MTP-GGUF
Image-Text-to-Text • 0.8B • Updated • 16.6k • 52
Would have liked to have seen Zipf weighting and PCA used for token level models like GloVe and BPEmb for a fair baseline (I assume for these models, just a mean sentence vector was computed) or is M2V_base_glove_subword just that?
Is M2V_base_output also using tokenlearn fine-tuning?