Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yeah, the standard SPLADE model trained from BERT typically already has a vocabulary/vector size of 30,552. If the SPLADE model is based on a multilingual version of BERT, such as mBERT or XLM-R, the vocabulary size could inherently expand to approximately 100,000, as does the vector size.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: