Login / Signup

Large Pre-Trained Models with Extra-Large Vocabularies: A Contrastive Analysis of Hebrew BERT Models and a New One to Outperform Them All.

Eylon GuettaAvi ShmidmanShaltiel ShmidmanCheyn Shmuel ShmidmanJoshua GuedaliaMoshe KoppelDan BareketAmit SekerReut Tsarfaty
Published in: CoRR (2022)
Keyphrases
  • probabilistic model
  • prior knowledge
  • dimensionality reduction