• search
    search
  • reviewers
    reviewers
  • feeds
    feeds
  • assignments
    assignments
  • settings
  • logout

Large Pre-Trained Models with Extra-Large Vocabularies: A Contrastive Analysis of Hebrew BERT Models and a New One to Outperform Them All.

Eylon GuettaAvi ShmidmanShaltiel ShmidmanCheyn Shmuel ShmidmanJoshua GuedaliaMoshe KoppelDan BareketAmit SekerReut Tsarfaty
Published in: CoRR (2022)
Keyphrases
  • probabilistic model
  • prior knowledge
  • dimensionality reduction