mDeBERTa-v3-base-MNLI-XNLI

by Vrije Universiteit Amsterdam

mDeBERTa-v3-base-MNLI-XNLI

by Vrije Universiteit Amsterdam

Main use cases: A multilingual model based on a further development of the BERT architecture. It was explicitly trained for concern recognition and text classification without examples (zero shot).  

 
Input length: 512 tokens (approx. 384 words) - theoretically 24,528 tokens, but considerable speed losses are to be expected above 512 tokens.  

 
Languages: Evaluated for 15 languages, including English and German. To a lesser extent 85 other languages.  

 
Model size: ~86 million parameters

Main use cases: A multilingual model based on a further development of the BERT architecture. It was explicitly trained for concern recognition and text classification without examples (zero shot).  

 
Input length: 512 tokens (approx. 384 words) - theoretically 24,528 tokens, but considerable speed losses are to be expected above 512 tokens.  

 
Languages: Evaluated for 15 languages, including English and German. To a lesser extent 85 other languages.  

 
Model size: ~86 million parameters