Results for "fill-mask"
100 matches found.
google-bert/bert-base-uncased
BERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the ...
FacebookAI/roberta-large
RoBERTa is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on t...
FacebookAI/xlm-roberta-base
XLM-RoBERTa is a multilingual version of RoBERTa. It is pre-trained on 2.5TB of filtered CommonCrawl data containing 100 languages. RoBERTa ...
FacebookAI/roberta-base
RoBERTa is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on t...
distilbert/distilbert-base-uncased
DistilBERT is a transformers model, smaller and faster than BERT, which was pretrained on the same corpus in a self-supervised fashion, usin...
FacebookAI/xlm-roberta-large
XLM-RoBERTa is a multilingual version of RoBERTa. It is pre-trained on 2.5TB of filtered CommonCrawl data containing 100 languages. RoBERTa ...
google-bert/bert-base-multilingual-uncased
BERT is a transformers model pretrained on a large corpus of multilingual data in a self-supervised fashion. This means it was pretrained on...
google-bert/bert-base-multilingual-cased
BERT is a transformers model pretrained on a large corpus of multilingual data in a self-supervised fashion. This means it was pretrained on...
google-bert/bert-base-cased
BERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the ...
emilyalsentzer/Bio_ClinicalBERT
- fill-mask...
microsoft/deberta-v3-base
No description available.
distilbert/distilroberta-base
No description available.
facebook/esm2_t33_650M_UR50D
No description available.
microsoft/deberta-base
No description available.
thomas-sounack/BioClinical-ModernBERT-base
No description available.
google-bert/bert-base-chinese
No description available.
distilbert/distilbert-base-multilingual-cased
No description available.
answerdotai/ModernBERT-base
No description available.
dccuchile/bert-base-spanish-wwm-uncased
No description available.
kakaobank/kf-deberta-base
KF-DeBERTaλ λ²μ© λλ©μΈ λ§λμΉμ κΈμ΅ λλ©μΈ λ§λμΉλ₯Ό ν¨κ» νμ΅ν μΈμ΄λͺ¨λΈ μ λλ€. λͺ¨λΈ μν€ν μ³λ DeBERTa-v2λ₯Ό κΈ°λ°μΌλ‘ νμ΅νμμ΅λλ€. ELECTRAμ RTDλ₯Ό training objectiveλ‘ μ¬μ©ν DeBERTa-v3λ μΌλΆ...
microsoft/deberta-v3-large
No description available.
microsoft/mdeberta-v3-base
No description available.
almanach/camembert-base
No description available.
microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract
No description available.
neuralmind/bert-base-portuguese-cased
No description available.
facebook/esm2_t6_8M_UR50D
No description available.
google-bert/bert-large-uncased
BERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the ...
facebook/esm2_t12_35M_UR50D
No description available.
microsoft/graphcodebert-base
GraphCodeBERT is a graph-based pre-trained model based on the Transformer architecture for programming language, which also considers data-f...
neuralmind/bert-large-portuguese-cased
No description available.
facebook/esm2_t36_3B_UR50D
No description available.
aubmindlab/bert-base-arabertv02
No description available.
tohoku-nlp/bert-base-japanese-whole-word-masking
No description available.
julien-c/dummy-unknown
No description available.
google-bert/bert-base-german-cased
No description available.
nlpaueb/bert-base-greek-uncased-v1
No description available.
emilyalsentzer/Bio_Discharge_Summary_BERT
No description available.
albert/albert-base-v2
ALBERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on th...
microsoft/deberta-v3-small
No description available.
beomi/kcbert-base
No description available.
vinai/bertweet-base
No description available.
distilbert/distilbert-base-german-cased
No description available.
vinai/phobert-base
Pre-trained PhoBERT models are the state-of-the-art language models for Vietnamese (Pho, i.e. "Phα»", is a popular food in Vietnam):...
jhu-clsp/mmBERT-base
mmBERT represents the first significant advancement over XLM-R for massively multilingual encoder models. Key features include: 1. Massive L...
nlpaueb/legal-bert-base-uncased
No description available.
yikuan8/Clinical-Longformer
- longformer - clinical...
DeepChem/ChemBERTa-77M-MLM
No description available....
answerdotai/ModernBERT-large
No description available.
tohoku-nlp/bert-base-japanese
No description available.
Shushant/nepaliBERT
Pretraining done on bert base architecture....
microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract-fulltext
No description available.
vinai/phobert-base-v2
No description available.
hfl/chinese-roberta-wwm-ext
No description available.
google-bert/bert-large-cased
BERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the ...
microsoft/deberta-v2-xlarge
No description available.
distilbert/distilbert-base-cased
DistilBERT is a transformers model, smaller and faster than BERT, which was pretrained on the same corpus in a self-supervised fashion, usin...
seyonec/ChemBERTa-zinc-base-v1
No description available.
Rostlab/prot_bert_bfd
ProtBert-BFD is based on Bert model which pretrained on a large corpus of protein sequences in a self-supervised fashion. This means it was ...
Rostlab/prot_bert
ProtBert is based on Bert model which pretrained on a large corpus of protein sequences in a self-supervised fashion. This means it was pret...
UBC-NLP/MARBERTv2
MARBERTv2 is one of three models described in our ACL 2021 paper "ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic". We find tha...
Derify/ChemBERTa_augmented_pubchem_13m
- Derify/augmentedcanonicalpubchem13m - rocauc - rmse libraryname: transformers - ChemBERTa - cheminformatics pipelinetag: fill-mask - name:...
microsoft/mpnet-base
No description available....
microsoft/infoxlm-large
InfoXLM (NAACL 2021, paper, repo, model) InfoXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training....
facebook/esm1v_t33_650M_UR90S_1
No description available....
neulab/codebert-python
This is a `microsoft/codebert-base-mlm` model, trained for 1,000,000 steps (with `batchsize=32`) on Python code from the `codeparrot/github-...
FacebookAI/xlm-mlm-en-2048
The XLM model was proposed in Cross-lingual Language Model Pretraining by Guillaume Lample and Alexis Conneau. Itβs a transformer pretrained...
albert/albert-base-v1
ALBERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on th...
kuleshov-group/mdlm-owt
The model, which has a context length of `1024` and is similar in size to GPT2-medium with approximately `130 million` non-embedding paramet...
weiweishi/roc-bert-base-zh
No description available.
flaubert/flaubert_base_cased
No description available.
studio-ousia/luke-base
No description available.
junnyu/roformer_chinese_small
https://github.com/ZhuiyiTechnology/roformer...
faisalq/bert-base-arapoembert
- ar - Arabic BERT - Poetry - Masked Langauge Model...
moussaKam/mbarthez
- summarization...
deepmind/language-perceiver
Perceiver IO is a transformer encoder model that can be applied on any modality (text, images, audio, video, ...). The core idea is to emplo...
facebook/esm2_t30_150M_UR50D
No description available.
tohoku-nlp/bert-base-japanese-char
No description available.
nreimers/MiniLMv2-L6-H384-distilled-from-BERT-Large
No description found....
novelcore/gem-roberta
GEM-RoBERTa HQ Legal is a RoBERTa-base model pre-trained from scratch on a strategically curated 21GB corpus of Greek legal, parliamentary, ...
dbmdz/bert-base-italian-xxl-cased
No description available.
monologg/distilkobert
No description available.
alabnii/jmedroberta-base-sentencepiece
This is a Japanese RoBERTa base model pre-trained on academic articles in medical sciences collected by Japan Science and Technology Agency ...
LazarusNLP/NusaBERT-base
No description available.
facebook/esm1b_t33_650M_UR50S
No description available.
dmis-lab/biobert-base-cased-v1.2
No description available....
facebook/xlm-roberta-xl
XLM-RoBERTa-XL is a extra large multilingual version of RoBERTa. It is pre-trained on 2.5TB of filtered CommonCrawl data containing 100 lang...
kykim/bert-kor-base
No description available.
DeepChem/MoLFormer-c3-1.1B
No description available.
nlpaueb/bert-base-uncased-contracts
No description available.
dandelin/vilt-b32-mlm
No description available.
tohoku-nlp/bert-base-japanese-char-v2
No description available.
anferico/bert-for-patents
- en - masked-lm - pytorch pipeline-tag: "fill-mask" mask-token: "[MASK]" - text: "The present [MASK] provides a torque sensor that is small...
PlanTL-GOB-ES/bsc-bio-ehr-es
Biomedical pretrained language model for Spanish. For more details about the corpus, the pretraining and the evaluation, check the official ...
westlake-repl/SaProt_650M_AF2
No description available.
airesearch/wangchanberta-base-att-spm-uncased
The architecture of the pretrained model is based on RoBERTa [[Liu et al., 2019]](https://arxiv.org/abs/1907.11692)....
medicalai/ClinicalBERT
No description available.
klue/roberta-large
No description available.
aubmindlab/bert-base-arabertv2
No description available.
microsoft/BiomedVLP-CXR-BERT-specialized
No description available.
huggingface/CodeBERTa-small-v1
No description available.