Malteos
Rank:Average Model Cost: $0.0000
Number of Runs: 45,165
Models by this creator
bloom-6b4-clp-german
$-/run
888
Huggingface
bloom-1b5-clp-german
bloom-1b5-clp-german
The multilingual bloom-1b7 model adapted to the German language using the CLP-Transfer method. Tokenizer from https://huggingface.co/malteos/gpt2-xl-wechsel-german Trained on the German subset of OSCAR https://huggingface.co/datasets/oscar See also malteos/bloom-6b4-clp-german malteos/bloom-6b4-clp-german-oasst-v0.1
$-/run
748
Huggingface
gpt2-wechsel-german-ds-meg
gpt2-wechsel-german-ds-meg
Replication of gpt2-wechsel-german trained with BigScience's DeepSpeed-Megatron-LM code base 22hrs on 4xA100 GPUs (~ 80 TFLOPs / GPU) stopped after 100k steps less than a single epoch on oscar_unshuffled_deduplicated_de (excluding validation set; original model was trained for 75 epochs on less data) bf16 zero stage 1 tp/pp = 1 Evaluation License MIT
$-/run
430
Huggingface
gpt2-xl-wechsel-german
$-/run
306
Huggingface
bloom-350m-german
$-/run
162
Huggingface
scincl-wol
$-/run
121
Huggingface
PubMedNCL
PubMedNCL
PubMedNCL A pretrained language model for document representations of biomedical papers. PubMedNCL is based on PubMedBERT, which is a BERT model pretrained on abstracts and full-texts from PubMedCentral, and fine-tuned via citation neighborhood contrastive learning, as introduced by SciNCL. How to use the pretrained model Citation Neighborhood Contrastive Learning for Scientific Document Representations with Citation Embeddings (EMNLP 2022 paper). Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing. License MIT
$-/run
105
Huggingface
bloom-6b4-clp-german-oasst-v0.1
bloom-6b4-clp-german-oasst-v0.1
Instruction-fine-tuned German language model (6B parameters; early alpha version) Base model: malteos/bloom-6b4-clp-german (Ostendorff and Rehm, 2023) Trained on: 20B additional German tokens (Wikimedia dumps and OSCAR 2023) OpenAssistant/oasst1 (German subset) LEL-A/translated_german_alpaca_validation LEL-A's version of deepset/germandpr Chat demo https://opengptx.dfki.de/chat/ Please note that this a research prototype and may not be suitable for extensive use. How to cite If you are using our code or models, please cite our paper: License BigScience BLOOM RAIL 1.0 Acknowledgements This model was trained during the Helmholtz GPU Hackathon 2023. We gratefully thank the organizers for hosting this event and the provided computing resources.
$-/run
45
Huggingface
aspect-acl-scibert-scivocab-uncased
aspect-acl-scibert-scivocab-uncased
Platform did not provide a description for this model.
$-/run
41
Huggingface