Tsmatz
Rank:Average Model Cost: $0.0000
Number of Runs: 150,190
Models by this creator
xlm-roberta-ner-japanese
$-/run
146.9K
Huggingface
mt5_summarize_japanese
$-/run
2.5K
Huggingface
roberta_qa_japanese
roberta_qa_japanese
roberta_qa_japanese (Japanese caption : 日本語の (抽出型) 質問応答のモデル) This model is a fine-tuned version of rinna/japanese-roberta-base (pre-trained RoBERTa model provided by rinna Co., Ltd.) trained for extractive question answering. The model is fine-tuned on JaQuAD dataset provided by Skelter Labs, in which data is collected from Japanese Wikipedia articles and annotated by a human. Intended uses When running with a dedicated pipeline : When manually running through forward pass : Training procedure You can download the source code for fine-tuning from here. Training hyperparameters The following hyperparameters were used during training: learning_rate: 7e-05 train_batch_size: 2 eval_batch_size: 1 seed: 42 gradient_accumulation_steps: 16 total_train_batch_size: 32 optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 lr_scheduler_type: linear lr_scheduler_warmup_steps: 100 num_epochs: 3 Training results Framework versions Transformers 4.23.1 Pytorch 1.12.1+cu102 Datasets 2.6.1 Tokenizers 0.13.1
$-/run
753
Huggingface