Aliosm

Rank:

Average Model Cost: $0.0000

Number of Runs: 4,795

Models by this creator

ComVE-distilgpt2

ComVE-distilgpt2

aliosm

ComVE-distilgpt2 Model description Finetuned model on Commonsense Validation and Explanation (ComVE) dataset introduced in SemEval2020 Task4 using a causal language modeling (CLM) objective. The model is able to generate a reason why a given natural language statement is against commonsense. Intended uses & limitations You can use the raw model for text generation to generate reasons why natural language statements are against commonsense. You can use this model directly to generate reasons why the given statement is against commonsense using generate.sh script. Note: make sure that you are using version 2.4.1 of transformers package. Newer versions has some issue in text generation and the model repeats the last token generated again and again. The model biased to negate the entered sentence usually instead of producing a factual reason. Training data The model is initialized from the distilgpt2 model and finetuned using ComVE dataset which contains 10K against commonsense sentences, each of them is paired with three reference reasons. Training procedure Each natural language statement that against commonsense is concatenated with its reference reason with <|continue|> as a separator, then the model finetuned using CLM objective. The model trained on Nvidia Tesla P100 GPU from Google Colab platform with 5e-5 learning rate, 15 epochs, 128 maximum sequence length and 64 batch size. Eval results The model achieved 13.7582/13.8026 BLEU scores on SemEval2020 Task4: Commonsense Validation and Explanation development and testing dataset. BibTeX entry and citation info

Read more

$-/run

4.6K

Huggingface

ComVE-gpt2

ComVE-gpt2

ComVE-gpt2 Model description Finetuned model on Commonsense Validation and Explanation (ComVE) dataset introduced in SemEval2020 Task4 using a causal language modeling (CLM) objective. The model is able to generate a reason why a given natural language statement is against commonsense. Intended uses & limitations You can use the raw model for text generation to generate reasons why natural language statements are against commonsense. You can use this model directly to generate reasons why the given statement is against commonsense using generate.sh script. Note: make sure that you are using version 2.4.1 of transformers package. Newer versions has some issue in text generation and the model repeats the last token generated again and again. The model biased to negate the entered sentence usually instead of producing a factual reason. Training data The model is initialized from the gpt2 model and finetuned using ComVE dataset which contains 10K against commonsense sentences, each of them is paired with three reference reasons. Training procedure Each natural language statement that against commonsense is concatenated with its reference reason with <|continue|> as a separator, then the model finetuned using CLM objective. The model trained on Nvidia Tesla P100 GPU from Google Colab platform with 5e-5 learning rate, 5 epochs, 128 maximum sequence length and 64 batch size. Eval results The model achieved 14.0547/13.6534 BLEU scores on SemEval2020 Task4: Commonsense Validation and Explanation development and testing dataset. BibTeX entry and citation info

Read more

$-/run

21

Huggingface

ComVE-gpt2-medium

ComVE-gpt2-medium

ComVE-gpt2-medium Model description Finetuned model on Commonsense Validation and Explanation (ComVE) dataset introduced in SemEval2020 Task4 using a causal language modeling (CLM) objective. The model is able to generate a reason why a given natural language statement is against commonsense. Intended uses & limitations You can use the raw model for text generation to generate reasons why natural language statements are against commonsense. You can use this model directly to generate reasons why the given statement is against commonsense using generate.sh script. Note: make sure that you are using version 2.4.1 of transformers package. Newer versions has some issue in text generation and the model repeats the last token generated again and again. The model biased to negate the entered sentence usually instead of producing a factual reason. Training data The model is initialized from the gpt2-medium model and finetuned using ComVE dataset which contains 10K against commonsense sentences, each of them is paired with three reference reasons. Training procedure Each natural language statement that against commonsense is concatenated with its reference reason with <|continue|> as a separator, then the model finetuned using CLM objective. The model trained on Nvidia Tesla P100 GPU from Google Colab platform with 5e-5 learning rate, 5 epochs, 128 maximum sequence length and 64 batch size. Eval results The model achieved fifth place with 16.7153/16.1187 BLEU scores and third place with 1.94 Human Evaluation score on SemEval2020 Task4: Commonsense Validation and Explanation development and testing dataset. These are some examples generated by the model: BibTeX entry and citation info

Read more

$-/run

20

Huggingface

Similar creators