"NoSleep" Writing Prompt Generator

Finetuned version of GPT2 to facilitate generation of Writing Prompts for the GPT-NoSleep-355m model

You can use the space linked on the right to use this model, then use the NoSleep model in tandem to generate stories!

Training Procedure

This was trained on the 'reddt-nosleep-posts' dataset, using the "HappyTransformers" library on Google Colab. This model was trained for X epochs with learning rate 1e-2.

Biases & Limitations

This likely contains the same biases and limitations as the original GPT2 that it is based on, and additionally heavy biases from the dataset. It likely will generate offensive output.

Intended Use

This model is meant for fun, nothing else.

Sample Use

from happytransformer import GENSettings
args_top_k = GENSettings(no_repeat_ngram_size=1, do_sample=True, top_k=80, temperature=0.4, max_length=25, early_stopping=True)

result = happy_gen.generate_text("[WP] \"", args=args_top_k)
print(result.text)
Downloads last month
112
Safetensors
Model size
137M params
Tensor type
F32
·
BOOL
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Dataset used to train DarwinAnim8or/NoSleepPromptGen

Space using DarwinAnim8or/NoSleepPromptGen 1