Spaces:
Paused
Paused
| # Models | |
| With the `AutoModelForCausalLMWithValueHead` class TRL supports all decoder model architectures in transformers such as GPT-2, OPT, and GPT-Neo. In addition, with `AutoModelForSeq2SeqLMWithValueHead` you can use encoder-decoder architectures such as T5. TRL also requires reference models which are frozen copies of the model that is trained. With `create_reference_model` you can easily create a frozen copy and also share layers between the two models to save memory. | |
| ## PreTrainedModelWrapper | |
| [[autodoc]] PreTrainedModelWrapper | |
| ## AutoModelForCausalLMWithValueHead | |
| [[autodoc]] AutoModelForCausalLMWithValueHead | |
| - __init__ | |
| - forward | |
| - generate | |
| - _init_weights | |
| ## AutoModelForSeq2SeqLMWithValueHead | |
| [[autodoc]] AutoModelForSeq2SeqLMWithValueHead | |
| - __init__ | |
| - forward | |
| - generate | |
| - _init_weights | |
| ## create_reference_model | |
| [[autodoc]] create_reference_model |