RoBERTa Large OpenAI Detector

This model is a FP16 optimized version of openai-community/roberta-large-openai-detector. It runs exclusively on the GPU. The speedup compared to the base ONNX and pytorch versions depends chiefly on your GPU's FP16:FP32 ratio. For more comparison benchmarks and sample code of a related model, check here: https://github.com/joaopn/gpu_benchmark_goemotions.

You will need the GPU version of the ONNX Runtime. It can be installed with

pip install optimum[onnxruntime-gpu] --extra-index-url https://aiinfra.pkgs.visualstudio.com/PublicPackages/_packaging/onnxruntime-cuda-12/pypi/simple/

For convenience, this benchmark repo provides an environment.yml file to create a conda env with all the requirements.

Downloads last month
12
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for joaopn/roberta-large-openai-detector-onnx-fp16

Quantized
(1)
this model

Datasets used to train joaopn/roberta-large-openai-detector-onnx-fp16