| # HPSv2-hf | |
| This is a Huggingface CLIPModel flavor of the [HPSv2](https://github.com/tgxs002/HPSv2/) model, which is trained to predict human preferences over AI generated images. | |
| I converted the model weights from the openclip format to huggingface CLIPModel. | |
| The two text and image embeddings were tested to be equal before and after conversion. | |
| You can load the model the same as any huggingface clip model: | |
| ```python | |
| from transformers import CLIPProcessor, CLIPModel | |
| model = CLIPModel.from_pretrained("adams-story/HPSv2-hf") | |
| processor = CLIPProcessor.from_pretrained("openai/clip-vit-base-patch32") # uses the same exact vanilla clip processor | |
| ``` | |
| All credit goes to the original authors of HPSv2 |