manandey commited on
Commit
113c045
·
1 Parent(s): ecc7ad5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -13,10 +13,10 @@ st.markdown('''
13
 
14
  Hi! This is the demo for the [flax sentence embeddings](https://huggingface.co/flax-sentence-embeddings) created for the **Flax/JAX community week 🤗**.
15
  We trained three general-purpose flax-sentence-embeddings models: a **distilroberta base**, a **mpnet base** and a **minilm-l6**.
16
- All were trained on all the datasets of the [1B+ train corpus](https://huggingface.co/flax-sentence-embeddings/all_datasets_v4_MiniLM-L6#training-data) with the v3 setup.
17
 
18
  In addition, we trained [20 models](https://huggingface.co/flax-sentence-embeddings) focused on general-purpose, QuestionAnswering and Code search.
19
- View our models [here](https://huggingface.co/flax-sentence-embeddings)
20
 
21
  To evaluate the [flax-sentence-embeddings](https://huggingface.co/flax-sentence-embeddings) *Sentence Similarity models* for *Gender Bias based on stereotypical occupations*, we created an *evaluation set* which can be found [here](https://huggingface.co/datasets/manandey/Gender_Bias_Evaluation)
22
 
@@ -63,8 +63,8 @@ For more cool information on sentence embeddings, see the [sBert project](https:
63
  elif menu == "Asymmetric QA":
64
  st.header('Asymmetric QA')
65
  st.markdown('''
66
- **Instructions**: You can compare the Answer likeliness of a given Query with answer candidates of your choice. In the background, we'll create an embedding for each answers, and then we'll use the cosine similarity function to calculate a similarity metric between our query sentence and the others.
67
- `mpnet_asymmetric_qa` model works best for hard negative answers or distinguishing similar queries due to separate models applied for encoding questions and answers.
68
 
69
  For more cool information on sentence embeddings, see the [sBert project](https://www.sbert.net/examples/applications/computing-embeddings/README.html).
70
  ''')
 
13
 
14
  Hi! This is the demo for the [flax sentence embeddings](https://huggingface.co/flax-sentence-embeddings) created for the **Flax/JAX community week 🤗**.
15
  We trained three general-purpose flax-sentence-embeddings models: a **distilroberta base**, a **mpnet base** and a **minilm-l6**.
16
+ The models were trained on a dataset comprising of [1 Billion+ training corpus](https://huggingface.co/flax-sentence-embeddings/all_datasets_v4_MiniLM-L6#training-data) with the v3 setup.
17
 
18
  In addition, we trained [20 models](https://huggingface.co/flax-sentence-embeddings) focused on general-purpose, QuestionAnswering and Code search.
19
+ You can view our models [here](https://huggingface.co/flax-sentence-embeddings)
20
 
21
  To evaluate the [flax-sentence-embeddings](https://huggingface.co/flax-sentence-embeddings) *Sentence Similarity models* for *Gender Bias based on stereotypical occupations*, we created an *evaluation set* which can be found [here](https://huggingface.co/datasets/manandey/Gender_Bias_Evaluation)
22
 
 
63
  elif menu == "Asymmetric QA":
64
  st.header('Asymmetric QA')
65
  st.markdown('''
66
+ **Instructions**: You can compare the Answer likeliness of a given Query with answer candidates of your choice. In the background, we'll create an embedding for each answer, and then we'll use the cosine similarity function to calculate a similarity metric between our query sentence and the others.
67
+ `mpnet_asymmetric_qa` model works best for hard-negative answers or distinguishing similar queries due to separate models applied for encoding questions and answers.
68
 
69
  For more cool information on sentence embeddings, see the [sBert project](https://www.sbert.net/examples/applications/computing-embeddings/README.html).
70
  ''')