BoyuNLP commited on
Commit
2a24238
·
verified ·
1 Parent(s): a1dcbe1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -12
README.md CHANGED
@@ -30,28 +30,25 @@ UGround is a strong GUI visual grounding model trained with a simple recipe. Che
30
  - [UGround-V1-2B (Qwen2-VL)](https://huggingface.co/osunlp/UGround-V1-2B)
31
  - [UGround-V1-7B (Qwen2-VL)](https://huggingface.co/osunlp/UGround-V1-7B)
32
  - [UGround-V1-72B (Qwen2-VL)](https://huggingface.co/osunlp/UGround-V1-72B)
33
- - [Training Data](https://huggingface.co/osunlp/UGround)
34
 
35
  ## Release Plan
36
 
37
  - [x] [Model Weights](https://huggingface.co/collections/osunlp/uground-677824fc5823d21267bc9812)
38
  - [x] Initial Version (the one used in the paper)
39
- - [x] Qwen2-VL-Based V1
40
- - [x] 2B
41
- - [x] 7B
42
- - [x] 72B
43
  - [x] Code
44
- - [x] Inference Code of UGround (Initial & Qwen2-VL-Based
45
  - [x] Offline Experiments (Code, Results, and Useful Resources)
46
- - [x] ScreenSpot (along with referring expressions generated by GPT-4/4o)
47
- - [x] Multimodal-Mind2Web
48
- - [x] OmniAct
49
- - [x] Android Control
50
  - [x] Online Experiments
51
- - [x] Mind2Web-Live-SeeAct-V
52
  - [x] [AndroidWorld-SeeAct-V](https://github.com/boyugou/android_world_seeact_v)
53
  - [ ] Data Synthesis Pipeline (Coming Soon)
54
- - [x] Training-Data (V1)
55
  - [x] Online Demo (HF Spaces)
56
 
57
 
 
30
  - [UGround-V1-2B (Qwen2-VL)](https://huggingface.co/osunlp/UGround-V1-2B)
31
  - [UGround-V1-7B (Qwen2-VL)](https://huggingface.co/osunlp/UGround-V1-7B)
32
  - [UGround-V1-72B (Qwen2-VL)](https://huggingface.co/osunlp/UGround-V1-72B)
33
+ - [Training Data](https://huggingface.co/datasets/osunlp/UGround-V1-Data)
34
 
35
  ## Release Plan
36
 
37
  - [x] [Model Weights](https://huggingface.co/collections/osunlp/uground-677824fc5823d21267bc9812)
38
  - [x] Initial Version (the one used in the paper)
39
+ - [x] Qwen2-VL-Based V1 (2B, 7B, 72B)
 
 
 
40
  - [x] Code
41
+ - [x] [Inference Code of UGround (Initial & Qwen2-VL-Based)](https://github.com/boyugou/llava_uground/)
42
  - [x] Offline Experiments (Code, Results, and Useful Resources)
43
+ - [x] [ScreenSpot](https://github.com/OSU-NLP-Group/UGround/tree/main/offline_evaluation/ScreenSpot)
44
+ - [x] [Multimodal-Mind2Web](https://github.com/OSU-NLP-Group/UGround/tree/main/offline_evaluation/Multimodal-Mind2Web)
45
+ - [x] [OmniAct](https://github.com/OSU-NLP-Group/UGround/tree/main/offline_evaluation/OmniACT)
46
+ - [x] [Android Control](https://github.com/OSU-NLP-Group/UGround/tree/main/offline_evaluation/AndroidControl)
47
  - [x] Online Experiments
48
+ - [x] [Mind2Web-Live-SeeAct-V](https://github.com/boyugou/Mind2Web_Live_SeeAct_V)
49
  - [x] [AndroidWorld-SeeAct-V](https://github.com/boyugou/android_world_seeact_v)
50
  - [ ] Data Synthesis Pipeline (Coming Soon)
51
+ - [x] [Training-Data (V1)](https://huggingface.co/datasets/osunlp/UGround-V1-Data)
52
  - [x] Online Demo (HF Spaces)
53
 
54