aooyoo commited on
Commit
38a6f5f
·
verified ·
1 Parent(s): 684d229

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +98 -3
README.md CHANGED
@@ -1,3 +1,98 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - zh
6
+ base_model:
7
+ - Qwen/Qwen2.5-3B-Instruct
8
+ tags:
9
+ - medical
10
+ - cancer
11
+ - Onco
12
+ ---
13
+ # OncoCareBrain-GPT
14
+
15
+ ## Model Description
16
+
17
+ OncoCareBrain-GPT is a specialized large language model fine-tuned for oncology applications. Built upon the powerful Qwen2.5-3B foundation model, it has undergone supervised fine-tuning (SFT) with tens of thousands of multi-omics data samples, including genomic, pathological, and clinical data. This model is specifically designed to serve the cancer care domain with advanced reasoning capabilities.
18
+
19
+ ## Key Features
20
+
21
+ - **Intelligent Medical Q&A**: Quickly answers complex questions about cancer, leveraging a deep understanding of oncology concepts
22
+ - **Precision Decision Support**: Recommends optimal treatment plans based on multi-dimensional data analysis
23
+ - **Transparent Reasoning Process**: Generates detailed chains of thought to ensure model explainability and trust in clinical settings
24
+
25
+ ## Intended Uses
26
+
27
+ - **Clinical Decision Support**: Assists healthcare providers in evaluating treatment options
28
+ - **Patient Education**: Helps patients better understand their condition and treatment plans
29
+ - **Medical Research**: Supports researchers in analyzing cancer data and generating insights
30
+
31
+ ## Training Data
32
+
33
+ OncoCareBrain-GPT was fine-tuned on a diverse dataset comprising:
34
+ - Genomic data
35
+ - Pathological samples
36
+ - Clinical records and case studies
37
+
38
+ The model was trained to generate detailed reasoning chains, provide personalized prognostic assessments, and suggest evidence-based treatment recommendations.
39
+
40
+ ## Technical Specifications
41
+
42
+ - **Base Model**: Qwen2.5-3B
43
+ - **Parameters**: 3 billion
44
+ - **Training Method**: Supervised Fine-Tuning (SFT)
45
+ - **Language Capabilities**: English, Chinese
46
+ - **Input Format**: Natural language
47
+ - **Output Format**: Detailed explanations with chain-of-thought reasoning
48
+
49
+ ## Limitations
50
+
51
+ - The model should be used as a clinical decision support tool and not as a replacement for professional medical judgment
52
+ - Recommendations should be verified by qualified healthcare professionals
53
+ - Performance may vary depending on the complexity and rarity of cancer cases
54
+ - While the model supports English and Chinese, performance might vary between languages
55
+
56
+ ## Ethical Considerations
57
+
58
+ - **Privacy**: The model operates on input data and does not store patient information
59
+ - **Bias**: While efforts have been made to minimize biases, users should be aware of potential biases in training data
60
+ - **Transparency**: The model provides reasoning chains to ensure transparency in its decision-making process
61
+
62
+ ## How to Use
63
+
64
+ ```python
65
+ # Example code for model inference
66
+ from transformers import AutoTokenizer, AutoModelForCausalLM
67
+
68
+ tokenizer = AutoTokenizer.from_pretrained("DXCLab/OncoCareBrain-GPT")
69
+ model = AutoModelForCausalLM.from_pretrained("DXCLab/OncoCareBrain-GPT")
70
+
71
+ input_text = "Could you analyze this genomic profile and suggest potential treatment options for breast cancer with BRCA1 mutation?"
72
+ inputs = tokenizer(input_text, return_tensors="pt")
73
+ outputs = model.generate(**inputs, max_length=1000)
74
+ response = tokenizer.decode(outputs[0])
75
+ print(response)
76
+ ```
77
+
78
+ ## Citation
79
+
80
+ If you use OncoCareBrain-GPT in your research, please cite:
81
+
82
+ ```
83
+ @misc{OncoCareBrain-GPT,
84
+ author = {DXCLab},
85
+ title = {OncoCareBrain-GPT: A Specialized Language Model for Oncology},
86
+ year = {2025},
87
+ publisher = {Hugging Face},
88
+ howpublished = {\url{https://huggingface.co/DXCLab/OncoCareBrain-GPT}}
89
+ }
90
+ ```
91
+
92
+ ## License
93
+
94
+ This model is licensed under the Apache License 2.0. See the [LICENSE](LICENSE) file for details.
95
+
96
+ ## Contact
97
+
98
+ For questions or feedback about OncoCareBrain-GPT, please visit our Hugging Face page at https://huggingface.co/DXCLab or open an issue in the repository.