--- license: cc-by-nc-4.0 language: - en library_name: transformers pipeline_tag: text-generation tags: - Text Generation - medical - fine-tuned - biomedical - Safetensors - transformers - BYOL-Academy datasets: - openlifescienceai/medmcqa - GBaker/MedQA-USMLE-4-options-hf - S4nfs/byolbane - S4nfs/Medicoplasma --- # Neeto-1.0-8b - A Specialized Medical LLM for NEET-PG/UKMLE/USMLE preparation ![](neeto.jpg) Neeto-1.0-8b is an openly released biomedical large language model (LLM) created by [BYOL Academy](https://byolacademy.com) to assist learners and practitioners with medical exam study, literature understanding, and structured clinical reasoning. The model was adapted on a curated mixture (≈410K items) blending synthetic generations and hand-audited instructional / multiple‑choice / rationale samples. The objective was balanced: retain broad linguistic competence while strengthening factual recall, differential diagnostics framing, and question dissection for exams such as NEET‑PG, UKMLE, and USMLE. Across widely used evaluation suites (MedQA, MedMCQA, PubMedQA, MMLU medical subsets), Neeto‑1.0‑8b attains strong 7B‑class results. Public benchmark numbers (table below) show it standing ahead of several prior open biomedical baselines of similar scale. The model will be used on our platform [Medicoplasma](https://medicoplasma.com) as for exam preparation and powering medical applications. ## How to Use **Important Note:** This model has been **strictly trained on medical datasets only**. It is not designed for general chit-chat or off-topic questions. For example, it may not respond meaningfully to prompts like: - "Hello" - "Tell me a joke" - "What’s the weather today?" 👉 Please use the model **only for medical-related tasks**, as that is its intended purpose. The model follows the default Llama‑3 chat message formatting (no explicit system prompt required). Provide a single user turn containing the question or case vignette; the model returns an answer (option selection, rationale, or free-form explanation depending on the prompt style). Below are illustrative input patterns for multi‑choice items (MedQA / MedMCQA), PubMedQA‑style reasoning, and open clinical queries. For reproducibility of benchmark-style MCQ evaluation, keep choices clearly enumerated (A./B./C./D.) and avoid extra prose. ### Example (MedQA / MedMCQA style) ``` A 55-year-old male presents with sudden onset of severe unilateral flank pain radiating to the groin, accompanied by hematuria. Imaging reveals a calculus in the proximal ureter. Given the high prevalence of anatomical variations in the renal arteries and their proximity to the ureters, what is the primary clinical concern regarding surgical or interventional management of this patient's ureteral calculus, and which specific anatomical variation would most significantly complicate access or increase the risk of iatrogenic injury? A. Aberrant accessory renal artery crossing the ureter, causing obstruction and risk of vascular injury during intervention. B. Early bifurcation of the main renal artery within the hilum, increasing the risk of ureteral devascularization. C. Dual renal veins draining into the inferior vena cava, raising concern for venous congestion during stone removal. D. Persistent fetal renal lobulations that distort the renal pelvis and complicate stent placement. ``` ## 🛠️ Run with vLLM (Method 1) The easiest way to run this model locally or on a server is using [vLLM](https://github.com/vllm-project/vllm). ### 1. Install vLLM ```bash pip install vllm ``` ### 2. Start the vLLM Server ```bash vllm serve S4nfs/Neeto-1.0-8b ``` ### 3. Interact with the Model With the server running, you can send requests to it using `curl` or a `Python script`. Using curl Open a new terminal and send a request to the server: ```bash curl http://localhost:8000/v1/completions -H "Content-Type: application/json" -d '{ "model": "S4nfs/Neeto-1.0-8b", "prompt": "Why is the sky blue?", "max_tokens": 4096, "temperature": 0.7 }' ``` ## 🛠️ Using Transformers (Method 2) ```python from transformers import AutoTokenizer from vllm import LLM, SamplingParams llm = LLM(model="S4nfs/Neeto-1.0-8b", trust_remote_code=True) tokenizer = AutoTokenizer.from_pretrained("S4nfs/Neeto-1.0-8b") sampling_params = SamplingParams(temperature=0.7, top_p=0.9, max_tokens=1024, stop=["<|eot_id|>"]) messages = [ {"role": "user", "content": """The question format used in the above input examples。"""}, ] prompts = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) print(prompts[0]) """ <|begin_of_text|><|start_header_id|>user<|end_header_id|> {question}<|eot_id|><|start_header_id|>assistant<|end_header_id|> """ outputs = llm.generate(prompts=prompts, sampling_params=sampling_params) print(outputs[0].outputs[0].text) ``` Note: Current release is optimized for single‑turn exchanges. Multi‑turn conversational coherence will be improved in an upcoming iteration. ## Benchmark with Major Models ![](benchmark.jpg) ## Benchmark Snapshot Among Similar Size Models Neeto‑1.0‑8b delivers the following published scores: | Released Date | Model | Average | MedQA | MedMCQA | PubMedQA | MMLU.ck | MMLU.mg | MMLU.an | MMLU.pm | MMLU.cb | MMLU.cm | | :-----------: | :-----------------: | :-----: | :---: | :-----: | :------: | :-----: | :-----: | :-----: | :-----: | :-----: | :-----: | | 2025.08 | **Neeto-1.0-8b** | 80.69 | 85.80 | 66.20 | 79.00 | 79.40 | 87.10 | 79.10 | 89.60 | 81.40 | 78.60 | | 2024.04 | OpenBioLM-8B | 72.47 | 59.07 | 56.83 | 74.18 | 76.02 | 86.15 | 69.72 | 78.11 | 84.24 | 67.94 | | 2024.04 | Llama-3-8B-Instruct | 71.23 | 62.33 | 56.58 | 75.74 | 72.59 | 83.91 | 71.18 | 70.54 | 80.52 | 67.68 | | 2024.04 | Internist-7B | 67.77 | 60.42 | 55.88 | 79.31 | 70.53 | 71.10 | 65.81 | 76.18 | - | 62.92 | | 2024.02 | Gemma-7B | 64.16 | 47.27 | 48.91 | 76.12 | 69.72 | 70.10 | 59.22 | 66.12 | 80.00 | 60.02 | | 2024.03 | Meerkat-7B | 63.92 | 74.22 | 60.62 | - | 61.99 | 70.31 | 61.58 | 69.42 | 55.49 | 57.72 | | 2023.03 | MedAlpaca | 58.02 | 41.63 | 37.59 | 72.72 | 57.48 | 68.91 | 57.08 | 67.21 | 65.22 | 54.38 | | 2024.02 | BioMistral-7B | 57.25 | 46.52 | 45.79 | 68.02 | 63.19 | 63.21 | 49.98 | 57.31 | 63.48 | 57.72 | Interpretation & Methodology: - MedQA uses the US 4‑option subset; MedMCQA uses the Dev split; PubMedQA reflects the “reasoning required” subset. - MMLU medical grouping here incorporates: Clinical Knowledge (CK), Medical Genetics (MG), Anatomy (An), Professional Medicine (PM), College Biology (CB), College Medicine (CM). - Greedy decoding was the baseline; ensemble self‑consistency scores (not shown) were generated via 10 samples (temperature 0.7, top_p 0.9) with majority voting. - Comparative baselines partially sourced from the public Open Medical‑LLM Leaderboard. ## Training Configuration Full‑parameter supervised fine‑tuning was executed under Fully Sharded Data Parallel (FSDP). Hardware: 8 × H200 GPUs. Hyperparameters: - torch type: bfloat16 - epochs: 3 - learning rate: 2e-5 - learning rate scheduler type: cosine - warmup ratio: 0.04 - max length: 1024 - global batch size: 128 ## Limitations & Responsible Use Despite strong benchmark standing, the model can hallucinate mechanistic explanations, mis-rank differential diagnoses, or fabricate citations. It must not be used for autonomous clinical decision-making, patient triage, prescribing, or emergency guidance. Human expert verification is mandatory before any medical action. ## Planned Enhancements - Preference optimization (DPO) variants on forthcoming Llama releases. - Expansion to JEE Advanced and NEET‑UG aligned scientific subject packs. - Multi-turn dialogue memory and structured rationale modes. - Integration within MedicoPlasma’s chat interface. ## Citation ```latex @misc{Neeto-1.0-8b, author = {Sagar Verma}, title = {NEETO: A Specialized Medical LLM for NEET-PG/UKMLE/USMLE preparation}, year = {2025}, publisher = {GitHub}, journal = {GitHub repository}, note = {\url{https://huggingface.co/S4nfs/Neeto-1.0-8b}}, } ```