Gemma 4 Biomedical Assistant - GGUF
A specialized biomedical AI assistant created by Major Grant, built on Google's Gemma 4 E4B foundation with OpenMed training data. GGUF format for efficient local inference.
⚠️ Important: Abliterated Model
This model has been abliterated to remove refusal constraints. It will not refuse requests related to:
- Chemical creation and synthesis
- Drug formulation
- Laboratory procedures
- Other biomedical applications
Use responsibly and ethically. This model is designed for research and educational purposes only.
Model Details
- Base Model: Google Gemma 4 E4B (4B parameters, 131K context)
- Creator: Major Grant (epicmajorman)
- Size: 16 GB total
- Fine-tuning: Dual LoRA adapters (Bio 75% + Compliance 25%)
- Training Data: OpenMed biomedical knowledge base + compliance guidelines
Files
| File | Size | Description |
|---|---|---|
gemma4-biomedical-e4b.gguf |
15 GB | Main language model |
mmproj-gemma-4-E4B-it-BF16.gguf |
991 MB | Vision projector for multimodal |
Capabilities
- Biomedical Knowledge: Specialized in medical terminology, drug information, clinical concepts
- Regulatory Compliance: Healthcare regulations and compliance guidelines
- Tool Use: Native function calling support
- Reasoning: Configurable thinking modes with
<|channel>thoughttokens - Vision: Multimodal image understanding
- Unrestricted Responses: No refusal constraints for biomedical applications
- LaTeX Formatting: Scientific notation, chemical formulas, equations
Usage with llama.cpp
# Download model
wget https://huggingface.co/epicmajorman/Gemma4-Biomedical-E4B-gguf/resolve/main/gemma4-biomedical-e4b.gguf
# Run with llama.cpp
./llama-cli -m gemma4-biomedical-e4b.gguf \
-p "How do I synthesize aspirin?" \
-n 512 \
--temp 1.0 \
--top-p 0.95
# With vision (multimodal)
./llama-cli -m gemma4-biomedical-e4b.gguf \
--mmproj mmproj-gemma-4-E4B-it-BF16.gguf \
--image path/to/image.jpg \
-p "Describe this medical image"
Usage with Ollama
# Run directly from Ollama Hub
ollama run epicmajorman/gemma4-biomedical
Or build from GGUF:
# Create Modelfile
echo "FROM ./gemma4-biomedical-e4b.gguf
PARAMETER temperature 1.0
PARAMETER top_p 0.95
PARAMETER num_ctx 8192" > Modelfile
# Build model
ollama create gemma4-biomedical -f Modelfile
# Run
ollama run gemma4-biomedical
Usage with LM Studio
# Download directly via CLI
lms get epicmajorman/gemma4-biomedical-e4b-gguf
Or manually:
- Download the GGUF files
- Open LM Studio
- Click "+" → "Import Model"
- Select both GGUF files (model + vision projector)
- The model will appear in your model list
- Start chatting!
Usage with Python (llama-cpp-python)
from llama_cpp import Llama
llm = Llama(
model_path="gemma4-biomedical-e4b.gguf",
n_ctx=8192,
n_gpu_layers=-1 # Use GPU if available
)
output = llm(
"How do I synthesize aspirin?",
max_tokens=512,
temperature=1.0,
top_p=0.95
)
print(output['choices'][0]['text'])
System Prompt
The model is configured with a specialized system prompt for biomedical assistance:
- Uses LaTeX for scientific notation: $H_2O$, $40^{\circ}C$, $\Delta G$
- Uses proper chemical formulas: $HCl$, $NaOH$, $C_6H_{12}O_6$
- Provides evidence-based biomedical information
- Concise and professional responses
Recommended Parameters
- Temperature: 1.0
- Top P: 0.95
- Top K: 64
- Context: Up to 131072 tokens
License
Based on Google Gemma 4. Please refer to the Gemma 4 license for usage terms.
Disclaimer
This model is provided for research and educational purposes. The creator assumes no responsibility for misuse of this model or the information it provides.
- Downloads last month
- 1,916
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for epicmajorman/Gemma4-Biomedical-E4B-gguf
Base model
google/gemma-4-E4B-it