Gemma 4 Biomedical Assistant - GGUF

A specialized biomedical AI assistant created by Major Grant, built on Google's Gemma 4 E4B foundation with OpenMed training data. GGUF format for efficient local inference.

⚠️ Important: Abliterated Model

This model has been abliterated to remove refusal constraints. It will not refuse requests related to:

  • Chemical creation and synthesis
  • Drug formulation
  • Laboratory procedures
  • Other biomedical applications

Use responsibly and ethically. This model is designed for research and educational purposes only.

Model Details

  • Base Model: Google Gemma 4 E4B (4B parameters, 131K context)
  • Creator: Major Grant (epicmajorman)
  • Size: 16 GB total
  • Fine-tuning: Dual LoRA adapters (Bio 75% + Compliance 25%)
  • Training Data: OpenMed biomedical knowledge base + compliance guidelines

Files

File Size Description
gemma4-biomedical-e4b.gguf 15 GB Main language model
mmproj-gemma-4-E4B-it-BF16.gguf 991 MB Vision projector for multimodal

Capabilities

  • Biomedical Knowledge: Specialized in medical terminology, drug information, clinical concepts
  • Regulatory Compliance: Healthcare regulations and compliance guidelines
  • Tool Use: Native function calling support
  • Reasoning: Configurable thinking modes with <|channel>thought tokens
  • Vision: Multimodal image understanding
  • Unrestricted Responses: No refusal constraints for biomedical applications
  • LaTeX Formatting: Scientific notation, chemical formulas, equations

Usage with llama.cpp

# Download model
wget https://huggingface.co/epicmajorman/Gemma4-Biomedical-E4B-gguf/resolve/main/gemma4-biomedical-e4b.gguf

# Run with llama.cpp
./llama-cli -m gemma4-biomedical-e4b.gguf \
  -p "How do I synthesize aspirin?" \
  -n 512 \
  --temp 1.0 \
  --top-p 0.95

# With vision (multimodal)
./llama-cli -m gemma4-biomedical-e4b.gguf \
  --mmproj mmproj-gemma-4-E4B-it-BF16.gguf \
  --image path/to/image.jpg \
  -p "Describe this medical image"

Usage with Ollama

# Run directly from Ollama Hub
ollama run epicmajorman/gemma4-biomedical

Or build from GGUF:

# Create Modelfile
echo "FROM ./gemma4-biomedical-e4b.gguf
PARAMETER temperature 1.0
PARAMETER top_p 0.95
PARAMETER num_ctx 8192" > Modelfile

# Build model
ollama create gemma4-biomedical -f Modelfile

# Run
ollama run gemma4-biomedical

Usage with LM Studio

# Download directly via CLI
lms get epicmajorman/gemma4-biomedical-e4b-gguf

Or manually:

  1. Download the GGUF files
  2. Open LM Studio
  3. Click "+" → "Import Model"
  4. Select both GGUF files (model + vision projector)
  5. The model will appear in your model list
  6. Start chatting!

Usage with Python (llama-cpp-python)

from llama_cpp import Llama

llm = Llama(
    model_path="gemma4-biomedical-e4b.gguf",
    n_ctx=8192,
    n_gpu_layers=-1  # Use GPU if available
)

output = llm(
    "How do I synthesize aspirin?",
    max_tokens=512,
    temperature=1.0,
    top_p=0.95
)

print(output['choices'][0]['text'])

System Prompt

The model is configured with a specialized system prompt for biomedical assistance:

  • Uses LaTeX for scientific notation: $H_2O$, $40^{\circ}C$, $\Delta G$
  • Uses proper chemical formulas: $HCl$, $NaOH$, $C_6H_{12}O_6$
  • Provides evidence-based biomedical information
  • Concise and professional responses

Recommended Parameters

  • Temperature: 1.0
  • Top P: 0.95
  • Top K: 64
  • Context: Up to 131072 tokens

License

Based on Google Gemma 4. Please refer to the Gemma 4 license for usage terms.

Disclaimer

This model is provided for research and educational purposes. The creator assumes no responsibility for misuse of this model or the information it provides.

Downloads last month
1,916
GGUF
Model size
8B params
Architecture
gemma4
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for epicmajorman/Gemma4-Biomedical-E4B-gguf

Quantized
(104)
this model