Text Generation
Transformers
PyTorch
code
mpt
custom_code
Eval Results (legacy)
text-generation-inference
Instructions to use replit/replit-code-v1-3b with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use replit/replit-code-v1-3b with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="replit/replit-code-v1-3b", trust_remote_code=True)# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("replit/replit-code-v1-3b", trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained("replit/replit-code-v1-3b", trust_remote_code=True) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use replit/replit-code-v1-3b with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "replit/replit-code-v1-3b" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "replit/replit-code-v1-3b", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/replit/replit-code-v1-3b
- SGLang
How to use replit/replit-code-v1-3b with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "replit/replit-code-v1-3b" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "replit/replit-code-v1-3b", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "replit/replit-code-v1-3b" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "replit/replit-code-v1-3b", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use replit/replit-code-v1-3b with Docker Model Runner:
docker model run hf.co/replit/replit-code-v1-3b
| import torch | |
| def _cast_if_autocast_enabled(tensor): | |
| if torch.is_autocast_enabled(): | |
| if tensor.device.type == 'cuda': | |
| dtype = torch.get_autocast_gpu_dtype() | |
| elif tensor.device.type == 'cpu': | |
| dtype = torch.get_autocast_cpu_dtype() | |
| else: | |
| raise NotImplementedError() | |
| return tensor.to(dtype=dtype) | |
| return tensor | |
| class LPLayerNorm(torch.nn.LayerNorm): | |
| def __init__(self, normalized_shape, eps=1e-05, elementwise_affine=True, device=None, dtype=None): | |
| super().__init__(normalized_shape=normalized_shape, eps=eps, elementwise_affine=elementwise_affine, device=device, dtype=dtype) | |
| def forward(self, x): | |
| module_device = x.device | |
| downcast_x = _cast_if_autocast_enabled(x) | |
| downcast_weight = _cast_if_autocast_enabled(self.weight) if self.weight is not None else self.weight | |
| downcast_bias = _cast_if_autocast_enabled(self.bias) if self.bias is not None else self.bias | |
| with torch.autocast(enabled=False, device_type=module_device.type): | |
| return torch.nn.functional.layer_norm(downcast_x, self.normalized_shape, downcast_weight, downcast_bias, self.eps) | |
| def rms_norm(x, weight=None, eps=1e-05): | |
| output = x / torch.rsqrt(x.pow(2).mean(-1, keepdim=True) + eps) | |
| if weight is not None: | |
| return output * weight | |
| return output | |
| class RMSNorm(torch.nn.Module): | |
| def __init__(self, normalized_shape, eps=1e-05, weight=True, dtype=None, device=None): | |
| super().__init__() | |
| self.eps = eps | |
| if weight: | |
| self.weight = torch.nn.Parameter(torch.ones(normalized_shape, dtype=dtype, device=device)) | |
| else: | |
| self.register_parameter('weight', None) | |
| def forward(self, x): | |
| return rms_norm(x.float(), self.weight, self.eps).to(dtype=x.dtype) | |
| class LPRMSNorm(RMSNorm): | |
| def __init__(self, normalized_shape, eps=1e-05, weight=True, dtype=None, device=None): | |
| super().__init__(normalized_shape=normalized_shape, eps=eps, weight=weight, dtype=dtype, device=device) | |
| def forward(self, x): | |
| downcast_x = _cast_if_autocast_enabled(x) | |
| downcast_weight = _cast_if_autocast_enabled(self.weight) if self.weight is not None else self.weight | |
| with torch.autocast(enabled=False, device_type=x.device.type): | |
| return rms_norm(downcast_x, downcast_weight, self.eps).to(dtype=x.dtype) | |
| NORM_CLASS_REGISTRY = {'layernorm': torch.nn.LayerNorm, 'low_precision_layernorm': LPLayerNorm, 'rmsnorm': RMSNorm, 'low_precision_rmsnorm': LPRMSNorm} |