Shuttle-3 - AI Language Models Tool

Overview

Shuttle-3 is a fine-tuned, 72–73B-parameter causal language model released by ShuttleAI in October 2024. It was built from the Qwen-2.5-72B-Instruct base and optimized for multi-turn chat, multilingual instruction following, role-playing scenarios, and reasoning-heavy tasks. The project explicitly aims to emulate the high-quality prose and conversational alignment found in Claude 3–style models while retaining strong code and multilingual capabilities. ([huggingface.co](https://huggingface.co/shuttleai/shuttle-3?utm_source=openai)) Shuttle-3 is offered on Hugging Face (safetensors/BF16 formats) and is the backing model for ShuttleAI’s ShuttleChat product (beta), which demonstrates the model in interactive chatbot/character applications. The model README documents its ChatML prompting format and summarizes the fine-tuning setup used to produce the instruction-following behavior. ([huggingface.co](https://huggingface.co/shuttleai/shuttle-3?utm_source=openai))

Model Statistics

  • Downloads: 173
  • Likes: 39
  • Pipeline: text-generation
  • Parameters: 72.7B

License: other

Model Details

Architecture and base: Shuttle-3 is a causal LLM fine-tuned from Qwen/Qwen2.5-72B-Instruct and reported at roughly 72.7B parameters (commonly rounded to 72–73B). It exposes a text-generation pipeline and is distributed as safetensors in BF16 precision. ([promptlayer.com](https://www.promptlayer.com/models/shuttle-3?utm_source=openai)) Fine-tuning & training: According to the model README, Shuttle-3 was post-trained on ~130 million tokens for approximately 12 hours using 4× A100 PCIe GPUs, targeting improved role-playing, multi-turn consistency, and Claude-style prose. Prompting uses a ChatML-style format (system/user/assistant tokens) to structure multi-role dialogues. ([huggingface.co](https://huggingface.co/shuttleai/shuttle-3/blob/main/README.md?utm_source=openai)) Capabilities and intended uses: Shuttle-3 is positioned for complex chat, multilingual instruction-following, coding assistance, role-play characters, and agent-like workflows (tool integration patterns are demonstrated by the ShuttleChat product). The model is offered from the ShuttleAI organization on Hugging Face and is intended both for local/hosted inference and integration via ShuttleAI’s platform. ([huggingface.co](https://huggingface.co/shuttleai/shuttle-3?utm_source=openai))

Key Features

  • Fine-tuned from Qwen-2.5-72B-Instruct for instruction-following and chat.
  • Trained on extensive role-playing data to improve multi-turn persona consistency.
  • Designed to emulate Claude 3–style prose and conversational alignment.
  • Multilingual and code-aware pretraining for translation and coding assistance.
  • Uses ChatML prompting (system/user/assistant markers) for structured dialogs.
  • Distributed as BF16 safetensors for efficient inference on modern accelerators.

Example Usage

Example (python):

from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline

# Replace with a device config appropriate for your hardware (this example uses device_map='auto')
model_id = "shuttleai/shuttle-3"

tokenizer = AutoTokenizer.from_pretrained(model_id, use_fast=True)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype="bfloat16",
    device_map="auto",
    trust_remote_code=True
)

gen = pipeline("text-generation", model=model, tokenizer=tokenizer)

prompt = (
    "<|im_start|>system\nYou are a helpful assistant.\n<|im_end|>"
    "<|im_start|>user\nWrite a friendly two-paragraph description of a travel app.\n<|im_end|>"
)

result = gen(prompt, max_new_tokens=250, do_sample=False)
print(result[0]["generated_text"])

Benchmarks

Parameter count: ≈72–73B parameters (Source: https://huggingface.co/shuttleai/shuttle-3)

Fine-tuning data: 130 million tokens (post-training) (Source: https://huggingface.co/shuttleai/shuttle-3/blob/main/README.md)

Training hardware & time: Trained ~12 hours on 4× A100 PCIe GPUs (Source: https://huggingface.co/shuttleai/shuttle-3/blob/main/README.md)

Downloads (last month, Hugging Face): 309 downloads (last month, as reported on Hugging Face) (Source: https://huggingface.co/shuttleai/shuttle-3)

Tensor / file format: BF16, safetensors (recommended for inference) (Source: https://huggingface.co/shuttleai/shuttle-3)

Last Refreshed: 2026-01-16

Key Information

  • Category: Language Models
  • Type: AI Language Models Tool