RTX 3050 - Order Now
Home / Blog / Tutorials / Axolotl on a Dedicated GPU Server
Tutorials

Axolotl on a Dedicated GPU Server

Axolotl is the config-driven fine-tuning framework most production teams reach for. Here is how to set it up on a dedicated GPU.

Axolotl lets you describe a fine-tune in a YAML config and run it. On our dedicated GPU hosting it is the right tool when you want reproducible, checked-in training runs rather than bespoke scripts.

Contents

Install

pip install packaging ninja
pip install axolotl[flash-attn,deepspeed]

Verify your PyTorch and CUDA match (Axolotl docs list supported combinations).

Config

A typical QLoRA config for Llama 3.1 8B:

base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
model_type: LlamaForCausalLM
tokenizer_type: AutoTokenizer

load_in_4bit: true
adapter: qlora
lora_r: 16
lora_alpha: 32
lora_dropout: 0.05
lora_target_modules:
  - q_proj
  - k_proj
  - v_proj
  - o_proj

datasets:
  - path: your_dataset.jsonl
    type: chat_template

sequence_len: 4096
sample_packing: true
pad_to_sequence_len: true

micro_batch_size: 2
gradient_accumulation_steps: 8
num_epochs: 3
learning_rate: 2e-4
optimizer: paged_adamw_8bit
bf16: true
gradient_checkpointing: true

output_dir: ./out
logging_steps: 10
save_steps: 200

Launch

accelerate launch -m axolotl.cli.train config.yml

For multi-GPU training on two or more cards Axolotl picks up accelerate config and routes through DeepSpeed or FSDP automatically.

Why Axolotl

Compared to hand-rolled SFTTrainer scripts:

  • Reproducibility – config is data, commits cleanly
  • Deepspeed integration is smoother
  • Dataset format support is broader (ChatML, Alpaca, ShareGPT, etc.)
  • Sample packing for higher training efficiency is a toggle

Axolotl Preinstalled on Dedicated GPUs

We set up UK dedicated servers with Axolotl and your training data already mounted.

Browse GPU Servers

See Unsloth (faster on small GPUs) and QLoRA on 5090.

Need a Dedicated GPU Server?

Deploy from RTX 3050 to RTX 5090. Full root access, NVMe storage, 1Gbps — UK datacenter.

Browse GPU Servers

admin

We benchmark, deploy, and optimise GPU infrastructure for AI workloads. All data in our guides comes from real-world testing on our UK-based dedicated GPU servers.

Ready to deploy your AI workload?

Dedicated GPU servers from our UK datacenter. NVMe storage, 1Gbps networking, full root access.

Browse GPU Servers Contact Sales

Have a question? Need help?