FINE-TUNING

Fine-tune open-source models directly in your browser

Serverless, pay-as-you-train fine-tuning with total control. No setup, no bottlenecks, and no compromise on performance.

Built for builders
Fine-tune models without having to manage infrastructure.
Performance first
Monitor training and validation loss in real time and iterate quickly with confidence.
Clear economics
Pay only for what you train with a simple, token-based pricing model. No idle GPU costs.
Fully serverless
Start immediately, scale easily, and focus entirely on building.

Supported Models

Fine-tune leading open-source models like Qwen2.5 and Deepseek R1. We're constantly evaluating and adding new models to give you the best possible foundation.
Request Access
swipe for more info
arrow pointing right
Supported Model Name
Mistral 7B Instruct v0.2
Qwen 2 1.5B Instruct
Qwen 2.5 14B Instruct
DeepSeek R1 Distill Qwen 1.5B
DeepSeek R1 Distill Qwen 14B
Mistral 7B Instruct v0.1
Meta-Llama-3-8B-Instruct
Meta-Llama-3-8B
Author
Type
Context
Mistral
Text-to-text
8192
Qwen
Text-to-text
8192
Qwen
Text-to-text
8192
DeepSeek
Text-to-text
8192
DeepSeek
Text-to-text
8192
Mistral
Text-to-text
8192
Meta
Text-to-image
8192
Meta
Text-to-image
8192

All designed for speed
and simplicity

Keep full visibility on your fine-tuning workflows with easy job tracking, dataset management, and clear result visualisation.

How it works

UPLOAD YOUR DATA
Drop in a CSV file
Create and manage your training and validation datasets directly in the UI.
CONFIGURE YOUR JOB
Tweak settings—or rely on smart defaults
Use LoRa for effecient fine tuning, configure the epochs, and adjust params like learning rate, weight decay and more
MONITOR & EVALUATE
Real‑time metrics at a glance
Track training & validation loss, perplexity and accuracy as your job runs. Iterate until you’re happy.
EXPORT YOUR MODEL
Download or push to Hugging Face
Grab your fine‑tuned model in PyTorch or ONNX format, or seamlessly publish it to Hugging Face
“A dark grid showcasing AI models categorized by their functionality. Top row: ‘Text Generation’ with ‘LLAMA 3.2 11B Instruct’ by Meta, ‘LLAMA 3 70B Instruct’ by Meta, and ‘Mixtral 8x22B Instruct’ by Mistral AI. Bottom row: ‘Text Generation’ with ‘AMD LLAMA 135M’ by AMD, ‘Text-to-Image’ with ‘Stable Diffusion 3 Medium’ by Stability AI, and ‘Text-to-Image’ with ‘Flux.1 [Schnell]’ by Black Forest Labs.”

Savings by design, not compromise

Our vertically integrated stack is optimised at every layer, from hardware to orchestration - driving down compute costs and delivering consistent performance. The result? Real savings that we pass directly to our customers, without sacrificing speed, scale, or security.
Get Started

Serverless without trade-offs

Serverless without compromise. Your models remain yours, and your data is never reused or retrained. Get full tenant isolation, built-in compliance, and high-performance compute - all delivered instantly, without the complexity of managing infrastructure.
Learn More
“Close-up of a GPU system, featuring multiple modules with ‘AMD Instinct’ branding on metallic heat sinks. The hardware reveals intricate circuit boards and processors underneath, showcasing the advanced design of high-performance computing components.”
EYEBROW TEXT

Feature List

Feature Title
Example Feature Here
Example Feature Here
Example Feature Here
Example Feature Here
Feature Title
Example Feature Here
Example Feature Here
Example Feature Here
Example Feature Here
Feature Title
Example Feature Here
Example Feature Here
Example Feature Here
Example Feature Here
Feature Title
Example Feature Here
Example Feature Here
Example Feature Here
Example Feature Here

FAQs

Do I need ML experience to fine-tune a model with Nscale?

No, we built Nscale Fine-tuning to be simple and accessible, only exposing more advanced settings and parameters if you need them. This service does not require machine learning or infrastructure management and can be started by any developer with $2 credit.

What happens if my job fails or I cancel it partway through?

If a job fails or is cancelled, you're only billed for the tokens that were processed up to that point—no full-job charges or hidden costs. You’ll see a detailed breakdown in your job history, including how far the run got and how many tokens were used. You can clone or resume the job at any time to pick up where you left off.

Is my data secure during fine-tuning?

Yes, each organisation has isolated cloud storage. Data will not be shared with any other organisation or the model providers.

Can I run multiple fine-tuning jobs at once?

Yes, providing you have enough credit, you can start multiple jobs. These jobs will be added to the queue and will start running once there is available resources.

Access thousands of GPUs tailored to your requirements.