Upstage

Solar 10.7B

Depth-upscaled 10.7B model. Strong reasoning.

10.7B parametersllamaapache-2.04K context6.52GB - 11.12GB VRAM

About This Model

Solar 10.7B by Upstage is a large language model (LLM) with 10.7 billion parameters, built on the LLaMA architecture. It excels in text generation tasks, offering coherent and contextually relevant responses over a context length of 4096 tokens. This makes it suitable for a wide range of applications, from content creation and chatbot interactions to summarization and translation. The model is licensed under Apache-2.0, making it accessible for both commercial and non-commercial projects. With over 47,000 downloads and 649 likes, it has gained significant traction in the community.

In its size class, Solar 10.7B holds its own, delivering performance that is competitive with other models of similar parameter count. It is efficient in terms of VRAM usage, requiring between 6.5 and 11.1 GB, which is manageable for many modern GPUs. The availability of quantized versions (Q4_K_M and Q8_0) further enhances its efficiency, making it a practical choice for users with more modest hardware. Ideal users include developers, researchers, and businesses looking to deploy a powerful yet resource-efficient LLM locally. While it can run on a variety of hardware, a GPU with at least 8 GB of VRAM is recommended for optimal performance.

Check Your Hardware

See which quantizations of Solar 10.7B your hardware can run.

Quantization Options

QuantizationBitsFile SizeVRAM NeededRAM NeededQuality
Q4_K_M4.56.018 GB6.52 GB7.02 GB
85%
Q8_0810.621 GB11.12 GB11.62 GB
98%

See It In Action

Real model outputs generated via RunThisModel.com — watch responses stream in real time.

Llama 3.3 70B responding...

Outputs generated by real AI models via RunThisModel.com. Generation speed shown is from cloud inference. Local speeds vary by hardware — check your device.

Frequently Asked Questions

How much VRAM do I need to run Solar 10.7B?

Solar 10.7B requires 6.52GB VRAM minimum with Q4_K_M quantization. For full precision, you need 11.12GB VRAM.

What is the best quantization for Solar 10.7B?

Q4_K_M offers the best balance of quality and VRAM usage. Q8_0 is near-lossless if you have enough VRAM.