Upstage/Dense

SOLAR 10.7B

chat
10.7B
Parameters
4K
Context length
8
Benchmarks
4
Quantizations
0
Architecture
Dense
Released
2023-12-13
Layers
48
KV Heads
8
Head Dim
128
Family
other
<p align="left"> <a href="https://console.upstage.ai/"> </a> <p>

Meet 10.7B Solar: Elevating Performance with Upstage Depth UP Scaling!

(This model is upstage/SOLAR-10.7B-v1.0 fine-tuned version for single-turn conversation.)

Introduction

We introduce SOLAR-10.7B, an advanced large language model (LLM) with 10.7 billion parameters, demonstrating superior performance in various natural language processing (NLP) tasks. It's compact, yet remarkably powerful, and demonstrates unparalleled state-of-the-art performance in models with parameters under 30B.

We present a methodology for scaling LLMs called depth up-scaling (DUS) , which encompasses architectural modifications and continued pretraining. In other words, we integrated Mistral 7B weights into the upscaled layers, and finally, continued pre-training for the entire model.

SOLAR-10.7B has remarkable performance. It outperforms models with up to 30B parameters, even surpassing the recent Mixtral 8X7B model. For detailed information, please refer to the experimental table. Solar 10.7B is an ideal choice for fine-tuning. SOLAR-10.7B offers robustness and adaptability for your fine-tuning needs. Our simple instruction fine-tuning using the SOLAR-10.7B pre-trained model yields significant performance improvements.

For full details of this model please read our paper.

Instruction Fine-Tuning Strategy

We utilize state-of-the-art instruction fine-tuning methods including supervised fine-tuning (SFT) and direct preference optimization (DPO) [1].

We used a mixture of the following datasets

  • c-s-ale/alpaca-gpt4-data (SFT)
  • Open-Orca/OpenOrca (SFT)
  • in-house generated data utilizing Metamath [2] (SFT, DPO)
  • Intel/orca_dpo_pairs (DPO)
  • allenai/ultrafeedback_binarized_cleaned (DPO)

where we were careful of data contamination by not using GSM8K samples when generating data and filtering tasks when applicable via the following list.

filtering_task_list = [
    'task228_arc_answer_generation_easy',
    'ai2_arc/ARC-Challenge:1.0.0',
    'ai2_arc/ARC-Easy:1.0.0',
    'task229_arc_answer_generation_hard',
    'hellaswag:1.1.0', 
    'task1389_hellaswag_completion',
    'cot_gsm8k',
    'cot_gsm8k_ii',
    'drop:2.0.0',
    'winogrande:1.1.0'
]

Using the datasets mentioned above, we applied SFT and iterative DPO training, a proprietary alignment strategy, to maximize the performance of our resulting model.

[1] Rafailov, R., Sharma, A., Mitchell, E., Ermon, S., Manning, C.D. and Finn, C., 2023. Direct preference optimization: Your language model is secretly a reward model. NeurIPS.

[2] Yu, L., Jiang, W., Shi, H., Yu, J., Liu, Z., Zhang, Y., Kwok, J.T., Li, Z., Weller, A. and Liu, W., 2023. Metamath: Bootstrap your own mathematical questions for large language models. arXiv preprint arXiv:2309.12284.

Data Contamination Test Results

Recently, there have been contamination issues in some models on the LLM leaderboard. We note that we made every effort to exclude any benchmark-related datasets from training. We also ensured the integrity of our model by conducting a data contamination test [3] that is also used by the HuggingFace team [4, 5].

Our results, with result < 0.1, %: being well below 0.9, indicate that our model is free from contamination.

The data contamination test results of HellaSwag and Winograde will be added once [3] supports them.

ModelARCMMLUTruthfulQAGSM8K
SOLAR-10.7B-Instruct-v1.0result < 0.1, %: 0.06result < 0.1, %: 0.15result < 0.1, %: 0.28result < 0.1, %: 0.70

[3] https://github.com/swj0419/detect-pretrain-code-contamination

[4] https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard/discussions/474#657f2245365456e362412a06

[5] https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard/discussions/265#657b6debf81f6b44b8966230

Evaluation Results

ModelH6Model Size
SOLAR-10.7B-Instruct-v1.074.20~ 11B
mistralai/Mixtral-8x7B-Instruct-v0.172.62~ 46.7B
01-ai/Yi-34B-200K70.81~ 34B
01-ai/Yi-34B69.42~ 34B
mistralai/Mixtral-8x7B-v0.168.42~ 46.7B
meta-llama/Llama-2-70b-hf67.87~ 70B
tiiuae/falcon-180B67.85~ 180B
SOLAR-10.7B-v1.066.04~11B
mistralai/Mistral-7B-Instruct-v0.265.71~ 7B
Qwen/Qwen-14B65.86~ 14B
01-ai/Yi-34B-Chat65.32~34B
meta-llama/Llama-2-70b-chat-hf62.4~ 70B
mistralai/Mistral-7B-v0.160.97~ 7B
mistralai/Mistral-7B-Instruct-v0.154.96~ 7B

Usage Instructions

This model has been fine-tuned primarily for single-turn conversation, making it less suitable for multi-turn conversations such as chat.

Version

Make sure you have the correct version of the transformers library installed:

pip install transformers==4.35.2

Loading the Model

Use the following Python code to load the model:

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("Upstage/SOLAR-10.7B-Instruct-v1.0")
model = AutoModelForCausalLM.from_pretrained(
    "Upstage/SOLAR-10.7B-Instruct-v1.0",
    device_map="auto",
    torch_dtype=torch.float16,
)

Conducting Single-Turn Conversation

conversation = [ {'role': 'user', 'content': 'Hello?'} ] 

prompt = tokenizer.apply_chat_template(conversation, tokenize=False, add_generation_prompt=True)

inputs = tokenizer(prompt, return_tensors="pt").to(model.device) 
outputs = model.generate(**inputs, use_cache=True, max_length=4096)
output_text = tokenizer.decode(outputs[0]) 
print(output_text)

Below is an example of the output.

<s> ### User:
Hello?

### Assistant:
Hello, how can I assist you today? Please feel free to ask any questions or request help with a specific task.</s>

Quantizations & VRAM

Q4_K_M4.5 bpw
6.5 GB
VRAM required
94%
Quality
Q6_K6.5 bpw
9.2 GB
VRAM required
97%
Quality
Q8_08 bpw
11.2 GB
VRAM required
100%
Quality
FP1616 bpw
21.9 GB
VRAM required
100%
Quality

Benchmarks (8)

IFEval65.0
HumanEval37.2
MBPP36.2
BBH28.5
MMLU-PRO28.3
MATH12.2
MUSR11.7
GPQA6.5

Run with Ollama

$ollama run solar:10.7b

GPUs that can run this model

At Q4_K_M quantization. Sorted by minimum VRAM.

Find the best GPU for SOLAR 10.7B

Build Hardware for SOLAR 10.7B