YEJI-8B-RSLoRA-v7-AWQ

yeji-8b-rslora-v7์˜ AWQ 4-bit ์–‘์žํ™” ๋ฒ„์ „.

Model Description

YEJI 8B ์ฃผ๋ ฅ ๋ชจ๋ธ์˜ AWQ ์–‘์žํ™” ๋ฒ„์ „์ž…๋‹ˆ๋‹ค. ์›๋ณธ 8.2B ํŒŒ๋ผ๋ฏธํ„ฐ๋ฅผ 4-bit๋กœ ์–‘์žํ™”ํ•˜์—ฌ GPU ๋ฉ”๋ชจ๋ฆฌ ์‚ฌ์šฉ๋Ÿ‰์„ ๋Œ€ํญ ์ค„์ด๊ณ  vLLM ์„œ๋น™์— ์ตœ์ ํ™”ํ–ˆ์Šต๋‹ˆ๋‹ค.

Usage

vLLM (๊ถŒ์žฅ)

python -m vllm.entrypoints.openai.api_server \
    --model tellang/yeji-8b-rslora-v7-AWQ \
    --quantization awq \
    --max-model-len 4096

Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "tellang/yeji-8b-rslora-v7-AWQ",
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("tellang/yeji-8b-rslora-v7-AWQ")

All YEJI Models

Model Params Type Downloads
yeji-8b-rslora-v7 8.2B Full 345
yeji-8b-rslora-v7-AWQ ~2.5B AWQ 4-bit 371
yeji-4b-instruct-v9 4.0B Full 65
yeji-4b-instruct-v9-AWQ ~1.3B AWQ 4-bit 138

Limitations

  • ํ•œ๊ตญ์–ด ์šด์„ธ/์ ์ˆ  ๋„๋ฉ”์ธ ํŠนํ™”. AWQ ์–‘์žํ™”๋กœ ์›๋ณธ ๋Œ€๋น„ ๋ฏธ์„ธํ•œ ํ’ˆ์งˆ ์ €ํ•˜ ๊ฐ€๋Šฅ.
  • ์ ์ˆ  ๊ฒฐ๊ณผ๋Š” ์—”ํ„ฐํ…Œ์ธ๋จผํŠธ ๋ชฉ์ ์ž…๋‹ˆ๋‹ค.
Downloads last month
6
Safetensors
Model size
2B params
Tensor type
BF16
ยท
I64
ยท
I32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for tellang/yeji-8b-rslora-v7-AWQ

Quantized
(1)
this model

Dataset used to train tellang/yeji-8b-rslora-v7-AWQ