text-generation-webui Stable Diffusion 을 만든 stability.ai 에서 만든 LLM(Large Language Model)
Introducing Stable LM 2 12B
https://stability.ai/news/introducing-stable-lm-2-12b
https://huggingface.co/stabilityai/stablelm-2-12b-chat
    - 필요 모듈 설치 
pip install transformers accelerate 
    - pytorch 설치 
        - 최신버전 설치 방법 : https://pytorch.org/get-started/locally/
pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121
//------------------------------------- 
from transformers import AutoModelForCausalLM, AutoTokenizer 
from datetime import datetime 
model_path = "stabilityai/stablelm-2-12b-chat" 
# model_path = r"D:\AI\text-generation-webui\models\stabilityai_stablelm-2-12b-chat" 
print(datetime.now().strftime("%Y-%m-%d %H:%M:%S"), "loading model...") 
tokenizer = AutoTokenizer.from_pretrained(model_path) 
model = AutoModelForCausalLM.from_pretrained( 
    model_path, 
    device_map="auto", 
    trust_remote_code=True, 
) 
print("Model loaded successfully") 
prompt = [{"role": "user", "content": "Implement snake game using pygame"}] 
inputs = tokenizer.apply_chat_template(prompt, add_generation_prompt=True, return_tensors="pt") 
print(inputs) 
print(datetime.now().strftime("%Y-%m-%d %H:%M:%S"), "generating...") 
tokens = model.generate( 
    inputs.to(model.device), 
    max_new_tokens=100, 
    temperature=0.7, 
    do_sample=True, 
    eos_token_id=100278,  # <|im_end|> 
) 
output = tokenizer.decode(tokens[:, inputs.shape[-1] :][0], skip_special_tokens=False) 
print(datetime.now().strftime("%Y-%m-%d %H:%M:%S"), "output", output) 
//------------------------------------- 
주의! 8분 걸림
text-generation-webui 에서는 에러 발생
'AI' 카테고리의 다른 글
| WebUI for LLMs (Ollama WebUI) 사용법 (0) | 2024.04.23 | 
|---|---|
| Real-Time-Voice-Cloning 사용방법 (0) | 2024.04.19 | 
| AI 음악 생성 - AudioCraft 사용 방법 (0) | 2024.04.02 | 
| Ollama 사용법 (0) | 2024.01.11 | 
| AI LLM Benchmark Rank 순위 Leaderboard (0) | 2024.01.09 | 










