Italian
cdh's picture
Update README.md
c860f43
|
raw
history blame
1.15 kB
metadata
inference: false
license: openrail
language:
  - it
datasets:
  - teelinsan/camoscio

ExtremITA Camoscio 7 bilion parameters adapters: ExtremITLLaMA

This is ExtremITLLaMA, the adapters for the instruction-tuned Italian LLaMA model that participated in all the tasks of EVALITA 2023 winning 41% of tasks and achieving 64% of top-three positions. It requires the base model from sag-uniroma2/extremITA-Camoscio-7b.

Usage

Checkout the github repository for more insights and codes: https://github.com/crux82/ExtremITA

from peft import PeftModel
from transformers import LLaMATokenizer, LLaMAForCausalLM
import torch

tokenizer = LLaMATokenizer.from_pretrained("yahma/llama-7b-hf")
model = LlamaForCausalLM.from_pretrained(
        "sag-uniroma2/extremITA-Camoscio-7b",
        load_in_8bit=True,
        torch_dtype=torch.float16,
        device_map="auto",
    )
model = PeftModel.from_pretrained(
    model,
    "sag-uniroma2/extremITA-Camoscio-7b-adapters",
    torch_dtype=torch.float16,
    device_map="auto",
)