mod allegro

This commit is contained in:
l.gabrysiak 2025-02-28 21:40:42 +01:00
parent 03faf77ee4
commit cd535b4fe3
2 changed files with 6 additions and 4 deletions

View File

@ -9,7 +9,7 @@ import numpy as np
from sentence_transformers import SentenceTransformer from sentence_transformers import SentenceTransformer
from datasets import Dataset from datasets import Dataset
from peft import LoraConfig, get_peft_model from peft import LoraConfig, get_peft_model
from transformers import AutoModelForCausalLM, AutoTokenizer, TrainingArguments, Trainer, DataCollatorForLanguageModeling from transformers import AutoModelForCausalLM, AutoTokenizer, TrainingArguments, Trainer, DataCollatorForLanguageModeling, MarianForCausalLM, MarianTokenizer
embed_model = SentenceTransformer("all-MiniLM-L6-v2") embed_model = SentenceTransformer("all-MiniLM-L6-v2")
@ -64,8 +64,8 @@ eval_dataset = split_dataset["test"]
device = "cuda" if torch.cuda.is_available() else "cpu" device = "cuda" if torch.cuda.is_available() else "cpu"
model_name = "allegro/multislav-5lang" model_name = "allegro/multislav-5lang"
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16).to(device) model = MarianForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16).to(device)
tokenizer = AutoTokenizer.from_pretrained(model_name) tokenizer = MarianForCausalLM.from_pretrained(model_name)
lora_config = LoraConfig( lora_config = LoraConfig(
r=8, lora_alpha=32, lora_dropout=0.1, bias="none", task_type="CAUSAL_LM" r=8, lora_alpha=32, lora_dropout=0.1, bias="none", task_type="CAUSAL_LM"

View File

@ -12,3 +12,5 @@ peft
weaviate-client weaviate-client
sentence_transformers sentence_transformers
faiss-gpu faiss-gpu
sentencepiece
sacremoses