mod allegro
This commit is contained in:
parent
03faf77ee4
commit
cd535b4fe3
|
|
@ -9,7 +9,7 @@ import numpy as np
|
||||||
from sentence_transformers import SentenceTransformer
|
from sentence_transformers import SentenceTransformer
|
||||||
from datasets import Dataset
|
from datasets import Dataset
|
||||||
from peft import LoraConfig, get_peft_model
|
from peft import LoraConfig, get_peft_model
|
||||||
from transformers import AutoModelForCausalLM, AutoTokenizer, TrainingArguments, Trainer, DataCollatorForLanguageModeling
|
from transformers import AutoModelForCausalLM, AutoTokenizer, TrainingArguments, Trainer, DataCollatorForLanguageModeling, MarianForCausalLM, MarianTokenizer
|
||||||
|
|
||||||
embed_model = SentenceTransformer("all-MiniLM-L6-v2")
|
embed_model = SentenceTransformer("all-MiniLM-L6-v2")
|
||||||
|
|
||||||
|
|
@ -64,8 +64,8 @@ eval_dataset = split_dataset["test"]
|
||||||
|
|
||||||
device = "cuda" if torch.cuda.is_available() else "cpu"
|
device = "cuda" if torch.cuda.is_available() else "cpu"
|
||||||
model_name = "allegro/multislav-5lang"
|
model_name = "allegro/multislav-5lang"
|
||||||
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16).to(device)
|
model = MarianForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16).to(device)
|
||||||
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
tokenizer = MarianForCausalLM.from_pretrained(model_name)
|
||||||
|
|
||||||
lora_config = LoraConfig(
|
lora_config = LoraConfig(
|
||||||
r=8, lora_alpha=32, lora_dropout=0.1, bias="none", task_type="CAUSAL_LM"
|
r=8, lora_alpha=32, lora_dropout=0.1, bias="none", task_type="CAUSAL_LM"
|
||||||
|
|
|
||||||
|
|
@ -12,3 +12,5 @@ peft
|
||||||
weaviate-client
|
weaviate-client
|
||||||
sentence_transformers
|
sentence_transformers
|
||||||
faiss-gpu
|
faiss-gpu
|
||||||
|
sentencepiece
|
||||||
|
sacremoses
|
||||||
Loading…
Reference in New Issue