mod gpt
This commit is contained in:
parent
01005c52c5
commit
992b55745e
10
gpt.py
10
gpt.py
|
|
@ -37,19 +37,17 @@ def main():
|
|||
tokenized_dataset = dataset.map(tokenize_function, batched=True)
|
||||
|
||||
# Model
|
||||
model = AutoModelForCausalLM.from_pretrained(
|
||||
MODEL_NAME,
|
||||
mean_resizing=False # Wyłączenie ostrzeżenia
|
||||
)
|
||||
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
|
||||
model.resize_token_embeddings(len(tokenizer))
|
||||
|
||||
# Konfiguracja treningu
|
||||
training_args = TrainingArguments(
|
||||
output_dir="./results",
|
||||
num_train_epochs=1, # Poprawiona nazwa parametru
|
||||
num_train_epochs=1,
|
||||
per_device_train_batch_size=2,
|
||||
remove_unused_columns=True,
|
||||
logging_steps=1
|
||||
logging_steps=1,
|
||||
report_to="none" # Wyłączenie raportowania
|
||||
)
|
||||
|
||||
# Trainer
|
||||
|
|
|
|||
Loading…
Reference in New Issue