fine tunning

This commit is contained in:
Alex
2026-02-11 17:02:29 +01:00
parent d5313fb143
commit 71e595a966
3 changed files with 2 additions and 4 deletions

View File

@@ -145,7 +145,7 @@ training_args = TrainingArguments(
per_device_train_batch_size=1,
gradient_accumulation_steps=16,
learning_rate=1e-4,
num_train_epochs=3,
num_train_epochs=2,
max_steps=1000,
fp16=False, # ⚠ disable AMP

View File

@@ -8,7 +8,7 @@ from nltk.translate.bleu_score import corpus_bleu
# Configuration
# ----------------------------
BASE_MODEL = "Qwen/Qwen2.5-7B-Instruct" # base model
LORA_DIR = "./qwen2.5-7b-uk-fr-lora" # fine-tuned LoRA
LORA_DIR = "./qwen2.5-7b-uk-fr-lora-2epoch" # fine-tuned LoRA
VALIDATION_FILE = "validation.jsonl" # small validation subset
MAX_INPUT_LENGTH = 1024
DEVICE = "cuda" if torch.cuda.is_available() else "cpu"