fine tunning
This commit is contained in:
parent
d5313fb143
commit
71e595a966
@ -145,7 +145,7 @@ training_args = TrainingArguments(
|
|||||||
per_device_train_batch_size=1,
|
per_device_train_batch_size=1,
|
||||||
gradient_accumulation_steps=16,
|
gradient_accumulation_steps=16,
|
||||||
learning_rate=1e-4,
|
learning_rate=1e-4,
|
||||||
num_train_epochs=3,
|
num_train_epochs=2,
|
||||||
max_steps=1000,
|
max_steps=1000,
|
||||||
|
|
||||||
fp16=False, # ⚠ disable AMP
|
fp16=False, # ⚠ disable AMP
|
||||||
|
|||||||
@ -8,7 +8,7 @@ from nltk.translate.bleu_score import corpus_bleu
|
|||||||
# Configuration
|
# Configuration
|
||||||
# ----------------------------
|
# ----------------------------
|
||||||
BASE_MODEL = "Qwen/Qwen2.5-7B-Instruct" # base model
|
BASE_MODEL = "Qwen/Qwen2.5-7B-Instruct" # base model
|
||||||
LORA_DIR = "./qwen2.5-7b-uk-fr-lora" # fine-tuned LoRA
|
LORA_DIR = "./qwen2.5-7b-uk-fr-lora-2epoch" # fine-tuned LoRA
|
||||||
VALIDATION_FILE = "validation.jsonl" # small validation subset
|
VALIDATION_FILE = "validation.jsonl" # small validation subset
|
||||||
MAX_INPUT_LENGTH = 1024
|
MAX_INPUT_LENGTH = 1024
|
||||||
DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
|
DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
|
||||||
|
|||||||
@ -11,7 +11,6 @@ from reportlab.pdfbase.ttfonts import TTFont
|
|||||||
import os, time
|
import os, time
|
||||||
|
|
||||||
# Configuration
|
# Configuration
|
||||||
DEBUG = True
|
|
||||||
PDF_PATH = "Traduction/TaniaBorecMemoir(Ukr).pdf"
|
PDF_PATH = "Traduction/TaniaBorecMemoir(Ukr).pdf"
|
||||||
OLLAMA_MODEL = "traductionUkrainienVersFrancais:latest"
|
OLLAMA_MODEL = "traductionUkrainienVersFrancais:latest"
|
||||||
OLLAMA_URL = "http://localhost:11434/api/generate"
|
OLLAMA_URL = "http://localhost:11434/api/generate"
|
||||||
@ -174,7 +173,6 @@ def load_checkpoint():
|
|||||||
return json.load(f)
|
return json.load(f)
|
||||||
return {"last_processed_index": -1, "results": {}}
|
return {"last_processed_index": -1, "results": {}}
|
||||||
|
|
||||||
# Sauvegarde le checkpoint
|
|
||||||
# Sauvegarde le checkpoint
|
# Sauvegarde le checkpoint
|
||||||
def save_checkpoint(last_index, results):
|
def save_checkpoint(last_index, results):
|
||||||
# Trier les clés du dictionnaire results
|
# Trier les clés du dictionnaire results
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user