Skip to content

Commit

Permalink
אצווה 32 קבועה
Browse files Browse the repository at this point in the history
  • Loading branch information
NHLOCAL committed Jul 29, 2024
1 parent ae615d6 commit 386a140
Show file tree
Hide file tree
Showing 16 changed files with 25 additions and 545 deletions.
10 changes: 8 additions & 2 deletions machine-learn/compar_f1score.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
import spacy
from spacy.training.example import Example
from tabulate import tabulate
import os

# Function to evaluate model and return evaluation metrics
def evaluate_model(model_name, data):
Expand Down Expand Up @@ -192,8 +193,13 @@ def evaluate_model(model_name, data):
models_metrics = []

# Iterate over model versions
for i in range(1, 10):
model_name = f"custom_ner_model23-{i}git"
model_names = [i for i in os.listdir() if "custom_ner_model" in i]





for model_name in model_names:
print(f"Evaluating {model_name}...")
f1_score, precision, recall = evaluate_model(model_name, data)
models_metrics.append({
Expand Down
11 changes: 6 additions & 5 deletions machine-learn/creating_model_git.py
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,6 @@ def custom_tokenizer(nlp):
example = Example.from_dict(nlp.make_doc(example_text), {'entities': entities})
training_data.append(example)

random.shuffle(training_data)

nlp.begin_training()

Expand All @@ -75,18 +74,20 @@ def custom_tokenizer(nlp):

n_iter = 100
batch_sizes = compounding(16.0, 64.0, 1.001)
batch_size = 32
drop_size = 0.4
iteration_data = {}
#initial_lr = 0.001 # שיעור למידה התחלתי
#lr_decay = 0.95 # קצב דעיכת שיעור הלמידה
# optimizer.learn_rate = initial_lr

for itn in range(n_iter):
random.shuffle(training_data)
batches = minibatch(training_data, size=batch_sizes)
losses = {}
for batch in batches:
nlp.update(batch, drop=0.4, losses=losses)
print(f"Iteration {itn}, Losses: {losses}")
for i in range(0, len(training_data), batch_size):
batch = training_data[i:i + batch_size]
nlp.update(batch, drop=drop_size, losses=losses)
print(f"Iteration {itn}: {losses}")
iteration_data[itn] = losses.copy()

current_loss = losses.get('ner', float('inf'))
Expand Down
129 changes: 0 additions & 129 deletions machine-learn/custom_ner_model/config.cfg

This file was deleted.

34 changes: 0 additions & 34 deletions machine-learn/custom_ner_model/meta.json

This file was deleted.

13 changes: 0 additions & 13 deletions machine-learn/custom_ner_model/ner/cfg

This file was deleted.

Binary file removed machine-learn/custom_ner_model/ner/model
Binary file not shown.
1 change: 0 additions & 1 deletion machine-learn/custom_ner_model/ner/moves

This file was deleted.

3 changes: 0 additions & 3 deletions machine-learn/custom_ner_model/tokenizer

This file was deleted.

1 change: 0 additions & 1 deletion machine-learn/custom_ner_model/vocab/key2row

This file was deleted.

1 change: 0 additions & 1 deletion machine-learn/custom_ner_model/vocab/lookups.bin

This file was deleted.

Loading

0 comments on commit 386a140

Please sign in to comment.