File size: 3,733 Bytes
bfd5fab
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94


# Check if the training sentences are made up entirely of (test sentences or their repetitions)

# Observation : Some of the sentences in original dataset comprised of other sentences in the dataset
    # For example, One sentence would be "A B C D" and another sentence would be "A B C D A B C D"
    # This is not a good thing as the model can easily overfit on the training data and 
    # if the other sentence thats being repeated is in the test set, the model will perform very well on the test set but not on the real world data

import pandas as pd
import ahocorasick
from tqdm import tqdm
import json

# File paths
training_set_path = '../train_set.csv'
test_set_path = '../small_validation_set.csv'

print("Loading the CSV files...")
# Load the CSV files into Pandas DataFrames
test_set = pd.read_csv(test_set_path, encoding='utf-8')
training_set = pd.read_csv(training_set_path, encoding='utf-8')

print("CSV files loaded successfully.")

# Prepare the test sentences
test_sentences = test_set['Urdu text'].dropna()
# Remove extra spaces and standardize
test_sentences = test_sentences.apply(lambda x: ' '.join(x.strip().split()))
# Skip sentences with only one word if desired
test_sentences = test_sentences[test_sentences.str.split().str.len() > 1].unique()
test_sentences_set = set(test_sentences)

print(f"Number of test sentences: {len(test_sentences_set)}")

# Build the Aho-Corasick automaton
print("Building the Aho-Corasick automaton with test sentences...")
A = ahocorasick.Automaton()

for idx, test_sentence in enumerate(test_sentences):
    A.add_word(test_sentence, (idx, test_sentence))

A.make_automaton()
print("Automaton built successfully.")

# Initialize matches dictionary
matches = {}

print("Processing training sentences...")
training_sentences = training_set['Urdu text'].dropna()
# Remove extra spaces and standardize
training_sentences = training_sentences.apply(lambda x: ' '.join(x.strip().split()))
training_sentences = training_sentences.unique()

for training_sentence in tqdm(training_sentences):
    s = training_sentence
    s_length = len(s)
    matches_in_s = []
    for end_index, (insert_order, test_sentence) in A.iter(s):
        start_index = end_index - len(test_sentence) + 1
        matches_in_s.append((start_index, end_index, test_sentence))
    if not matches_in_s:
        continue
    # Sort matches by start_index
    matches_in_s.sort(key=lambda x: x[0])
    # Now check if matches cover the entire training sentence without gaps
    # And all matches are of the same test sentence
    covers_entire_sentence = True
    current_index = 0
    first_test_sentence = matches_in_s[0][2]
    all_same_test_sentence = True
    for start_index, end_index, test_sentence in matches_in_s:
        if start_index != current_index:
            covers_entire_sentence = False
            break
        if test_sentence != first_test_sentence:
            all_same_test_sentence = False
            break
        current_index = end_index + 1
    if covers_entire_sentence and current_index == s_length and all_same_test_sentence:
        # Training sentence is made up entirely of repetitions of test_sentence
        if test_sentence not in matches:
            matches[test_sentence] = []
        matches[test_sentence].append(s)
print("Processing completed.")
print("Number of matches: ", sum(len(v) for v in matches.values()))

# Optionally, save matches to a JSON file
output_file_path = '/netscratch/butt/Transliterate/RUP/finetuning/scripts/one_time_usage/test_training_matches.json'
with open(output_file_path, 'w', encoding='utf-8') as json_file:
    json.dump(matches, json_file, ensure_ascii=False, indent=4)

print(f"Matches have been written to {output_file_path}")