Newer
Older
from utils import isclose, DataFile
P1 = UnigramTextModel(wordseq)
P2 = NgramTextModel(2, wordseq)
P3 = NgramTextModel(3, wordseq)
# The most frequent entries in each model
assert P1.top(10) == [(2081, 'the'), (1479, 'of'), (1021, 'and'),
(1008, 'to'), (850, 'a'), (722, 'i'), (640, 'in'),
(478, 'that'), (399, 'is'), (348, 'you')]
assert P2.top(10) == [(368, ('of', 'the')), (152, ('to', 'the')),
(152, ('in', 'the')), (86, ('of', 'a')),
(80, ('it', 'is')),
(71, ('by', 'the')), (68, ('for', 'the')),
(68, ('and', 'the')), (62, ('on', 'the')),
(60, ('to', 'be'))]
assert P3.top(10) == [(30, ('a', 'straight', 'line')),
(19, ('of', 'three', 'dimensions')),
(16, ('the', 'sense', 'of')),
(13, ('by', 'the', 'sense')),
(13, ('as', 'well', 'as')),
(12, ('of', 'the', 'circles')),
(12, ('of', 'sight', 'recognition')),
(11, ('the', 'number', 'of')),
(11, ('that', 'i', 'had')), (11, ('so', 'as', 'to'))]
assert isclose(P1['the'], 0.0611, rel_tol=0.001)
assert isclose(P2['of', 'the'], 0.0108, rel_tol=0.01)
assert isclose(P3['', '', 'but'], 0.0, rel_tol=0.001)
assert isclose(P3['', '', 'but'], 0.0, rel_tol=0.001)
assert isclose(P3['so', 'as', 'to'], 0.000323, rel_tol=0.001)
assert P2.cond_prob.get(('went',)) is None
assert P3.cond_prob['in', 'order'].dictionary == {'to': 6}
test_string = 'unigram'
wordseq = words(test_string)
P1 = UnigramTextModel(wordseq)
assert P1.dictionary == {('unigram'): 1}
test_string = 'bigram text'
wordseq = words(test_string)
P2 = NgramTextModel(2, wordseq)
assert (P2.dictionary == {('', 'bigram'): 1, ('bigram', 'text'): 1} or
P2.dictionary == {('bigram', 'text'): 1, ('', 'bigram'): 1})
test_string = 'test trigram text'
wordseq = words(test_string)
P3 = NgramTextModel(3, wordseq)
assert ('', '', 'test') in P3.dictionary
assert ('', 'test', 'trigram') in P3.dictionary
assert ('test', 'trigram', 'text') in P3.dictionary
assert len(P3.dictionary) == 3
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
def test_char_models():
test_string = 'unigram'
wordseq = words(test_string)
P1 = NgramCharModel(1, wordseq)
assert len(P1.dictionary) == len(test_string)
for char in test_string:
assert tuple(char) in P1.dictionary
test_string = 'a b c'
wordseq = words(test_string)
P1 = NgramCharModel(1, wordseq)
assert len(P1.dictionary) == len(test_string.split())
for char in test_string.split():
assert tuple(char) in P1.dictionary
test_string = 'bigram'
wordseq = words(test_string)
P2 = NgramCharModel(2, wordseq)
expected_bigrams = {(' ', 'b'): 1, ('b', 'i'): 1, ('i', 'g'): 1, ('g', 'r'): 1, ('r', 'a'): 1, ('a', 'm'): 1}
assert len(P2.dictionary) == len(expected_bigrams)
for bigram, count in expected_bigrams.items():
assert bigram in P2.dictionary
assert P2.dictionary[bigram] == count
test_string = 'bigram bigram'
wordseq = words(test_string)
P2 = NgramCharModel(2, wordseq)
expected_bigrams = {(' ', 'b'): 2, ('b', 'i'): 2, ('i', 'g'): 2, ('g', 'r'): 2, ('r', 'a'): 2, ('a', 'm'): 2}
assert len(P2.dictionary) == len(expected_bigrams)
for bigram, count in expected_bigrams.items():
assert bigram in P2.dictionary
assert P2.dictionary[bigram] == count
test_string = 'trigram'
wordseq = words(test_string)
P3 = NgramCharModel(3, wordseq)
expected_trigrams = {(' ', ' ', 't'): 1, (' ', 't', 'r'): 1, ('t', 'r', 'i'): 1,
('r', 'i', 'g'): 1, ('i', 'g', 'r'): 1, ('g', 'r', 'a'): 1,
('r', 'a', 'm'): 1}
assert len(P3.dictionary) == len(expected_trigrams)
for bigram, count in expected_trigrams.items():
assert bigram in P3.dictionary
assert P3.dictionary[bigram] == count
test_string = 'trigram trigram trigram'
wordseq = words(test_string)
P3 = NgramCharModel(3, wordseq)
expected_trigrams = {(' ', ' ', 't'): 3, (' ', 't', 'r'): 3, ('t', 'r', 'i'): 3,
('r', 'i', 'g'): 3, ('i', 'g', 'r'): 3, ('g', 'r', 'a'): 3,
('r', 'a', 'm'): 3}
assert len(P3.dictionary) == len(expected_trigrams)
for bigram, count in expected_trigrams.items():
assert bigram in P3.dictionary
assert P3.dictionary[bigram] == count
def test_viterbi_segmentation():
flatland = DataFile("EN-text/flatland.txt").read()
wordseq = words(flatland)
P = UnigramTextModel(wordseq)
text = "itiseasytoreadwordswithoutspaces"
s, p = viterbi_segment(text, P)
assert s == [
'it', 'is', 'easy', 'to', 'read', 'words', 'without', 'spaces']
def test_shift_encoding():
code = shift_encode("This is a secret message.", 17)
assert code == 'Kyzj zj r jvtivk dvjjrxv.'
ring = ShiftDecoder(flatland)
msg = ring.decode('Kyzj zj r jvtivk dvjjrxv.')
assert msg == 'This is a secret message.'
def test_permutation_decoder():
gutenberg = DataFile("EN-text/gutenberg.txt").read()
flatland = DataFile("EN-text/flatland.txt").read()
pd = PermutationDecoder(canonicalize(gutenberg))
assert pd.decode('aba') in ('ece', 'ete', 'tat', 'tit', 'txt')
pd = PermutationDecoder(canonicalize(flatland))
assert pd.decode('aba') in ('ded', 'did', 'ece', 'ele', 'eme', 'ere', 'eve', 'eye', 'iti', 'mom', 'ses', 'tat', 'tit')
def test_rot13_encoding():
code = rot13('Hello, world!')
assert code == 'Uryyb, jbeyq!'
flatland = DataFile("EN-text/flatland.txt").read()
ring = ShiftDecoder(flatland)
msg = ring.decode(rot13('Hello, world!'))
assert msg == 'Hello, world!'
def test_counting_probability_distribution():
D = CountingProbDist()
for i in range(10000):
D.add(random.choice('123456'))
ps = [D[n] for n in '123456']
assert 1 / 7 <= min(ps) <= max(ps) <= 1 / 5
def test_ir_system():
from collections import namedtuple
Results = namedtuple('IRResults', ['score', 'url'])
uc = UnixConsultant()
def verify_query(query, expected):
assert len(expected) == len(query)
for expected, (score, d) in zip(expected, query):
doc = uc.documents[d]
assert "{0:.2f}".format(
expected.score) == "{0:.2f}".format(score * 100)
assert os.path.basename(expected.url) == os.path.basename(doc.url)
q1 = uc.query("how do I remove a file")
assert verify_query(q1, [
Results(76.83, "aima-data/MAN/rm.txt"),
Results(67.83, "aima-data/MAN/tar.txt"),
Results(67.79, "aima-data/MAN/cp.txt"),
Results(66.58, "aima-data/MAN/zip.txt"),
Results(64.58, "aima-data/MAN/gzip.txt"),
Results(63.74, "aima-data/MAN/pine.txt"),
Results(62.95, "aima-data/MAN/shred.txt"),
Results(57.46, "aima-data/MAN/pico.txt"),
Results(43.38, "aima-data/MAN/login.txt"),
Results(41.93, "aima-data/MAN/ln.txt"),
])
q2 = uc.query("how do I delete a file")
assert verify_query(q2, [
Results(75.47, "aima-data/MAN/diff.txt"),
Results(69.12, "aima-data/MAN/pine.txt"),
Results(63.56, "aima-data/MAN/tar.txt"),
Results(60.63, "aima-data/MAN/zip.txt"),
Results(57.46, "aima-data/MAN/pico.txt"),
Results(51.28, "aima-data/MAN/shred.txt"),
Results(26.72, "aima-data/MAN/tr.txt"),
])
q3 = uc.query("email")
assert verify_query(q3, [
Results(18.39, "aima-data/MAN/pine.txt"),
Results(12.01, "aima-data/MAN/info.txt"),
Results(9.89, "aima-data/MAN/pico.txt"),
Results(8.73, "aima-data/MAN/grep.txt"),
Results(8.07, "aima-data/MAN/zip.txt"),
Results(128.15, "aima-data/MAN/grep.txt"),
Results(94.20, "aima-data/MAN/find.txt"),
Results(81.71, "aima-data/MAN/du.txt"),
Results(55.45, "aima-data/MAN/ps.txt"),
Results(53.42, "aima-data/MAN/more.txt"),
Results(42.00, "aima-data/MAN/dd.txt"),
Results(12.85, "aima-data/MAN/who.txt"),
])
q5 = uc.query("learn: date")
assert verify_query(q5, [])
q6 = uc.query("2003")
assert verify_query(q6, [
Results(14.58, "aima-data/MAN/pine.txt"),
Results(11.62, "aima-data/MAN/jar.txt"),
def test_words():
assert words("``EGAD!'' Edgar cried.") == ['egad', 'edgar', 'cried']
def test_canonicalize():
assert canonicalize("``EGAD!'' Edgar cried.") == 'egad edgar cried'
def test_translate():
text = 'orange apple lemon '
func = lambda x: ('s ' + x) if x ==' ' else x
assert translate(text, func) == 'oranges apples lemons '
def test_bigrams():
assert bigrams('this') == ['th', 'hi', 'is']
assert bigrams(['this', 'is', 'a', 'test']) == [['this', 'is'], ['is', 'a'], ['a', 'test']]
# TODO: for .ipynb
"""
>>> P1.samples(20)
'you thought known but were insides of see in depend by us dodecahedrons just but i words are instead degrees'
>>> P2.samples(20)
'flatland well then can anything else more into the total destruction and circles teach others confine women must be added'
>>> P3.samples(20)
'flatland by edwin a abbott 1884 to the wake of a certificate from nature herself proving the equal sided triangle'
if __name__ == '__main__':
pytest.main()