forked from aimacode/aima-python
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathtest_learning4e.py
More file actions
127 lines (97 loc) · 4.03 KB
/
test_learning4e.py
File metadata and controls
127 lines (97 loc) · 4.03 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
import pytest
from deep_learning4e import PerceptronLearner
from learning4e import *
random.seed("aima-python")
def test_exclude():
iris = DataSet(name='iris', exclude=[3])
assert iris.inputs == [0, 1, 2]
def test_parse_csv():
iris = open_data('iris.csv').read()
assert parse_csv(iris)[0] == [5.1, 3.5, 1.4, 0.2, 'setosa']
def test_weighted_mode():
assert weighted_mode('abbaa', [1, 2, 3, 1, 2]) == 'b'
def test_weighted_replicate():
assert weighted_replicate('ABC', [1, 2, 1], 4) == ['A', 'B', 'B', 'C']
def test_means_and_deviation():
iris = DataSet(name='iris')
means, deviations = iris.find_means_and_deviations()
assert round(means['setosa'][0], 3) == 5.006
assert round(means['versicolor'][0], 3) == 5.936
assert round(means['virginica'][0], 3) == 6.588
assert round(deviations['setosa'][0], 3) == 0.352
assert round(deviations['versicolor'][0], 3) == 0.516
assert round(deviations['virginica'][0], 3) == 0.636
def test_plurality_learner():
zoo = DataSet(name='zoo')
pl = PluralityLearner(zoo)
assert pl.predict([1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 0, 4, 1, 0, 1]) == 'mammal'
def test_k_nearest_neighbors():
iris = DataSet(name='iris')
knn = NearestNeighborLearner(iris, k=3)
assert knn.predict([5, 3, 1, 0.1]) == 'setosa'
assert knn.predict([6, 5, 3, 1.5]) == 'versicolor'
assert knn.predict([7.5, 4, 6, 2]) == 'virginica'
def test_decision_tree_learner():
iris = DataSet(name='iris')
dtl = DecisionTreeLearner(iris)
assert dtl.predict([5, 3, 1, 0.1]) == 'setosa'
assert dtl.predict([6, 5, 3, 1.5]) == 'versicolor'
assert dtl.predict([7.5, 4, 6, 2]) == 'virginica'
def test_svc():
iris = DataSet(name='iris')
classes = ['setosa', 'versicolor', 'virginica']
iris.classes_to_numbers(classes)
n_samples, n_features = len(iris.examples), iris.target
X, y = (np.array([x[:n_features] for x in iris.examples]),
np.array([x[n_features] for x in iris.examples]))
svm = MultiClassLearner(SVC()).fit(X, y)
assert svm.predict([[5.0, 3.1, 0.9, 0.1]]) == 0
assert svm.predict([[5.1, 3.5, 1.0, 0.0]]) == 0
assert svm.predict([[4.9, 3.3, 1.1, 0.1]]) == 0
assert svm.predict([[6.0, 3.0, 4.0, 1.1]]) == 1
assert svm.predict([[6.1, 2.2, 3.5, 1.0]]) == 1
assert svm.predict([[5.9, 2.5, 3.3, 1.1]]) == 1
assert svm.predict([[7.5, 4.1, 6.2, 2.3]]) == 2
assert svm.predict([[7.3, 4.0, 6.1, 2.4]]) == 2
assert svm.predict([[7.0, 3.3, 6.1, 2.5]]) == 2
def test_information_content():
assert information_content([]) == 0
assert information_content([4]) == 0
assert information_content([5, 4, 0, 2, 5, 0]) > 1.9
assert information_content([5, 4, 0, 2, 5, 0]) < 2
assert information_content([1.5, 2.5]) > 0.9
assert information_content([1.5, 2.5]) < 1.0
def test_random_forest():
iris = DataSet(name='iris')
rf = RandomForest(iris)
tests = [([5.0, 3.0, 1.0, 0.1], 'setosa'),
([5.1, 3.3, 1.1, 0.1], 'setosa'),
([6.0, 5.0, 3.0, 1.0], 'versicolor'),
([6.1, 2.2, 3.5, 1.0], 'versicolor'),
([7.5, 4.1, 6.2, 2.3], 'virginica'),
([7.3, 3.7, 6.1, 2.5], 'virginica')]
assert grade_learner(rf, tests) >= 1 / 3
def test_random_weights():
min_value = -0.5
max_value = 0.5
num_weights = 10
test_weights = random_weights(min_value, max_value, num_weights)
assert len(test_weights) == num_weights
for weight in test_weights:
assert min_value <= weight <= max_value
def test_ada_boost():
iris = DataSet(name='iris')
classes = ['setosa', 'versicolor', 'virginica']
iris.classes_to_numbers(classes)
wl = WeightedLearner(PerceptronLearner(iris))
ab = ada_boost(iris, wl, 5)
tests = [([5, 3, 1, 0.1], 0),
([5, 3.5, 1, 0], 0),
([6, 3, 4, 1.1], 1),
([6, 2, 3.5, 1], 1),
([7.5, 4, 6, 2], 2),
([7, 3, 6, 2.5], 2)]
assert grade_learner(ab, tests) > 2 / 3
assert err_ratio(ab, iris) < 0.25
if __name__ == "__main__":
pytest.main()