Skip to content

Commit 2378011

Browse files
committed
Style fixes
1 parent ffbe408 commit 2378011

25 files changed

+31
-28
lines changed

mla/datasets/base.py

+1
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,6 @@
11
# coding:utf-8
22
import os
3+
34
import numpy as np
45

56

mla/ensemble/gbm.py

-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,5 @@
11
# coding:utf-8
22
import numpy as np
3-
43
# logistic function
54
from scipy.special import expit
65

mla/ensemble/random_forest.py

-1
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,6 @@
44
from mla.base import BaseEstimator
55
from mla.ensemble.base import information_gain, mse_criterion
66
from mla.ensemble.tree import Tree
7-
from six.moves import range
87

98

109
class RandomForest(BaseEstimator):

mla/fm.py

+3-2
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,11 @@
11
# coding:utf-8
22

3-
from mla.base import BaseEstimator
4-
from mla.metrics import mean_squared_error, binary_crossentropy
53
import autograd.numpy as np
64
from autograd import elementwise_grad
75

6+
from mla.base import BaseEstimator
7+
from mla.metrics import mean_squared_error, binary_crossentropy
8+
89
np.random.seed(9999)
910

1011
"""

mla/gaussian_mixture.py

+3-1
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,11 @@
11
# coding:utf-8
22

33
import random
4+
5+
import matplotlib.pyplot as plt
46
import numpy as np
57
from scipy.stats import multivariate_normal
6-
import matplotlib.pyplot as plt
8+
79
from mla.base import BaseEstimator
810
from mla.kmeans import KMeans
911

mla/kmeans.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
# coding:utf-8
22

33
import random
4-
import seaborn as sns
5-
import matplotlib.pyplot as plt
64

5+
import matplotlib.pyplot as plt
76
import numpy as np
7+
import seaborn as sns
88

99
from mla.base import BaseEstimator
1010
from mla.metrics.distance import euclidean_distance

mla/metrics/distance.py

+2-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,8 @@
11
# coding:utf-8
2-
import numpy as np
32
import math
43

4+
import numpy as np
5+
56

67
def euclidean_distance(a, b):
78
if isinstance(a, list) and isinstance(b, list):

mla/naive_bayes.py

+1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
# coding:utf-8
22

33
import numpy as np
4+
45
from mla.base import BaseEstimator
56
from mla.neuralnet.activations import softmax
67

mla/neuralnet/layers/normalization.py

+2-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,8 @@
11
# coding:utf-8
2+
import numpy as np
3+
24
from mla.neuralnet.layers import Layer, PhaseMixin, ParamMixin
35
from mla.neuralnet.parameters import Parameters
4-
import numpy as np
56

67
"""
78
References:
+1-1
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
11
# coding:utf-8
2-
from .rnn import *
32
from .lstm import *
3+
from .rnn import *

mla/neuralnet/layers/recurrent/lstm.py

-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
# coding:utf-8
22
import autograd.numpy as np
33
from autograd import elementwise_grad
4-
from six.moves import range
54

65
from mla.neuralnet.activations import sigmoid
76
from mla.neuralnet.initializations import get_initializer

mla/neuralnet/layers/recurrent/rnn.py

-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
# coding:utf-8
22
import autograd.numpy as np
33
from autograd import elementwise_grad
4-
from six.moves import range
54

65
from mla.neuralnet.initializations import get_initializer
76
from mla.neuralnet.layers import Layer, get_activation, ParamMixin

mla/neuralnet/loss.py

-1
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,4 @@
11
from ..metrics import mse, logloss, mae, hinge, binary_crossentropy
2-
32
categorical_crossentropy = logloss
43

54

mla/neuralnet/optimizers.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -215,7 +215,7 @@ def update(self, network):
215215
self.ms[i][n] = self.beta_1 * self.ms[i][n] + (1.0 - self.beta_1) * grad
216216
self.us[i][n] = np.maximum(self.beta_2 * self.us[i][n], np.abs(grad))
217217

218-
step = self.lr / (1 - self.beta_1 ** self.t) * self.ms[i][n]/(self.us[i][n] + self.epsilon)
218+
step = self.lr / (1 - self.beta_1 ** self.t) * self.ms[i][n] / (self.us[i][n] + self.epsilon)
219219
layer.parameters.step(n, -step)
220220
self.t += 1
221221

mla/neuralnet/regularizers.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# coding:utf-8
2-
from autograd import elementwise_grad
32
import numpy as np
3+
from autograd import elementwise_grad
44

55

66
class Regularizer(object):

mla/neuralnet/tests/test_activations.py

+1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,5 @@
11
import sys
2+
23
import numpy as np
34

45
from mla.neuralnet.activations import *

mla/neuralnet/tests/test_optimizers.py

+3-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
1-
from sklearn.model_selection import train_test_split
21
from sklearn.datasets import make_classification
32
from sklearn.metrics import roc_auc_score
3+
from sklearn.model_selection import train_test_split
44

55
from mla.neuralnet import NeuralNet
66
from mla.neuralnet.layers import Dense, Activation, Dropout, Parameters
@@ -46,9 +46,11 @@ def test_adadelta():
4646
def test_adam():
4747
assert clasifier(Adam()) > 0.9
4848

49+
4950
def test_adamax():
5051
assert clasifier(Adamax()) > 0.9
5152

53+
5254
def test_rmsprop():
5355
assert clasifier(RMSprop()) > 0.9
5456

mla/pca.py

+3-2
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,9 @@
11
# coding:utf-8
2-
from scipy.linalg import svd
3-
import numpy as np
42
import logging
53

4+
import numpy as np
5+
from scipy.linalg import svd
6+
67
from mla.base import BaseEstimator
78

89
np.random.seed(1000)

mla/rbm.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
# coding:utf-8
22
import logging
33

4-
from mla.base import BaseEstimator
5-
from scipy.special import expit
64
import numpy as np
5+
from scipy.special import expit
76

7+
from mla.base import BaseEstimator
88
from mla.utils import batch_iterator
99

1010
np.random.seed(9999)

mla/rl/dqn.py

-1
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55
import gym
66
import numpy as np
77
from gym import wrappers
8-
from six.moves import range
98

109
np.random.seed(9999)
1110

mla/svm/svm.py

+4-2
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,10 @@
11
# coding:utf-8
2+
import logging
3+
4+
import numpy as np
5+
26
from mla.base import BaseEstimator
37
from mla.svm.kernerls import Linear
4-
import numpy as np
5-
import logging
68

79
np.random.seed(9999)
810

mla/tests/test_reduction.py

+1-2
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,7 @@
11
# coding=utf-8
22
import pytest
3-
4-
from sklearn.metrics import roc_auc_score
53
from sklearn.datasets import make_classification
4+
from sklearn.metrics import roc_auc_score
65

76
try:
87
from sklearn.model_selection import train_test_split

mla/tests/test_regression_accuracy.py

-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,6 @@
1212
from mla.neuralnet.optimizers import Adam
1313
from mla.neuralnet.parameters import Parameters
1414

15-
1615
# Generate a random regression problem
1716
X, y = make_regression(
1817
n_samples=1000, n_features=10, n_informative=10, n_targets=1, noise=0.05, random_state=1111, bias=0.5

mla/tsne.py

-1
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,6 @@
22
import logging
33

44
import numpy as np
5-
from six.moves import range
65

76
from mla.base import BaseEstimator
87
from mla.metrics.distance import l2_distance

requirements.txt

-1
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,5 @@ numpy>=1.11.1
44
scikit-learn>=0.18
55
scipy>=0.18.0
66
seaborn>=0.7.1
7-
six>=1.10.0
87
autograd>=1.1.7
98
gym

0 commit comments

Comments
 (0)