Update data augmentation comments in largedataset_cnn example
diff --git a/examples/largedataset_cnn/train_largedata.py b/examples/largedataset_cnn/train_largedata.py
old mode 100644
new mode 100755
index 81f3e72..adba9ce
--- a/examples/largedataset_cnn/train_largedata.py
+++ b/examples/largedataset_cnn/train_largedata.py
@@ -27,7 +27,7 @@
from PIL import Image
import process_data
-# Data Augmentation
+# Data augmentation
def augmentation(x, batch_size):
xpad = np.pad(x, [[0, 0], [0, 0], [4, 4], [4, 4]], 'symmetric')
for data_num in range(0, batch_size):
@@ -41,7 +41,7 @@
return x
-# Calculate Accuracy
+# Calculate accuracy
def accuracy(pred, target):
# y is network output to be compared with ground truth (int)
y = np.argmax(pred, axis=1)
@@ -68,7 +68,7 @@
return train_x, train_y, val_x, val_y
-# Function to all reduce NUMPY Accuracy and Loss from Multiple Devices
+# Function to all reduce NUMPY accuracy and loss from multiple devices
def reduce_variable(variable, dist_opt, reducer):
reducer.copy_from_numpy(variable)
dist_opt.all_reduce(reducer.data)
@@ -130,7 +130,7 @@
model = alexnet.create_model(num_channels=num_channels,
num_classes=num_classes)
- # For distributed training, sequential gives better performance
+ # For distributed training, sequential method gives better performance
if hasattr(sgd, "communicator"):
DIST = True
sequential = True