# Multiclass logistic regression with `gluon`

¶

Now that we’ve built a logistic regression model from
scratch, let’s make this more
efficient with `gluon`

. If you completed the corresponding chapters on
linear regression, you might be tempted rest your eyes a little in this
one. We’ll be using `gluon`

in a rather similar way and since the
interface is reasonably well designed, you won’t have to do much work.
To keep you awake we’ll introduce a few subtle tricks.

Let’s start by importing the standard packages.

```
In [52]:
```

```
from __future__ import print_function
import mxnet as mx
from mxnet import nd, autograd
from mxnet import gluon
import numpy as np
```

## Set the context¶

Now, let’s set the context. In the linear regression tutorial we did all
of our computation on the cpu (`mx.cpu()`

) just to keep things simple.
When you’ve got 2-dimensional data and scalar labels, a smartwatch can
probably handle the job. Already, in this tutorial we’ll be working with
a considerably larger dataset. If you happen to be running this code on
a server with a GPU and installed the GPU-enabled version of MXNet (or
remembered to build MXNet with `CUDA=1`

), you might want to substitute
the following line for its commented-out counterpart.

```
In [53]:
```

```
data_ctx = mx.cpu()
model_ctx = mx.cpu()
# model_ctx = mx.gpu()
```

## The MNIST Dataset¶

We won’t suck up too much wind describing the MNIST dataset for a second time. If you’re unfamiliar with the dataset and are reading these chapters out of sequence, take a look at the data section in the previous chapter on softmax regression from scratch.

We’ll load up data iterators corresponding to the training and test splits of MNIST dataset.

```
In [54]:
```

```
batch_size = 64
num_inputs = 784
num_outputs = 10
num_examples = 60000
def transform(data, label):
return data.astype(np.float32)/255, label.astype(np.float32)
train_data = mx.gluon.data.DataLoader(mx.gluon.data.vision.MNIST(train=True, transform=transform),
batch_size, shuffle=True)
test_data = mx.gluon.data.DataLoader(mx.gluon.data.vision.MNIST(train=False, transform=transform),
batch_size, shuffle=False)
```

We’re also going to want to load up an iterator with *test* data. After
we train on the training dataset we’re going to want to test our model
on the test data. Otherwise, for all we know, our model could be doing
something stupid (or treacherous?) like memorizing the training examples
and regurgitating the labels on command.

## Multiclass Logistic Regression¶

Now we’re going to define our model. Remember from our tutorial on
linear regression with ``gluon` <./P02-C02-linear-regression-gluon>`__
that we add `Dense`

layers by calling
`net.add(gluon.nn.Dense(num_outputs))`

. This leaves the parameter
shapes under-specified, but `gluon`

will infer the desired shapes the
first time we pass real data through the network.

```
In [55]:
```

```
net = gluon.nn.Dense(num_outputs)
```

## Parameter initialization¶

As before, we’re going to register an initializer for our parameters.
Remember that `gluon`

doesn’t even know what shape the parameters have
because we never specified the input dimension. The parameters will get
initialized during the first call to the forward method.

```
In [56]:
```

```
net.collect_params().initialize(mx.init.Normal(sigma=1.), ctx=model_ctx)
```

## Softmax Cross Entropy Loss¶

Note, we didn’t have to include the softmax layer because MXNet’s has an efficient function that simultaneously computes the softmax activation and cross-entropy loss. However, if ever need to get the output probabilities,

```
In [57]:
```

```
softmax_cross_entropy = gluon.loss.SoftmaxCrossEntropyLoss()
```

## Optimizer¶

And let’s instantiate an optimizer to make our updates

```
In [58]:
```

```
trainer = gluon.Trainer(net.collect_params(), 'sgd', {'learning_rate': 0.1})
```

## Evaluation Metric¶

This time, let’s simplify the evaluation code by relying on MXNet’s
built-in `metric`

package.

```
In [59]:
```

```
def evaluate_accuracy(data_iterator, net):
acc = mx.metric.Accuracy()
for i, (data, label) in enumerate(data_iterator):
data = data.as_in_context(model_ctx).reshape((-1,784))
label = label.as_in_context(model_ctx)
output = net(data)
predictions = nd.argmax(output, axis=1)
acc.update(preds=predictions, labels=label)
return acc.get()[1]
```

Because we initialized our model randomly, and because roughly one tenth of all examples belong to each of the ten classes, we should have an accuracy in the ball park of .10.

```
In [60]:
```

```
evaluate_accuracy(test_data, net)
```

```
Out[60]:
```

```
0.1154
```

## Execute training loop¶

```
In [61]:
```

```
epochs = 10
moving_loss = 0.
niter = 0
for e in range(epochs):
cumulative_loss = 0
for i, (data, label) in enumerate(train_data):
data = data.as_in_context(model_ctx).reshape((-1,784))
label = label.as_in_context(model_ctx)
with autograd.record():
output = net(data)
loss = softmax_cross_entropy(output, label)
loss.backward()
trainer.step(batch_size)
cumulative_loss += nd.sum(loss).asscalar()
test_accuracy = evaluate_accuracy(test_data, net)
train_accuracy = evaluate_accuracy(train_data, net)
print("Epoch %s. Loss: %s, Train_acc %s, Test_acc %s" % (e, cumulative_loss/num_examples, train_accuracy, test_accuracy))
```

```
Epoch 0. Loss: 0.000342435105642, Train_acc 0.793733333333, Test_acc 0.809
Epoch 1. Loss: 0.000266353193919, Train_acc 0.83805, Test_acc 0.8477
Epoch 2. Loss: 0.000140365982056, Train_acc 0.856316666667, Test_acc 0.8648
Epoch 3. Loss: 0.000119470739365, Train_acc 0.86695, Test_acc 0.874
Epoch 4. Loss: 0.000254932610194, Train_acc 0.8731, Test_acc 0.8796
Epoch 5. Loss: 0.000143766593933, Train_acc 0.879266666667, Test_acc 0.8847
Epoch 6. Loss: 0.000247673273087, Train_acc 0.882366666667, Test_acc 0.8863
Epoch 7. Loss: 0.000343579641978, Train_acc 0.88615, Test_acc 0.8896
Epoch 8. Loss: 0.000479016272227, Train_acc 0.88865, Test_acc 0.8911
Epoch 9. Loss: 0.000274674447378, Train_acc 0.8905, Test_acc 0.8919
```

## Visualize predictions¶

```
In [62]:
```

```
import matplotlib.pyplot as plt
def model_predict(net,data):
output = net(data.as_in_context(model_ctx))
return nd.argmax(output, axis=1)
# let's sample 10 random data points from the test set
sample_data = mx.gluon.data.DataLoader(mx.gluon.data.vision.MNIST(train=False, transform=transform),
10, shuffle=True)
for i, (data, label) in enumerate(sample_data):
data = data.as_in_context(ctx)
print(data.shape)
im = nd.transpose(data,(1,0,2,3))
im = nd.reshape(im,(28,10*28,1))
imtiles = nd.tile(im, (1,1,3))
plt.imshow(imtiles.asnumpy())
plt.show()
pred=model_predict(net,data.reshape((-1,784)))
print('model predictions are:', pred)
break
```

```
(10, 28, 28, 1)
```

```
model predictions are:
[ 9. 9. 0. 4. 7. 6. 8. 2. 7. 3.]
<NDArray 10 @cpu(0)>
```

## Next¶

Overfitting and regularization from scratch

For whinges or inquiries, open an issue on GitHub.