TensorFlow Logistic Regression
Logistic regression is borrowed from statistics. You can use this for classification problems. Given an image, is it class 0 or class 1?
The word “logistic regression” is named after its function “the logistic”. You may know this function as the sigmoid function.
Related Course:
Deep Learning for Computer Vision with Tensor Flow and Keras
Introduction
Sigmund function
Logisitic regression uses the sigmund function for classification problems. What is this function exactly?
The sigmund function is:
1 / (1 + e^t) 
It’s an sshaped curve.
Why use the signmund function for prediction? The sshaped curve is kind of strange, isn’t it?
Classifications in prediction problems are probabilistic. The model shouldn’t be below zero or higher than one, the sshaped curve helps to create that. Because of the limits, it can be used for binary classification.
Sigmund function in logistic regression
The function can be used to make predictions.
p(X) = e^(b0 + b1*X) / (1 + e^(b0 + b1*X)) 
The variable b0 is the bias and b1 is the coefficient for the single input value (x)
This can be rewritten as
ln(odds) = b0 + b1 * X 
or
odds = e^(b0 + b1 * X) 
To make predictions, you need b0 and b1.
These values are found with the training data.
Initially we set them to zero:
W = tf.Variable(tf.zeros([784, 10])) 
Tensorflow will take care of that.
Then the model (based on formula) is:
pred = tf.nn.softmax(tf.matmul(x, W) + b) # Softmax 
Where’s the exponent?
The softmax function does the equivalent of:
softmax = tf.exp(logits) / tf.reduce_sum(tf.exp(logits), axis) 
Logistic regression with handwriting recognition
Lets use logistic regression for handwriting recognition. The MNIST datset contains 28x28 images of handwritten numbers. Each of those is flattened to be a 784 size 1d vector.
The problem is:
 X: image of a handwritten digit
 Y: the digit value
 Recognize the digit in the image
The model:
 logits = X * w + b
 Y_predicted = softmax(logits)
 loss = cross_entropy(Y, Y_predicted)
The same in code:
pred = tf.nn.softmax(tf.matmul(x, W) + b) # Softmax 
Loss is sometimes called cost.
The code below runs the logistic regression model on the handwriting set. Surprisingly the accuracy is 91.43% for this model. Simply copy and run!

© 2007  2019 pythonprogramminglanguage.com.
Cookie Policy & Privacy Policy.