Deep learning with Tensorflow and Keras

Deep learning with Tensorflow and Keras
Deep learning with Tensorflow and Keras

Deep Learning is nowadays on the boom because of the frameworks like Tensorflow and Keras. It has become easy to make Machine Learning model without actually knowing the working beneath it.

Making a machine learning model using basic libraries is a nightmare for someone mostly if they are in learning time, so the framework comes in picture.

Google builds a framework called TensorFlow to let researchers and developers work together on an AI model. Once developed and scaled, it allows lots of people to use it. TensorFlow uses the full computational power of modern hardware to train or run the model in terms of graphs. Keras on other hand provides another layer of API over Tensorflow, thus making the model without knowing the actual implementation of the model or more precisely layer.

Machine learning frameworks like TensorFlow, Paddle Paddle, Torch, Caffe, Keras and many others can speed up your machine learning development significantly all of these frameworks also have a lot of documentation. We will discuss Tensorflow and Keras and make our first machine learning model for hand digit recognition using Keras.


Tensorflow: Programming frameworks can not only shorten your coding time but sometimes also perform optimisations that speed up your code. Tensorflow is a numerical computation library that was created on C++, CUDA, Python and it is an open-source library that is maintained by Google Brain Team so we can be reliable on the framework.

It is supported on Linux, Windows and macOS. Well in recent time if we are machine learning enthusiastically then we already know that there is TensorFlow lite is launched for the small computation hardware i.e. raspberry pie. So, this framework can be pretty reliable and efficient.

Advantages
We know will go through the advantages of using Tensorflow:

  • Uses Computation Graphs for computation and stores variable as nodes.
  • TensorFlow is fast, as it uses Computational Graphs.
  • Can use GPU/CPU for computing, where the same code can be executed on both architectures.
  • Efficiently works with mathematical expressions involving multi-dimensional arrays.
  • TensorFlow has Energetic Community.

Steps
Writing and running programs in TensorFlow has the following steps:

  • Create Tensors (variables) that are not yet executed/evaluated.
  • Write operations between those Tensors.
  • Initialise your Tensors.
  • Create a Session.
  • Run the Session.

Implementation: Below is the basic implementation of implementing TensorFlow with very basic python code.

import tensorflow as tf
y_hat = tf.constant(36, name=’y_hat’) # Define y_hat constant. Set to 36.
y = tf.constant(39, name=’y’) # Define y. Set to 39
loss = tf.Variable((y – y_hat)**2, name=’loss’) # Create a variable for the loss
init = tf.global_variables_initialiser() # When init is run later (session.run(init)),
# the loss variable will be initialised and ready to be computed
with tf.Session() as session: # Create a session and print the output
session.run(init) # Initialises the variables
print(session.run(loss)) # Prints the loss

a = tf.constant(2)
b = tf.constant(10)
c = tf.multiply(a, b)
print(c)

The output here will not be 20. All we did was put in the ‘computation graph’, but we have not run this computation yet. In order to actually multiply the two numbers, we will have to create a session and run it.

sess = tf.Session()
print(sess.run(c))

A placeholder is an object whose value we can specify only later. To specify values for a placeholder, we can pass in values by using a “feed dictionary” (feed_dict variable). Below, we created a placeholder for x. This allows us to pass in a number later when we run the session.

x = tf.placeholder(tf.int64, name = ‘x’)
print(sess.run(2 * x, feed_dict = {x: 3}))
sess.close()

When we specify the operations needed for a computation, we are telling TensorFlow how to construct a computation graph. Finally, when we run the session, we are telling TensorFlow to execute the computation graph.

From above python code I have tried to provide with basic ideas of working with TensorFlow, Implementing Neural Network here will take a lot of time and we will get diverted from the basic idea of the topic, so we could always refer here for more information and reading.

Keras: It is a powerful and easy-to-use free open-source Python library for developing and evaluating deep learning models. It was developed under project ONEIROS (Open-ended Neuro-Electronic Intelligent Robot Operating System) where the primary author was François Chollet. It lay on top of the efficient numerical computation libraries like Theano and TensorFlow. It also allows you to define and train neural network models in just a few lines of code.

Advantages of using Keras:

  • Allows users to productise deep models on smartphones (Android and iOS).
  • Contains numerous implementations of commonly used neural-network building blocks.
  • Keras contains some good Prelabelled Datasets.
  • Multiple methods for Data Pre-processing.
  • Modularity.

Writing and running programs in TensorFlow has the following steps:

  • Load Data.
  • Define Keras Model.
  • Compile Keras Model.
  • Fit Keras Model.
  • Evaluate Keras Model.
  • Tie It All Together.
  • Make Predictions.

Implementation
In this section, we will discover how to create your first deep learning neural network model in Python using Keras. Here we would be using Keras API through TensorFlow. We will here build a simple model for handwritten digit recognition and take the accuracy till 99% using two dense layers with 512 and 10 neurons, the first layer would be using relu activation function and the last layer with 10 neurons would be for classification of 10 digits. I have used Jupyter notebook for making the model as it is easier to visualise the project, you could also use google colab. We will see step by step process below.

import tensorflow as tf
path = “ “ #path of the dataset

def train_mnist():

class myCallback(tf.keras.callbacks.Callback):                 #callback to stop the training when a certain accuracy is reached
  def on_epoch_end(self, epoch, logs={}):                
    if(logs.get('acc')>0.99):
      print("Reached 99% accuracy so cancelling training!")
      self.model.stop_training = True
mnist = tf.keras.datasets.mnist

(x_train, y_train),(x_test, y_test) = mnist.load_data(path=path)    #Loading the dataset

x_train, x_test = x_train/255, x_test/255                #normal distribution of the dataset
callbacks = myCallback()                        #making a callback object
#creating model
model = tf.keras.models.Sequential([                    
    tf.keras.layers.Flatten(input_shape=(28, 28)),            #making a tensor of shape 784 (28*28 pixel)
    tf.keras.layers.Dense(512, activation=tf.nn.relu),            #making a dense layer of 512 neuron with relu activation
    tf.keras.layers.Dense(10, activation=tf.nn.softmax)            #making a dense layer of 10 neuron with softmax activation
])

model.compile(optimizer='adam',                    #using adam optimizer
              loss='sparse_categorical_crossentropy',
              metrics=['accuracy'])
# model fitting
history = model.fit(
        x_train, y_train, epochs=10, callbacks=[callbacks]            #training for 10 epochs 
)
return history.epoch, history.history['acc'][-1]

train_mnist()

Training this can take time depending on the hardware (Processor and RAM) but you can always make use of virtualisation (online notebooks like google colab), that are free and can provide descent speed. Here from the model history we can see that the model has stopped training when it reached 99% accuracy because of callback class. From each epoch you can easily see that the loss is decreasing and the accuracy is increasing as the model learns. In other cases where the accuracy is not that good you can always tune your hyper parameter’s and can make your training more accurate i.e. you can try for changing the number of epochs.

Conclusion

You have now seen that with a few lines of code you can create a CNN model that can reach an accuracy of 99% in training without the actual implementation of the hard-core mathematics been used, this all is because of frameworks like Keras and TensorFlow. Well, you are always free to implement the core ideas and the algorithms of machine learning from the basic using TensorFlow. Tensorflow is also used to perform scientific calculation cause of its computational graphs. TensorFlow can provide an edge with the rich libraries and can make use of hardware acceleration very well and is also maintained by Google thus is reliable also.

To enrol in Machine Learning & Deep Learning course, click here.

By Gaurav Verma

Exit mobile version