Tensorflow...
.jpeg)
Let's discuss little bit about TensorFlow ...It is a libre open source software which
is also said to be "FLOSS" and originated via Google researchers and scientists.
Generally we
utilized it to speed the machine learning , deep learning and other various
arithmetical forecasting or
prescriptive analytics consignments.
Tensorflow also
depicted to prune the procedure over generating and implementing proceeded data interpretation approach
like prediction, forecasting , arithematical and data science modellers .A modeller that create
theoretical descriptions of techniques or
procedure for the prediction and development.
Here, TensorFlow software manipulate a collection of data which generally
positioned even as a computational vertex in tabulation or graphical shape, we can say.
Let's understand, What is tensor? In general, Tensor is a tabulation in which we produce
the outline and vertex that can depict the multidimensional matrices or vertex especially where very
large dataset
computations accomplished like neural network.
If I talk about, what is Tensorflow? Then,we probably say, Its flexible architecture allows for the easy deployment of computation across a variety of platforms (CPUs, GPUs, TPUs), and from desktops to clusters of servers to mobile and edge devices.
Numpy
Now let's gain some info regarding the strongest library of python i.e numPy. So Numpy, we can say one of the very influencive libraries in Python libraries, and afterthat consistency and unification in the data structure provided via Tensorflow efficiently. .Accordingly, we can say integrated Tensorflow enables us to build complex input pipelines,null chunks. For example NDarrays, the library's initial datatype, which spontaneously converted into TensorFlow Tensors in TF operations; similar we can say also vericious or else vice-versa. That permits as dual libraries can perform task simultaneously at a time in the absence of operator who ca withoutrequiring the user to write explicit data conversions. Moreover, the integration extends to memory optimization by having TF Tensors share the underlying memory representations of Numpy NDarrays whenever possibleGoogle Colab
As I am using Colab, released via Google, a TensorFlow Jupyter notebook environment which is not demanding any updation. We can use it in browser also. that provides free access to GPUs and the storage and also share notebooks on Google Drive.Let's perform some coding of tensorflow
from tensorflow import keras
import matplotlib.pyplot as plt
import numpy as np
Convolutional layer: CNN Layer is used to process the image — used for convolutional neural networks.
Recurrent layer: Recurrent layer is used to process sequencial form of data — used for recurrent neural networks.
MaxPooling layer:MaxPooling layer used to down-sampling attribute by taking the maximum value in non-overlapping rectangular blocks — used to sustain important attributes also reducing the probability overfitting.
Flatten layer: Flatten layer used to flattens the multi-dimensional input tensors into a single dimension — utilized as a transformation layer in between convolutional or recurrent layers and fully connected layers in a neural network.
Layer: A layer generally used to put down input units to 0 randomly (using a defined frequency) during training time —utilized as a regularization technique for preserving the overfitting in neural networks.
Embedding layer: Embedding layer: a layer that represents words or phrases in a high-dimensional vector space — used to map words or phrases to dense vectors for use as input to a neural network.
Now we compile our model ,here we used optimizer as
SGD : As Gradient Descent is an iterative optimization process which generally used to searche an objective function’s optimum value (Minimum/Maximum). It is one of the most used methods for changing a model’s parameters in order to reduce a cost function in machine learning, we ca say...
Categorical Cross Entropy :Categorical Cross Entropy is aloss function mainly used for multi-class classification model in which two or more than two outputs labels.It produces a one-hot array sustaining the expected match.
Accuracy:An accuracy i metrics can be determined as a simple comparison between how many target values matches the predicted values.
model.compile( optimizer='SGD', loss='categorical_crossentropy', metrics=['accuracy'])
model.fit(X_train_scaled, y_train_categorical, epochs=5)
model.compile( optimizer='SGD', loss='categorical_crossentropy', metrics=['accuracy'])
return model
1563/1563 [==============================] - 79s 50ms/step - loss: 1.8111 - accuracy: 0.3549 Epoch 2/5 1563/1563 [==============================] - 77s 49ms/step - loss: 1.6228 - accuracy: 0.4294 Epoch 3/5 1563/1563 [==============================] - 77s 49ms/step - loss: 1.5435 - accuracy: 0.4573 Epoch 4/5 1563/1563 [==============================] - 80s 51ms/step - loss: 1.4849 - accuracy: 0.4774 Epoch 5/5 1563/1563 [==============================] - 79s 50ms/step - loss: 1.4336 - accuracy: 0.4964 6min 32s ± 0 ns per loop (mean ± std. dev. of 1 run, 1 loop each)
1563/1563 [==============================] - 6s 4ms/step - loss: 1.8112 - accuracy: 0.3545 Epoch 2/5 1563/1563 [==============================] - 6s 4ms/step - loss: 1.6253 - accuracy: 0.4266 Epoch 3/5 1563/1563 [==============================] - 6s 4ms/step - loss: 1.5435 - accuracy: 0.4542 Epoch 4/5 1563/1563 [==============================] - 6s 4ms/step - loss: 1.4828 - accuracy: 0.4776 Epoch 5/5 1563/1563 [==============================] - 8s 5ms/step - loss: 1.4338 - accuracy: 0.4923 42.6 s ± 0 ns per loop (mean ± std. dev. of 1 run, 1 loop each)