What should know about Deep Learning
Hello I have wrote some questions about important concepts that every Data Scientist working in Deep Learning should know.
What type of predictive model correspond to the continuous regression outcome?
The linear regression is the regression needed.
What type of predictive model correspond to the class outcome?
Logistic regression, SVM, Naive Bayes
What types of Discriminative models correspond to the the table below:
Discriminative model  

Goal  Directly estimate P(yx) 
What’s learned  Decision boundary 
Illustration 
The Discriminative models needed are : Regressions, SVMs
What type of Generative model correspond to the the table below:
Generative model  

Goal  Estimate \(P(xy)\) to then deduce \(P(yx)\) 
What’s learned  Probability distributions of the data 
Illustration 
The Generative model are : GDA, Naive Bayes
What is Loss function?
A loss function is a function \(L:(z,y)\in\mathbb{R}\times Y\longmapsto L(z,y)\in\mathbb{R}\) that takes as inputs the predicted value z corresponding to the real data value y and outputs how different they are.
What are the most common loss functions ?
The common loss functions are summed up in the table below:
Least squared error  Logistic loss  Hinge loss  Crossentropy 

\(\displaystyle\frac{1}{2}(yz)^2\)  \(\displaystyle\log(1+\exp(yz))\)  \(\displaystyle\max(0,1yz)\)  \(\displaystyle\Big[y\log(z)+(1y)\log(1z)]\) 
Linear regression  Logistic regression  SVM  Neural Network 
What is a Cost function?
The cost function is commonly used to assess the performance of a model, and is defined with the loss function as follows:
\[J(\theta)=\sum_{i=1}^mL(h_\theta(x^{(i)}), y^{(i)})\]where \(h_\theta\) is the hypothesis and is the model that we choose. For a given input data \(x^{(i)}\) the model prediction output is \(h_\theta(x^{(i)})\)
What is Gradient Descent?
Gradient descent By noting \(\alpha\in\mathbb{R}\) the learning rate, the update rule for gradient descent is expressed with the learning rate and the cost function $J$ as follows:
\[\theta\longleftarrow \theta  \alpha \nabla J(\theta)\]What is Stochastic gradient?
Stochastic gradient descent (SGD) is updating the parameter based on each training example, and batch gradient descent is on a batch of training examples.
What is The likelihood of a model?
The likelihood of a model \(L(\theta)\) given parameters \(\theta\) is used to find the optimal parameters \(\theta\) through likelihood maximization. We have:
\[\theta^{\textrm{opt}}=\underset{\theta}{\textrm{arg max }}L(\theta)\]Remark: in practice, we use the loglikelihood \(\ell(\theta)=\log(L(\theta))\) which is easier to optimize.
What is Newton’s algorithm Newton’s algorithm ?
Newton’s algorithm Newton’s algorithm is a numerical method that finds \(\theta\) such that \(\ell'(\theta)=0\). Its update rule is as follows:
\[\theta\leftarrow\theta\frac{\ell'(\theta)}{\ell''(\theta)}\]Remark: the multidimensional generalization, also known as the NewtonRaphson method, has the following update rule:
\[\theta\leftarrow\theta\left(\nabla_\theta^2\ell(\theta)\right)^{1}\nabla_\theta\ell(\theta)\]What is the Least Mean Squares (LMS) algorithm ?
LMS algorithm By noting \(\alpha\) the learning rate, the update rule of the Least Mean Squares (LMS) algorithm for a training set of \(m\) data points, which is also known as the WidrowHoff learning rule, is as follows:
\[\displaystyle \forall j \quad \theta_j \leftarrow \theta_j+\alpha\sum_{i=1}^m[y^{(i)}h_\theta(x^{(i)})]x_j^{(i)}\]What is Sigmoid function in classification ?
Sigmoid function The sigmoid function $g$, also known as the logistic function, is defined as follows:
\[\forall z\in\mathbb{R},\quad g(z)=\frac{1}{1+e^{z}} \in [0,1]\]What is Logistic regresion?
Logistic regression is a statistical model that in its basic form uses a logistic function to model a binary dependent variable, although many more complex extensions exist. In regression analysis, logistic regression is estimating the parameters of a logistic model (a form of binary regression).
For example, assuming $$x  y ; \theta\sim\text{Bernoulli}(\phi)$$. 
We have the following form:
\[\phi=p(y=1x;\theta)=\frac{1}{1+\exp(\theta^Tx)}=g(\theta^Tx)\]For example, for a student who studies 2 hours, entering the value the equation gives the estimated probability of passing the exam of 0.26: \(\text { Probability of passing exam }=\frac{1}{1+\exp ((1.5046 \cdot 24.0777))}=0.26\)
Remark: logistic regressions do not have closed form solutions.
What is Softmax regression?
A softmax regression, also called a multiclass logistic regression, is used to generalize logistic regression when there are more than 2 outcome classes. By convention, we set \(\theta_K=0\), which makes the Bernoulli parameter \(\phi_i\) of each class \(i\) be such that:
\[\displaystyle\phi_i=\frac{\exp(\theta_i^Tx)}{\displaystyle\sum_{j=1}^K\exp(\theta_j^Tx)}\]Remark: the update rule is a particular case of the gradient ascent.
What is Support Vector Machines?
The goal of support vector machines is to find the line that maximizes the minimum distance to the line.
The optimal margin classifier h is such that:
\[\boxed{h(x)=\textrm{sign}(w^Txb)}\]where \((w, b)\in\mathbb{R}^n\times\mathbb{R}\) is the solution of the following optimization problem:
\[\boxed{\min\frac{1}{2}w^2}\quad\quad\textrm{such that }\quad \boxed{y^{(i)}(w^Tx^{(i)}b)\geqslant1}\]*Remark: the decision boundary is defined as \(\boxed{w^Txb=0}\)
What is Hinge loss?
The hinge loss is used in the setting of SVMs and is defined as follows:
\[\boxed{L(z,y)=[1yz]_+=\max(0,1yz)}\]What is Kernel in Support Vector?
Given a feature mapping \(\phi\), we define the kernel K as follows:
\[\boxed{K(x,z)=\phi(x)^T\phi(z)}\]In practice, the kernel K defined by
\[K(x,z)=\exp\left(\frac{\xz\^2}{2\sigma^2}\right)\]is called the Gaussian kernel and is commonly used.
Remark: we say that we use the “kernel trick” to compute the cost function using the kernel because we actually don’t need to know the explicit mapping \(\phi\), which is often very complicated. Instead, only the values \(K(x,z)\) are needed.
What is Lagrangian?
Lagrange’s method of multipliers is used to derive the local maxima and minima in a function subject to equality constraints.
We define the Lagrangian \(\mathcal{L}(w,b)\) as follows:
\[\boxed{\mathcal{L}(w,b)=f(w)+\sum_{i=1}^l\beta_ih_i(w)}\]Remark: the coefficients \(\beta_i\) are called the Lagrange multipliers.
What is Generative Learning
A generative model first tries to learn how the data is generated by estimating
P(x  y), which we can then use to estimate P(y  x)) by using Bayes’ rule. 
What is Gaussian Discriminant Analysis?
GDA, is a method for data classification commonly used when data can be approximated with a Normal distribution. As first step, you will need a training set, i.e. a bunch of data yet classified. These data are used to train your classifier, and obtain a discriminant function that will tell you to which class a data has higher probability to belong.
When you have your training set you need to compute the mean μ and the standard deviation σ2 These two variables, as you know, allow you to describe a Normal distribution.
Once you have computed the Normal distribution for each class, to classify a data you will need to compute, for each one, the probability that that data belongs to it. The class with the highest probability will be chosen as the affinity class.
The following table sums up the estimates that we find when maximizing the likelihood:
\(\widehat{\phi}\)  \(\widehat{\mu_j}\quad{\small(j=0,1)}\)  \(\widehat{\Sigma}\) 

\(\displaystyle\frac{1}{m}\sum_{i=1}^m1_{\{y^{(i)}=1\}}\)  \(\displaystyle\frac{\sum_{i=1}^m1_{\{y^{(i)}=j\}}x^{(i)}}{\sum_{i=1}^m1_{\{y^{(i)}=j\}}}\)  \(\displaystyle\frac{1}{m}\sum_{i=1}^m(x^{(i)}\mu_{y^{(i)}})(x^{(i)}\mu_{y^{(i)}})^T*m*1*i*=1\) 
What is Naive Bayes?
Naïve Bayes is a simple learning algorithm that utilizes Bayes rule together with a strong assumption that the attributes are conditionally independent, given the class.
The Naive Bayes model supposes that the features of each data point are all independent:
\[\boxed{P(xy)=P(x_1,x_2,...y)=P(x_1y)P(x_2y)...=\prod_{i=1}^nP(x_iy)}\]Remark: Naive Bayes is widely used for text classification and spam detection.
What is Bayes rule?
Describes the probability of an event, based on prior knowledge of conditions that might be related to the event. Bayes’ theorem is stated mathematically as the following equation: \(P(A \mid B)=\frac{P(B \mid A) P(A)}{P(B)}\) where \(A, B\) are events and \(P(B)\neq 0\)
What is decision trees?
A decision tree is a decision support tool that uses a treelike model of decisions and their possible consequences, including chance event outcomes, resource costs, and utility.
What is Random forest?
Random forests are an ensemble learning method for classification, regression and other tasks that operates by constructing a multitude of decision trees at training time.
It is a treebased technique that uses a high number of decision trees built out of randomly selected sets of features. Contrary to the simple decision tree, it is highly uninterpretable but its generally good performance makes it a popular algorithm.
What is the difference between decision tree and random forest?
A decision tree combines some decisions, whereas a random forest combines several decision trees. Thus, it is a long process, yet slow. Whereas, a decision tree is fast and operates easily on large data sets, especially the linear one. The random forest model needs rigorous training.
What are the boosting methods ?
The idea of boosting methods is to combine several weak learners to form a stronger one. The main ones are summed up in the table below:
Adaptive boosting  Gradient boosting 

• High weights are put on errors to improve at the next boosting step. Known as Adaboost  • Weak learners are trained on residuals. Examples include XGBoost 
What is the knearest neighbors?
The knearest neighbors algorithm, commonly known as kNN, is a nonparametric approach where the response of a data point is determined by the nature of its k neighbors from the training set. It can be used in both classification and regression settings. Remark: the higher the parameter k the higher the bias, and the lower the parameter k, the higher the variance.
What is the goal of Unsupervised Learning?
The goal of unsupervised learning is to find hidden patterns in unlabeled data \(\{x^{(1)},...,x^{(m)}\}\)
What is Latent variables in Clustering?
Latent variables are hidden/unobserved variables that make estimation problems difficult, and are often denoted z. Here are the most common settings where there are latent variables:
Setting  Latent variable z  \(xz\)  Comments 

Mixture of k Gaussians  \(\textrm{Multinomial}(\phi)\)  \(\mathcal{N}(\mu_j,\Sigma_j))\)  \(\mu_j\in\mathbb{R}^n\) 
Factor analysis  \(\mathcal{N}(0,I)\)  \(\mathcal{N}(\mu+\Lambda z,\psi)\) 
The ExpectationMaximization (EM) algorithm gives an efficient method at estimating the parameter \(\theta\) through maximum likelihood estimation by repeatedly constructing a lowerbound on the likelihood (Estep) and optimizing that lower bound (Mstep)
What is kmeans clustering?
We note \(c^{(i)}\) the cluster of data point i and \(\mu_j\) the center of cluster j After randomly initializing the cluster centroids \(\mu_1,\mu_2,...,\mu_k\in\mathbb{R}\) the kmeans algorithm repeats the following step until convergence:
What is Distortion function? In order to see if the algorithm converges, we look at the distortion function defined as follows:
\[J(c,\mu)=\sum_{i=1}^m\ x^{(i)}\mu_{c^{(i)}} \^2\]What is Hierarchical clustering?
It is a clustering algorithm with an agglomerative hierarchical approach that build nested clusters in a successive manner. There are different sorts of hierarchical clustering algorithms that aims at optimizing different objective functions, which is summed up in the table below:
Ward linkage  Average linkage  Complete linkage 

Minimize within cluster distance  Minimize average distance between cluster pairs  Minimize maximum distance of between cluster pairs 
In an unsupervised learning setting, it is often hard to assess the performance of a model since we don’t have the ground truth labels as was the case in the supervised learning setting.
What is Silhouette coefficient?
By noting a and b the mean distance between a sample and all other points in the same class, and between a sample and all other points in the next nearest cluster, the silhouette coefficient s for a single sample is defined as follows: \(s=\frac{ba}{\max(a,b)}\)
What is Principal component analysis?
It is a dimension reduction technique that finds the variance maximizing directions onto which to project the data.
What is Eigenvalue, eigenvector?
Given a matrix \(A\in\mathbb{R}^{n\times n}\), \(\lambda\) is said to be an eigenvalue of A if there exists a vector \(z\in\mathbb{R}^n\backslash\{0\}\), called eigenvector, such that we have: \(\boxed{Az=\lambda z}\)
What is the The Principal Component Analysis (PCA) algorithm?
The Principal Component Analysis (PCA) procedure is a dimension reduction technique that projects the data on kk dimensions by maximizing the variance of the data as follows:
This procedure maximizes the variance among all kdimensional spaces.
What is the difference between Machine Learning and Deep Learning?
Machine Learning forms a subset of Artificial Intelligence, where we use statistics and algorithms to train machines with data, thereby helping them improve with experience.
Deep Learning is a part of Machine Learning, which involves mimicking the human brain in terms of structures called neurons, thereby forming neural networks.
What is a perceptron?
A perceptron is similar to the actual neuron in the human brain. It receives inputs from various entities and applies functions to these inputs, which transform them to be the output.
A perceptron is mainly used to perform binary classification where it sees an input, computes functions based on the weights of the input, and outputs the required transformation.
How is Deep Learning better than Machine Learning?
Machine Learning is powerful in a way that it is sufficient to solve most of the problems. However,[Deep Learning gets an upper hand when it comes to working with data that has a large number of dimensions. With data that is large in size, a Deep Learning model can easily work with it as it is built to handle this.
What are some of the most used applications of Deep Learning?
Deep Learning is used in a variety of fields today. The most used ones are as follows:
 Sentiment Analysis
 Computer Vision
 Automatic Text Generation
 Object Detection
 Natural Language Processing
 Image Recognition
What is the meaning of overfitting?
Overfitting is a very common issue when working with Deep Learning. It is a scenario where the Deep Learning algorithm vigorously hunts through the data to obtain some valid information. This makes the Deep Learning model pick up noise rather than useful data, causing very high variance and low bias. This makes the model less accurate, and this is an undesirable effect that can be prevented.
What are activation functions?
Activation functions are entities in Deep Learning that are used to translate inputs into a usable output parameter. It is a function that decides if a neuron needs activation or not by calculating the weighted sum on it with the bias.
Using an activation function makes the model output to be nonlinear. There are many types of activation functions:
 ReLU
 Softmax
 Sigmoid
 Linear
 Tanh
Why is Fourier transform used in Deep Learning?
Fourier transform is an effective package used for analyzing and managing large amounts of data present in a database. It can take in realtime array data and process it quickly. This ensures that high efficiency is maintained and also makes the model more open to processing a variety of signals.
What are the steps involved in training a perception in Deep Learning?
There are five main steps that determine the learning of a perceptron:
 Initialize thresholds and weights
 Provide inputs
 Calculate outputs
 Update weights in each step
 Repeat steps 2 to 4
What is the use of the loss function?
The loss function is used as a measure of accuracy to see if a neural network has learned accurately from the training data or not. This is done by comparing the training dataset to the testing dataset. The loss function is a primary measure of the performance of the neural network. In Deep Learning, a good performing network will have a low loss function at all times when training.
What are some of the Deep Learning frameworks or tools that you have used?
This question is quite common in a Deep Learning interview. Make sure to answer based on the experience you have with the tools.
However, some of the top Deep Learning frameworks out there today are:
 TensorFlow
 Keras
 PyTorch
 Caffe2
 CNTK
 MXNet
 Theano
What is the use of the swish function?
The swish function is a selfgated activation function developed by Google. It is now a popular activation function used by many as Google claims that it outperforms all of the other activation functions in terms of computational efficiency.
What are autoencoders?
Autoencoders are artificial neural networks that learn without any supervision. Here, these networks have the ability to automatically learn by mapping the inputs to the corresponding outputs.
Autoencoders, as the name suggests, consist of two entities:
 Encoder: Used to fit the input into an internal computation state
 Decoder: Used to convert the computational state back into the output
What are the steps to be followed to use the gradient descent algorithm?
There are five main steps that are used to initialize and use the gradient descent algorithm:
 Initialize biases and weights for the network
 Send input data through the network (the input layer)
 Calculate the difference (the error) between expected and predicted values
 Change values in neurons to minimize the loss function
 Multiple iterations to determine the best weights for efficient working
Differentiate between a singlelayer perceptron and a multilayer perceptron.
Singlelayer Perceptron  Multilayer Perceptron 

Cannot classify nonlinear data points  Can classify nonlinear data 
Takes in a limited amount of parameters  Withstands a lot of parameters 
Less efficient with large data  Highly efficient with large datasets 
What is data normalization in Deep Learning?
Data normalization is a preprocessing step that is used to refit the data into a specific range. This ensures that the network can learn effectively as it has better convergence when performing backpropagation.
What is forward propagation?
Forward propagation is the scenario where inputs are passed to the hidden layer with weights. In every single hidden layer, the output of the activation function is calculated until the next layer can be processed. It is called forward propagation as the process begins from the input layer and moves toward the final output layer.
What is backpropagation?
Backprobation is used to minimize the cost function by first seeing how the value changes when weights and biases are tweaked in the neural network. This change is easily calculated by understanding the gradient at every hidden layer. It is called backpropagation as the process begins from the output layer, moving backward to the input layers.
What are hyperparameters in Deep Learning?
Hyperparameters are variables used to determine the structure of a neural network. They are also used to understand parameters, such as the learning rate and the number of hidden layers and more, present in the neural network.
How can hyperparameters be trained in neural networks?
Hyperparameters can be trained using four components as shown below:
 Batch size: This is used to denote the size of the input chunk. Batch sizes can be varied and cut into subbatches based on the requirement.
 Epochs: An epoch denotes the number of times the training data is visible to the neural network so that it can train. Since the process is iterative, the number of epochs will vary based on the data.
 Momentum: Momentum is used to understand the next consecutive steps that occur with the current data being executed at hand. It is used to avoid oscillations when training.
 Learning rate: Learning rate is used as a parameter to denote the time required for the network to update the parameters and learn.
Next up on this top Deep Learning interview questions and answers blog, let us take a look at the intermediate questions.
What is the meaning of dropout in Deep Learning?
Dropout is a technique that is used to avoid overfitting a model in Deep Learning. If the dropout value is too low, then it will have minimal effect on learning. If it is too high, then the model can underlearn, thereby causing lower efficiency.
What are tensors?
Tensors are multidimensional arrays in Deep Learning that are used to represent data. They represent the data with higher dimensions. Due to the highlevel nature of the programming languages, the syntax of tensors are easily understood and broadly used.
What is the meaning of model capacity in Deep Learning?
In Deep Learning, model capacity refers to the capacity of the model to take in a variety of mapping functions. Higher model capacity means a large amount of information can be stored in the network.
We will check out neural network interview questions alongside as it is also a vital part of Deep Learning.
What is a Boltzmann machine?
A Boltzmann machine is a type of recurrent neural network that uses binary decisions, alongside biases, to function. These neural networks can be hooked up together to create deep belief networks, which are very sophisticated and used to solve the most complex problems out there.
What are some of the advantages of using TensorFlow?
TensorFlow has numerous advantages, and some of them are as follows:
 High amount of flexibility and platform independence
 Trains using CPU and GPU
 Supports auto differentiation and its features
 Handles threads and asynchronous computation easily
 Opensource
 Has a large community
What is a computational graph in Deep Learning?
A computation graph is a series of operations that are performed to take in inputs and arrange them as nodes in a graph structure. It can be considered as a way of implementing mathematical calculations into a graph. This helps in parallel processing and provides high performance in terms of computational capability.
What is a CNN?
CNNs are convolutional neural networks that are used to perform analysis on images and visuals. These classes of neural networks can input a multichannel image and work on it easily.
These Deep Learning questions must be answered in a concise way. So make sure to understand them and revisit them if necessary.
What are the various layers present in a CNN?
There are four main layers that form a convolutional neural network:
 Convolution: These are layers consisting of entities called filters that are used as parameters to train the network.
 ReLu: It is used as the activation function and used always with the convolution layer.
 Pooling: Pooling is the concept of shrinking the complex data entities that form after convolution and is primarily used to maintain the size of an image after shrinkage.
 Connectedness: This is used to ensure that all of the layers in the neural network are fully connected and activation can be computed using the bias easily.
What is Convolutional layers?
Convolutional layers in a convolutional neural network systematically apply learned filters to input images in order to create feature maps that summarize the presence of those features in the input.
Convolutional layers prove very effective, and stacking convolutional layers in deep models allows layers close to the input to learn lowlevel features (e.g. lines) and layers deeper in the model to learn highorder or more abstract features, like shapes or specific objects.
What is down sampling?
A limitation of the feature map output of convolutional layers is that they record the precise position of features in the input. This means that small movements in the position of the feature in the input image will result in a different feature map. This can happen with recropping, rotation, shifting, and other minor changes to the input image.
A common approach to addressing this problem from signal processing is called down sampling. This is where a lower resolution version of an input signal is created that still contains the large or important structural elements, without the fine detail that may not be as useful to the task.
Down sampling can be achieved with convolutional layers by changing the stride of the convolution across the image. A more robust and common approach is to use a pooling layer.
What is pooling layer?
A pooling layer is a new layer added after the convolutional layer. Specifically, after a nonlinearity (e.g. ReLU) has been applied to the feature maps output by a convolutional layer; for example the layers in a model may look as follows:
 Input Image
 Convolutional Layer
 Nonlinearity
 Pooling Layer
The addition of a pooling layer after the convolutional layer is a common pattern used for ordering layers within a convolutional neural network that may be repeated one or more times in a given model.
The pooling layer operates upon each feature map separately to create a new set of the same number of pooled feature maps.
In what consists pooling operation?
Pooling involves selecting a pooling operation, much like a filter to be applied to feature maps. The size of the pooling operation or filter is smaller than the size of the feature map; specifically, it is almost always 2×2 pixels applied with a stride of 2 pixels.
This means that the pooling layer will always reduce the size of each feature map by a factor of 2, e.g. each dimension is halved, reducing the number of pixels or values in each feature map to one quarter the size. For example, a pooling layer applied to a feature map of 6×6 (36 pixels) will result in an output pooled feature map of 3×3 (9 pixels).
What are the common functions used in the pooling operation?
The pooling operation is specified, rather than learned. Two common functions used in the pooling operation are:
 Average Pooling: Calculate the average value for each patch on the feature map.
 Maximum Pooling (or Max Pooling): Calculate the maximum value for each patch of the feature map.
The result of using a pooling layer and creating down sampled or pooled feature maps is a summarized version of the features detected in the input. They are useful as small changes in the location of the feature in the input detected by the convolutional layer will result in a pooled feature map with the feature in the same location. This capability added by pooling is called the model’s invariance to local translation.
What is an RNN in Deep Learning?
RNNs stand for recurrent neural networks, which form to be a popular type of artificial neural network. They are used to process sequences of data, text, genomes, handwriting, and more. RNNs make use of backpropagation for the training requirements.
What is a vanishing gradient when using RNNs?
Vanishing gradient is a scenario that occurs when we use RNNs. Since RNNs make use of backpropagation, gradients at every step of the way will tend to get smaller as the network traverses through backward iterations. This equates to the model learning very slowly, thereby causing efficiency problems in the network.
What is exploding gradient descent in Deep Learning?
Exploding gradients are an issue causing a scenario that clumps up the gradients. This creates a large number of updates of the weights in the model when training.
The working of gradient descent is based on the condition that the updates are small and controlled. Controlling the updates will directly affect the efficiency of the model.
What is the use of LSTM?
LSTM stands for long shortterm memory. It is a type of RNN that is used to sequence a string of data. It consists of feedback chains that give it the ability to perform like a generalpurpose computational entity.
Where are autoencoders used?
Autoencoders have a wide variety of usage in the real world. The following are some of the popular ones:
 Adding color to black–white images
 Removing noise from images
 Dimensionality reduction
 Feature removal and variation
What are the types of autoencoders?
There are four main types of autoencoders:
 Deep autoencoders
 Convolutional autoencoders
 Sparse autoencoders
 Contractive autoencoders
What is a Restricted Boltzmann Machine?
A Restricted Boltzmann Machine, or RBM for short, is an undirected graphical model that is popularly used in Deep Learning today. It is an algorithm that is used to perform:
 Dimensionality reduction
 Regression
 Classification
 Collaborative filtering
 Topic modeling
Next up on this top Deep Learning interview questions and answers blog, let us take a look at the advanced questions.
What are some of the limitations of Deep Learning?
There are a few disadvantages of Deep Learning as mentioned below:
 Networks in Deep Learning require a huge amount of data to train well.
 Deep Learning concepts can be complex to implement sometimes.
 Achieving a high amount of model efficiency is difficult in many cases.
These are some of the vital advanced deep learning interview questions that you have to know about!
What are the variants of gradient descent?
There are three variants of gradient descent as shown below:
 Stochastic gradient descent: A single training example is used for the calculation of gradient and for updating parameters.
 Batch gradient descent: Gradient is calculated for the entire dataset, and parameters are updated at every iteration.
 Minibatch gradient descent: Samples are broken down into smallersized batches and then worked on as in the case of stochastic gradient descent.
Why is minibatch gradient descent so popular?
Minibatch gradient descent is popular as:
 It is more efficient when compared to stochastic gradient descent.
 Generalization is done by finding the flat minima.
 It helps avoid the local minima by allowing the approximation of the gradient for the entire dataset.
What are deep autoencoders?
Deep autoencoders are an extension of the regular autoencoders. Here, the first layer is responsible for the firstorder function execution of the input. The second layer will take care of the secondorder functions, and it goes on.
Usually, a deep autoencoder is a combination of two or more symmetrical deepbelief networks where:
 The first five shallow layers consist of the encoding part
 The other layers take care of the decoding part
On the next set of Deep Learning questions, let us look further into the topic.
Why is the Leaky ReLU function used in Deep Learning?
Leaky ReLU, also called LReL, is used to manage a function to allow the passing of smallsized negative values if the input value to the network is less than zero.
What are some of the examples of supervised learning algorithms in Deep Learning?
There are three main supervised learning algorithms in Deep Learning:
 Artificial neural networks
 Convolutional neural networks
 Recurrent neural networks
What are some of the examples of unsupervised learning algorithms in Deep Learning?**
There are three main unsupervised learning algorithms in Deep Learning:
 Autoencoders
 Boltzmann machines
 Selforganizing maps
Next up, let us look at more neural network interview questions that will help you ace the interviews.
Can we initialize the weights of a network to start from zero?
Yes, it is possible to begin with zero initialization. However, it is not recommended to use because setting up the weights to zero initially will cause all of the neurons to produce the same output and the same gradients when performing backpropagation. This means that the network will not have the ability to learn at all due to the absence of asymmetry between each of the neurons.
What is the meaning of valid padding and same padding in CNN?
 Valid padding: It is used when there is no requirement for padding. The output matrix will have the dimensions (n – f + 1) X (n – f + 1) after convolution.
 Same padding: Here, padding elements are added all around the output matrix. It will have the same dimensions as the input matrix.
What are some of the applications of transfer learning in Deep Learning?
Transfer learning is a scenario where a large model is trained on a dataset with a large amount of data and this model is used on simpler datasets, thereby resulting in extremely efficient and accurate neural networks.
The popular examples of transfer learning are in the case of:
 BERT
 ResNet
 GPT2
 VGG16
How is the transformer architecture better than RNNs in Deep Learning?
With the use of sequential processing, programmers were up against:
 The usage of high processing power
 The difficulty of parallel execution
This caused the rise of the transformer architecture. Here, there is a mechanism called attention mechanism, which is used to map all of the dependencies between sentences, thereby making huge progress in the case of NLP models.
What are the steps involved in the working of an LSTM network?
There are three main steps involved in the working of an LSTM network:
 The network picks up the information that it has to remember and identifies what to forget.
 Cell state values are updated based on Step 1.
 The network calculates and analyzes which part of the current state should make it to the output.
What are the elements in TensorFlow that are programmable?
In TensorFlow, users can program three elements:
 Constants
 Variables
 Placeholders
What is the meaning of bagging and boosting in Deep Learning?
Bagging is the concept of splitting a dataset and randomly placing it into bags for training the model.
Boosting is the scenario where incorrect data points are used to force the model to produce the wrong output. This is used to retrain the model and increase accuracy.
What are generative adversarial networks (GANs)?
Generative adversarial networks are used to achieve generative modeling in Deep Learning. It is an unsupervised task that involves the discovery of patterns in the input data to generate the output.
The generator is used to generate new examples, while the discriminator is used to classify the examples generated by the generator.
Why are generative adversarial networks (GANs) so popular?
Generative adversarial networks are used for a variety of purposes. In the case of working with images, they have a high amount of traction and efficient working.
 Creation of art: GANs are used to create artistic images, sketches, and paintings.
 Image enhancement: They are used to greatly enhance the resolution of the input images.
 Image translation: They are also used to change certain aspects, such as day to night and summer to winter, in images easily.
What are Neural Networks?
Neural networks are a class of models that are built with layers. Commonly used types of neural networks include convolutional and recurrent neural networks. The vocabulary around neural networks architectures is described in the figure below: By noting \(i\) the \(i^{th}\) layer of the network and \(j\) the \(j^{th}\) hidden unit of the layer, we have:
\(\boxed{z_j^{[i]}={w_j^{[i]}}^Tx+b_j^{[i]}}\) where we note w, b, z the weight, bias and output respectively.
What is a neuron in Neural Networks?
The individual neuron. As a diagram, a neuron (or unit) with one input looks like:
The Linear Unit: \(y=wx+by=wx+b\) The input is \(x\) . Its connection to the neuron has a weight which is \(w\) . Whenever a value flows through a connection, you multiply the value by the connection’s weight. For the input \(x\), what reaches the neuron is \(wx\).
How the neuron learn new data?
A neural network “learns” by modifying its weights.
The \(b\) is a special kind of weight we call the bias.
The bias doesn’t have any input data associated with it; instead, we put a \(1\) in the diagram so that the value that reaches the neuron is just \(b\) (since \(1 \times b = b\)).
The bias enables the neuron to modify the output independently of its inputs.
The \(y\) is the value the neuron ultimately outputs. To get the output, the neuron sums up all the values it receives through its connections.
This neuron’s activation is \(y = w x + b\) or as a formula \(y=wx+by=wx+b\).
Give me one neuron in Keras?
The easiest way to create a model in Keras is through keras.Sequential
, which creates a neural network as a stack of layers. We can create models like those above using a dense layer (which we’ll learn more about in the next lesson).
We could define a linear model accepting three input features for example ('sugars'
, 'fiber'
, and 'protein'
) and producing a single output ('calories'
) like so:
from tensorflow import keras
from tensorflow.keras import layers
# Create a network with 1 linear unit
model = keras.Sequential([
layers.Dense(units=1, input_shape=[3])
])
What is Activation function?
Activation functions are used at the end of a hidden unit to introduce nonlinear complexities to the model. Here are the most common ones:
Sigmoid  Tanh  ReLU  Leaky ReLU 

\(g(z)=\displaystyle\frac{1}{1+e^{z}}\)  \(g(z)=\displaystyle\frac{e^{z}e^{z}}{e^{z}+e^{z}}\)  \(g(z)=\textrm{max}(0,z)\)  \(g(z)=\textrm{max}(\epsilon z,z)\) 
What is Crossentropy loss ?
In the context of neural networks, the crossentropy loss \(L(z,y)\) is commonly used and is defined as follows: \(\boxed{L(z,y)=\Big[y\log(z)+(1y)\log(1z)\Big]}\)
What is Learning rate?
The learning rate, often noted \(\alpha\) or sometimes \(\eta\), indicates at which pace the weights get updated. This can be fixed or adaptively changed. The current most popular method is called Adam, which is a method that adapts the learning rate.
What is Backpropagation?
Backpropagation is a method to update the weights in the neural network by taking into account the actual output and the desired output. The derivative with respect to weight w is computed using chain rule and is of the following form:
\[\boxed{\frac{\partial L(z,y)}{\partial w}=\frac{\partial L(z,y)}{\partial a}\times\frac{\partial a}{\partial z}\times\frac{\partial z}{\partial w}}\]As a result, the weight is updated as follows:
\[\boxed{w\longleftarrow w\alpha\frac{\partial L(z,y)}{\partial w}}\]What is Updating weights In a neural network?
In a neural network, weights are updated as follows:
 Step 1: Take a batch of training data.
 Step 2: Perform forward propagation to obtain the corresponding loss.
 Step 3: Backpropagate the loss to get the gradients.
 Step 4: Use the gradients to update the weights of the network.
What is Dropout?
Dropout is a technique meant to prevent overfitting the training data by dropping out units in a neural network. In practice, neurons are either dropped with probability p or kept with probability \(1p\).
What is Convolutional layer requirement?
By noting W the input volume size, F the size of the convolutional layer neurons, P the amount of zero padding, then the number of neurons N that fit in a given volume is such that:
\[\boxed{N=\frac{WF+2P}{S}+1}\]What is Batch normalization?
It is a step of hyperparameter \(\gamma, \beta\) that normalizes the batch \(x_i\). It is usually done after a fully connected/convolutional layer and before a nonlinearity layer and aims at allowing higher learning rates and reducing the strong dependence on initialization.
What is a Types of gates in Recurrent Neural Networks?
Here are the different types of gates that we encounter in a typical recurrent neural network:
Input gate  Forget gate  Gate  Output gate 

Write to cell or not?  Erase a cell or not?  How much to write to cell?  How much to reveal cell? 
What is LSTM?
A long shortterm memory (LSTM) network is a type of RNN model that avoids the vanishing gradient problem by adding ‘forget’ gates.
What is the goal of Reinforcement Learning ?
The goal of reinforcement learning is for an agent to learn how to evolve in an environment. Reinforcement learning involves an agent, a set of states \(S\), and a set \(A\) of actions per state. By performing an action \(a \in A\), the agent transitions from state to state. Executing an action in a specific state provides the agent with a reward (a numerical score). The goal of the agent is to maximize its total reward. It does this by adding the maximum reward attainable from future states to the reward for achieving its current state, effectively influencing the current action by the potential future reward. This potential reward is a weighted sum of the expected values of the rewards of all future steps starting from the current state
What is Markov decision processes?
A Markov decision process (MDP) is a 5tuple \((\mathcal{S},\mathcal{A},\{P_{sa}\},\gamma,R)\) where:

\(\mathcal{S}\)is the set of states

\(\mathcal{A}\)is the set of actions

\(\{P_{sa}\}\)are the state transition probabilities for s\in\mathcal{S}and a\in\mathcal{A}

\(\gamma\in[0,1]\) is the discount factor

\(R:\mathcal{S}\times\mathcal{A}\longrightarrow\mathbb{R}\)or \(R:\mathcal{S}\longrightarrow\mathbb{R}\) is the reward function that the algorithm wants to maximize
What is Policy in Reinforcement Learning?
A policy \(\pi\) is a function \(\pi:\mathcal{S}\longrightarrow\mathcal{A}\) that maps states to actions.
Remark: we say that we execute a given policy \(\pi\) if given a state s we take the action \(a=\pi(s)\)
What is Value function in a policy?
For a given policy \(\pi\) and a given state \(s\), we define the value function \(V^{\pi}\) as follows:
\[\boxed{V^\pi(s)=E\Big[R(s_0)+\gamma R(s_1)+\gamma^2 R(s_2)+...s_0=s,\pi\Big]}\]What is the Bellman equation?
The optimal Bellman equations characterizes the value function \(V^{\pi^*}\) of the optimal policy \(\pi\):
\[\boxed{V^{\pi^*}(s)=R(s)+\max_{a\in\mathcal{A}}\gamma\sum_{s'\in S}P_{sa}(s')V^{\pi^*}(s')}\]Remark: we note that the optimal policy \(\pi\) for a given state s is such that:
\[\boxed{\pi^*(s)=\underset{a\in\mathcal{A}}{\textrm{argmax}}\sum_{s'\in\mathcal{S}}P_{sa}(s')V^*(s')}\]Value iteration algorithm?
The value iteration algorithm is in two steps:
1) We initialize the value:
\[\boxed{V_0(s)=0}\]2) We iterate the value based on the values before:
\[\boxed{V_{i+1}(s)=R(s)+\max_{a\in\mathcal{A}}\left[\sum_{s'\in\mathcal{S}}\gamma P_{sa}(s')V_i(s')\right]}\]What is Qlearning?
Qlearning is a modelfree reinforcement learning algorithm to learn the value of an action in a particular state. It does not require a model of the environment (hence “modelfree”), and it can handle problems with stochastic transitions and rewards without requiring adaptations. Qlearning is a modelfree estimation of Q, which is done as follows:
\[\boxed{Q(s,a)\leftarrow Q(s,a)+\alpha\Big[R(s,a,s')+\gamma\max_{a'}Q(s',a')Q(s,a)\Big]}\]What is ARIMA ?
ARIMA is a method for forecasting or predicting future outcomes based on a historical time series. It is based on the statistical concept of serial correlation, where past data points influence future data points.
What are the differences between autoregressive and moving average models?
ARIMA combines autoregressive features with those of moving averages. An AR(1) autoregressive process, for instance, is one in which the current value is based on the immediately preceding value, while an AR(2) process is one in which the current value is based on the previous two values. A moving average is a calculation used to analyze data points by creating a series of averages of different subsets of the full data set in order to smooth out the influence of outliers. As a result of this combination of techniques, ARIMA models can take into account trends, cycles, seasonality, and other nonstatic types of data when making forecasts.
How does ARIMA forecasting work?
ARIMA forecasting is achieved by plugging in time series data for the variable of interest. Statistical software will identify the appropriate number of lags or amount of differencing to be applied to the data and check for stationarity. It will then output the results, which are often interpreted similarly to that of a multiple linear regression model.
What are the components of ARIMA?
 Autoregression(AR): refers to a model that shows a changing variable that regresses on its own lagged, or prior, values.
 Integrated (I): represents the differencing of raw observations to allow for the time series to become stationary (i.e., data values are replaced by the difference between the data values and the previous values).
 Moving average (MA): incorporates the dependency between an observation and a residual error from a moving average model applied to lagged observations.
Congratulations you have some of the concepts that any Data Scientist about Deep Learning should know. For sure there are a lot of more concepts that are missing here but I wanted only to summarize some of basic elements that are needed to begin to work in this area.
Leave a comment