InterviewSolution
This section includes InterviewSolutions, each offering curated multiple-choice questions to sharpen your knowledge and support exam preparation. Choose a topic below to get started.
| 1. |
What do you mean by an epochs in the context of deep learning? |
|
Answer» An EPOCH is a TERMINOLOGY used in deep learning that refers to the number of passes the deep learning algorithm has made across the full training dataset. Batches are commonly used to GROUP DATA sets (especially when the amount of data is very large). The term "iteration" refers to the process of running one batch through the model. The number of epochs equals the number of iterations if the batch size is the entire training dataset. This is frequently not the case for practical reasons. Several epochs are used in the creation of many models. There is a general relation which is given by:- d * e = i * b where, d is the dataset size e is the number of epochs i is the number of iterations b is the batch size |
|
| 2. |
What is an activation function? What is the use of an activation function? |
|
Answer» An artificial neural network's activation function is a function that is introduced to help the network learn complex patterns in the data. When compared to a neuron-based model seen in our brains, the activation function is responsible for determining what is to be fired to the next neuron at the end of the process. In an ANN, an activation function performs the same job. It takes the preceding cell's output signal and turns it into a FORMAT that may be used as input to the next cell. Here, x0 and x1 are the inputs. W1 is the WEIGHT and a is the activation function. The activation function introduces non-linearity into the neural network, allowing it to learn more complex functions. The neural network would only be able to learn a function that is a linear combination of its input data if it didn't have the Activation function. The activation function converts inputs to outputs. The activation function is in charge of determining WHETHER or not a neuron should be stimulated. It arrives at a decision by calculating the weighted total and then adds bias. The activation function's main goal is to introduce non-linearity into a neuron's output. |
|
| 3. |
Which deep learning algorithm is the best for face detection? |
|
Answer» Face identification may be accomplished using a variety of machine learning METHODS, but the best ONES use Convolutional NEURAL Networks and deep learning. The following are some NOTABLE face detection algorithms: FaceNet, Probablisit, Face Embedding, ArcFace, Cosface, and Spherface. |
|
| 4. |
Explain Stochastic Gradient Descent. How is it different from Batch Gradient Descent? |
||||||||||||||||||
|
Answer» Stochastic Gradient Descent: Stochastic Gradient Descent seeks to tackle the major difficulty with Batch Gradient Descent, which is the use of the entire training set to CALCULATE gradients at each STEP. It is stochastic in nature, which means it chooses up a "random" instance of training data at each step and then computes the gradient, which is SIGNIFICANTLY faster than Batch Gradient Descent because there are much fewer data to modify at once. Stochastic Gradient Descent is best suited for unconstrained optimization problems. The stochastic nature of SGD has a drawback in that once it gets close to the MINIMUM value, it doesn't settle down and instead bounces around, giving us a GOOD but not optimal value for model parameters. This can be solved by lowering the learning rate at each step, which will reduce the bouncing and allow SGD to settle down at the global minimum after some time. Following are the differences between the two:-
|
|||||||||||||||||||
| 5. |
Explain Batch Gradient Descent. |
|
Answer» BATCH Gradient Descent: Batch Gradient Descent entails computation (involved in each step of gradient descent) over the entire TRAINING SET at each step and hence it is highly slow on very big training sets. As a result, Batch Gradient Descent becomes extremely computationally expensive. This is ideal for error manifolds that are convex or somewhat smooth. Batch Gradient Descent also scales nicely as the NUMBER of features grows. |
|
| 6. |
In a Convolutional Neural Network (CNN), how can you fix the constant validation accuracy? |
|
Answer» When training any neural network, constant validation accuracy is a common issue because the network just REMEMBERS the SAMPLE, resulting in an over-fitting problem. Over-fitting a model indicates that the neural network model performs ADMIRABLY on the training sample, but the model's PERFORMANCE deteriorates on the validation set. Following are some ways for improving CNN's constant validation accuracy:
|
|
| 7. |
Explain the difference between a shallow network and a deep network. |
|
Answer» A HIDDEN layer, as well as INPUT and output layers, are present in every neural network. Shallow neural networks are those that have only one hidden layer, whereas DEEP neural networks include numerous hidden layers. Both shallow and deep networks can fit into any function, however, shallow networks require a large number of input parameters, whereas deep networks, because of their several layers, can fit FUNCTIONS with a small number of input parameters. Deep networks are currently favored over shallow networks because the model learns a new and abstract representation of the input at each layer. In COMPARISON to shallow networks, they are also far more efficient in terms of the number of parameters and computations. |
|
| 8. |
What is a tensor in deep learning? |
|
Answer» A tensor is a multidimensional array that represents a generalization of vectors and matrices. It is one of the KEY data structures used in DEEP learning. Tensors are represented as n-dimensional arrays of base data types. The data TYPE of each element in the Tensor is the same, and the data type is always known. It's possible that only a portion of the shape (that is, the number of dimensions and the SIZE of each dimension) is known. Most operations yield fully-known tensors if their inputs are likewise fully known, however, in other circumstances, the shape of a tensor can only be determined at graph execution TIME. |
|
| 9. |
Is it possible to train a neural network model by setting all biases to 0? Also, is it possible to train a neural network model by setting all of the weights to 0? |
|
Answer» Yes, even if all of the biases are set to zero, the neural network model has a chance of learning. No, TRAINING a model by setting all of the weights to 0 is impossible SINCE the neural network will never learn to complete a task. When all weights are set to zero, the DERIVATIVES for each w remain constant, resulting in NEURONS learning the same features in each iteration. Any constant initialization of weights, not simply zero, is likely to generate a poor result. |
|
| 10. |
What are the advantages of transfer learning? |
|
Answer» Following are the advantages of transfer LEARNING :
|
|
| 11. |
Explain transfer learning in the context of deep learning. |
|
Answer» Transfer learning is a learning technique that allows DATA scientists to use what they've learned from a previous machine learning model that was used for a similar task. The ability of humans to transfer their knowledge is used as an example in this learning. You can LEARN to operate other two-wheeled vehicles more simply if you learn to ride a bicycle. A model trained for autonomous automobile driving can also be used for autonomous TRUCK driving. The features and weights can be used to train the new model, allowing it to be reused. When there is limited data, transfer learning works effectively for quickly training a model. In the above image, the FIRST diagram represents training a model from scratch while the second diagram represents using a model already trained on cats and dogs to classify the different class of vehicles, thereby REPRESENTING transfer learning. |
|
| 12. |
Difference between multi-class and multi-label classification problems. |
|
Answer» The classification task in a multi-class classification problem has more than two mutually exclusive classes (classes that have no intersection or no attributes in common), whereas in a multi-LABEL classification problem, each label has a different classification task, although the tasks are related in some way. For example, classifying a group of photographs of ANIMALS that could be cats, dogs, or bears is a multi-class classification problem that assumes each sample can be of only one TYPE, implying that an image can be categorized as either a cat or a DOG, but not both at the same time. Now let us assume you wish to manipulate the image below. The image above must be categorized as both a cat and a dog because it DEPICTS both creatures. A set of labels is allocated to each sample in a multi-label classification issue, and the classes are not mutually exclusive. In a multi-label classification problem, a pattern can belong to one or more classes. |
|
| 13. |
What do you mean by hyperparameters in the context of deep learning? |
|
Answer» Hyperparameters are variables that determine the network topology (for example, the number of hidden units) and how the network is trained (Eg: Learning RATE). They are SET before training the MODEL, that is, before optimizing the weights and the bias. Following are some of the examples of hyperparameters:-
|
|
| 14. |
What are the different techniques to achieve data normalization? |
|
Answer» Following are the different techniques employed to achieve data normalization:-
|
|
| 15. |
Explain Data Normalisation. What is the need for it? |
|
Answer» Data Normalisation is a technique in which data is transformed in such a WAY that they are EITHER dimensionless or have a SIMILAR distribution. It is also known as standardization and feature SCALING. It's a pre-processing procedure for the input data that removes redundant data from the dataset. Normalization provides each variable equal weights/importance, ensuring that no single variable biases model performance in its favour SIMPLY because it is larger. It vastly improves model precision by converting the values of numeric columns in a dataset to a similar scale without distorting the range of values. |
|
| 16. |
Explain Forward and Back Propagation in the context of deep learning. |
Answer»
|
|
| 17. |
What do you understand about gradient clipping in the context of deep learning? |
|
Answer» Gradient Clipping is a technique for dealing with the problem of exploding gradients (a situation in which huge error gradients build up over time, resulting in massive modifications to neural network model weights during training) that happens during backpropagation. The problem of exploding gradients occurs when the gradients get excessively big during training, causing the model to become unstable. If the gradient has CROSSED the anticipated range, the gradient VALUES are driven element-by-element to a specific minimum or MAXIMUM VALUE. Gradient clipping improves numerical STABILITY while training a neural network, but it has little effect on the performance of the model. |
|
| 18. |
What do you mean by end-to-end learning? |
|
Answer» It's a deep learning procedure in which a model is fed raw data and the entire data is trained at the same time to create the desired result with no INTERMEDIATE steps. It is a deep learning method in which all of the different steps are trained SIMULTANEOUSLY rather than sequentially. End-to-end learning has the advantage of eliminating the requirement for implicit feature engineering, which usually results in lower bias. Driverless automobiles are an excellent example that you MAY use in your end-to-end learning content. They are guided by human input and are programmed to learn and interpret information automatically using a CNN to FULFILL tasks. Another good example is the generation of a written TRANSCRIPT (output) from a recorded audio clip (input). The model here skips all of the steps in the middle, focusing instead on the fact that it can manage the entire sequence of steps and tasks. |
|
| 19. |
What are the different types of deep neural networks? |
|
Answer» FOLLOWING are the different types of deep neural networks:-
|
|
| 20. |
Explain what a deep neural network is. |
|
Answer» An ARTIFICIAL neural network (ANN) having numerous layers between the input and output layers is known as a deep neural network (DNN). Deep neural networks are neural networks that use deep architectures. The term "deep" refers to functions that have a HIGHER number of layers and units in a single layer. It is POSSIBLE to create more ACCURATE models by adding more and larger layers to capture higher levels of patterns. The below IMAGE depicts a deep neural network. |
|
| 21. |
What are the disadvantages of neural networks? |
|
Answer» Following are the disadvantages of NEURAL networks:-
|
|
| 22. |
What are the advantages of neural networks? |
|
Answer» Following are the advantages of neural networks:
|
|
| 23. |
Explain learning rate in the context of neural network models. What happens if the learning rate is too high or too low? |
|
Answer» Learning rate is a number that ranges from 0 to 1. It is one of the most important tunable hyperparameters in neural network training models. The learning rate determines how quickly or slowly a neural network model adapts to a given situation and learns. A higher learning rate value indicates that the model only needs a few training epochs and produces rapid changes, whereas a lower learning rate indicates that the model MAY take a long time to converge or may never converge and BECOME stuck on a poor solution. As a result, it is recommended that a GOOD learning rate value be established by TRIAL and error rather than using a learning rate that is too low or too high. In the above IMAGE, we can clearly see that a big learning rate leads us to move away from the desired output. However, having a small learning rate leads us to the desired output eventually. |
|
| 24. |
What are the applications of deep learning? |
|
Answer» Following are some of the applications of deep LEARNING:-
|
|
| 25. |
What do you understand about Neural Networks in the context of Deep Learning? |
|
Answer» Neural Networks are artificial systems that have a lot of resemblance to the biological neural networks in the human body. A neural network is a set of algorithms that attempts to recognize UNDERLYING relationships in a batch of data using a method that mimics how the human brain works. Without any task-specific rules, these systems learn to do tasks by being exposed to a variety of DATASETS and examples. The NOTION is that instead of being programmed with a pre-coded understanding of these datasets, the system derives identifying TRAITS from the data it is fed to. Neural networks are built on threshold logic computational models. Because neural networks can adapt to changing input, they can produce the best possible outcome without requiring the output CRITERIA to be redesigned. |
|