Online Exam Quiz

questionQuestion  1 to 30

Which type of neural network architecture is used for anomaly detection?

  • Convolutional Neural Network (CNN)
  • Long Short-Term Memory Network (LSTM)
  • Autoencoder
  • Recurrent Neural Network (RNN)
Mark for review

Which of the following is a common application of recurrent neural networks (RNNs)?

  • Image classification
  • Sequence prediction
  • Object detection
  • Text generation
Mark for review

Which of the following is not a commonly used deep learning framework?

  • TensorFlow
  • Keras
  • PyTorch
  • Scikit-learn
Mark for review

In deep learning, what does the term "backpropagation" refer to?

  • The process of passing data through the network
  • The process of adjusting model parameters based on the error
  • The process of optimizing hyperparameters
  • The process of selecting the best architecture for a problem
Mark for review

Which technique is used to normalize the input to a neural network, leading to faster training and better generalization?

  • Dropout
  • Batch Normalization
  • L2 Regularization
  • Weight Initialization
Mark for review

Which activation function is typically used for the output layer in a binary classification problem?

  • ReLU
  • Tanh
  • Sigmoid
  • Leaky ReLU
Mark for review

Which type of neural network is used for sequence prediction tasks?

  • Convolutional Neural Network (CNN)
  • Long Short-Term Memory Network (LSTM)
  • Restricted Boltzmann Machine (RBM)
  • Autoencoder
Mark for review

Which of the following is a drawback of using deep neural networks?

  • Requires large amounts of labeled data
  • Prone to overfitting
  • Computationally expensive
  • All of the above
Mark for review

Which type of layer is used to increase the non-linearity of a neural network?

  • Fully Connected Layer
  • Convolutional Layer
  • Pooling Layer
  • Activation Layer
Mark for review

Which technique is used to adjust the learning rate during training based on the validation performance?

  • Learning rate decay
  • Gradient clipping
  • Dropout
  • Batch normalization
Mark for review

What is the purpose of the softmax function in deep learning?

  • Introduce non-linearity
  • Normalize the output probabilities
  • Prevent overfitting
  • Reduce computational complexity
Mark for review

Which of the following activation functions is generally not used in hidden layers due to vanishing gradient problems?

  • ReLU
  • Tanh
  • Sigmoid
  • Leaky ReLU
Mark for review

Which layer in a deep neural network is responsible for reducing the dimensionality of the input data?

  • Convolutional Layer
  • Activation Layer
  • Pooling Layer
  • Fully Connected Layer
Mark for review

Which of the following is a drawback of using unsupervised pre-training in deep learning?

  • Requires more labeled data
  • Increases computational cost
  • Leads to overfitting
  • None of the above
Mark for review

Which activation function is preferred in the output layer of a multi-class classification problem?

  • ReLU
  • Sigmoid
  • Tanh
  • Softmax
Mark for review

Which of the following techniques is used to reduce the computational cost of training deep neural networks?

  • Gradient Descent
  • Stochastic Gradient Descent
  • Mini-batch Gradient Descent
  • All of the above
Mark for review

Which of the following techniques is used to prevent overfitting in deep learning?

  • Data augmentation
  • Dropout
  • Early stopping
  • All of the above
Mark for review

What is the purpose of data augmentation in deep learning?

  • Increase model capacity
  • Reduce model complexity
  • Increase training data diversity
  • Decrease training time
Mark for review

Which technique is used to reduce the dimensionality of the input data in deep learning?

  • Feature scaling
  • Principal Component Analysis (PCA)
  • Dropout
  • Batch Normalization
Mark for review

Which technique is used to prevent the weights of a neural network from becoming too large during training?

  • L1 Regularization
  • L2 Regularization
  • Dropout
  • Batch Normalization
Mark for review

What is the purpose of the "momentum" parameter in optimization algorithms such as SGD and Adam?

  • To speed up convergence by incorporating past gradients
  • To control the learning rate during training
  • To add noise to the gradient updates
  • To adjust the size of the weight updates
Mark for review

Which of the following techniques is used to improve the convergence of stochastic gradient descent?

  • Learning rate scheduling
  • Momentum
  • Adam optimization
  • All of the above
Mark for review

Which layer is responsible for introducing non-linearity into the neural network?

  • Fully Connected Layer
  • Activation Layer
  • Pooling Layer
  • Convolutional Layer
Mark for review

Which of the following techniques is used for regularization in deep learning?

  • Dropout
  • Batch Normalization
  • L1 Regularization
  • All of the above
Mark for review

Which deep learning architecture is commonly used for image classification tasks?

  • Recurrent Neural Networks (RNNs)
  • Convolutional Neural Networks (CNNs)
  • Long Short-Term Memory Networks (LSTMs)
  • Autoencoders
Mark for review

In deep learning, what is the purpose of the "padding" parameter in convolutional neural networks?

  • To control the size of the filter
  • To specify the number of filters in each layer
  • To add zeros around the input data to maintain spatial dimensions
  • To control the stride of the filter
Mark for review

What is the purpose of dropout in deep learning?

  • Speed up training
  • Prevent overfitting
  • Increase model complexity
  • Reduce model size
Mark for review

In deep learning, what does the term "epoch" refer to?

  • A complete pass through the entire dataset during training
  • A single forward pass through the neural network
  • A single backward pass through the neural network
  • A measure of model complexity
Mark for review

What is the purpose of the learning rate scheduler in deep learning?

  • To adjust the learning rate during training based on the validation performance
  • To control the number of epochs in training
  • To initialize the weights of the neural network
  • To normalize the input data
Mark for review

Which type of layer is commonly used to connect different parts of a neural network, enabling gradients to flow during backpropagation?

  • Fully Connected Layer
  • Convolutional Layer
  • Recurrent Layer
  • Skip Connection
Mark for review

blinkerOnline

  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16
  • 17
  • 18
  • 19
  • 20
  • 21
  • 22
  • 23
  • 24
  • 25
  • 26
  • 27
  • 28
  • 29
  • 30
  • Answered
  • Not Answered
  • Marked for Review