Explore foundational concepts of deep learning and neural networks, including basic components, activation functions, and how to build a simple model.
Which statement best describes deep learning in the field of artificial intelligence?
Explanation: Deep learning utilizes artificial neural networks that can learn and extract important features from raw data automatically, making it powerful for tasks like image or speech recognition. Manual feature selection (option B) is typically associated with traditional machine learning. Shallow models (option C) do not represent deep learning's key advantage. Option D is incorrect because deep learning has significantly advanced fields such as computer vision and language processing.
What is the primary role of hidden layers in a neural network?
Explanation: Hidden layers are responsible for the core computations in a neural network, transforming and learning patterns from the input data. Input layers (option B) simply receive data, while output layers (option C) provide results. Option D is not correct since hidden layers do not store labels; labels are part of dataset annotations.
Why are activation functions like ReLU and Sigmoid important in neural networks?
Explanation: Activation functions such as ReLU and Sigmoid provide non-linearity, enabling neural networks to solve complex, non-linear problems. Optimizers (option B) and normalization (option C) serve different purposes in model training. Option D is inaccurate since activation functions are used throughout the network, not just in the input layer.
What does training a neural network typically involve?
Explanation: Training a neural network means iteratively updating its weights to reduce prediction errors using training data. Just loading/normalizing data (option B) is preparation, not training. Fixing weights before training (option C) defeats the learning process. Option D describes evaluation, but does not capture the essence of training.
When building a basic neural network to classify handwritten digits, why is the output layer often set to have 10 units with a softmax activation?
Explanation: Setting 10 output units aligns with the 10 possible digit classes (0–9), and softmax provides a probability for each class enabling multiclass classification. A single output unit (option B) is insufficient for multiclass problems. Option C is incorrect; softmax is designed for multiclass outputs. Option D is not true; this design allows and does not hinder learning.