Explore key concepts related to machine learning bias, fairness, and ethical considerations with this quiz. Deepen your understanding of responsible ML deployment, common pitfalls, and best practices to ensure equitable and ethical outcomes.
Which scenario best demonstrates selection bias in a machine learning dataset intended to predict housing prices?
Explanation: Selection bias occurs when the data sampled is not representative of the desired population, such as only including urban property listings while excluding rural areas. This skews the model's predictions and reduces fairness. Overfitting due to too many hidden layers is a training issue, not a data bias. Incorrect normalization affects feature scaling, not selection bias. Unbalanced numerical ranges might impact model performance but do not constitute selection bias.
Which approach is most effective in reducing algorithmic discrimination when training a machine learning model for loan approvals?
Explanation: Eliminating sensitive attributes like race or gender can help reduce direct discrimination and make the model fairer. Simply increasing model complexity does not address fairness concerns. Gathering larger datasets is beneficial in general but without attention to attribute distribution might perpetuate existing biases. Treating outliers separately addresses rare cases, not discrimination against protected groups.
Why is transparency important when deploying machine learning applications in healthcare decision-making?
Explanation: Transparency leads to better understanding and trust of ML model decisions by clinicians and patients, which is critical in healthcare. Improving model precision is unrelated to transparency. Hardware acceleration is a technical efficiency concern, not an ethical one. Input feature types are separate from the need for transparent explanations.
What is a likely outcome if a computer vision system for facial recognition is trained primarily on images of adults but rarely on those of children?
Explanation: If training data underrepresents children, the model may not generalize well and thus misidentify or inaccurately recognize children’s faces. Overfitting does not necessarily mean balanced error across all groups. Increased robustness would require diverse data, not a lack of it. Automatic normalization does not compensate for insufficient representation of specific groups.
Which fairness metric compares the rate of positive outcomes between different demographic groups in a binary classification problem?
Explanation: Demographic parity measures fairness by comparing the rates of favorable outcomes among groups, ensuring no group is unduly advantaged. Mean absolute error and root mean square error are prediction accuracy metrics, not fairness-related. The area under the ROC curve assesses overall model discrimination ability, not parity between groups.