Explore the essential concepts and workflow of Natural Language Processing, including data preparation, model evaluation, and key text classification steps. This quiz focuses on foundational NLP practices used in modern machine learning projects.
Which of these is a typical sequence of stages in a standard NLP pipeline for building a machine learning model?
Explanation: The commonly accepted sequence starts with acquiring data, followed by cleaning, pre-processing, feature engineering, modelling, evaluation, and finally deployment. The other options mix up these steps, which can disrupt the logical flow necessary for effective NLP projects.
What is the main purpose of feature engineering in Natural Language Processing tasks?
Explanation: Feature engineering involves transforming textual data into numerical vectors that algorithms can process. Visualizing performance and stopword removal are parts of evaluation and pre-processing, whereas storing raw text is not feature engineering.
What is the difference between intrinsic and extrinsic evaluation methods in NLP?
Explanation: Intrinsic evaluation measures the performance of a model using quantitative metrics, while extrinsic evaluation assesses how well the model accomplishes external or business tasks. The other options confuse evaluation with unrelated processes.
Which of the following is commonly performed during pre-processing of text for NLP projects?
Explanation: Pre-processing typically involves steps like stopword removal, lemmatization, lowercasing, and filtering unwanted characters. Training models and deployment are later stages, while feature importance plots relate to model interpretation.
Why is model updating important in an NLP workflow that classifies new documents over time?
Explanation: Regularly updating the model ensures it learns from new data and maintains its performance. Cleaning the dataset and feature engineering are separate tasks, and disabling monitoring would reduce effectiveness rather than improve the workflow.