Data Science MCQ Test: Data Science MCQs - Practice Questions
1. What is the difference between supervised and unsupervised learning?
2. Define the concept of ROC-AUC in the context of binary classification models and its significance.
3. What is the purpose of regularization in machine learning, and why is it important?
4. What is 'ANOVA' (Analysis of Variance) and when is it used in statistical analysis?
5. In data science, what does 'outlier detection' involve, and why is it important?
6. What are the key considerations when dealing with imbalanced datasets in machine learning?
7. In data science, what is 'cross-domain analysis,' and how does it contribute to understanding patterns?
8. What is the 'area under the curve' (AUC) in the context of receiver operating characteristic (ROC) analysis?
9. Explain the concept of 'logistic regression' and its applications in binary classification problems.
10. In data science, what is the purpose of feature scaling, and how does it impact machine learning models?
11. Discuss the concept of transfer learning and its applications in machine learning.
12. Explain the concept of 'bootstrapping' in statistics and its use in estimating sample distributions.
13. In regression analysis, what does the 'coefficient of determination (R-squared)' indicate?
14. Explain the concept of regularization in machine learning and its significance.
15. Explain the curse of dimensionality and its impact on machine learning algorithms.
16. Define A/B testing and explain its significance in the field of data science.
17. Examine the concept of overfitting in machine learning and its relationship with model complexity.
18. Explain the concept of 'bagging' in ensemble learning and provide an example of a bagging algorithm.
19. What is 'feature scaling' in machine learning, and why is it important for certain algorithms?
20. What is the significance of 'cross-entropy loss' in machine learning, especially in classification tasks?
21. What is the purpose of 'k-fold cross-validation' in machine learning, and how does it differ from simple cross-validation?
22. What is 'feature extraction' in machine learning, and how does it differ from feature selection?
23. What is 'hyperparameter tuning' in machine learning, and why is it crucial for model optimization?
24. What is the role of 'p-values' in hypothesis testing, and how are they interpreted?
25. What is 'L1 regularization' in machine learning, and how does it contribute to model sparsity?
26. Examine the differences between bagging and boosting algorithms in ensemble learning.
27. Explain the concept of cross-validation and its role in model evaluation.
28. What are the key considerations when selecting an appropriate evaluation metric for a machine learning problem?
29. Explain the concept of 'statistical power' in hypothesis testing and its importance in study design.
30. Examine the concept of feature importance in machine learning and its implications for model interpretability.
31. What is 'skewness' in probability distributions, and how does it impact data analysis?
32. Explain the purpose of 'A/B testing' in data science and its application in experimentation.
33. Explain the concept of 'one-sample t-test' and its application in hypothesis testing.
34. In statistical hypothesis testing, what is a 'Type II error,' and how does it impact decision-making?
35. Explain the concept of ensemble learning and provide an example of an ensemble method.
36. What is the 'central limit theorem' in statistics, and how does it impact hypothesis testing?
37. Explain the concept of 'word embeddings' in natural language processing (NLP) and its applications.
38. What is 'correlation' in statistics, and how does it differ from causation?
39. Define the concept of bias in machine learning models and discuss how it can impact decision-making.
40. Explain the concept of 'precision' and 'recall' in the context of classification metrics.