[Experimentation]
You have access to training data but no access to test dat
a. What evaluation method can you use to assess the performance of your AI model?
When test data is unavailable, cross-validation is the most effective method to assess an AI model's performance using only the training dataset. Cross-validation involves splitting the training data into multiple subsets (folds), training the model on some folds, and validating it on others, repeating this process to estimate generalization performance. NVIDIA's documentation on machine learning workflows, particularly in the NeMo framework for model evaluation, highlights k-fold cross-validation as a standard technique for robust performance assessment when a separate test set is not available. Option B (randomized controlled trial) is a clinical or experimental method, not typically used for model evaluation. Option C (average entropy approximation) is not a standard evaluation method. Option D (greedy decoding) is a generation strategy for LLMs, not an evaluation technique.
NVIDIA NeMo Documentation: https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/stable/nlp/model_finetuning.html
Goodfellow, I., et al. (2016). 'Deep Learning.' MIT Press.
Eveline
2 days agoDelila
9 days agoShaun
4 days agoLeonor
12 days agoNovella
20 days agoLauran
1 months agoTwana
1 months agoLourdes
2 hours agoLilli
24 days agoKimbery
1 months agoBurma
8 days agoWinfred
15 days agoTasia
25 days agoLeonor
1 months agoNovella
1 months ago