Based on the example shown in the lecture (Training R²: 0.95, Test R²: 0.61), what does this indicate?, The model is underfitting, The model is overfitting, The model has perfect generalization, The regularization parameter is too high, What does Elastic Net combine?, L1 and L2 penalties, Gradient descent and coordinate descent, Classification and regression, Linear and polynomial features, What are support vectors in an SVM?, All training samples, Points that lie on or within the margin boundaries, Only misclassified points, The centroid of each class, How does SGDClassifier relate to SVMs?, It cannot approximate SVM behavior, It's only for logistic regression, It's slower than standard SVM for all dataset sizes, It can approximate SVM with appropriate loss function (hinge loss), Which regularization technique performs automatic feature selection?, Ridge (L2), Lasso (L1), Neither, Both equally, What does the gradient descent algorithm minimize in linear regression?, The number of iterations, The number of features, The mean squared error loss function, The learning rate, As we increase the regularization parameter λ in Ridge regression, what happens to bias and variance?, Both bias and variance increase, Both bias and variance decrease, Bias increases, variance decreases, Bias decreases, variance increases, You're training with SGD. What's the tradeoff when choosing batch size?, Large batch → noisy gradients, faster; small batch → stable gradients, slower, Small batch → noisy gradients, slower; large batch → accurate gradients, faster, Large batch → better generalization; small batch → worse accuracy, Batch size only affects epochs, not training.
0%
Tryout
共享
共享
共享
由
Yildirimcerenel
编辑内容
打印
嵌入
更多
作业
排行榜
显示更多
显示更少
此排行榜当前是私人享有。单击
,共享
使其公开。
资源所有者已禁用此排行榜。
此排行榜被禁用,因为您的选择与资源所有者不同。
还原选项
随机轮盘
是一个开放式模板。它不会为排行榜生成分数。
需要登录
视觉风格
字体
需要订阅
选项
切换模板
显示所有
播放活动时将显示更多格式。
)
打开成绩
复制链接
QR 代码
删除
恢复自动保存:
?