No-Free-Lunch Theorems for Transfer Learning
en-de
en-es
en-fr
en-sl
en
en-zh
0.25
0.5
0.75
1.25
1.5
1.75
2
I will present a formal framework for transfer learning and investigate under which conditions is it possible to provide performance guarantees for such scenarios. I will address two key issues: *1) Which notions of task-similarity suffice to provide meaningful error bounds on a target task, for a predictor trained on a (different) source task? *2) Can we do better than just train a hypothesis on the source task and analyze its performance on the target task? Can the use of unlabeled target samples reduce the target prediction error?