Description
Transfer Learning through Embedding Spaces
Author: Rostami Mohammad
Language: EnglishSubjects for Transfer Learning through Embedding Spaces:
Keywords
Embedding Space; Labeled Data Points; Catastrophic Forgetting; Deep Neural Networks; Domain Adaptation; Sparse Vector; Deep Network; Consensus Tracking; RL Task; Multi-task Learning; Target Domain; Source Domain; Past Tasks; Complex Network Systems; Deep Nets; Lasso Problem; Single Learning Agent; STL; Machine Learning Problems; Adversarial Learning; Dictionary Learning; ERM; Multiple Lyapunov Function; Multi-class Classification; SAR Data
Publication date: 06-2023
· 17.8x25.4 cm · Paperback
Publication date: 06-2021
· 17.8x25.4 cm · Hardback
Description
/li>Contents
/li>Biography
/li>
Recent progress in artificial intelligence (AI) has revolutionized our everyday life. Many AI algorithms have reached human-level performance and AI agents are replacing humans in most professions. It is predicted that this trend will continue and 30% of work activities in 60% of current occupations will be automated.
This success, however, is conditioned on availability of huge annotated datasets to training AI models. Data annotation is a time-consuming and expensive task which still is being performed by human workers. Learning efficiently from less data is a next step for making AI more similar to natural intelligence. Transfer learning has been suggested a remedy to relax the need for data annotation. The core idea in transfer learning is to transfer knowledge across similar tasks and use similarities and previously learned knowledge to learn more efficiently.
In this book, we provide a brief background on transfer learning and then focus on the idea of transferring knowledge through intermediate embedding spaces. The idea is to couple and relate different learning through embedding spaces that encode task-level relations and similarities. We cover various machine learning scenarios and demonstrate that this idea can be used to overcome challenges of zero-shot learning, few-shot learning, domain adaptation, continual learning, lifelong learning, and collaborative learning.
Mohammad Rostami is a computer scientist at USC Information Sciences Institute. He is a graduate of the University of Pennsylvania, University of Waterloo, and Sharif University of Technology. His research area includes continual machine learning and learning in data scarce regimes.