Transfer Learning through Embedding Spaces

Author:

Language: English

56.31 €

In Print (Delivery period: 14 days).

Add to cartAdd to cart
Transfer Learning through Embedding Spaces
Publication date:
· 17.8x25.4 cm · Paperback

142.05 €

In Print (Delivery period: 14 days).

Add to cartAdd to cart
Transfer Learning through Embedding Spaces
Publication date:
· 17.8x25.4 cm · Hardback

Recent progress in artificial intelligence (AI) has revolutionized our everyday life. Many AI algorithms have reached human-level performance and AI agents are replacing humans in most professions. It is predicted that this trend will continue and 30% of work activities in 60% of current occupations will be automated.

This success, however, is conditioned on availability of huge annotated datasets to training AI models. Data annotation is a time-consuming and expensive task which still is being performed by human workers. Learning efficiently from less data is a next step for making AI more similar to natural intelligence. Transfer learning has been suggested a remedy to relax the need for data annotation. The core idea in transfer learning is to transfer knowledge across similar tasks and use similarities and previously learned knowledge to learn more efficiently.

In this book, we provide a brief background on transfer learning and then focus on the idea of transferring knowledge through intermediate embedding spaces. The idea is to couple and relate different learning through embedding spaces that encode task-level relations and similarities. We cover various machine learning scenarios and demonstrate that this idea can be used to overcome challenges of zero-shot learning, few-shot learning, domain adaptation, continual learning, lifelong learning, and collaborative learning.

Introduction. Background and Related Work. Zero-Shot Image Classification through Coupled Visual and Semantic Embedding Spaces. Learning a Discriminative Embedding for Unsupervised Domain Adaptation. Few-Shot Image Classification through Coupled Embedding Spaces. Cross-Task Knowledge Transfer. Lifelong Zero-Shot Learning Using High-Level Task Descriptors. Complementary Learning Systems Theory for Tackling Catastrophic Forgetting. Continual Concept Learning. Collective Lifelong Learning for Multi-Agent Networks. Concluding Remarks and Potential Future Research Directions.

Mohammad Rostami is a computer scientist at USC Information Sciences Institute. He is a graduate of the University of Pennsylvania, University of Waterloo, and Sharif University of Technology. His research area includes continual machine learning and learning in data scarce regimes.