Publication Type
Edited Volumes
Journals and Highly Selective Conferences
Preprints
Refereed Workshops and Symposia
Refereed Short Papers and Abstracts
Dissertation and Thesis
Technical Reports
Software
Edited Volumes
| AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
|---|---|---|---|
| Eric Eaton, Carla Gomes, & Brian Williams, editors |
Special Issue of AI Magazine on Computational Sustainability
Download:
[ Website ]
|
2014 | AAAI Press |
| Eric Eaton, Carla Gomes, & Brian Williams, editors. Special Issue of AI Magazine on Computational Sustainability, AI Magazine 35(2-3) AAAI Press, June 2014. | |||
@book{Eaton2014AIMagazine,
editor = {Eric Eaton and Carla Gomes and Brian Williams},
title = {Special Issue of AI Magazine on Computational Sustainability},
series = {AI Magazine},
volume = {35},
number = {2-3},
month = {June},
year = {2014},
publisher = {AAAI Press},
url = {http://www.aaai.org/ojs/index.php/aimagazine/issue/view/206}
}
| |||
| Eric Eaton (chair) |
Lifelong Machine Learning: Proceedings of the 2013 AAAI Spring Symposium
Download:
[ Website ]
|
2013 | AAAI Press |
| Eric Eaton (chair). Lifelong Machine Learning: Proceedings of the 2013 AAAI Spring Symposium, AAAI Technical Report SS-13-05, AAAI Press, May 2013. | |||
@book{Eaton2013AAAISSS,
author = {Eric Eaton (chair)},
title = {Lifelong Machine Learning: Proceedings of the 2013 AAAI Spring Symposium},
series = {AAAI Technical Report SS-13-05}
month = {May},
year = {2013},
publisher = {AAAI Press},
isbn = {ISBN 978-1-57735-602-8}
url = {http://www.aaai.org/Press/Reports/Symposia/Spring/ss-13-05.php}
}
| |||
Journals and Highly Selective Conferences
| AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
|---|---|---|---|
| Jorge Mendez & Eric Eaton |
Lifelong learning of compositional structures
Download:
[ PDF ]
|
2021 | International Conference on Learning Representations |
| Jorge Mendez & Eric Eaton. Lifelong learning of compositional structures. In International Conference on Learning Representations, 2021. | |||
@inproceedings{Mendez2021Lifelong,
title = {Lifelong learning of compositional structures},
author = {Jorge Mendez and Eric Eaton},
booktitle = {International Conference on Learning Representations},
year = {2021},
abstract = {
A hallmark of human intelligence is the ability to construct
self-contained chunks of knowledge and adequately reuse them
in novel combinations for solving different yet structurally
related problems. Learning such compositional structures has
been a significant challenge for artificial systems, due to
the combinatorial nature of the underlying search problem.
To date, research into compositional learning has largely
proceeded separately from work on lifelong or continual
learning. We integrate these two lines of work to present a
general-purpose framework for lifelong learning of
compositional structures that can be used for solving a
stream of related tasks. Our framework separates the learning
process into two broad stages: learning how to best combine
existing components in order to assimilate a novel problem,
and learning how to adapt the set of existing components to
accommodate the new problem. This separation explicitly
handles the trade-off between the stability required to
remember how to solve earlier tasks and the flexibility
required to solve new tasks, as we show empirically in an
extensive evaluation.
},
}
| |||
Abstract:A hallmark of human intelligence is the ability to construct self-contained chunks of knowledge and adequately reuse them in novel combinations for solving different yet structurally related problems. Learning such compositional structures has been a significant challenge for artificial systems, due to the combinatorial nature of the underlying search problem. To date, research into compositional learning has largely proceeded separately from work on lifelong or continual learning. We integrate these two lines of work to present a general-purpose framework for lifelong learning of compositional structures that can be used for solving a stream of related tasks. Our framework separates the learning process into two broad stages: learning how to best combine existing components in order to assimilate a novel problem, and learning how to adapt the set of existing components to accommodate the new problem. This separation explicitly handles the trade-off between the stability required to remember how to solve earlier tasks and the flexibility required to solve new tasks, as we show empirically in an extensive evaluation. | |||
| Efstathios D. Gennatas, Jerome H. Friedman, Lyle H. Ungar, Romain Pirracchio, Eric Eaton, Lara G. Reichmann, Yannet Interian, Jose Marcio Luna, Charles B. Simone, Andrew Auerbach, Elier Delgado, Mark J. van der Laan, Timothy D. Solberg, & Gilmer Valdes | Expert-augmented machine learning | 2020 | Proceedings of the National Academy of Sciences 117(9) |
| Efstathios D. Gennatas, Jerome H. Friedman, Lyle H. Ungar, Romain Pirracchio, Eric Eaton, Lara G. Reichmann, Yannet Interian, Jose Marcio Luna, Charles B. Simone, Andrew Auerbach, Elier Delgado, Mark J. van der Laan, Timothy D. Solberg, & Gilmer Valdes. Expert-augmented machine learning. Proceedings of the National Academy of Sciences, 117(9):4571–4577, National Academy of Sciences, 2020. | |||
@article {Gennatas2020ExpertAugmented,
author = {Gennatas, Efstathios D. and Friedman, Jerome H. and Ungar, Lyle H. and Pirracchio, Romain and Eaton, Eric and Reichmann, Lara G. and Interian, Yannet and Luna, Jose Marcio and Simone, Charles B. and Auerbach, Andrew and Delgado, Elier and van der Laan, Mark J. and Solberg, Timothy D. and Valdes, Gilmer},
title = {Expert-augmented machine learning},
volume = {117},
number = {9},
pages = {4571--4577},
year = {2020},
publisher = {National Academy of Sciences},
abstract = {
Machine learning is increasingly used across fields to derive insights from data,
which further our understanding of the world and help us anticipate the future. The
performance of predictive modeling is dependent on the amount and quality of available
data. In practice, we rely on human experts to perform certain tasks and on machine
learning for others. However, the optimal learning strategy may involve combining the
complementary strengths of humans and machines. We present expert-augmented machine
learning, an automated way to automatically extract problem-specific human expert
knowledge and integrate it with machine learning to build robust, dependable, and
data-efficient predictive models.Machine learning is proving invaluable across
disciplines. However, its success is often limited by the quality and quantity of
available data, while its adoption is limited by the level of trust afforded by given
models. Human vs. machine performance is commonly compared empirically to decide
whether a certain task should be performed by a computer or an expert. In reality, the
optimal learning strategy may involve combining the complementary strengths of humans
and machines. Here, we present expert-augmented machine learning (EAML), an automated
method that guides the extraction of expert knowledge and its integration into
machine-learned models. We used a large dataset of intensive-care patient data to
derive 126 decision rules that predict hospital mortality. Using an online platform,
we asked 15 clinicians to assess the relative risk of the subpopulation defined by
each rule compared to the total sample. We compared the clinician-assessed risk to the
empirical risk and found that, while clinicians agreed with the data in most cases,
there were notable exceptions where they overestimated or underestimated the true
risk. Studying the rules with greatest disagreement, we identified problems with the
training data, including one miscoded variable and one hidden confounder. Filtering
the rules based on the extent of disagreement between clinician-assessed risk and
empirical risk, we improved performance on out-of-sample data and were able to train
with less data. EAML provides a platform for automated creation of problem-specific
priors, which help build robust and dependable machine-learning models in critical
applications.
},
journal = {Proceedings of the National Academy of Sciences},
}
| |||
Abstract:Machine learning is increasingly used across fields to derive insights from data, which further our understanding of the world and help us anticipate the future. The performance of predictive modeling is dependent on the amount and quality of available data. In practice, we rely on human experts to perform certain tasks and on machine learning for others. However, the optimal learning strategy may involve combining the complementary strengths of humans and machines. We present expert-augmented machine learning, an automated way to automatically extract problem-specific human expert knowledge and integrate it with machine learning to build robust, dependable, and data-efficient predictive models.Machine learning is proving invaluable across disciplines. However, its success is often limited by the quality and quantity of available data, while its adoption is limited by the level of trust afforded by given models. Human vs. machine performance is commonly compared empirically to decide whether a certain task should be performed by a computer or an expert. In reality, the optimal learning strategy may involve combining the complementary strengths of humans and machines. Here, we present expert-augmented machine learning (EAML), an automated method that guides the extraction of expert knowledge and its integration into machine-learned models. We used a large dataset of intensive-care patient data to derive 126 decision rules that predict hospital mortality. Using an online platform, we asked 15 clinicians to assess the relative risk of the subpopulation defined by each rule compared to the total sample. We compared the clinician-assessed risk to the empirical risk and found that, while clinicians agreed with the data in most cases, there were notable exceptions where they overestimated or underestimated the true risk. Studying the rules with greatest disagreement, we identified problems with the training data, including one miscoded variable and one hidden confounder. Filtering the rules based on the extent of disagreement between clinician-assessed risk and empirical risk, we improved performance on out-of-sample data and were able to train with less data. EAML provides a platform for automated creation of problem-specific priors, which help build robust and dependable machine-learning models in critical applications. | |||
| Jorge Mendez, Boyu Wang, & Eric Eaton |
Lifelong policy gradient learning of factored policies for faster training without forgetting Earlier version was awarded best paper at the ICML'20 Workshop on Lifelong Learning
Download:
[ PDF ]
|
2020 | Advances in Neural Information Processing Systems |
| Jorge Mendez, Boyu Wang, & Eric Eaton. Lifelong policy gradient learning of factored policies for faster training without forgetting. In Advances in Neural Information Processing Systems, 2020. | |||
@inproceedings{Mendez2020LifelongPG,
title = {Lifelong policy gradient learning of factored policies for faster training without forgetting},
author = {Jorge Mendez and Boyu Wang and Eric Eaton},
booktitle = {Advances in Neural Information Processing Systems},
year = {2020},
abstract = {
Policy gradient methods have shown success in learning control
policies for highdimensional dynamical systems. Their biggest
downside is the amount of exploration they require before
yielding high-performing policies. In a lifelong learning
setting, in which an agent is faced with multiple consecutive
tasks over its lifetime, reusing information from previously
seen tasks can substantially accelerate the learning of new
tasks. We provide a novel method for lifelong policy gradient
learning that trains lifelong function approximators directly
via policy gradients, allowing the agent to benefit from
accumulated knowledge throughout the entire training process.
We show empirically that our algorithm learns faster and
converges to better policies than single-task and lifelong
learning baselines, and completely avoids catastrophic forgetting
on a variety of challenging domains.
},
}
| |||
Abstract:Policy gradient methods have shown success in learning control policies for highdimensional dynamical systems. Their biggest downside is the amount of exploration they require before yielding high-performing policies. In a lifelong learning setting, in which an agent is faced with multiple consecutive tasks over its lifetime, reusing information from previously seen tasks can substantially accelerate the learning of new tasks. We provide a novel method for lifelong policy gradient learning that trains lifelong function approximators directly via policy gradients, allowing the agent to benefit from accumulated knowledge throughout the entire training process. We show empirically that our algorithm learns faster and converges to better policies than single-task and lifelong learning baselines, and completely avoids catastrophic forgetting on a variety of challenging domains. | |||
| Mohammad Rostami, David Isele, & Eric Eaton |
Using task descriptions in lifelong machine learning for improved performance and zero-shot transfer
Download:
[ PDF ] [ Website ]
|
2020 | Journal of Artificial Intelligence Research |
| Mohammad Rostami, David Isele, & Eric Eaton. Using task descriptions in lifelong machine learning for improved performance and zero-shot transfer. Journal of Artificial Intelligence Research, 67:673–704, 2020. | |||
@article{Rostami2020Using,
title = {Using task descriptions in lifelong machine learning for improved performance and zero-shot transfer},
author = {Mohammad Rostami and David Isele and Eric Eaton},
journal = {Journal of Artificial Intelligence Research},
volume = {67},
pages = {673--704},
year = {2020},
abstract = {
Knowledge transfer between tasks can improve the performance of learned models, but
requires an accurate estimate of inter-task relationships to identify the relevant
knowledge to transfer. These inter-task relationships are typically estimated based
on training data for each task, which is inefficient in lifelong learning settings
where the goal is to learn each consecutive task rapidly from as little data as
possible. To reduce this burden, we develop a lifelong learning method based on
coupled dictionary learning that utilizes high-level task descriptions to model
inter-task relationships. We show that using task descriptors improves the performance
of the learned task policies, providing both theoretical justification for the benefit
and empirical demonstration of the improvement across a variety of learning problems.
Given only the descriptor for a new task, the lifelong learner is also able to
accurately predict a model for the new task through zero-shot learning using the
coupled dictionary, eliminating the need to gather training data before addressing the
task.
},
}
| |||
Abstract:Knowledge transfer between tasks can improve the performance of learned models, but requires an accurate estimate of inter-task relationships to identify the relevant knowledge to transfer. These inter-task relationships are typically estimated based on training data for each task, which is inefficient in lifelong learning settings where the goal is to learn each consecutive task rapidly from as little data as possible. To reduce this burden, we develop a lifelong learning method based on coupled dictionary learning that utilizes high-level task descriptions to model inter-task relationships. We show that using task descriptors improves the performance of the learned task policies, providing both theoretical justification for the benefit and empirical demonstration of the improvement across a variety of learning problems. Given only the descriptor for a new task, the lifelong learner is also able to accurately predict a model for the new task through zero-shot learning using the coupled dictionary, eliminating the need to gather training data before addressing the task. | |||
| Pengyuan Shen, William Braham, Yun Kyu Yi, & Eric Eaton |
Rapid multi-objective optimization with multi-year future weather condition and decision-making support for building retrofit
Download:
[ PDF ] [ Website ]
|
2019 | Energy |
| Pengyuan Shen, William Braham, Yun Kyu Yi, & Eric Eaton. Rapid multi-objective optimization with multi-year future weather condition and decision-making support for building retrofit. Energy, 172:892–912, Elsevier Limited, April 2019. | |||
@article{Shen2019Rapid,
title = "Rapid multi-objective optimization with multi-year future weather condition and decision-making support for building retrofit",
keywords = "Building retrofit, Climate change, Heuristic method, Hierarchical clustering, Optimization, Pareto fronts",
author = "Pengyuan Shen and William Braham and Yi, {Yun Kyu} and Eric Eaton",
year = "2019",
month = {April},
volume = "172",
pages = "892--912",
journal = "Energy",
publisher = "Elsevier Limited",
abstract = {
A method of fast multi-objective optimization and decision-making support for building
retrofit planning is developed, and lifecycle cost analysis method taking into account
of future climate condition is used in evaluating the retrofit performance. In order to
resolve the optimization problem in a fast manner with recourse to non-dominate sorting
differential evolution algorithm, the simplified hourly dynamic simulation modeling
tool SimBldPy is used as the simulator for objective function evaluation. Moreover, the
generated non-dominated solutions are treated and rendered by a layered scheme using
agglomerative hierarchical clustering technique to make it more intuitive and sense
making during the decision-making process as well as to be better presented. The
suggested optimization method is implemented to the retrofit planning of a campus
building in UPenn with various energy conservation measures (ECM) and costs, and more
than one thousand Pareto fronts are obtained and being analyzed according to the
proposed decision-making framework. Twenty ECM combinations are eventually selected
from all generated Pareto fronts. It is manifested that the developed decision-making
support scheme shows robustness in dealing with retrofit optimization problem and is
able to provide support for brainstorming and enumerating various possibilities during
the decision-making process.},
}
| |||
Abstract:A method of fast multi-objective optimization and decision-making support for building retrofit planning is developed, and lifecycle cost analysis method taking into account of future climate condition is used in evaluating the retrofit performance. In order to resolve the optimization problem in a fast manner with recourse to non-dominate sorting differential evolution algorithm, the simplified hourly dynamic simulation modeling tool SimBldPy is used as the simulator for objective function evaluation. Moreover, the generated non-dominated solutions are treated and rendered by a layered scheme using agglomerative hierarchical clustering technique to make it more intuitive and sense making during the decision-making process as well as to be better presented. The suggested optimization method is implemented to the retrofit planning of a campus building in UPenn with various energy conservation measures (ECM) and costs, and more than one thousand Pareto fronts are obtained and being analyzed according to the proposed decision-making framework. Twenty ECM combinations are eventually selected from all generated Pareto fronts. It is manifested that the developed decision-making support scheme shows robustness in dealing with retrofit optimization problem and is able to provide support for brainstorming and enumerating various possibilities during the decision-making process. | |||
| Seungwon Lee, James Stokes, & Eric Eaton |
Learning shared knowledge for deep lifelong learning using deconvolutional networks
Download:
[ PDF ] [ Website ]
|
2019 | Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) |
| Seungwon Lee, James Stokes, & Eric Eaton. Learning shared knowledge for deep lifelong learning using deconvolutional networks. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19), pp. 2837–2844, 7 2019. | |||
@inproceedings{Lee2019Learning,
title = {Learning shared knowledge for deep lifelong learning using deconvolutional networks},
author = {Lee, Seungwon and Stokes, James and Eaton, Eric},
booktitle = {Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19)},
pages = {2837--2844},
year = {2019},
month = {7},
abstract = {
Current mechanisms for knowledge transfer in deep networks tend to either share the
lower layers between tasks, or build upon representations trained on other tasks.
However, existing work in non-deep multi-task and lifelong learning has shown success
with using factorized representations of the model parameter space for transfer,
permitting more flexible construction of task models. Inspired by this idea, we
introduce a novel architecture for sharing latent factorized representations in
convolutional neural networks (CNNs). The proposed approach, called a deconvolutional
factorized CNN, uses a combination of deconvolutional factorization and tensor
contraction to perform flexible transfer between tasks. Experiments on two computer
vision data sets show that the DF-CNN achieves superior performance in challenging
lifelong learning settings, resists catastrophic forgetting, and exhibits reverse
transfer to improve previously learned tasks from subsequent experience without
retraining.
},
}
| |||
Abstract:Current mechanisms for knowledge transfer in deep networks tend to either share the lower layers between tasks, or build upon representations trained on other tasks. However, existing work in non-deep multi-task and lifelong learning has shown success with using factorized representations of the model parameter space for transfer, permitting more flexible construction of task models. Inspired by this idea, we introduce a novel architecture for sharing latent factorized representations in convolutional neural networks (CNNs). The proposed approach, called a deconvolutional factorized CNN, uses a combination of deconvolutional factorization and tensor contraction to perform flexible transfer between tasks. Experiments on two computer vision data sets show that the DF-CNN achieves superior performance in challenging lifelong learning settings, resists catastrophic forgetting, and exhibits reverse transfer to improve previously learned tasks from subsequent experience without retraining. | |||
| Jose Marcio Luna, Efstathios D. Gennatas, Lyle H. Ungar, Eric Eaton, Eric S. Diffenderfer, Shane T. Jensen, Charles B. Simone, Jerome H. Friedman, Timothy D. Solberg, & Gilmer Valdes | Building more accurate decision trees with the additive tree | 2019 | Proceedings of the National Academy of Sciences 116(40) |
| Jose Marcio Luna, Efstathios D. Gennatas, Lyle H. Ungar, Eric Eaton, Eric S. Diffenderfer, Shane T. Jensen, Charles B. Simone, Jerome H. Friedman, Timothy D. Solberg, & Gilmer Valdes. Building more accurate decision trees with the additive tree. Proceedings of the National Academy of Sciences, 116(40):19887–19893, National Academy of Sciences, 2019. | |||
@article{Luna2019Building,
author = {Luna, Jose Marcio and Gennatas, Efstathios D. and Ungar, Lyle H. and Eaton, Eric and Diffenderfer, Eric S. and Jensen, Shane T. and Simone, Charles B. and Friedman, Jerome H. and Solberg, Timothy D. and Valdes, Gilmer},
title = {Building more accurate decision trees with the additive tree},
volume = {116},
number = {40},
pages = {19887--19893},
year = {2019},
publisher = {National Academy of Sciences},
journal = {Proceedings of the National Academy of Sciences},
abstract = {
As machine learning applications expand to high-stakes areas such as criminal justice,
finance, and medicine, legitimate concerns emerge about high-impact effects of
individual mispredictions on people{\textquoteright}s lives. As a result, there has
been increasing interest in understanding general machine learning models to overcome
possible serious risks. Current decision trees, such as Classification and Regression
Trees (CART), have played a predominant role in fields such as medicine, due to their
simplicity and intuitive interpretation. However, such trees suffer from intrinsic
limitations in predictive power. We developed the additive tree, a theoretical
approach to generate a more accurate and interpretable decision tree, which reveals
connections between CART and gradient boosting. The additive tree exhibits superior
predictive performance to CART, as validated on 83 classification tasks.The expansion
of machine learning to high-stakes application domains such as medicine, finance, and
criminal justice, where making informed decisions requires clear understanding of the
model, has increased the interest in interpretable machine learning. The widely used
Classification and Regression Trees (CART) have played a major role in health
sciences, due to their simple and intuitive explanation of predictions. Ensemble
methods like gradient boosting can improve the accuracy of decision trees, but at the
expense of the interpretability of the generated model. Additive models, such as those
produced by gradient boosting, and full interaction models, such as CART, have been
investigated largely in isolation. We show that these models exist along a spectrum,
revealing previously unseen connections between these approaches. This paper
introduces a rigorous formalization for the additive tree, an empirically validated
learning technique for creating a single decision tree, and shows that this method can
produce models equivalent to CART or gradient boosted stumps at the extremes by
varying a single parameter. Although the additive tree is designed primarily to
provide both the model interpretability and predictive performance needed for
high-stakes applications like medicine, it also can produce decision trees represented
by hybrid models between CART and boosted stumps that can outperform either of these
approaches.},
}
| |||
Abstract:As machine learning applications expand to high-stakes areas such as criminal justice, finance, and medicine, legitimate concerns emerge about high-impact effects of individual mispredictions on people\textquoterights lives. As a result, there has been increasing interest in understanding general machine learning models to overcome possible serious risks. Current decision trees, such as Classification and Regression Trees (CART), have played a predominant role in fields such as medicine, due to their simplicity and intuitive interpretation. However, such trees suffer from intrinsic limitations in predictive power. We developed the additive tree, a theoretical approach to generate a more accurate and interpretable decision tree, which reveals connections between CART and gradient boosting. The additive tree exhibits superior predictive performance to CART, as validated on 83 classification tasks.The expansion of machine learning to high-stakes application domains such as medicine, finance, and criminal justice, where making informed decisions requires clear understanding of the model, has increased the interest in interpretable machine learning. The widely used Classification and Regression Trees (CART) have played a major role in health sciences, due to their simple and intuitive explanation of predictions. Ensemble methods like gradient boosting can improve the accuracy of decision trees, but at the expense of the interpretability of the generated model. Additive models, such as those produced by gradient boosting, and full interaction models, such as CART, have been investigated largely in isolation. We show that these models exist along a spectrum, revealing previously unseen connections between these approaches. This paper introduces a rigorous formalization for the additive tree, an empirically validated learning technique for creating a single decision tree, and shows that this method can produce models equivalent to CART or gradient boosted stumps at the extremes by varying a single parameter. Although the additive tree is designed primarily to provide both the model interpretability and predictive performance needed for high-stakes applications like medicine, it also can produce decision trees represented by hybrid models between CART and boosted stumps that can outperform either of these approaches. | |||
| Julia E. Reid & Eric Eaton |
Artificial intelligence for pediatric ophthalmology
Download:
[ PDF ] [ Website ]
|
2019 | Current Opinion in Ophthalmology 30(5) |
| Julia E. Reid & Eric Eaton. Artificial intelligence for pediatric ophthalmology. Current Opinion in Ophthalmology, 30(5):337–346, 2019. | |||
@article{Reid2019Artificial,
title = {Artificial intelligence for pediatric ophthalmology},
author = {Reid, Julia E. and Eaton, Eric},
journal = {Current Opinion in Ophthalmology},
volume = {30},
number = {5},
pages = {337-346},
year = {2019},
abstract = {
PURPOSE OF REVIEW
Despite the impressive results of recent artificial intelligence applications to
general ophthalmology, comparatively less progress has been made toward solving
problems in pediatric ophthalmology using similar techniques. This article discusses
the unique needs of pediatric patients and how artificial intelligence techniques can
address these challenges, surveys recent applications to pediatric ophthalmology, and
discusses future directions.
RECENT FINDINGS
The most significant advances involve the automated detection of retinopathy of
prematurity, yielding results that rival experts. Machine learning has also been
applied to the classification of pediatric cataracts, prediction of postoperative
complications following cataract surgery, detection of strabismus and refractive
error, prediction of future high myopia, and diagnosis of reading disability. In
addition, machine learning techniques have been used for the study of visual
development, vessel segmentation in pediatric fundus images, and ophthalmic image
synthesis.
SUMMARY
Artificial intelligence applications could significantly benefit clinical care by
optimizing disease detection and grading, broadening access to care, furthering
scientific discovery, and improving clinical efficiency. These methods need to match
or surpass physician performance in clinical trials before deployment with patients.
Owing to the widespread use of closed-access data sets and software implementations,
it is difficult to directly compare the performance of these approaches, and
reproducibility is poor. Open-access data sets and software could alleviate these
issues and encourage further applications to pediatric ophthalmology.
},
}
| |||
Abstract:PURPOSE OF REVIEW Despite the impressive results of recent artificial intelligence applications to general ophthalmology, comparatively less progress has been made toward solving problems in pediatric ophthalmology using similar techniques. This article discusses the unique needs of pediatric patients and how artificial intelligence techniques can address these challenges, surveys recent applications to pediatric ophthalmology, and discusses future directions. RECENT FINDINGS The most significant advances involve the automated detection of retinopathy of prematurity, yielding results that rival experts. Machine learning has also been applied to the classification of pediatric cataracts, prediction of postoperative complications following cataract surgery, detection of strabismus and refractive error, prediction of future high myopia, and diagnosis of reading disability. In addition, machine learning techniques have been used for the study of visual development, vessel segmentation in pediatric fundus images, and ophthalmic image synthesis. SUMMARY Artificial intelligence applications could significantly benefit clinical care by optimizing disease detection and grading, broadening access to care, furthering scientific discovery, and improving clinical efficiency. These methods need to match or surpass physician performance in clinical trials before deployment with patients. Owing to the widespread use of closed-access data sets and software implementations, it is difficult to directly compare the performance of these approaches, and reproducibility is poor. Open-access data sets and software could alleviate these issues and encourage further applications to pediatric ophthalmology. | |||
| Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim |
Deep transfer learning for few-shot SAR image classification
Download:
[ PDF ] [ Website ]
|
2019 | Remote Sensing |
| Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim. Deep transfer learning for few-shot SAR image classification. Remote Sensing, 11:1374, 2019. | |||
@article{Rostami2019Deep,
title = {Deep transfer learning for few-shot {SAR} image classification},
author = {Mohammad Rostami and Soheil Kolouri and Eric Eaton and Kyungnam Kim},
journal = {Remote Sensing},
volume = {11},
pages = {1374},
year = {2019},
abstract = {
The emergence of Deep Neural Networks (DNNs) has lead to high-performance
supervised learning algorithms for the Electro-Optical (EO) domain classification and detection
problems. This success is because generating huge labeled datasets has become possible using
modern crowdsourcing labeling platforms such as Amazon’s Mechanical Turk that recruit ordinary
people to label data. Unlike the EO domain, labeling the Synthetic Aperture Radar (SAR) domain data
can be much more challenging, and for various reasons, using crowdsourcing platforms is not feasible
for labeling the SAR domain data. As a result, training deep networks using supervised learning is
more challenging in the SAR domain. In the paper, we present a new framework to train a deep neural
network for classifying Synthetic Aperture Radar (SAR) images by eliminating the need for a huge
labeled dataset. Our idea is based on transferring knowledge from a related EO domain problem,
where labeled data are easy to obtain. We transfer knowledge from the EO domain through learning
a shared invariant cross-domain embedding space that is also discriminative for classification. To this
end, we train two deep encoders that are coupled through their last year to map data points from
the EO and the SAR domains to the shared embedding space such that the distance between the
distributions of the two domains is minimized in the latent embedding space. We use the Sliced
Wasserstein Distance (SWD) to measure and minimize the distance between these two distributions
and use a limited number of SAR label data points to match the distributions class-conditionally. As a
result of this training procedure, a classifier trained from the embedding space to the label space using
mostly the EO data would generalize well on the SAR domain. We provide a theoretical analysis
to demonstrate why our approach is effective and validate our algorithm on the problem of ship
classification in the SAR domain by comparing against several other competing learning approaches.}
}
| |||
Abstract:The emergence of Deep Neural Networks (DNNs) has lead to high-performance supervised learning algorithms for the Electro-Optical (EO) domain classification and detection problems. This success is because generating huge labeled datasets has become possible using modern crowdsourcing labeling platforms such as Amazon’s Mechanical Turk that recruit ordinary people to label data. Unlike the EO domain, labeling the Synthetic Aperture Radar (SAR) domain data can be much more challenging, and for various reasons, using crowdsourcing platforms is not feasible for labeling the SAR domain data. As a result, training deep networks using supervised learning is more challenging in the SAR domain. In the paper, we present a new framework to train a deep neural network for classifying Synthetic Aperture Radar (SAR) images by eliminating the need for a huge labeled dataset. Our idea is based on transferring knowledge from a related EO domain problem, where labeled data are easy to obtain. We transfer knowledge from the EO domain through learning a shared invariant cross-domain embedding space that is also discriminative for classification. To this end, we train two deep encoders that are coupled through their last year to map data points from the EO and the SAR domains to the shared embedding space such that the distance between the distributions of the two domains is minimized in the latent embedding space. We use the Sliced Wasserstein Distance (SWD) to measure and minimize the distance between these two distributions and use a limited number of SAR label data points to match the distributions class-conditionally. As a result of this training procedure, a classifier trained from the embedding space to the label space using mostly the EO data would generalize well on the SAR domain. We provide a theoretical analysis to demonstrate why our approach is effective and validate our algorithm on the problem of ship classification in the SAR domain by comparing against several other competing learning approaches. | |||
| Boyu Wang, Jorge Mendez, Mingbo Cai, & Eric Eaton | Transfer learning via minimizing the performance gap between domains | 2019 | Advances in Neural Information Processing Systems |
| Boyu Wang, Jorge Mendez, Mingbo Cai, & Eric Eaton. Transfer learning via minimizing the performance gap between domains. In Advances in Neural Information Processing Systems, pp. 10645–10655, 2019. | |||
@incollection{Wang2019Transfer,
title = {Transfer learning via minimizing the performance gap between domains},
author = {Wang, Boyu and Mendez, Jorge and Cai, Mingbo and Eaton, Eric},
booktitle = {Advances in Neural Information Processing Systems},
volume = {32},
pages = {10645--10655},
year = {2019},
abstract = {
We propose a new principle for transfer learning, based on a straightforward intuition:
if two domains are similar to each other, the model trained on one domain should also
perform well on the other domain, and vice versa. To formalize this intuition, we
define the performance gap as a measure of the discrepancy between the source and
target domains. We derive generalization bounds for the instance weighting approach to
transfer learning, showing that the performance gap can be viewed as an
algorithm-dependent regularizer, which controls the model complexity. Our theoretical
analysis provides new insight into transfer learning and motivates a set of general,
principled rules for designing new instance weighting schemes for transfer learning.
These rules lead to gapBoost, a novel and principled boosting approach for transfer
learning. Our experimental evaluation on benchmark data sets shows that gapBoost
significantly outperforms previous boosting-based transfer learning algorithms.
}
}
| |||
Abstract:We propose a new principle for transfer learning, based on a straightforward intuition: if two domains are similar to each other, the model trained on one domain should also perform well on the other domain, and vice versa. To formalize this intuition, we define the performance gap as a measure of the discrepancy between the source and target domains. We derive generalization bounds for the instance weighting approach to transfer learning, showing that the performance gap can be viewed as an algorithm-dependent regularizer, which controls the model complexity. Our theoretical analysis provides new insight into transfer learning and motivates a set of general, principled rules for designing new instance weighting schemes for transfer learning. These rules lead to gapBoost, a novel and principled boosting approach for transfer learning. Our experimental evaluation on benchmark data sets shows that gapBoost significantly outperforms previous boosting-based transfer learning algorithms. | |||
| Jorge A. Mendez, Shashank Shivkumar, & Eric Eaton |
Lifelong inverse reinforcement learning
Download:
[ PDF ]
|
2018 | Neural Information Processing Systems |
| Jorge A. Mendez, Shashank Shivkumar, & Eric Eaton. Lifelong inverse reinforcement learning. Neural Information Processing Systems, 2018. | |||
@article{Mendez2018Lifelong,
title = {Lifelong inverse reinforcement learning},
author = {Jorge A. Mendez and Shashank Shivkumar and Eric Eaton},
journal = {Neural Information Processing Systems},
year = {2018},
abstract = {
Methods for learning from demonstration (LfD) have shown
success in acquiring behavior policies by imitating a user.
However, even for a single task, LfD may require numerous
demonstrations. For versatile agents that must learn many
tasks via demonstration, this process would substantially
burden the user if each task were learned in isolation. To
address this challenge, we introduce the novel problem of
lifelong learning from demonstration, which allows the
agent to continually build upon knowledge learned from
previously demonstrated tasks to accelerate the learning
of new tasks, reducing the amount of demonstrations
required. As one solution to this problem, we propose the
first lifelong learning approach to inverse reinforcement
learning, which learns consecutive tasks via demonstration,
continually transferring knowledge between tasks to improve
performance.}
}
| |||
Abstract:Methods for learning from demonstration (LfD) have shown success in acquiring behavior policies by imitating a user. However, even for a single task, LfD may require numerous demonstrations. For versatile agents that must learn many tasks via demonstration, this process would substantially burden the user if each task were learned in isolation. To address this challenge, we introduce the novel problem of lifelong learning from demonstration, which allows the agent to continually build upon knowledge learned from previously demonstrated tasks to accelerate the learning of new tasks, reducing the amount of demonstrations required. As one solution to this problem, we propose the first lifelong learning approach to inverse reinforcement learning, which learns consecutive tasks via demonstration, continually transferring knowledge between tasks to improve performance. | |||
| Mohammad Rostami, Soheil Kolouri, Kyungnam Kim, & Eric Eaton |
Multi-agent distributed lifelong learning for collective knowledge acquisition
Download:
[ PDF ]
|
2018 | Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18) |
| Mohammad Rostami, Soheil Kolouri, Kyungnam Kim, & Eric Eaton. Multi-agent distributed lifelong learning for collective knowledge acquisition. In Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18), 2018. | |||
@inproceedings{Rostami2018MultiAgent,
title = {Multi-agent distributed lifelong learning for collective knowledge acquisition},
author = {Mohammad Rostami and Soheil Kolouri and Kyungnam Kim and Eric Eaton},
booktitle = {Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18)},
year = {2018},
abstract = {
Lifelong machine learning methods acquire knowledge over
a series of consecutive tasks, continually building upon
their experience. Current lifelong learning algorithms
rely upon a single learning agent that has centralized
access to all data. In this paper, we extend the idea of
lifelong learning from a single agent to a network of
multiple agents that collectively learn a series of
tasks. Each agent faces some (potentially unique) set of
tasks; the key idea is that knowledge learned from these
tasks may benefit other agents trying to learn different
(but related) tasks. Our Collective Lifelong Learning
Algorithm (CoLLA) provides an efficient way for a network
of agents to share their learned knowledge in a
distributed and decentralized manner, while eliminating
the need to share locally observed data. We provide
theoretical guarantees for robust performance of the
algorithm and empirically demonstrate that CoLLA
outperforms existing approaches for distributed
multi-task learning on a variety of datasets.}
}
| |||
Abstract:Lifelong machine learning methods acquire knowledge over a series of consecutive tasks, continually building upon their experience. Current lifelong learning algorithms rely upon a single learning agent that has centralized access to all data. In this paper, we extend the idea of lifelong learning from a single agent to a network of multiple agents that collectively learn a series of tasks. Each agent faces some (potentially unique) set of tasks; the key idea is that knowledge learned from these tasks may benefit other agents trying to learn different (but related) tasks. Our Collective Lifelong Learning Algorithm (CoLLA) provides an efficient way for a network of agents to share their learned knowledge in a distributed and decentralized manner, while eliminating the need to share locally observed data. We provide theoretical guarantees for robust performance of the algorithm and empirically demonstrate that CoLLA outperforms existing approaches for distributed multi-task learning on a variety of datasets. | |||
| Decebal Constantin Mocanu, Haitham Bou Ammar, Luis Puig, Eric Eaton, & Antonio Liotta |
Estimating 3D trajectories from 2D projections via disjunctive factored four-way conditional restricted Boltzmann machines
Download:
[ Website ]
|
2017 | Pattern Recognition |
| Decebal Constantin Mocanu, Haitham Bou Ammar, Luis Puig, Eric Eaton, & Antonio Liotta. Estimating 3D trajectories from 2D projections via disjunctive factored four-way conditional restricted Boltzmann machines. Pattern Recognition, 69:325–335, September 2017. | |||
@article{Mocanu2017Estimating,
title = {Estimating 3D trajectories from 2D projections via disjunctive factored four-way conditional restricted Boltzmann machines},
author = {Decebal Constantin Mocanu and Haitham Bou Ammar and Luis Puig and Eric Eaton and Antonio Liotta},
journal = {Pattern Recognition},
volume = {69},
pages = {325--335},
month = {September},
year = {2017},
abstract = {
Estimation, recognition, and near-future prediction of 3D trajectories based on their
two dimensional projections available from one camera source is an exceptionally
difficult problem due to uncertainty in the trajectories and environment, high
dimensionality of the specific trajectory states, lack of enough labeled data and so
on. In this article, we propose a solution to solve this problem based on a novel
deep learning model dubbed disjunctive factored four-way conditional restricted
Boltzmann machine (DFFW-CRBM). Our method improves state-of-the-art deep learning
techniques for high dimensional time-series modeling by introducing a novel tensor
factorization capable of driving forth order Boltzmann machines to considerably lower
energy levels, at no computational costs. DFFW-CRBMs are capable of accurately
estimating, recognizing, and performing near-future prediction of three-dimensional
trajectories from their 2D projections while requiring limited amount of labeled
data. We evaluate our method on both simulated and real-world data, showing its
effectiveness in predicting and classifying complex ball trajectories and human
activities.
}
}
| |||
Abstract:Estimation, recognition, and near-future prediction of 3D trajectories based on their two dimensional projections available from one camera source is an exceptionally difficult problem due to uncertainty in the trajectories and environment, high dimensionality of the specific trajectory states, lack of enough labeled data and so on. In this article, we propose a solution to solve this problem based on a novel deep learning model dubbed disjunctive factored four-way conditional restricted Boltzmann machine (DFFW-CRBM). Our method improves state-of-the-art deep learning techniques for high dimensional time-series modeling by introducing a novel tensor factorization capable of driving forth order Boltzmann machines to considerably lower energy levels, at no computational costs. DFFW-CRBMs are capable of accurately estimating, recognizing, and performing near-future prediction of three-dimensional trajectories from their 2D projections while requiring limited amount of labeled data. We evaluate our method on both simulated and real-world data, showing its effectiveness in predicting and classifying complex ball trajectories and human activities. | |||
| Christopher Clingerman & Eric Eaton |
Lifelong machine learning with Gaussian processes
Download:
[ PDF ]
|
2017 | European Conference on Machine Learning & Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD-17) |
| Christopher Clingerman & Eric Eaton. Lifelong machine learning with Gaussian processes. In Proceedings of the European Conference on Machine Learning & Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD-17), 2017. | |||
@inproceedings{Clingerman2017Lifelong,
title = {Lifelong machine learning with Gaussian processes},
author = {Christopher Clingerman and Eric Eaton},
booktitle = {Proceedings of the European Conference on Machine Learning \& Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD-17)},
year = {2017},
abstract = {
Recent developments in lifelong machine learning have demonstrated
that it is possible to learn multiple tasks consecutively, transferring
knowledge between those tasks to accelerate learning and improve performance.
However, these methods are limited to using linear parametric
base learners, substantially restricting the predictive power of the resulting
models. We present a lifelong learning algorithm that can support nonparametric
models, focusing on Gaussian processes. To enable efficient
online transfer between Gaussian process models, our approach assumes
a factorized formulation of the covariance functions, and incrementally
learns a shared sparse basis for the models’ parameterizations. We show
that this lifelong learning approach is highly computationally efficient,
and outperforms existing methods on a variety of data sets.
}
}
| |||
Abstract:Recent developments in lifelong machine learning have demonstrated that it is possible to learn multiple tasks consecutively, transferring knowledge between those tasks to accelerate learning and improve performance. However, these methods are limited to using linear parametric base learners, substantially restricting the predictive power of the resulting models. We present a lifelong learning algorithm that can support nonparametric models, focusing on Gaussian processes. To enable efficient online transfer between Gaussian process models, our approach assumes a factorized formulation of the covariance functions, and incrementally learns a shared sparse basis for the models’ parameterizations. We show that this lifelong learning approach is highly computationally efficient, and outperforms existing methods on a variety of data sets. | |||
| Eric Eaton |
Teaching integrated AI through interdisciplinary project-driven courses
Download:
[ PDF ] [ Website ]
|
2017 | AI Magazine 38(2) |
| Eric Eaton. Teaching integrated AI through interdisciplinary project-driven courses. AI Magazine, 38(2):13–21, 2017. | |||
@article{Eaton2017Teaching,
title = {Teaching integrated {AI} through interdisciplinary project-driven courses},
author = {Eric Eaton},
journal = {AI Magazine},
volume = {38},
number = {2},
pages = {13--21},
year = {2017},
abstract = {
Different subfields of AI (such as vision, learning, reasoning, planning, and others)
are often studied in isolation, both in individual courses and in the research
literature. This promulgates the idea that these different AI capabilities can easily
be integrated later, whereas, in practice, developing integrated AI systems remains
an open challenge for both research and industry. Interdisciplinary project-driven
courses can fill this gap in AI education, providing challenging problems that
require the integration of multiple AI methods. This article explores teaching
integrated AI through two project-driven courses: a capstone-style graduate course in
advanced robotics, and an undergraduate course on computational sustainability and
assistive computing. In addition to studying the integration of AI techniques, these
courses provide students with practical applications experience and exposure to
social issues of AI and computing. My hope is that other instructors find these
courses as useful examples for constructing their own project-driven courses to teach
integrated AI.
}
}
| |||
Abstract:Different subfields of AI (such as vision, learning, reasoning, planning, and others) are often studied in isolation, both in individual courses and in the research literature. This promulgates the idea that these different AI capabilities can easily be integrated later, whereas, in practice, developing integrated AI systems remains an open challenge for both research and industry. Interdisciplinary project-driven courses can fill this gap in AI education, providing challenging problems that require the integration of multiple AI methods. This article explores teaching integrated AI through two project-driven courses: a capstone-style graduate course in advanced robotics, and an undergraduate course on computational sustainability and assistive computing. In addition to studying the integration of AI techniques, these courses provide students with practical applications experience and exposure to social issues of AI and computing. My hope is that other instructors find these courses as useful examples for constructing their own project-driven courses to teach integrated AI. | |||
| Gilmer Valdes, Jose Marcio Luna, Eric Eaton, Charles B. Simone II, Lyle H. Ungar, & Timothy D. Solberg | MediBoost: a Patient Stratification Tool for Interpretable Decision Making in the Era of Precision Medicine | 2016 | Scientific Reports |
| Gilmer Valdes, Jose Marcio Luna, Eric Eaton, Charles B. Simone II, Lyle H. Ungar, & Timothy D. Solberg. MediBoost: a Patient Stratification Tool for Interpretable Decision Making in the Era of Precision Medicine. Scientific Reports, 6:37854, November 2016. | |||
@article{Valdes2016MediBoost,
title = {MediBoost: a Patient Stratification Tool for Interpretable Decision Making in the Era of Precision Medicine},
author = {Gilmer Valdes and Jose Marcio Luna and Eric Eaton and Charles B. {Simone II} and Lyle H. Ungar and Timothy D. Solberg},
journal = {Scientific Reports},
volume = {6},
pages = {37854},
year = {2016},
month = {November},
abstract = {
Machine learning algorithms that are both interpretable
and accurate are essential in applications such as
medicine where errors can have a dire consequence.
Unfortunately, there is currently a tradeoff between
accuracy and interpretability among state-of-the-art
methods. Decision trees are interpretable and are
therefore used extensively throughout medicine for
stratifying patients. Current decision tree algorithms,
however, are consistently outperformed in accuracy by
other, less-interpretable machine learning models, such
as ensemble methods. We present MediBoost, a novel
framework for constructing decision trees that retain
interpretability while having accuracy similar to
ensemble methods, and compare MediBoost’s performance
to that of conventional decision trees and ensemble
methods on 13 medical classification problems. MediBoost
significantly outperformed current decision tree
algorithms in 11 out of 13 problems, giving accuracy
comparable to ensemble methods. The resulting trees are
of the same type as decision trees used throughout
clinical practice but have the advantage of improved
accuracy. Our algorithm thus gives the best of both
worlds: it grows a single, highly interpretable tree
that has the high accuracy of ensemble methods.}
}
| |||
Abstract:Machine learning algorithms that are both interpretable and accurate are essential in applications such as medicine where errors can have a dire consequence. Unfortunately, there is currently a tradeoff between accuracy and interpretability among state-of-the-art methods. Decision trees are interpretable and are therefore used extensively throughout medicine for stratifying patients. Current decision tree algorithms, however, are consistently outperformed in accuracy by other, less-interpretable machine learning models, such as ensemble methods. We present MediBoost, a novel framework for constructing decision trees that retain interpretability while having accuracy similar to ensemble methods, and compare MediBoost’s performance to that of conventional decision trees and ensemble methods on 13 medical classification problems. MediBoost significantly outperformed current decision tree algorithms in 11 out of 13 problems, giving accuracy comparable to ensemble methods. The resulting trees are of the same type as decision trees used throughout clinical practice but have the advantage of improved accuracy. Our algorithm thus gives the best of both worlds: it grows a single, highly interpretable tree that has the high accuracy of ensemble methods. | |||
| David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor | Lifelong Learning for Disturbance Rejection on Mobile Robots | 2016 | International Conference on Intelligent Robots and Systems (IROS-16) |
| David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor. Lifelong Learning for Disturbance Rejection on Mobile Robots. In Proceedings of the International Conference on Intelligent Robots and Systems (IROS-16), IEEE/RSJ, October 2016. | |||
@INPROCEEDINGS{Isele2016Lifelong,
author = {David Isele and Jose Marcio Luna and Eric Eaton and
Gabriel V. {de la Cruz} and James Irwin and Brandon Kallaher and Matthew E. Taylor},
year = {2016},
title = {Lifelong Learning for Disturbance Rejection on Mobile Robots},
booktitle = {Proceedings of the International Conference on Intelligent Robots and Systems (IROS-16)},
month = {October},
publisher = {IEEE/RSJ},
abstract = {
No two robots are exactly the same---even for a given
model of robot, different units will require slightly
different controllers. Furthermore, because robots
change and degrade over time, a controller will need
to change over time to remain optimal. This paper
leverages lifelong learning in order to learn
controllers for different robots. In particular,
we show that by learning a set of control policies
over robots with different (unknown) motion models,
we can quickly adapt to changes in the robot, or
learn a controller for a new robot with a unique
set of disturbances. Furthermore, the approach is
completely model-free, allowing us to apply this
method to robots that have not, or cannot, be
fully modeled.}
}
| |||
Abstract:No two robots are exactly the same---even for a given model of robot, different units will require slightly different controllers. Furthermore, because robots change and degrade over time, a controller will need to change over time to remain optimal. This paper leverages lifelong learning in order to learn controllers for different robots. In particular, we show that by learning a set of control policies over robots with different (unknown) motion models, we can quickly adapt to changes in the robot, or learn a controller for a new robot with a unique set of disturbances. Furthermore, the approach is completely model-free, allowing us to apply this method to robots that have not, or cannot, be fully modeled. | |||
| David Isele, Mohammad Rostami, & Eric Eaton |
Using task features for zero-shot knowledge transfer in lifelong learning Awarded sole IJCAI-16 Distinguished Student Paper |
2016 | International Joint Conference on Artificial Intelligence (IJCAI-16) |
| David Isele, Mohammad Rostami, & Eric Eaton. Using task features for zero-shot knowledge transfer in lifelong learning. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-16), July 2016. | |||
@INPROCEEDINGS{Isele2016Using,
author = {David Isele and Mohammad Rostami and Eric Eaton},
year = {2016},
title = {Using task features for zero-shot knowledge transfer in lifelong learning},
booktitle = {Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-16)},
month = {July},
abstract = {
Knowledge transfer between tasks can improve the performance of learned
models, but requires an accurate estimate of the inter-task relationships
to identify the relevant knowledge to transfer. These inter-task
relationships are typically estimated based on training data for each
task, which is inefficient in lifelong learning settings where the goal
is to learn each consecutive task rapidly from as little data as possible.
To reduce this burden, we develop a lifelong reinforcement learning method
based on coupled dictionary learning that incorporates high-level task
descriptors to model the inter-task relationships. We show that using
task descriptors improves the performance of the learned task policies,
providing both theoretical justification for the benefit and empirical
demonstration of the improvement across a variety of dynamical control
problems. Given only the descriptor for a new task, the lifelong learner
is also able to accurately predict the task policy through zero-shot
learning using the coupled dictionary, eliminating the need to pause to
gather training data before addressing the task.}
}
| |||
Abstract:Knowledge transfer between tasks can improve the performance of learned models, but requires an accurate estimate of the inter-task relationships to identify the relevant knowledge to transfer. These inter-task relationships are typically estimated based on training data for each task, which is inefficient in lifelong learning settings where the goal is to learn each consecutive task rapidly from as little data as possible. To reduce this burden, we develop a lifelong reinforcement learning method based on coupled dictionary learning that incorporates high-level task descriptors to model the inter-task relationships. We show that using task descriptors improves the performance of the learned task policies, providing both theoretical justification for the benefit and empirical demonstration of the improvement across a variety of dynamical control problems. Given only the descriptor for a new task, the lifelong learner is also able to accurately predict the task policy through zero-shot learning using the coupled dictionary, eliminating the need to pause to gather training data before addressing the task. | |||
| Haitham Bou Ammar, Rasul Tutunov, & Eric Eaton | Safe policy search for lifelong reinforcement learning with sublinear regret | 2015 | International Conference on Machine Learning (ICML-15) |
| Haitham Bou Ammar, Rasul Tutunov, & Eric Eaton. Safe policy search for lifelong reinforcement learning with sublinear regret. In Proceedings of the 32nd International Conference on Machine Learning (ICML-15), July 2015. | |||
@INPROCEEDINGS{BouAmmar2015Safe,
author = {Haitham {Bou Ammar} and Rasul Tutunov and Eric Eaton},
year = {2015},
title = {Safe policy search for lifelong reinforcement learning with sublinear regret},
booktitle = {Proceedings of the 32nd International Conference on Machine Learning (ICML-15)},
month = {July},
abstract = {
Lifelong reinforcement learning provides a promising framework for
developing versatile agents that can accumulate knowledge over a
lifetime of experience and rapidly learn new tasks by building
upon prior knowledge. However, current lifelong learning methods
exhibit non-vanishing regret as the amount of experience increases,
and include limitations that can lead to suboptimal or unsafe
control policies. To address these issues, we develop a lifelong
policy gradient learner that operates in an adversarial setting to
learn multiple tasks online while enforcing safety constraints on
the learned policies. We demonstrate, for the first time,
sublinear regret for lifelong policy search, and validate our
algorithm on several benchmark dynamical systems and an application
to quadrotor control.}
}
| |||
Abstract:Lifelong reinforcement learning provides a promising framework for developing versatile agents that can accumulate knowledge over a lifetime of experience and rapidly learn new tasks by building upon prior knowledge. However, current lifelong learning methods exhibit non-vanishing regret as the amount of experience increases, and include limitations that can lead to suboptimal or unsafe control policies. To address these issues, we develop a lifelong policy gradient learner that operates in an adversarial setting to learn multiple tasks online while enforcing safety constraints on the learned policies. We demonstrate, for the first time, sublinear regret for lifelong policy search, and validate our algorithm on several benchmark dynamical systems and an application to quadrotor control. | |||
| Haitham Bou Ammar, Eric Eaton, Jose Marcio Luna, & Paul Ruvolo |
Autonomous cross-domain knowledge transfer in lifelong policy gradient reinforcement learning Finalist for IJCAI-15 Distinguished Paper award
Download:
[ PDF ]
|
2015 | International Joint Conference on Artificial Intelligence (IJCAI-15) |
| Haitham Bou Ammar, Eric Eaton, Jose Marcio Luna, & Paul Ruvolo. Autonomous cross-domain knowledge transfer in lifelong policy gradient reinforcement learning. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-15), July 2015. | |||
@INPROCEEDINGS{BouAmmar2015Autonomous,
author = {Haitham {Bou Ammar} and Eric Eaton and Jose Marcio Luna and Paul Ruvolo},
year = {2015},
title = {Autonomous cross-domain knowledge transfer in lifelong policy gradient reinforcement learning},
booktitle = {Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-15)},
month = {July},
abstract = {
Online multi-task learning is an important capability for lifelong
learning agents, enabling them to acquire models for diverse tasks
over time and rapidly learn new tasks by building upon prior experience.
However, recent progress toward lifelong reinforcement learning (RL)
has been limited to learning from within a single task domain. For
truly versatile lifelong learning, the agent must be able to autonomously
transfer knowledge between different task domains. A few methods for
cross-domain transfer have been developed, but these methods are
computationally inefficient for scenarios where the agent must learn
tasks consecutively.
In this paper, we develop the first cross-domain lifelong RL framework.
Our approach efficiently optimizes a shared repository of transferable
knowledge and learns projection matrices that specialize that knowledge
to different task domains. We provide rigorous theoretical guarantees on
the stability of this approach, and empirically evaluate its performance
on diverse dynamical systems. Our results show that the proposed method
can learn effectively from interleaved task domains and rapidly acquire
high performance in new domains.}
}
| |||
Abstract:Online multi-task learning is an important capability for lifelong learning agents, enabling them to acquire models for diverse tasks over time and rapidly learn new tasks by building upon prior experience. However, recent progress toward lifelong reinforcement learning (RL) has been limited to learning from within a single task domain. For truly versatile lifelong learning, the agent must be able to autonomously transfer knowledge between different task domains. A few methods for cross-domain transfer have been developed, but these methods are computationally inefficient for scenarios where the agent must learn tasks consecutively. In this paper, we develop the first cross-domain lifelong RL framework. Our approach efficiently optimizes a shared repository of transferable knowledge and learns projection matrices that specialize that knowledge to different task domains. We provide rigorous theoretical guarantees on the stability of this approach, and empirically evaluate its performance on diverse dynamical systems. Our results show that the proposed method can learn effectively from interleaved task domains and rapidly acquire high performance in new domains. | |||
| Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor | Unsupervised cross-domain transfer in policy gradient reinforcement learning via manifold alignment | 2015 | AAAI Conference on Artificial Intelligence (AAAI-15) |
| Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor. Unsupervised cross-domain transfer in policy gradient reinforcement learning via manifold alignment. In Proceedings of the 29th AAAI Conference on Artificial Intelligence (AAAI-15), January 2015. Acceptance rate: 27% | |||
@inproceedings{BouAmmar2015Unsupervised,
author={Haitham {Bou Ammar} and Eric Eaton and Paul Ruvolo and Matthew E. Taylor},
title={Unsupervised cross-domain transfer in policy gradient reinforcement learning via manifold alignment},
booktitle={Proceedings of the 29th AAAI Conference on Artificial Intelligence (AAAI-15)},
month={January},
year={2015},
note={Acceptance rate: 27%},
abstract={
The success of applying policy gradient reinforcement learning (RL) to
difficult control tasks hinges crucially on the ability to determine a
sensible initialization for the policy. Transfer learning methods tackle
this problem by reusing knowledge gleaned from solving other related tasks.
In the case of multiple task domains, these algorithms require an inter-task
mapping to facilitate knowledge transfer across domains. However, there are
currently no general methods to learn an inter-task mapping without requiring
either background knowledge that is not typically present in RL settings, or
an expensive analysis of an exponential number of inter-task mappings in the
size of the state and action spaces. This paper introduces an autonomous
framework that uses unsupervised manifold alignment to learn intertask mappings
and effectively transfer samples between different task domains. Empirical
results on diverse dynamical systems, including an application to quadrotor
control, demonstrate its effectiveness for cross-domain transfer in the context
of policy gradient RL.},
}
| |||
Abstract:The success of applying policy gradient reinforcement learning (RL) to difficult control tasks hinges crucially on the ability to determine a sensible initialization for the policy. Transfer learning methods tackle this problem by reusing knowledge gleaned from solving other related tasks. In the case of multiple task domains, these algorithms require an inter-task mapping to facilitate knowledge transfer across domains. However, there are currently no general methods to learn an inter-task mapping without requiring either background knowledge that is not typically present in RL settings, or an expensive analysis of an exponential number of inter-task mappings in the size of the state and action spaces. This paper introduces an autonomous framework that uses unsupervised manifold alignment to learn intertask mappings and effectively transfer samples between different task domains. Empirical results on diverse dynamical systems, including an application to quadrotor control, demonstrate its effectiveness for cross-domain transfer in the context of policy gradient RL. | |||
| Paul Ruvolo & Eric Eaton | Online Multi-Task Learning via Sparse Dictionary Optimization | 2014 | AAAI Conference on Artificial Intelligence (AAAI-14) |
| Paul Ruvolo & Eric Eaton. Online Multi-Task Learning via Sparse Dictionary Optimization. In Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14), July 2014. | |||
@INPROCEEDINGS{Ruvolo2014Online,
author = {Paul Ruvolo and Eric Eaton},
title = {Online Multi-Task Learning via Sparse Dictionary Optimization},
booktitle = {Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14)},
year = {2014},
month = {July},
abstract = {
This paper develops an efficient online algorithm
for learning multiple consecutive tasks based on
the K-SVD algorithm for sparse dictionary optimization.
We first derive a batch multi-task learning method that
builds upon K-SVD, and then extend the batch algorithm
to train models online in a lifelong learning setting.
The resulting method has lower computational
complexity than other current lifelong learning
algorithms while maintaining nearly identical model
performance. Additionally, the proposed method offers
an alternate formulation for lifelong learning that
supports both task and feature similarity matrices.}
}
| |||
Abstract:This paper develops an efficient online algorithm for learning multiple consecutive tasks based on the K-SVD algorithm for sparse dictionary optimization. We first derive a batch multi-task learning method that builds upon K-SVD, and then extend the batch algorithm to train models online in a lifelong learning setting. The resulting method has lower computational complexity than other current lifelong learning algorithms while maintaining nearly identical model performance. Additionally, the proposed method offers an alternate formulation for lifelong learning that supports both task and feature similarity matrices. | |||
| Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor |
Online Multi-Task Learning for Policy Gradient Methods
Download:
[ PDF ]
|
2014 | International Conference on Machine Learning (ICML-14) |
| Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor. Online Multi-Task Learning for Policy Gradient Methods. In Proceedings of the 31st International Conference on Machine Learning (ICML-14), June 2014. | |||
@INPROCEEDINGS{BouAmmar2014Online,
author = {Haitham Bou Ammar and Eric Eaton and Paul Ruvolo and Matthew E. Taylor},
title = {Online Multi-Task Learning for Policy Gradient Methods},
booktitle = {Proceedings of the 31st International Conference on Machine Learning (ICML-14)},
year = {2014},
month = {June},
abstract = {
Policy gradient algorithms have shown considerable recent
success in solving high-dimensional sequential
decision-making (SDM) tasks, particularly in robotics.
However, these methods often require extensive experience
in a domain to achieve high performance. To make agents
more sample-efficient, we developed a multi-task policy
gradient method to learn SDM tasks consecutively,
transferring knowledge between tasks to accelerate
learning. Our approach provides robust theoretical
guarantees and we show empirically that it dramatically
accelerates learning on a variety of dynamical systems,
including an application to quadcopter control.}
}
| |||
Abstract:Policy gradient algorithms have shown considerable recent success in solving high-dimensional sequential decision-making (SDM) tasks, particularly in robotics. However, these methods often require extensive experience in a domain to achieve high performance. To make agents more sample-efficient, we developed a multi-task policy gradient method to learn SDM tasks consecutively, transferring knowledge between tasks to accelerate learning. Our approach provides robust theoretical guarantees and we show empirically that it dramatically accelerates learning on a variety of dynamical systems, including an application to quadcopter control. | |||
| Eric Eaton, Carla Gomes, & Brian Williams |
Computational Sustainability
Download:
[ PDF ]
|
2014 | AI Magazine 35(2) |
| Eric Eaton, Carla Gomes, & Brian Williams. Computational Sustainability. AI Magazine, 35(2):3–7, June 2014. | |||
@article{Eaton2014CompSustainability,
author = {Eric Eaton and Carla Gomes and Brian Williams},
title = {Computational Sustainability},
journal = {AI Magazine},
volume = {35},
number = {2},
pages = {3--7},
year = {2014},
month = {June},
abstract = {
Computational sustainability problems, which exist in
dynamic environments with high amounts of uncertainty,
provide a variety of unique challenges to artificial
intelligence research and the opportunity for
significant impact upon our collective future. This
editorial provides an overview of artificial
intelligence for computational sustainability, and
introduces this special issue of AI Magazine.}
}
| |||
Abstract:Computational sustainability problems, which exist in dynamic environments with high amounts of uncertainty, provide a variety of unique challenges to artificial intelligence research and the opportunity for significant impact upon our collective future. This editorial provides an overview of artificial intelligence for computational sustainability, and introduces this special issue of AI Magazine. | |||
| Eric Eaton, Marie desJardins, & Sara Jacob |
Multi-view constrained clustering with an incomplete mapping between views
Download:
[ PDF ]
|
2014 | Knowledge and Information Systems 38(1) |
| Eric Eaton, Marie desJardins, & Sara Jacob. Multi-view constrained clustering with an incomplete mapping between views. Knowledge and Information Systems, 38(1):231–257, January 2014. | |||
@ARTICLE{Eaton2012MultiView,
author = {Eric Eaton and Marie desJardins and Sara Jacob},
title = {Multi-view constrained clustering with an incomplete mapping between views},
journal = {Knowledge and Information Systems},
volume = {38},
number = {1},
pages = {231--257},
month = {January},
year = {2014},
notes = {Published online November 2012, <a href="http://link.springer.com/content/pdf/10.1007%2Fs10115-012-0577-7">doi 10.1007/s10115-012-0577-7</a>.},
abstract = {
Multi-view learning algorithms typically assume a
complete bipartite mapping between the different
views in order to exchange information during the
learning process. However, many applications
provide only a partial mapping between the views,
creating a challenge for current methods. To
address this problem, we propose a multi-view
algorithm based on constrained clustering that can
operate with an incomplete mapping. Given a set
of pairwise constraints in each view, our approach
propagates these constraints using a local
similarity measure to those instances that can be
mapped to the other views, allowing the propagated
constraints to be transferred across views via the
partial mapping. It uses co-EM to iteratively
estimate the propagation within each view based on
the current clustering model, transfer the
constraints across views, and then update the
clustering model. By alternating the learning
process between views, this approach produces a
unified clustering model that is consistent with
all views. We show that this approach
significantly improves clustering performance over
several other methods for transferring constraints
and allows multi-view clustering to be reliably
applied when given a limited mapping between the
views. Our evaluation reveals that the propagated
constraints have high precision with respect to
the true clusters in the data, explaining their
benefit to clustering performance in both single-
and multi-view learning scenarios.}
}
| |||
Abstract:Multi-view learning algorithms typically assume a complete bipartite mapping between the different views in order to exchange information during the learning process. However, many applications provide only a partial mapping between the views, creating a challenge for current methods. To address this problem, we propose a multi-view algorithm based on constrained clustering that can operate with an incomplete mapping. Given a set of pairwise constraints in each view, our approach propagates these constraints using a local similarity measure to those instances that can be mapped to the other views, allowing the propagated constraints to be transferred across views via the partial mapping. It uses co-EM to iteratively estimate the propagation within each view based on the current clustering model, transfer the constraints across views, and then update the clustering model. By alternating the learning process between views, this approach produces a unified clustering model that is consistent with all views. We show that this approach significantly improves clustering performance over several other methods for transferring constraints and allows multi-view clustering to be reliably applied when given a limited mapping between the views. Our evaluation reveals that the propagated constraints have high precision with respect to the true clusters in the data, explaining their benefit to clustering performance in both single- and multi-view learning scenarios. | |||
| Paul Ruvolo & Eric Eaton | Active Task Selection for Lifelong Machine Learning | 2013 | AAAI Conference on Artificial Intelligence (AAAI-13) |
| Paul Ruvolo & Eric Eaton. Active Task Selection for Lifelong Machine Learning. In Proceedings of the 27th AAAI Conference on Artificial Intelligence (AAAI-13), July 2013. | |||
@INPROCEEDINGS{Ruvolo2013Active,
author = {Paul Ruvolo and Eric Eaton},
title = {Active Task Selection for Lifelong Machine Learning},
booktitle = {Proceedings of the 27th AAAI Conference on Artificial Intelligence (AAAI-13)},
year = {2013},
month = {July},
abstract = {
In a lifelong learning framework, an agent acquires
knowledge incrementally over consecutive learning
tasks, continually building upon its experience.
Recent lifelong learning algorithms have achieved
nearly identical performance to batch multi-task
learning methods while reducing learning time by
three orders of magnitude. In this paper, we
further improve the scalability of lifelong learning
by developing curriculum selection methods that
enable an agent to actively select the next task to
learn in order to maximize performance on future
learning tasks. We demonstrate that active task
selection is highly reliable and effective, allowing
an agent to learn high performance models using up
to 50% fewer tasks than when the agent has no
control over the task order. We also explore a
variant of transfer learning in the lifelong learning
setting in which the agent can focus knowledge
acquisition toward a particular target task.}
}
| |||
Abstract:In a lifelong learning framework, an agent acquires knowledge incrementally over consecutive learning tasks, continually building upon its experience. Recent lifelong learning algorithms have achieved nearly identical performance to batch multi-task learning methods while reducing learning time by three orders of magnitude. In this paper, we further improve the scalability of lifelong learning by developing curriculum selection methods that enable an agent to actively select the next task to learn in order to maximize performance on future learning tasks. We demonstrate that active task selection is highly reliable and effective, allowing an agent to learn high performance models using up to 50% fewer tasks than when the agent has no control over the task order. We also explore a variant of transfer learning in the lifelong learning setting in which the agent can focus knowledge acquisition toward a particular target task. | |||
| Paul Ruvolo & Eric Eaton | ELLA: An Efficient Lifelong Learning Algorithm | 2013 | International Conference on Machine Learning (ICML-13) |
| Paul Ruvolo & Eric Eaton. ELLA: An Efficient Lifelong Learning Algorithm. In Proceedings of the 30th International Conference on Machine Learning (ICML-13), June 2013. | |||
@INPROCEEDINGS{Ruvolo2013ELLA,
author = {Paul Ruvolo and Eric Eaton},
title = {ELLA: An Efficient Lifelong Learning Algorithm},
booktitle = {Proceedings of the 30th International Conference on Machine Learning (ICML-13)},
year = {2013},
month = {June},
abstract = {
The problem of learning multiple consecutive tasks,
known as \emph{lifelong learning}, is of great
importance to the creation of intelligent,
general-purpose, and flexible machines. In this
paper, we develop a method for online multi-task
learning in the lifelong learning setting. The
proposed Efficient Lifelong Learning Algorithm
(ELLA) maintains a sparsely shared basis for all
task models, transfers knowledge from the basis
to learn each new task, and refines the basis
over time to maximize performance across all
tasks. We show that ELLA has strong connections
to both online dictionary learning for sparse
coding and state-of-the-art batch multi-task
learning methods, and provide robust theoretical
performance guarantees. We show empirically that
ELLA yields nearly identical performance to batch
multi-task learning while learning tasks
sequentially in three orders of magnitude (over
1,000x) less time.}
}
| |||
Abstract:The problem of learning multiple consecutive tasks, known as <i>lifelong learning</i>, is of great importance to the creation of intelligent, general-purpose, and flexible machines. In this paper, we develop a method for online multi-task learning in the lifelong learning setting. The proposed Efficient Lifelong Learning Algorithm (ELLA) maintains a sparsely shared basis for all task models, transfers knowledge from the basis to learn each new task, and refines the basis over time to maximize performance across all tasks. We show that ELLA has strong connections to both online dictionary learning for sparse coding and state-of-the-art batch multi-task learning methods, and provide robust theoretical performance guarantees. We show empirically that ELLA yields nearly identical performance to batch multi-task learning while learning tasks sequentially in three orders of magnitude (over 1,000x) less time. | |||
| Eric Eaton & Rachael Mansbach | A spin-glass model for semi-supervised community detection | 2012 | AAAI Conference on Artificial Intelligence (AAAI-12) |
| Eric Eaton & Rachael Mansbach. A spin-glass model for semi-supervised community detection. In Proceedings of the 26th AAAI Conference on Artificial Intelligence (AAAI-12), pp. 900–906, AAAI Press, July 22--26 2012. | |||
@INPROCEEDINGS{Eaton2012SpinGlass,
author = {Eric Eaton and Rachael Mansbach},
title = {A spin-glass model for semi-supervised community detection},
booktitle = {Proceedings of the 26th AAAI Conference on Artificial Intelligence (AAAI-12)},
month = {July 22--26},
location = {Toronto, Canada},
publisher = {AAAI Press},
pages = {900--906},
year = {2012},
abstract = {
Current modularity-based community detection methods
show decreased performance as relational networks
become increasingly noisy. These methods also yield
a large number of diverse community structures as
solutions, which is problematic for applications
that impose constraints on the acceptable solutions
or in cases where the user is focused on specific
communities of interest. To address both of these
problems, we develop a semi-supervised spin-glass
model that enables current community detection
methods to incorporate background knowledge in the
forms of individual labels and pairwise constraints.
Unlike current methods, our approach shows robust
performance in the presence of noise in the
relational network, and the ability to guide the
discovery process toward specific community
structures. We evaluate our algorithm on several
benchmark networks and a new political sentiment
network representing cooperative events between
nations that was mined from news articles over
six years.}
}
| |||
Abstract:Current modularity-based community detection methods show decreased performance as relational networks become increasingly noisy. These methods also yield a large number of diverse community structures as solutions, which is problematic for applications that impose constraints on the acceptable solutions or in cases where the user is focused on specific communities of interest. To address both of these problems, we develop a semi-supervised spin-glass model that enables current community detection methods to incorporate background knowledge in the forms of individual labels and pairwise constraints. Unlike current methods, our approach shows robust performance in the presence of noise in the relational network, and the ability to guide the discovery process toward specific community structures. We evaluate our algorithm on several benchmark networks and a new political sentiment network representing cooperative events between nations that was mined from news articles over six years. | |||
| Eric Eaton & Marie desJardins |
Selective Transfer Between Learning Tasks Using Task-Based Boosting
Download:
[ PDF ] [ Supplementary Materials ]
|
2011 | AAAI Conference on Artificial Intelligence (AAAI-11) |
| Eric Eaton & Marie desJardins. Selective Transfer Between Learning Tasks Using Task-Based Boosting. In Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11), pp. 337–342, AAAI Press, August 7--11 2011. | |||
@INPROCEEDINGS{Eaton2011Selective,
author = {Eric Eaton and Marie desJardins},
title = {Selective Transfer Between Learning Tasks Using Task-Based Boosting},
booktitle = {Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11)},
month = {August 7--11},
location = {San Francisco, CA},
publisher = {AAAI Press},
pages = {337--342},
year = {2011},
abstract = {
The success of transfer learning on a target task
is highly dependent on the selected source data.
Instance transfer methods reuse data from the
source tasks to augment the training data for the
target task. If poorly chosen, this source data
may inhibit learning, resulting in negative
transfer. The current most widely used algorithm
for instance transfer, TrAdaBoost, performs poorly
when given irrelevant source data.
We present a novel task-based boosting technique
for instance transfer that selectively chooses the
source knowledge to transfer to the target task.
Our approach performs boosting at both the instance
level and the task level, assigning higher weight to
those source tasks that show positive transferability
to the target task, and adjusting the weights of
individual instances within each source task via
AdaBoost. We show that this combination of task- and
instance-level boosting significantly improves
transfer performance over existing instance transfer
algorithms when given a mix of relevant and irrelevant
source data, especially for small amounts of data on
the target task.}
}
| |||
Abstract:The success of transfer learning on a target task is highly dependent on the selected source data. Instance transfer methods reuse data from the source tasks to augment the training data for the target task. If poorly chosen, this source data may inhibit learning, resulting in negative transfer. The current most widely used algorithm for instance transfer, TrAdaBoost, performs poorly when given irrelevant source data. We present a novel task-based boosting technique for instance transfer that selectively chooses the source knowledge to transfer to the target task. Our approach performs boosting at both the instance level and the task level, assigning higher weight to those source tasks that show positive transferability to the target task, and adjusting the weights of individual instances within each source task via AdaBoost. We show that this combination of task- and instance-level boosting significantly improves transfer performance over existing instance transfer algorithms when given a mix of relevant and irrelevant source data, especially for small amounts of data on the target task. | |||
| Eric Eaton, Marie desJardins, & Sara Jacob |
Multi-View Clustering with Constraint Propagation for Learning with an Incomplete Mapping Between Views
Download:
[ PDF ] [ Slides ]
|
2010 | Conference on Information and Knowledge Management (CIKM'10) |
| Eric Eaton, Marie desJardins, & Sara Jacob. Multi-View Clustering with Constraint Propagation for Learning with an Incomplete Mapping Between Views. In Proceedings of the Conference on Information and Knowledge Management (CIKM'10), pp. 389–398, ACM Press, October 26--30 2010. | |||
@INPROCEEDINGS{Eaton2010MultiView,
author = {Eric Eaton and Marie desJardins and Sara Jacob},
title = {Multi-View Clustering with Constraint Propagation for Learning with an Incomplete Mapping Between Views},
booktitle = {Proceedings of the Conference on Information and Knowledge Management (CIKM'10)},
month = {October 26--30},
location = {Toronto, Ontario, Canada},
publisher = {ACM Press},
pages = {389--398},
year = {2010},
abstract = {
Multi-view learning algorithms typically assume a complete bipartite mapping between the different views in order to exchange information during the learning process. However, many applications provide only a partial mapping between the views, creating a challenge for current methods. To address this problem, we propose a multi-view algorithm based on constrained clustering that can operate with an incomplete mapping. Given a set of pairwise constraints in each view, our approach propagates these constraints using a local similarity measure to those instances that can be mapped to the other views, allowing the propagated constraints to be transferred across views via the partial mapping. It uses co-EM to iteratively estimate the propagation within each view based on the current clustering model, transfer the constraints across views, and update the clustering model, thereby learning a unified model for all views. We show that this approach significantly improves clustering performance over several other methods for transferring constraints and allows multi-view clustering to be reliably applied when given a limited mapping between the views.
}
}
| |||
Abstract:Multi-view learning algorithms typically assume a complete bipartite mapping between the different views in order to exchange information during the learning process. However, many applications provide only a partial mapping between the views, creating a challenge for current methods. To address this problem, we propose a multi-view algorithm based on constrained clustering that can operate with an incomplete mapping. Given a set of pairwise constraints in each view, our approach propagates these constraints using a local similarity measure to those instances that can be mapped to the other views, allowing the propagated constraints to be transferred across views via the partial mapping. It uses co-EM to iteratively estimate the propagation within each view based on the current clustering model, transfer the constraints across views, and update the clustering model, thereby learning a unified model for all views. We show that this approach significantly improves clustering performance over several other methods for transferring constraints and allows multi-view clustering to be reliably applied when given a limited mapping between the views. | |||
| Kiri Wagstaff, Marie desJardins, & Eric Eaton |
Modeling and learning user preferences over sets
Download:
[ PDF ]
|
2010 | Journal of Experimental & Theoretical Artificial Intelligence 22(3) |
| Kiri Wagstaff, Marie desJardins, & Eric Eaton. Modeling and learning user preferences over sets. Journal of Experimental & Theoretical Artificial Intelligence, 22(3):237–268, September 2010. | |||
@ARTICLE{Wagstaff2010Modeling,
author = {Kiri Wagstaff and Marie desJardins and Eric Eaton},
year = {2010},
title = {Modeling and learning user preferences over sets},
journal = {Journal of Experimental & Theoretical Artificial Intelligence},
volume = {22},
number = {3},
pages = {237--268},
month = {September},
abstract = {
Although there has been significant research on modeling and learning user preferences for various types of objects, there has been relatively little work on the problem of representing and learning preferences over <i>sets</i> of objects. We introduce a representation language, DD-PREF, that balances preferences for particular objects with preferences about the properties of the set. Specifically, we focus on the <i>depth</i> of objects (i.e. preferences for specific attribute values over others) and on the <i>diversity</i> of sets (i.e. preferences for broad vs. narrow distributions of attribute values). The DD-PREF framework is general and can incorporate additional object- and set-based preferences. We describe a greedy algorithm, DD-Select, for selecting satisfying sets from a collection of new objects, given a preference in this language. We show how preferences represented in DD-PREF can be learned from training data. Experimental results are given for three domains: a blocks world domain with several different task-based preferences, a real-world music playlist collection, and rover image data gathered in desert training exercises.
}
}
| |||
Abstract:Although there has been significant research on modeling and learning user preferences for various types of objects, there has been relatively little work on the problem of representing and learning preferences over <i>sets</i> of objects. We introduce a representation language, DD-PREF, that balances preferences for particular objects with preferences about the properties of the set. Specifically, we focus on the <i>depth</i> of objects (i.e. preferences for specific attribute values over others) and on the <i>diversity</i> of sets (i.e. preferences for broad vs. narrow distributions of attribute values). The DD-PREF framework is general and can incorporate additional object- and set-based preferences. We describe a greedy algorithm, DD-Select, for selecting satisfying sets from a collection of new objects, given a preference in this language. We show how preferences represented in DD-PREF can be learned from training data. Experimental results are given for three domains: a blocks world domain with several different task-based preferences, a real-world music playlist collection, and rover image data gathered in desert training exercises. | |||
| Eric Eaton, Gary Holness, & Daniel McFarlane |
Interactive Learning using Manifold Geometry
Download:
[ PDF ] [ Slides ]
|
2010 | AAAI Conference on Artificial Intelligence (AAAI-10) |
| Eric Eaton, Gary Holness, & Daniel McFarlane. Interactive Learning using Manifold Geometry. In Proceedings of the 24th AAAI Conference on Artificial Intelligence (AAAI-10), pp. 437–443, AAAI Press, July 11--15 2010. | |||
@INPROCEEDINGS{Eaton2010Interactive,
author = {Eric Eaton and Gary Holness and Daniel McFarlane},
title = {Interactive Learning using Manifold Geometry},
booktitle = {Proceedings of the 24th AAAI Conference on Artificial Intelligence (AAAI-10)},
month = {July 11--15},
location = {Atlanta, GA},
publisher = {AAAI Press},
pages = {437--443},
year = {2010},
abstract = {
We present an interactive learning method that enables
a user to iteratively refine a regression model. The
user examines the output of the model, visualized as
the vertical axis of a 2D scatterplot, and provides
corrections by repositioning individual data instances
to the correct output level. Each repositioned data
instance acts as a control point for altering the
learned model, using the geometry underlying the data.
We capture the underlying structure of the data as a
manifold, on which we compute a set of basis functions
as the foundation for learning. Our results show that
manifold-based interactive learning improves performance
monotonically with each correction, outperforming
alternative approaches.}
}
| |||
Abstract:We present an interactive learning method that enables a user to iteratively refine a regression model. The user examines the output of the model, visualized as the vertical axis of a 2D scatterplot, and provides corrections by repositioning individual data instances to the correct output level. Each repositioned data instance acts as a control point for altering the learned model, using the geometry underlying the data. We capture the underlying structure of the data as a manifold, on which we compute a set of basis functions as the foundation for learning. Our results show that manifold-based interactive learning improves performance monotonically with each correction, outperforming alternative approaches. | |||
| Eric Eaton, Marie desJardins, & Terran Lane |
Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer
Download:
[ PDF ]
|
2008 | European Conference on Machine Learning (ECML) |
| Eric Eaton, Marie desJardins, & Terran Lane. Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer. In Proceedings of the 19th European Conference on Machine Learning (ECML), pp. 317–332, Springer-Verlag, Berlin, Heidelberg, 2008. Acceptance rate: 20% | |||
@INPROCEEDINGS{Eaton2008Modeling,
author = {Eric Eaton and Marie desJardins and Terran Lane},
title = {Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer},
booktitle = ecml08,
year = {2008},
pages = {317--332},
publisher = {Springer-Verlag},
location = {Antwerp, Belgium},
address = {Berlin, Heidelberg},
note = {Acceptance rate: 20%},
abstract = { In this paper, we propose a novel graph-based method for
knowledge transfer. We model the transfer relationships between source
tasks by embedding the set of learned source models in a graph using
transferability as the metric. Transfer to a new problem proceeds by
mapping the problem into the graph, then learning a function on this
graph that automatically determines the parameters to transfer to the
new learning task. This method is analogous to inductive transfer along a
manifold that captures the transfer relationships between the tasks. We
demonstrate improved transfer performance using this method against
existing approaches in several real-world domains.}
}
| |||
Abstract:In this paper, we propose a novel graph-based method for knowledge transfer. We model the transfer relationships between source tasks by embedding the set of learned source models in a graph using transferability as the metric. Transfer to a new problem proceeds by mapping the problem into the graph, then learning a function on this graph that automatically determines the parameters to transfer to the new learning task. This method is analogous to inductive transfer along a manifold that captures the transfer relationships between the tasks. We demonstrate improved transfer performance using this method against existing approaches in several real-world domains. | |||
| Marie desJardins, Eric Eaton, & Kiri Wagstaff |
Learning user preferences for sets of objects Awarded recognition as a NASA Tech Brief in 2008 |
2006 | International Conference on Machine Learning (ICML) |
| Marie desJardins, Eric Eaton, & Kiri Wagstaff. Learning user preferences for sets of objects. In Proceedings of the 23rd International Conference on Machine Learning (ICML), Pittsburgh, PA, June 25--29 2006. Acceptance rate: 20% | |||
@INPROCEEDINGS{desJardins2006Learning,
author = {Marie desJardins and Eric Eaton and Kiri Wagstaff},
title = {Learning user preferences for sets of objects},
booktitle = icml06,
year = {2006},
month = {June 25--29},
address = {Pittsburgh, PA},
note = {Acceptance rate: 20%},
abstract = { Most work on preference learning has focused
on pairwise preferences or rankings over individual
items. In this paper, we present a
method for learning preferences over sets of
items. Our learning method takes as input a
collection of positive examples--that is, one
or more sets that have been identified by a
user as desirable. Kernel density estimation
is used to estimate the value function for individual
items, and the desired set diversity is
estimated from the average set diversity observed
in the collection. Since this is a new
learning problem, we introduce a new evaluation
methodology and evaluate the learning
method on two data collections: synthetic
blocks-world data and a new real-world music
data collection that we have gathered.}
}
| |||
Abstract:Most work on preference learning has focused on pairwise preferences or rankings over individual items. In this paper, we present a method for learning preferences over sets of items. Our learning method takes as input a collection of positive examples--that is, one or more sets that have been identified by a user as desirable. Kernel density estimation is used to estimate the value function for individual items, and the desired set diversity is estimated from the average set diversity observed in the collection. Since this is a new learning problem, we introduce a new evaluation methodology and evaluate the learning method on two data collections: synthetic blocks-world data and a new real-world music data collection that we have gathered. | |||
Preprints
| AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
|---|---|---|---|
| Rasul Tutunov, Haitham Bou Ammar, Ali Jadbabaie, & Eric Eaton |
On the degree distribution of Pólya urn graph processes
Download:
[ PDF ]
|
2014 | arXiv:1410.8515 Preprint |
| Rasul Tutunov, Haitham Bou Ammar, Ali Jadbabaie, & Eric Eaton. On the degree distribution of Pólya urn graph processes. arXiv:1410.8515 Preprint, October 2014. | |||
@ARTICLE{Tutunov2014DegreeDistribution,
author = {Rasul Tutunov and Haitham {Bou Ammar} and Ali Jadbabaie and Eric Eaton},
year = {2014},
title = {On the degree distribution of P\'{o}lya urn graph processes},
journal = {arXiv:1410.8515 Preprint},
month = {October},
abstract = {
This paper presents a tighter bound on the degree distribution of
arbitrary Polya urn graph processes, proving that the proportion of
vertices with degree d obeys a power-law distribution P(d) proportional
to d^(-gamma) for d <= n^(1/6 - epsilon) for any epsilon > 0, where n represents
the number of vertices in the network. Previous work by Bollobas et al.
formalized the well-known preferential attachment model of Barabasi and
Albert, and showed that the power-law distribution held for d <= n^(1/15)
with gamma = 3. Our revised bound represents a significant improvement
over existing models of degree distribution in scale-free networks,
where its tightness is restricted by the Azuma-Hoeffding concentration
inequality for martingales. We achieve this tighter bound through a
careful analysis of the first set of vertices in the network generation
process, and show that the newly acquired is at the edge of exhausting
Bollobas model in the sense that the degree expectation breaks down
for other powers.}
}
| |||
Abstract:This paper presents a tighter bound on the degree distribution of arbitrary Polya urn graph processes, proving that the proportion of vertices with degree d obeys a power-law distribution P(d) proportional to d^(-gamma) for d <= n^(1/6 - epsilon) for any epsilon > 0, where n represents the number of vertices in the network. Previous work by Bollobas et al. formalized the well-known preferential attachment model of Barabasi and Albert, and showed that the power-law distribution held for d <= n^(1/15) with gamma = 3. Our revised bound represents a significant improvement over existing models of degree distribution in scale-free networks, where its tightness is restricted by the Azuma-Hoeffding concentration inequality for martingales. We achieve this tighter bound through a careful analysis of the first set of vertices in the network generation process, and show that the newly acquired is at the edge of exhausting Bollobas model in the sense that the degree expectation breaks down for other powers. | |||
Refereed Workshops and Symposia
| AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
|---|---|---|---|
| Seungwon Lee, Sima Behpour, & Eric Eaton |
Sharing less is more: Lifelong learning in deep networks with selective layer transfer
Download:
[ PDF ]
|
2020 | Lifelong Learning Workshop at ICML |
| Seungwon Lee, Sima Behpour, & Eric Eaton. Sharing less is more: Lifelong learning in deep networks with selective layer transfer. In 4th Lifelong Learning Workshop at ICML, 2020. | |||
@inproceedings{Lee2020Sharing,
title = {Sharing less is more: Lifelong learning in deep networks with selective layer transfer},
author = {Seungwon Lee and Sima Behpour and Eric Eaton},
booktitle = {4th Lifelong Learning Workshop at ICML},
year = {2020},
abstract = {
Effective lifelong learning across diverse tasks
requires diverse knowledge, yet transferring irrelevant
knowledge may lead to interference and
catastrophic forgetting. In deep networks, transferring
the appropriate granularity of knowledge
is as important as the transfer mechanism, and
must be driven by the relationships among tasks.
We first show that the lifelong learning performance
of several current deep learning architectures
can be significantly improved by transfer
at the appropriate layers. We then develop an
expectation-maximization (EM) method to automatically
select the appropriate transfer configuration
and optimize the task network weights. This
EM-based selective transfer is highly effective,
as demonstrated on three algorithms in several
lifelong object classification scenarios.
},
}
| |||
Abstract:Effective lifelong learning across diverse tasks requires diverse knowledge, yet transferring irrelevant knowledge may lead to interference and catastrophic forgetting. In deep networks, transferring the appropriate granularity of knowledge is as important as the transfer mechanism, and must be driven by the relationships among tasks. We first show that the lifelong learning performance of several current deep learning architectures can be significantly improved by transfer at the appropriate layers. We then develop an expectation-maximization (EM) method to automatically select the appropriate transfer configuration and optimize the task network weights. This EM-based selective transfer is highly effective, as demonstrated on three algorithms in several lifelong object classification scenarios. | |||
| Jorge Mendez & Eric Eaton |
A general framework for continual learning of compositional structures Superseded by the ICLR-21 paper Lifelong learning of compositional structures.
Download:
[ PDF ] [ Supplementary Materials ]
|
2020 | Continual Learning Workshop at ICML |
| Jorge Mendez & Eric Eaton. A general framework for continual learning of compositional structures. In Continual Learning Workshop at ICML, 2020. | |||
@inproceedings{Mendez2020General,
title = {A general framework for continual learning of compositional structures},
author = {Jorge Mendez and Eric Eaton},
booktitle = {Continual Learning Workshop at ICML},
year = {2020},
abstract = {
A hallmark of human intelligence is the ability to construct self-contained chunks of
knowledge and adequately reuse them in novel combinations for solving different
yet structurally related problems. Learning such compositional structures has been
a significant challenge for artificial systems, due to the combinatorial nature of
the underlying search problem. To date, research into compositional learning
has largely proceeded separately from work on lifelong or continual learning.
We integrate these two lines of work to present a general-purpose framework for
lifelong learning of compositional structures that can be used for solving a stream of
related tasks. Our framework separates the learning process into two broad stages:
learning how to best combine existing components in order to assimilate a novel
problem, and learning how to adapt the set of existing components to accommodate
the new problem. This separation explicitly handles the trade-off between the
stability required to remember how to solve earlier tasks and the flexibility required
to solve new tasks, as we show empirically in an extensive evaluation.
},
}
| |||
Abstract:A hallmark of human intelligence is the ability to construct self-contained chunks of knowledge and adequately reuse them in novel combinations for solving different yet structurally related problems. Learning such compositional structures has been a significant challenge for artificial systems, due to the combinatorial nature of the underlying search problem. To date, research into compositional learning has largely proceeded separately from work on lifelong or continual learning. We integrate these two lines of work to present a general-purpose framework for lifelong learning of compositional structures that can be used for solving a stream of related tasks. Our framework separates the learning process into two broad stages: learning how to best combine existing components in order to assimilate a novel problem, and learning how to adapt the set of existing components to accommodate the new problem. This separation explicitly handles the trade-off between the stability required to remember how to solve earlier tasks and the flexibility required to solve new tasks, as we show empirically in an extensive evaluation. | |||
| Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim |
SAR image classification using few-shot cross-domain transfer learning
Download:
[ PDF ]
|
2019 | IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops |
| Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim. SAR image classification using few-shot cross-domain transfer learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2019. | |||
@inproceedings{Rostami2019SAR,
author = {Rostami, Mohammad and Kolouri, Soheil and Eaton, Eric and Kim, Kyungnam},
title = {{SAR} image classification using few-shot cross-domain transfer learning},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2019},
abstract = {
Data-driven classification algorithms based on deep convolutional neural networks (CNNs)
have reached human-level performance for many tasks within Electro-Optical (EO) computer
vision.Despite being the prevailing visual sensory data, EO imaging is not effective in
applications such as environmental monitoring at extended periods, where data collection
at occluded weather is necessary.Synthetic Aperture Radar (SAR) is an effective imaging
tool to circumvent these limitations and collect visual sensory information continually.
However, replicating the success of deep learning on SAR domains is not straightforward.
This is mainly because training deep networks requires huge labeled datasets anddata
labeling is a lot more challenging in SAR domains. We develop an algorithm to transfer
knowledge from EO domains to SAR domains to eliminate the need for huge labeled data
points in the SAR domains. Our idea is to learn a shared domain-invariant embedding for
cross-domain knowledge transfer such that the embedding is discriminative for two related
EO and SAR tasks, while the latent data distributions for both domains remain similar.
As a result, a classifier learned using mostly EO data can generalize well on the related
task for the EO domain.}
}
| |||
Abstract:Data-driven classification algorithms based on deep convolutional neural networks (CNNs) have reached human-level performance for many tasks within Electro-Optical (EO) computer vision.Despite being the prevailing visual sensory data, EO imaging is not effective in applications such as environmental monitoring at extended periods, where data collection at occluded weather is necessary.Synthetic Aperture Radar (SAR) is an effective imaging tool to circumvent these limitations and collect visual sensory information continually. However, replicating the success of deep learning on SAR domains is not straightforward. This is mainly because training deep networks requires huge labeled datasets anddata labeling is a lot more challenging in SAR domains. We develop an algorithm to transfer knowledge from EO domains to SAR domains to eliminate the need for huge labeled data points in the SAR domains. Our idea is to learn a shared domain-invariant embedding for cross-domain knowledge transfer such that the embedding is discriminative for two related EO and SAR tasks, while the latent data distributions for both domains remain similar. As a result, a classifier learned using mostly EO data can generalize well on the related task for the EO domain. | |||
| Eric Eaton |
A lightweight approach to academic research group management using online tools: Spend more time on research and less on management
Download:
[ PDF ] [ Website ]
|
2019 | Educational Advances in Artificial Intelligene (EAAI) Symposium |
| Eric Eaton. A lightweight approach to academic research group management using online tools: Spend more time on research and less on management. In Proceedings of the Educational Advances in Artificial Intelligene (EAAI) Symposium, pp. 9644–9647, 2019. | |||
@inproceedings{Eaton2019Lightweight,
title = {A lightweight approach to academic research group management using online tools: Spend more time on research and less on management},
author = {Eric Eaton},
booktitle = {Proceedings of the Educational Advances in Artificial Intelligene (EAAI) Symposium},
pages = {9644--9647},
year = {2019},
abstract = {
After years of taking a trial-and-error approach to managing a moderate-size academic
research group, I settled on using a set of online tools and protocols that seem
effective, require relatively little effort to use and maintain, and are inexpensive.
This paper discusses this approach to communication, project management, document and
code management, and logistics. It is my hope that other researchers, especially new
faculty and research scientists, might find this set of tools and protocols useful
when determining how to manage their own research group. This paper is targeted toward
research groups based in mathematics and engineering, although faculty in other
disciplines may find inspiration in some of these ideas.
},
}
| |||
Abstract:After years of taking a trial-and-error approach to managing a moderate-size academic research group, I settled on using a set of online tools and protocols that seem effective, require relatively little effort to use and maintain, and are inexpensive. This paper discusses this approach to communication, project management, document and code management, and logistics. It is my hope that other researchers, especially new faculty and research scientists, might find this set of tools and protocols useful when determining how to manage their own research group. This paper is targeted toward research groups based in mathematics and engineering, although faculty in other disciplines may find inspiration in some of these ideas. | |||
| Eric Eaton, Caio Mucchiani, Mayumi Mohan, David Isele, Jose Marcio Luna, & Christopher Clingerman | Design of a low-cost platform for autonomous mobile service robots | 2016 | IJCAI-16 Workshop on Autonomous Mobile Service Robots |
| Eric Eaton, Caio Mucchiani, Mayumi Mohan, David Isele, Jose Marcio Luna, & Christopher Clingerman. Design of a low-cost platform for autonomous mobile service robots. In IJCAI-16 Workshop on Autonomous Mobile Service Robots, July 2016. | |||
@INPROCEEDINGS{Eaton2016Design,
author = {Eric Eaton and Caio Mucchiani and Mayumi Mohan and
David Isele and Jose Marcio Luna and Christopher Clingerman},
year = {2016},
title = {Design of a low-cost platform for autonomous mobile service robots},
booktitle = {IJCAI-16 Workshop on Autonomous Mobile Service Robots},
month = {July},
abstract = {
Most current autonomous mobile service robots are
either expensive commercial platforms or custom
manufactured for research environments, limiting
their availability. We present the design for a low-cost
service robot based on the widely used TurtleBot 2
platform, with the goal of making service
robots affordable and accessible to the research, educational,
and hobbyist communities.
Our design uses a set of simple and inexpensive
modifications to transform the TurtleBot 2 into a
4.5ft (1.37m) tall tour-guide or telepresence-style
robot, capable of performing a wide variety of indoor
service tasks. The resulting platform provides
a shoulder-height touchscreen and 3D camera for
interaction, an optional low-cost arm for manipulation,
enhanced onboard computation, autonomous
charging, and up to 6 hours of runtime. The resulting
platform can support many of the tasks
performed by significantly more expensive service
robots. For compatibility with existing software
packages, the service robot runs the Robot Operating
System (ROS). }
}
| |||
Abstract:Most current autonomous mobile service robots are either expensive commercial platforms or custom manufactured for research environments, limiting their availability. We present the design for a low-cost service robot based on the widely used TurtleBot 2 platform, with the goal of making service robots affordable and accessible to the research, educational, and hobbyist communities. Our design uses a set of simple and inexpensive modifications to transform the TurtleBot 2 into a 4.5ft (1.37m) tall tour-guide or telepresence-style robot, capable of performing a wide variety of indoor service tasks. The resulting platform provides a shoulder-height touchscreen and 3D camera for interaction, an optional low-cost arm for manipulation, enhanced onboard computation, autonomous charging, and up to 6 hours of runtime. The resulting platform can support many of the tasks performed by significantly more expensive service robots. For compatibility with existing software packages, the service robot runs the Robot Operating System (ROS). | |||
| David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor |
Work in Progress: Lifelong Learning for Disturbance Rejection on Mobile Robots Superseded by the IROS-16 paper Lifelong Learning for Disturbance Rejection on Mobile Robots.
Download:
[ PDF ] [ Slides ]
|
2016 | AAMAS'16 Workshop on Adaptive Learning Agents |
| David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor. Work in Progress: Lifelong Learning for Disturbance Rejection on Mobile Robots. In Proceedings of the AAMAS'16 Workshop on Adaptive Learning Agents, May 2016. | |||
@INPROCEEDINGS{Isele2016Work,
author = {David Isele and Jose Marcio Luna and Eric Eaton and
Gabriel V. {de la Cruz} and James Irwin and Brandon Kallaher and Matthew E. Taylor},
year = {2016},
title = {Work in Progress: Lifelong Learning for Disturbance Rejection on Mobile Robots},
booktitle = {Proceedings of the AAMAS'16 Workshop on Adaptive Learning Agents},
month = {May},
abstract = {
No two robots are exactly the same -- even for a given model of robot,
different units will require slightly different controllers. Furthermore,
because robots change and degrade over time, a controller will need to
change over time to remain optimal. This paper leverages lifelong
learning in order to learn controllers for different robots. In particular,
we show that by learning a set of control policies over robots with
different (unknown) motion models, we can quickly adapt to changes in the
robot, or learn a controller for a new robot with a unique set of
disturbances. Further, the approach is completely model-free, allowing us
to apply this method to robots that have not, or cannot, be fully modeled.
These preliminary results are an initial step towards learning robust
fault-tolerant control for arbitrary robots.}
}
| |||
Abstract:No two robots are exactly the same -- even for a given model of robot, different units will require slightly different controllers. Furthermore, because robots change and degrade over time, a controller will need to change over time to remain optimal. This paper leverages lifelong learning in order to learn controllers for different robots. In particular, we show that by learning a set of control policies over robots with different (unknown) motion models, we can quickly adapt to changes in the robot, or learn a controller for a new robot with a unique set of disturbances. Further, the approach is completely model-free, allowing us to apply this method to robots that have not, or cannot, be fully modeled. These preliminary results are an initial step towards learning robust fault-tolerant control for arbitrary robots. | |||
| Haitham Bou Ammar, Eric Eaton, Matthew E. Taylor, Decebal Mocanu, Kurt Driessens, Gerhard Weiss, & Karl Tuyls | An automated measure of MDP similarity for transfer in reinforcement learning | 2014 | AAAI'14 Workshop on Machine Learning for Interactive Systems |
| Haitham Bou Ammar, Eric Eaton, Matthew E. Taylor, Decebal Mocanu, Kurt Driessens, Gerhard Weiss, & Karl Tuyls. An automated measure of MDP similarity for transfer in reinforcement learning. In Proceedings of the AAAI'14 Workshop on Machine Learning for Interactive Systems, July 2014. | |||
@INPROCEEDINGS{BouAmmar2014Automated,
author = {Haitham Bou Ammar and Eric Eaton and Matthew E. Taylor and
Decebal Mocanu and Kurt Driessens and Gerhard Weiss and Karl Tuyls},
year = {2014},
title = {An automated measure of {MDP} similarity for transfer in reinforcement learning},
booktitle = {Proceedings of the AAAI'14 Workshop on Machine Learning for Interactive Systems},
month = {July},
abstract = {
Transfer learning can improve the reinforcement learning
of a new task by allowing the agent to reuse knowledge
acquired from other source tasks. Despite their success,
transfer learning methods rely on having relevant source
tasks; transfer from inappropriate tasks can inhibit
performance on the new task. For fully autonomous
transfer, it is critical to have a method for automatically
choosing relevant source tasks, which requires a similarity
measure between Markov Decision Processes (MDPs). This
issue has received little attention, and is therefore still
a largely open problem.
This paper presents a data-driven automated similarity measure
for MDPs. This novel measure is a significant step toward
autonomous reinforcement learning transfer, allowing agents
to: (1) characterize when transfer will be useful and,
(2) automatically select tasks to use for transfer. The
proposed measure is based on the reconstruction error of a
restricted Boltzmann machine that attempts to model the
behavioral dynamics of the two MDPs being compared. Empirical
results illustrate that this measure is correlated with the
performance of transfer and therefore can be used to identify
similar source tasks for transfer learning.}
}
| |||
Abstract:Transfer learning can improve the reinforcement learning of a new task by allowing the agent to reuse knowledge acquired from other source tasks. Despite their success, transfer learning methods rely on having relevant source tasks; transfer from inappropriate tasks can inhibit performance on the new task. For fully autonomous transfer, it is critical to have a method for automatically choosing relevant source tasks, which requires a similarity measure between Markov Decision Processes (MDPs). This issue has received little attention, and is therefore still a largely open problem. This paper presents a data-driven automated similarity measure for MDPs. This novel measure is a significant step toward autonomous reinforcement learning transfer, allowing agents to: (1) characterize when transfer will be useful and, (2) automatically select tasks to use for transfer. The proposed measure is based on the reconstruction error of a restricted Boltzmann machine that attempts to model the behavioral dynamics of the two MDPs being compared. Empirical results illustrate that this measure is correlated with the performance of transfer and therefore can be used to identify similar source tasks for transfer learning. | |||
| Paul Ruvolo & Eric Eaton |
Online Multi-Task Learning based on K-SVD Superseded by the AAAI-14 paper Online Multi-Task Learning via Sparse Dictionary Optimization.
Download:
[ PDF ]
|
2013 | ICML 2013 Workshop on Theoretically Grounded Transfer Learning |
| Paul Ruvolo & Eric Eaton. Online Multi-Task Learning based on K-SVD. In Proceedings of the ICML 2013 Workshop on Theoretically Grounded Transfer Learning, June 2013. | |||
@INPROCEEDINGS{Ruvolo2013Online,
author = {Paul Ruvolo and Eric Eaton},
title = {Online Multi-Task Learning based on K-SVD},
booktitle = {Proceedings of the ICML 2013 Workshop on Theoretically Grounded Transfer Learning},
year = {2013},
month = {June},
abstract = {
This paper develops an efficient online algorithm
based on K-SVD for learning multiple consecutive
tasks. We first derive a batch multi-task learning
method that builds upon the K-SVD algorithm, and
then extend the batch algorithm to train models
online in a lifelong learning setting. The
resulting method has lower computational complexity
than other current lifelong learning algorithms
while maintaining nearly identical performance.
Additionally, the proposed method offers an
alternate formulation for lifelong learning that
supports both task and feature similarity matrices.}
}
| |||
Abstract:This paper develops an efficient online algorithm based on K-SVD for learning multiple consecutive tasks. We first derive a batch multi-task learning method that builds upon the K-SVD algorithm, and then extend the batch algorithm to train models online in a lifelong learning setting. The resulting method has lower computational complexity than other current lifelong learning algorithms while maintaining nearly identical performance. Additionally, the proposed method offers an alternate formulation for lifelong learning that supports both task and feature similarity matrices. | |||
| Paul Ruvolo & Eric Eaton |
Scalable Lifelong Learning with Active Task Selection Superseded by the AAAI-13 paper Active Task Selection for Lifelong Machine Learning.
Download:
[ PDF ]
|
2013 | AAAI 2013 Spring Symposium on Lifelong Machine Learning |
| Paul Ruvolo & Eric Eaton. Scalable Lifelong Learning with Active Task Selection. In Proceedings of the AAAI 2013 Spring Symposium on Lifelong Machine Learning, March 25-27 2013. | |||
@INPROCEEDINGS{Ruvolo2013Scalable,
author = {Paul Ruvolo and Eric Eaton},
title = {Scalable Lifelong Learning with Active Task Selection},
booktitle = {Proceedings of the AAAI 2013 Spring Symposium on Lifelong Machine Learning},
year = {2013},
location = {Stanford, CA},
month = {March 25-27},
abstract = {
The recently developed Efficient Lifelong Learning
Algorithm (ELLA) acquires knowledge incrementally
over a sequence of tasks, learning a repository of
latent model components that are sparsely shared
between models. ELLA shows strong performance in
comparison to other multi-task learning algorithms,
achieving nearly identical performance to batch
multi-task learning methods while learning tasks
sequentially in three orders of magnitude (over
1,000x) less time. In this paper, we evaluate
several curriculum selection methods that allow
ELLA to actively select the next task for learning
in order to maximize performance on future learning
tasks. Through experiments with three real and one
synthetic data set, we demonstrate that active
curriculum selection allows an agent to learn up to
50% more efficiently than when the agent has no
control over the task order.}
}
| |||
Abstract:The recently developed Efficient Lifelong Learning Algorithm (ELLA) acquires knowledge incrementally over a sequence of tasks, learning a repository of latent model components that are sparsely shared between models. ELLA shows strong performance in comparison to other multi-task learning algorithms, achieving nearly identical performance to batch multi-task learning methods while learning tasks sequentially in three orders of magnitude (over 1,000x) less time. In this paper, we evaluate several curriculum selection methods that allow ELLA to actively select the next task for learning in order to maximize performance on future learning tasks. Through experiments with three real and one synthetic data set, we demonstrate that active curriculum selection allows an agent to learn up to 50% more efficiently than when the agent has no control over the task order. | |||
| Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes |
Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text
Download:
[ PDF ] [ Poster ]
|
2012 | AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-12) |
| Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes. Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text. In Proceedings of the 3rd AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-12), AAAI Press, July 23--24 2012. | |||
@INPROCEEDINGS{Fisher2012IncorporatingEAAI,
author = {Douglas Fisher and Bistra Dilkina and Eric Eaton and Carla Gomes},
title = {Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text},
booktitle = {Proceedings of the 3rd AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-12)},
month = {July 23--24},
location = {Toronto, Canada},
publisher = {AAAI Press},
year = {2012},
abstract = {
We introduce a laboratory text on environmental
and societal sustainability applications that can
be a supplemental resource for any undergraduate
AI course. The lab text, entitled Artificial
Intelligence for Computational Sustainability:
A Lab Companion, is brand new and incomplete;
freely available through Wikibooks; and open to
community additions of projects, assignments,
and explanatory material on AI for sustainability.
The project adds to existing educational efforts
of the computational sustainability community,
encouraging the flow of knowledge from research to
education and public outreach. Besides summarizing
the laboratory book, this paper touches on its
implications for integration of research and
education, for communicating science to the public,
and other broader impacts.}
}
| |||
Abstract:We introduce a laboratory text on environmental and societal sustainability applications that can be a supplemental resource for any undergraduate AI course. The lab text, entitled Artificial Intelligence for Computational Sustainability: A Lab Companion, is brand new and incomplete; freely available through Wikibooks; and open to community additions of projects, assignments, and explanatory material on AI for sustainability. The project adds to existing educational efforts of the computational sustainability community, encouraging the flow of knowledge from research to education and public outreach. Besides summarizing the laboratory book, this paper touches on its implications for integration of research and education, for communicating science to the public, and other broader impacts. | |||
| Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes |
Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text [Oral Presentation]
Download:
[ PDF ]
|
2012 | International Conference on Computational Sustainability (CompSust'12) |
| Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes. Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text [Oral Presentation]. In Proceedings of the 3rd International Conference on Computational Sustainability (CompSust'12), July 5--6 2012. | |||
@INPROCEEDINGS{Fisher2012IncorporatingCompSust,
author = {Douglas Fisher and Bistra Dilkina and Eric Eaton and Carla Gomes},
title = {Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text [Oral Presentation]},
booktitle = {Proceedings of the 3rd International Conference on Computational Sustainability (CompSust'12)},
month = {July 5--6},
location = {Copenhagen, Denmark},
year = {2012},
}
| |||
| Diane Oyen, Eric Eaton, & Terran Lane |
Inferring tasks for improved network structure discovery
Download:
[ PDF ]
|
2012 | Snowbird Learning Workshop |
| Diane Oyen, Eric Eaton, & Terran Lane. Inferring tasks for improved network structure discovery. In Working Notes of the Snowbird Learning Workshop, April 3--6 2012. | |||
@INPROCEEDINGS{Oyen2012Inferring,
author = {Diane Oyen and Eric Eaton and Terran Lane},
title = {Inferring tasks for improved network structure discovery},
booktitle = {Working Notes of the Snowbird Learning Workshop},
month = {April 3--6},
location = {Snowbird, Utah},
year = {2012},
}
| |||
| Eric Eaton & Terran Lane |
The Importance of Selective Knowledge Transfer for Lifelong Learning
Download:
[ PDF ]
|
2011 | AAAI-11 Workshop on Lifelong Learning from Sensorimotor Data |
| Eric Eaton & Terran Lane. The Importance of Selective Knowledge Transfer for Lifelong Learning. In AAAI-11 Workshop on Lifelong Learning from Sensorimotor Data, AAAI Press, August 7 2011. | |||
@INPROCEEDINGS{Eaton2011Importance,
author = {Eric Eaton and Terran Lane},
title = {The Importance of Selective Knowledge Transfer for Lifelong Learning},
booktitle = {AAAI-11 Workshop on Lifelong Learning from Sensorimotor Data},
month = {August 7},
location = {San Francisco, CA},
publisher = {AAAI Press},
year = {2011},
abstract = {
As knowledge transfer research progresses from single
transfer to lifelong learning scenarios, it becomes
increasingly important to properly select the source
knowledge that would best transfer to the target task.
In this position paper, we describe our previous work
on selective knowledge transfer and relate it to
problems in lifelong learning. We also briefly discuss
our ongoing work to develop lifelong learning methods
capable of continual transfer between tasks and the
incorporation of guidance from an expert human user.}
}
| |||
Abstract:As knowledge transfer research progresses from single transfer to lifelong learning scenarios, it becomes increasingly important to properly select the source knowledge that would best transfer to the target task. In this position paper, we describe our previous work on selective knowledge transfer and relate it to problems in lifelong learning. We also briefly discuss our ongoing work to develop lifelong learning methods capable of continual transfer between tasks and the incorporation of guidance from an expert human user. | |||
| Eric Eaton & Marie desJardins |
Set-Based Boosting for Instance-level Transfer Superseded by the AAAI-11 paper Selective Transfer Between Learning Tasks Using Task-Based Boosting.
Download:
[ PDF ]
|
2009 | International Conference on Data Mining Workshop on Transfer Mining |
| Eric Eaton & Marie desJardins. Set-Based Boosting for Instance-level Transfer. In Proceedings of the International Conference on Data Mining Workshop on Transfer Mining, pp. 422–428, IEEE Press, December 2009. | |||
@INPROCEEDINGS{Eaton2009SetBased,
author = {Eric Eaton and Marie desJardins},
title = {Set-Based Boosting for Instance-level Transfer},
booktitle = {Proceedings of the International Conference on Data Mining Workshop on Transfer Mining},
location = {Miami, FL},
publisher = {IEEE Press},
pages = {422--428},
month = {December},
year = {2009},
abstract = { The success of transfer to improve learning on a
target task is highly dependent on the selected source data.
Instance-based transfer methods reuse data from the source
tasks to augment the training data for the target task. If
poorly chosen, this source data may inhibit learning, resulting
in negative transfer. The current best performing algorithm
for instance-based transfer, TrAdaBoost, performs poorly when
given irrelevant source data.
We present a novel set-based boosting technique for instance-based
transfer. The proposed algorithm, TransferBoost, boosts
both individual instances and collective sets of instances from
each source task. In effect, TransferBoost boosts each source
task, assigning higher weight to those source tasks which show
positive transferability to the target task, and then adjusts
the weights of the instances within each source task via
AdaBoost. The results demonstrate that TransferBoost significantly
improves transfer performance over existing instance-based
algorithms when given a mix of relevant and irrelevant source data.}
}
| |||
Abstract:The success of transfer to improve learning on a target task is highly dependent on the selected source data. Instance-based transfer methods reuse data from the source tasks to augment the training data for the target task. If poorly chosen, this source data may inhibit learning, resulting in negative transfer. The current best performing algorithm for instance-based transfer, TrAdaBoost, performs poorly when given irrelevant source data. We present a novel set-based boosting technique for instance-based transfer. The proposed algorithm, TransferBoost, boosts both individual instances and collective sets of instances from each source task. In effect, TransferBoost boosts each source task, assigning higher weight to those source tasks which show positive transferability to the target task, and then adjusts the weights of the instances within each source task via AdaBoost. The results demonstrate that TransferBoost significantly improves transfer performance over existing instance-based algorithms when given a mix of relevant and irrelevant source data. | |||
| Eric Eaton, Gary Holness, & Daniel McFarlane | Interactive Learning using Manifold Geometry Superseded by the AAAI-10 conference paper Interactive Learning using Manifold Geometry. |
2009 | AAAI Fall Symposium on Manifold Learning and Its Applications (AAAI Technical Report FS-09-04) |
| Eric Eaton, Gary Holness, & Daniel McFarlane. Interactive Learning using Manifold Geometry. In Proceedings of the AAAI Fall Symposium on Manifold Learning and Its Applications (AAAI Technical Report FS-09-04), pp. 10–17, AAAI Press, November 5--7 2009. | |||
@INPROCEEDINGS{Eaton2009Interactive,
author = {Eric Eaton and Gary Holness and Daniel McFarlane},
title = {Interactive Learning using Manifold Geometry},
booktitle = {Proceedings of the AAAI Fall Symposium on Manifold Learning and Its Applications (AAAI Technical Report FS-09-04)},
month = {November 5--7},
location = {Arlington, VA},
publisher = {AAAI Press},
pages = {10--17},
year = {2009},
}
| |||
| Eric Eaton |
Gridworld Search and Rescue: A Project Framework for a Course in Artificial Intelligence
Download:
[ PDF ] [ Website ]
|
2008 | AAAI-08 AI Education Colloquium |
| Eric Eaton. Gridworld Search and Rescue: A Project Framework for a Course in Artificial Intelligence. In Proceedings of the AAAI-08 AI Education Colloquium, Chicago, IL, July 13 2008. | |||
@INPROCEEDINGS{Eaton2008Gridworld,
author = {Eric Eaton},
title = {Gridworld Search and Rescue: A Project Framework for a Course in Artificial Intelligence},
booktitle = {Proceedings of the AAAI-08 AI Education Colloquium},
year = {2008},
month = {July 13},
address = {Chicago, IL},
keywords = {education, project framework, search and rescue, gridworld},
abstract = { This paper describes the Gridworld Search and Rescue simulator:
freely available educational software that allows students
to develop an intelligent agent for a search and rescue
application in a partially observable gridworld. It permits students
to focus on high-level AI issues for solving the problem
rather than low-level robotic navigation. The complexity of
the search and rescue problem supports a wide variety of solutions
and AI techniques, including search, logical reasoning,
planning, and machine learning, while the high-level GSAR
simulator makes the complex problem manageable. The simulator
represents a 2D disaster-stricken building for multiple
rescue agents to explore and rescue autonomous injured victims.
It was successfully used as the semester project for
CMSC 471 (Artificial Intelligence) in Fall 2007 at UMBC.}
}
| |||
Abstract:This paper describes the Gridworld Search and Rescue simulator: freely available educational software that allows students to develop an intelligent agent for a search and rescue application in a partially observable gridworld. It permits students to focus on high-level AI issues for solving the problem rather than low-level robotic navigation. The complexity of the search and rescue problem supports a wide variety of solutions and AI techniques, including search, logical reasoning, planning, and machine learning, while the high-level GSAR simulator makes the complex problem manageable. The simulator represents a 2D disaster-stricken building for multiple rescue agents to explore and rescue autonomous injured victims. It was successfully used as the semester project for CMSC 471 (Artificial Intelligence) in Fall 2007 at UMBC. | |||
| Eric Eaton, Marie desJardins, & Terran Lane |
Using functions on a model graph for inductive transfer Superseded by the ECML-08 paper Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer.
Download:
[ PDF ]
|
2008 | Northeast Student Colloquium on Artificial Intelligence (NESCAI-08) |
| Eric Eaton, Marie desJardins, & Terran Lane. Using functions on a model graph for inductive transfer. In Proceedings of the Northeast Student Colloquium on Artificial Intelligence (NESCAI-08), Ithaca, NY, May 2--4 2008. | |||
@inproceedings{Eaton2008Using,
author = {Eric Eaton and Marie desJardins and Terran Lane},
title = {Using functions on a model graph for inductive transfer},
booktitle = {Proceedings of the Northeast Student Colloquium on Artificial Intelligence (NESCAI-08)},
year = {2008},
month = {May 2--4},
address = {Ithaca, NY},
abstract = { In this paper, we propose a novel graph-based
method for knowledge transfer. We embed a set
of learned background models in a graph that
captures the transferability between the models.
We then learn a function on this graph that automatically
determines the parameters to transfer
to each learning task. Transfer to a new problem
proceeds by mapping the problem into the graph,
then using the function to determine the parameters
to transfer in learning the new model. This
method is analogous to inductive transfer along a
manifold that captures the transfer relationships
between the tasks.}
}
| |||
Abstract:In this paper, we propose a novel graph-based method for knowledge transfer. We embed a set of learned background models in a graph that captures the transferability between the models. We then learn a function on this graph that automatically determines the parameters to transfer to each learning task. Transfer to a new problem proceeds by mapping the problem into the graph, then using the function to determine the parameters to transfer in learning the new model. This method is analogous to inductive transfer along a manifold that captures the transfer relationships between the tasks. | |||
| Eric Eaton |
Multi-Resolution Learning for Knowledge Transfer
Download:
[ PDF ]
|
2006 | National Conference on Artificial Intelligence (AAAI) |
| Eric Eaton. Multi-Resolution Learning for Knowledge Transfer. In Proceedings of the 21st National Conference on Artificial Intelligence (AAAI), AAAI Press, Boston, MA, July 16--20 2006. [Doctoral Consortium] | |||
@INPROCEEDINGS{Eaton2006MultiResolution,
author = {Eric Eaton},
title = {Multi-Resolution Learning for Knowledge Transfer},
booktitle = aaai06,
year = {2006},
address = {Boston, MA},
month = {July 16--20},
publisher = {AAAI Press},
note = {[Doctoral Consortium]},
keywords = {knowledge transfer, multiresolution learning, aaai doctoral consortium},
abstract = { Related objects may look similar at low-resolutions;
differences begin to emerge naturally as the resolution
is increased. By learning across multiple resolutions of
input, knowledge can be transfered between related objects.
My dissertation develops this idea and applies it
to the problem of multitask transfer learning.}
}
| |||
Abstract:Related objects may look similar at low-resolutions; differences begin to emerge naturally as the resolution is increased. By learning across multiple resolutions of input, knowledge can be transfered between related objects. My dissertation develops this idea and applies it to the problem of multitask transfer learning. | |||
| Eric Eaton & Marie desJardins |
Knowledge Transfer with a Multiresolution Ensemble of Classifiers
Download:
[ PDF ]
|
2006 | ICML-06 Workshop on Structural Knowledge Transfer for Machine Learning |
| Eric Eaton & Marie desJardins. Knowledge Transfer with a Multiresolution Ensemble of Classifiers. In Proceedings of the ICML-06 Workshop on Structural Knowledge Transfer for Machine Learning, Pittsburgh, PA, June 29 2006. | |||
@INPROCEEDINGS{Eaton2006Knowledge,
author = {Eric Eaton and Marie desJardins},
title = {Knowledge Transfer with a Multiresolution Ensemble of Classifiers},
booktitle = {Proceedings of the ICML-06 Workshop on Structural Knowledge Transfer for Machine Learning},
year = {2006},
address = {Pittsburgh, PA},
month = {June 29},
keywords = {knowledge transfer, multiresolution learning},
abstract = { We demonstrate transfer via an ensemble of
classifiers, where each member focuses on one
resolution of data. Lower-resolution ensemble
members are shared between tasks, providing
a medium for knowledge transfer.}
}
| |||
Abstract:We demonstrate transfer via an ensemble of classifiers, where each member focuses on one resolution of data. Lower-resolution ensemble members are shared between tasks, providing a medium for knowledge transfer. | |||
| Marie desJardins, Eric Eaton, & Kiri Wagstaff |
A context-sensitive and user-centric approach to developing personal assistants
Download:
[ PDF ]
|
2005 | AAAI Spring Symposium on Persistent Assistants |
| Marie desJardins, Eric Eaton, & Kiri Wagstaff. A context-sensitive and user-centric approach to developing personal assistants. In Proceedings of the AAAI Spring Symposium on Persistent Assistants, pp. 98–100, Stanford, CA, March 21--23 2005. | |||
@INPROCEEDINGS{desJardins2005ContextSensitive,
author = {Marie desJardins and Eric Eaton and Kiri Wagstaff},
title = {A context-sensitive and user-centric approach to developing personal assistants},
booktitle = {Proceedings of the AAAI Spring Symposium on Persistent Assistants},
year = {2005},
month = {March 21--23},
address = {Stanford, CA},
pages = {98--100},
abstract = { Several ongoing projects in the MAPLE (Multi-Agent
Planning and LEarning) lab at UMBC and the Machine
Learning Systems Group at JPL focus on problems that
we view as central to the development of persistent
agents. This position paper describes our current research
in this area, focusing on four topics in particular:
effective use of observational and active learning,
utilizing repeated behavioral contexts, clustering with
annotated constraints, and learning user preferences.}
}
| |||
Abstract:Several ongoing projects in the MAPLE (Multi-Agent Planning and LEarning) lab at UMBC and the Machine Learning Systems Group at JPL focus on problems that we view as central to the development of persistent agents. This position paper describes our current research in this area, focusing on four topics in particular: effective use of observational and active learning, utilizing repeated behavioral contexts, clustering with annotated constraints, and learning user preferences. | |||
Refereed Short Papers and Abstracts
| AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
|---|---|---|---|
| David Isele, Eric Eaton, Mark Roberts, & David Aha |
Modeling consecutive task learning with task graph agendas
Download:
[ PDF ]
|
2018 | Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18) |
| David Isele, Eric Eaton, Mark Roberts, & David Aha. Modeling consecutive task learning with task graph agendas. In Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18), 2018. | |||
@inproceedings{Isele2018Modeling,
title = {Modeling consecutive task learning with task graph agendas},
author = {David Isele and Eric Eaton and Mark Roberts and David Aha},
booktitle = {Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18)},
year = {2018},
abstract = {
}
}
| |||
Abstract: | |||
| Vishnu Purushothaman Sreenivasan, Haitham Bou Ammar, & Eric Eaton |
Online Multi-Task Gradient Temporal-Difference Learning
Download:
[ PDF ]
|
2014 | AAAI Conference on Artificial Intelligence (AAAI-14) |
| Vishnu Purushothaman Sreenivasan, Haitham Bou Ammar, & Eric Eaton. Online Multi-Task Gradient Temporal-Difference Learning. In Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14), July 2014. [Student Abstract] | |||
@INPROCEEDINGS{VishnuPS2014Online,
author = {Vishnu Purushothaman Sreenivasan and Haitham Bou Ammar and Eric Eaton},
title = {Online Multi-Task Gradient Temporal-Difference Learning},
note = {[Student Abstract]},
booktitle = {Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14)},
year = {2014},
month = {July},
abstract = {
We develop an online multi-task formulation of
model-based gradient temporal-difference (GTD)
reinforcement learning. Our approach enables
an autonomous RL agent to accumulate knowledge
over its lifetime and efficiently share this
knowledge between tasks to accelerate learning.
Rather than learning a policy for an RL task
tabula rasa, as in standard GTD, our approach
rapidly learns a high performance policy by
building upon the agent's previously learned
knowledge. Our preliminary results on controlling
different mountain car tasks demonstrates that
GTD-ELLA significantly improves learning over
standard GTD(0).}
}
| |||
Abstract:We develop an online multi-task formulation of model-based gradient temporal-difference (GTD) reinforcement learning. Our approach enables an autonomous RL agent to accumulate knowledge over its lifetime and efficiently share this knowledge between tasks to accelerate learning. Rather than learning a policy for an RL task tabula rasa, as in standard GTD, our approach rapidly learns a high performance policy by building upon the agent's previously learned knowledge. Our preliminary results on controlling different mountain car tasks demonstrates that GTD-ELLA significantly improves learning over standard GTD(0). | |||
| Eric Eaton, Marie desJardins, & John Stevenson |
Using multiresolution learning for transfer in image classification
Download:
[ PDF ] [ Poster ]
|
2007 | National Conference on Artificial Intelligence (AAAI) |
| Eric Eaton, Marie desJardins, & John Stevenson. Using multiresolution learning for transfer in image classification. In Proceedings of the 22nd National Conference on Artificial Intelligence (AAAI), AAAI Press, Vancouver, British Columbia, Canada, 2007. [Student Abstract] | |||
@INPROCEEDINGS{Eaton2007Using,
author = {Eric Eaton and Marie desJardins and John Stevenson},
year = {2007},
title = {Using multiresolution learning for transfer in image classification},
booktitle = aaai07,
address = {Vancouver, British Columbia, Canada},
publisher = {AAAI Press},
note = {[Student Abstract]},
abstract = { Our work explores the transfer of knowledge at multiple
levels of abstraction to improve learning. By exploiting
the similarities between objects at various levels of detail,
multiresolution learning can facilitate transfer between
image classification tasks.
We extract features from images at multiple levels of
resolution, then use these features to create models
at different resolutions. Upon receiving a new task,
the closest-matching stored model can be generalized
(adapted to the appropriate resolution) and transferred
to the new task.}
}
| |||
Abstract:Our work explores the transfer of knowledge at multiple levels of abstraction to improve learning. By exploiting the similarities between objects at various levels of detail, multiresolution learning can facilitate transfer between image classification tasks. We extract features from images at multiple levels of resolution, then use these features to create models at different resolutions. Upon receiving a new task, the closest-matching stored model can be generalized (adapted to the appropriate resolution) and transferred to the new task. | |||
Dissertation and Thesis
| AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
|---|---|---|---|
| Eric Eaton | Selective Knowledge Transfer for Machine Learning | 2009 | Ph.D. Thesis, University of Maryland Baltimore County |
| Eric Eaton. Selective Knowledge Transfer for Machine Learning. Ph.D. Thesis, University of Maryland Baltimore County, 2009. | |||
@PHDTHESIS{Eaton2009Selective,
author = {Eric Eaton},
title = {Selective Knowledge Transfer for Machine Learning},
school = {University of Maryland Baltimore County},
year = {2009},
abstract = { Knowledge transfer from previously learned tasks to a new task is a fundamental component
of human learning. Recent work has shown that knowledge transfer can also improve
machine learning, enabling more rapid learning or higher levels of performance.
Transfer allows learning algorithms to reuse knowledge from a set of previously learned
source tasks to improve learning on new target tasks. Proper selection of the source
knowledge to transfer to a given target task is critical to the success of knowledge transfer.
Poorly chosen source knowledge may reduce the effectiveness of transfer, or hinder
learning through a phenomenon known as negative transfer.
This dissertation proposes several methods for source knowledge selection that are
based on the transferability between learning tasks. Transferability is introduced as the
change in performance on a target task between learning with and without transfer. These
methods show that transferability can be used to select source knowledge for two major
types of transfer: instance-based transfer, which reuses individual data instances from the
source tasks, and model-based transfer, which transfers components of previously learned
source models.
For selective instance-based transfer, the proposed TransferBoost algorithm uses a
novel form of set-based boosting to determine the individual source instances to transfer
in learning the target task. TransferBoost reweights instances from each source task based
on their collective transferability to the target task, and then performs regular boosting to
adjust individual instance weights.
For model-based transfer, the learning tasks are organized into a directed network
based on their transfer relationships to each other. Tasks that are close in this network
have high transferability, and tasks that are far apart have low transferability. Model-based
transfer is equivalent to learning a labeling function on this network. This dissertation
proposes the novel Spectral Graph Labeling algorithm that constrains the smoothness of
the learned function using the graph's Laplacian eigenvalues. This method is then applied
to the task transferability network to learn a transfer function that automatically determines
the model parameter values to transfer to a target task. Experiments validate the success
of these methods for selective knowledge transfer, demonstrating significantly improved
performance over existing methods.}
}
| |||
Abstract:Knowledge transfer from previously learned tasks to a new task is a fundamental component of human learning. Recent work has shown that knowledge transfer can also improve machine learning, enabling more rapid learning or higher levels of performance. Transfer allows learning algorithms to reuse knowledge from a set of previously learned source tasks to improve learning on new target tasks. Proper selection of the source knowledge to transfer to a given target task is critical to the success of knowledge transfer. Poorly chosen source knowledge may reduce the effectiveness of transfer, or hinder learning through a phenomenon known as negative transfer. This dissertation proposes several methods for source knowledge selection that are based on the transferability between learning tasks. Transferability is introduced as the change in performance on a target task between learning with and without transfer. These methods show that transferability can be used to select source knowledge for two major types of transfer: instance-based transfer, which reuses individual data instances from the source tasks, and model-based transfer, which transfers components of previously learned source models. For selective instance-based transfer, the proposed TransferBoost algorithm uses a novel form of set-based boosting to determine the individual source instances to transfer in learning the target task. TransferBoost reweights instances from each source task based on their collective transferability to the target task, and then performs regular boosting to adjust individual instance weights. For model-based transfer, the learning tasks are organized into a directed network based on their transfer relationships to each other. Tasks that are close in this network have high transferability, and tasks that are far apart have low transferability. Model-based transfer is equivalent to learning a labeling function on this network. This dissertation proposes the novel Spectral Graph Labeling algorithm that constrains the smoothness of the learned function using the graph's Laplacian eigenvalues. This method is then applied to the task transferability network to learn a transfer function that automatically determines the model parameter values to transfer to a target task. Experiments validate the success of these methods for selective knowledge transfer, demonstrating significantly improved performance over existing methods. | |||
| Eric Eaton |
Clustering with Propagated Constraints
Download:
[ PDF ]
|
2005 | Master's Thesis, University of Maryland Baltimore County |
| Eric Eaton. Clustering with Propagated Constraints. Master's Thesis, University of Maryland Baltimore County, 2005. | |||
@MASTERSTHESIS{Eaton2005MastersThesis,
author = {Eric Eaton},
title = {Clustering with Propagated Constraints},
school = {University of Maryland Baltimore County},
year = {2005},
abstract = { Background knowledge in the form of constraints can dramatically improve the quality
of generated clustering models. In constrained clustering, these constraints typically
specify the relative cluster membership of pairs of points. They are tedious to specify and
expensive from a user perspective, yet are very useful in large quantities. Existing constrained
clustering methods perform well when given large quantities of constraints, but do
not focus on performing well when given very small quantities.
This thesis focuses on providing a high-quality clustering with small quantities of
constraints. It proposes a method for propagating pairwise constraints to nearby instances
using a Gaussian function. This method takes a few easily specified constraints, and propagates
them to nearby pairs of points to constrain the local neighborhood. Clustering with
these propagated constraints can yield superior performance with fewer constraints than
clustering with only the original user-specified constraints. The experiments compare the
performance of clustering with propagated constraints to that of established constrained
clustering algorithms on several real-world data sets.}
}
| |||
Abstract:Background knowledge in the form of constraints can dramatically improve the quality of generated clustering models. In constrained clustering, these constraints typically specify the relative cluster membership of pairs of points. They are tedious to specify and expensive from a user perspective, yet are very useful in large quantities. Existing constrained clustering methods perform well when given large quantities of constraints, but do not focus on performing well when given very small quantities. This thesis focuses on providing a high-quality clustering with small quantities of constraints. It proposes a method for propagating pairwise constraints to nearby instances using a Gaussian function. This method takes a few easily specified constraints, and propagates them to nearby pairs of points to constrain the local neighborhood. Clustering with these propagated constraints can yield superior performance with fewer constraints than clustering with only the original user-specified constraints. The experiments compare the performance of clustering with propagated constraints to that of established constrained clustering algorithms on several real-world data sets. | |||
Technical Reports
Software
| AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
|---|---|---|---|
| Eric Eaton |
Gridworld Search and Rescue Software
Download:
[ Website ]
|
2008 | Available online at: http://maple.cs.umbc.edu/ ericeaton/searchandrescue/ |
| Eric Eaton. Gridworld Search and Rescue Software. Available online at: http://maple.cs.umbc.edu/ ericeaton/searchandrescue/, 2008. | |||
@misc{Eaton2008GridworldSoftware,
author = {Eric Eaton},
year = {2008},
title = {Gridworld Search and Rescue Software},
howpublished = {Available online at: http://maple.cs.umbc.edu/~ericeaton/searchandrescue/},
}
| |||
| Eric Eaton, Marie desJardins, & Kiri Wagstaff |
DDPref Software: Learning preferences for sets of objects
Download:
[ Code ]
|
2006 | Available online at: http://maple.cs.umbc.edu/ ericeaton/software/DDPref.zip |
| Eric Eaton, Marie desJardins, & Kiri Wagstaff. DDPref Software: Learning preferences for sets of objects. Available online at: http://maple.cs.umbc.edu/ ericeaton/software/DDPref.zip, 2006. | |||
@misc{Eaton2006DDPrefSoftware,
author = {Eric Eaton and Marie desJardins and Kiri Wagstaff},
year = {2006},
title = {DDPref Software: Learning preferences for sets of objects},
howpublished = {Available online at: http://maple.cs.umbc.edu/~ericeaton/software/DDPref.zip},
}
| |||
This material is presented to ensure timely dissemination of scholarly work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.
Generated by bib2html.pl (written by Patrick Riley ) on Tue Mar 16, 2021 18:23:17
Website Last Updated: 03/16/21
Copyright © 2002-2013 Eric Eaton.
Copyright restrictions apply to all files under the
website's directory. All rights reserved.
