Research Topic
Lifelong Learning and Multi-task Learning
Transfer Learning
Interactive and Interpretable Learning
Constrained Clustering
Relational Network Analysis
Computational Sustainability
Medicine
Education
Other
Unspecified
Lifelong Learning and Multi-task Learning
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Jorge Mendez & Eric Eaton |
Lifelong learning of compositional structures
Download:
[ PDF ]
|
2021 | International Conference on Learning Representations |
Jorge Mendez & Eric Eaton. Lifelong learning of compositional structures. In International Conference on Learning Representations, 2021. | |||
@inproceedings{Mendez2021Lifelong, title = {Lifelong learning of compositional structures}, author = {Jorge Mendez and Eric Eaton}, booktitle = {International Conference on Learning Representations}, year = {2021}, abstract = { A hallmark of human intelligence is the ability to construct self-contained chunks of knowledge and adequately reuse them in novel combinations for solving different yet structurally related problems. Learning such compositional structures has been a significant challenge for artificial systems, due to the combinatorial nature of the underlying search problem. To date, research into compositional learning has largely proceeded separately from work on lifelong or continual learning. We integrate these two lines of work to present a general-purpose framework for lifelong learning of compositional structures that can be used for solving a stream of related tasks. Our framework separates the learning process into two broad stages: learning how to best combine existing components in order to assimilate a novel problem, and learning how to adapt the set of existing components to accommodate the new problem. This separation explicitly handles the trade-off between the stability required to remember how to solve earlier tasks and the flexibility required to solve new tasks, as we show empirically in an extensive evaluation. }, } | |||
Abstract:A hallmark of human intelligence is the ability to construct self-contained chunks of knowledge and adequately reuse them in novel combinations for solving different yet structurally related problems. Learning such compositional structures has been a significant challenge for artificial systems, due to the combinatorial nature of the underlying search problem. To date, research into compositional learning has largely proceeded separately from work on lifelong or continual learning. We integrate these two lines of work to present a general-purpose framework for lifelong learning of compositional structures that can be used for solving a stream of related tasks. Our framework separates the learning process into two broad stages: learning how to best combine existing components in order to assimilate a novel problem, and learning how to adapt the set of existing components to accommodate the new problem. This separation explicitly handles the trade-off between the stability required to remember how to solve earlier tasks and the flexibility required to solve new tasks, as we show empirically in an extensive evaluation. | |||
Seungwon Lee, Sima Behpour, & Eric Eaton |
Sharing less is more: Lifelong learning in deep networks with selective layer transfer
Download:
[ PDF ]
|
2020 | Lifelong Learning Workshop at ICML |
Seungwon Lee, Sima Behpour, & Eric Eaton. Sharing less is more: Lifelong learning in deep networks with selective layer transfer. In 4th Lifelong Learning Workshop at ICML, 2020. | |||
@inproceedings{Lee2020Sharing, title = {Sharing less is more: Lifelong learning in deep networks with selective layer transfer}, author = {Seungwon Lee and Sima Behpour and Eric Eaton}, booktitle = {4th Lifelong Learning Workshop at ICML}, year = {2020}, abstract = { Effective lifelong learning across diverse tasks requires diverse knowledge, yet transferring irrelevant knowledge may lead to interference and catastrophic forgetting. In deep networks, transferring the appropriate granularity of knowledge is as important as the transfer mechanism, and must be driven by the relationships among tasks. We first show that the lifelong learning performance of several current deep learning architectures can be significantly improved by transfer at the appropriate layers. We then develop an expectation-maximization (EM) method to automatically select the appropriate transfer configuration and optimize the task network weights. This EM-based selective transfer is highly effective, as demonstrated on three algorithms in several lifelong object classification scenarios. }, } | |||
Abstract:Effective lifelong learning across diverse tasks requires diverse knowledge, yet transferring irrelevant knowledge may lead to interference and catastrophic forgetting. In deep networks, transferring the appropriate granularity of knowledge is as important as the transfer mechanism, and must be driven by the relationships among tasks. We first show that the lifelong learning performance of several current deep learning architectures can be significantly improved by transfer at the appropriate layers. We then develop an expectation-maximization (EM) method to automatically select the appropriate transfer configuration and optimize the task network weights. This EM-based selective transfer is highly effective, as demonstrated on three algorithms in several lifelong object classification scenarios. | |||
Jorge Mendez, Boyu Wang, & Eric Eaton |
Lifelong policy gradient learning of factored policies for faster training without forgetting Earlier version was awarded best paper at the ICML'20 Workshop on Lifelong Learning
Download:
[ PDF ]
|
2020 | Advances in Neural Information Processing Systems |
Jorge Mendez, Boyu Wang, & Eric Eaton. Lifelong policy gradient learning of factored policies for faster training without forgetting. In Advances in Neural Information Processing Systems, 2020. | |||
@inproceedings{Mendez2020LifelongPG, title = {Lifelong policy gradient learning of factored policies for faster training without forgetting}, author = {Jorge Mendez and Boyu Wang and Eric Eaton}, booktitle = {Advances in Neural Information Processing Systems}, year = {2020}, abstract = { Policy gradient methods have shown success in learning control policies for highdimensional dynamical systems. Their biggest downside is the amount of exploration they require before yielding high-performing policies. In a lifelong learning setting, in which an agent is faced with multiple consecutive tasks over its lifetime, reusing information from previously seen tasks can substantially accelerate the learning of new tasks. We provide a novel method for lifelong policy gradient learning that trains lifelong function approximators directly via policy gradients, allowing the agent to benefit from accumulated knowledge throughout the entire training process. We show empirically that our algorithm learns faster and converges to better policies than single-task and lifelong learning baselines, and completely avoids catastrophic forgetting on a variety of challenging domains. }, } | |||
Abstract:Policy gradient methods have shown success in learning control policies for highdimensional dynamical systems. Their biggest downside is the amount of exploration they require before yielding high-performing policies. In a lifelong learning setting, in which an agent is faced with multiple consecutive tasks over its lifetime, reusing information from previously seen tasks can substantially accelerate the learning of new tasks. We provide a novel method for lifelong policy gradient learning that trains lifelong function approximators directly via policy gradients, allowing the agent to benefit from accumulated knowledge throughout the entire training process. We show empirically that our algorithm learns faster and converges to better policies than single-task and lifelong learning baselines, and completely avoids catastrophic forgetting on a variety of challenging domains. | |||
Jorge Mendez & Eric Eaton |
A general framework for continual learning of compositional structures Superseded by the ICLR-21 paper Lifelong learning of compositional structures.
Download:
[ PDF ] [ Supplementary Materials ]
|
2020 | Continual Learning Workshop at ICML |
Jorge Mendez & Eric Eaton. A general framework for continual learning of compositional structures. In Continual Learning Workshop at ICML, 2020. | |||
@inproceedings{Mendez2020General, title = {A general framework for continual learning of compositional structures}, author = {Jorge Mendez and Eric Eaton}, booktitle = {Continual Learning Workshop at ICML}, year = {2020}, abstract = { A hallmark of human intelligence is the ability to construct self-contained chunks of knowledge and adequately reuse them in novel combinations for solving different yet structurally related problems. Learning such compositional structures has been a significant challenge for artificial systems, due to the combinatorial nature of the underlying search problem. To date, research into compositional learning has largely proceeded separately from work on lifelong or continual learning. We integrate these two lines of work to present a general-purpose framework for lifelong learning of compositional structures that can be used for solving a stream of related tasks. Our framework separates the learning process into two broad stages: learning how to best combine existing components in order to assimilate a novel problem, and learning how to adapt the set of existing components to accommodate the new problem. This separation explicitly handles the trade-off between the stability required to remember how to solve earlier tasks and the flexibility required to solve new tasks, as we show empirically in an extensive evaluation. }, } | |||
Abstract:A hallmark of human intelligence is the ability to construct self-contained chunks of knowledge and adequately reuse them in novel combinations for solving different yet structurally related problems. Learning such compositional structures has been a significant challenge for artificial systems, due to the combinatorial nature of the underlying search problem. To date, research into compositional learning has largely proceeded separately from work on lifelong or continual learning. We integrate these two lines of work to present a general-purpose framework for lifelong learning of compositional structures that can be used for solving a stream of related tasks. Our framework separates the learning process into two broad stages: learning how to best combine existing components in order to assimilate a novel problem, and learning how to adapt the set of existing components to accommodate the new problem. This separation explicitly handles the trade-off between the stability required to remember how to solve earlier tasks and the flexibility required to solve new tasks, as we show empirically in an extensive evaluation. | |||
Mohammad Rostami, David Isele, & Eric Eaton |
Using task descriptions in lifelong machine learning for improved performance and zero-shot transfer
Download:
[ PDF ] [ Website ]
|
2020 | Journal of Artificial Intelligence Research |
Mohammad Rostami, David Isele, & Eric Eaton. Using task descriptions in lifelong machine learning for improved performance and zero-shot transfer. Journal of Artificial Intelligence Research, 67:673–704, 2020. | |||
@article{Rostami2020Using, title = {Using task descriptions in lifelong machine learning for improved performance and zero-shot transfer}, author = {Mohammad Rostami and David Isele and Eric Eaton}, journal = {Journal of Artificial Intelligence Research}, volume = {67}, pages = {673--704}, year = {2020}, abstract = { Knowledge transfer between tasks can improve the performance of learned models, but requires an accurate estimate of inter-task relationships to identify the relevant knowledge to transfer. These inter-task relationships are typically estimated based on training data for each task, which is inefficient in lifelong learning settings where the goal is to learn each consecutive task rapidly from as little data as possible. To reduce this burden, we develop a lifelong learning method based on coupled dictionary learning that utilizes high-level task descriptions to model inter-task relationships. We show that using task descriptors improves the performance of the learned task policies, providing both theoretical justification for the benefit and empirical demonstration of the improvement across a variety of learning problems. Given only the descriptor for a new task, the lifelong learner is also able to accurately predict a model for the new task through zero-shot learning using the coupled dictionary, eliminating the need to gather training data before addressing the task. }, } | |||
Abstract:Knowledge transfer between tasks can improve the performance of learned models, but requires an accurate estimate of inter-task relationships to identify the relevant knowledge to transfer. These inter-task relationships are typically estimated based on training data for each task, which is inefficient in lifelong learning settings where the goal is to learn each consecutive task rapidly from as little data as possible. To reduce this burden, we develop a lifelong learning method based on coupled dictionary learning that utilizes high-level task descriptions to model inter-task relationships. We show that using task descriptors improves the performance of the learned task policies, providing both theoretical justification for the benefit and empirical demonstration of the improvement across a variety of learning problems. Given only the descriptor for a new task, the lifelong learner is also able to accurately predict a model for the new task through zero-shot learning using the coupled dictionary, eliminating the need to gather training data before addressing the task. | |||
Seungwon Lee, James Stokes, & Eric Eaton |
Learning shared knowledge for deep lifelong learning using deconvolutional networks
Download:
[ PDF ] [ Website ]
|
2019 | Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) |
Seungwon Lee, James Stokes, & Eric Eaton. Learning shared knowledge for deep lifelong learning using deconvolutional networks. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19), pp. 2837–2844, 7 2019. | |||
@inproceedings{Lee2019Learning, title = {Learning shared knowledge for deep lifelong learning using deconvolutional networks}, author = {Lee, Seungwon and Stokes, James and Eaton, Eric}, booktitle = {Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19)}, pages = {2837--2844}, year = {2019}, month = {7}, abstract = { Current mechanisms for knowledge transfer in deep networks tend to either share the lower layers between tasks, or build upon representations trained on other tasks. However, existing work in non-deep multi-task and lifelong learning has shown success with using factorized representations of the model parameter space for transfer, permitting more flexible construction of task models. Inspired by this idea, we introduce a novel architecture for sharing latent factorized representations in convolutional neural networks (CNNs). The proposed approach, called a deconvolutional factorized CNN, uses a combination of deconvolutional factorization and tensor contraction to perform flexible transfer between tasks. Experiments on two computer vision data sets show that the DF-CNN achieves superior performance in challenging lifelong learning settings, resists catastrophic forgetting, and exhibits reverse transfer to improve previously learned tasks from subsequent experience without retraining. }, } | |||
Abstract:Current mechanisms for knowledge transfer in deep networks tend to either share the lower layers between tasks, or build upon representations trained on other tasks. However, existing work in non-deep multi-task and lifelong learning has shown success with using factorized representations of the model parameter space for transfer, permitting more flexible construction of task models. Inspired by this idea, we introduce a novel architecture for sharing latent factorized representations in convolutional neural networks (CNNs). The proposed approach, called a deconvolutional factorized CNN, uses a combination of deconvolutional factorization and tensor contraction to perform flexible transfer between tasks. Experiments on two computer vision data sets show that the DF-CNN achieves superior performance in challenging lifelong learning settings, resists catastrophic forgetting, and exhibits reverse transfer to improve previously learned tasks from subsequent experience without retraining. | |||
David Isele, Eric Eaton, Mark Roberts, & David Aha |
Modeling consecutive task learning with task graph agendas
Download:
[ PDF ]
|
2018 | Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18) |
David Isele, Eric Eaton, Mark Roberts, & David Aha. Modeling consecutive task learning with task graph agendas. In Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18), 2018. | |||
@inproceedings{Isele2018Modeling, title = {Modeling consecutive task learning with task graph agendas}, author = {David Isele and Eric Eaton and Mark Roberts and David Aha}, booktitle = {Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18)}, year = {2018}, abstract = { } } | |||
Abstract: | |||
Jorge A. Mendez, Shashank Shivkumar, & Eric Eaton |
Lifelong inverse reinforcement learning
Download:
[ PDF ]
|
2018 | Neural Information Processing Systems |
Jorge A. Mendez, Shashank Shivkumar, & Eric Eaton. Lifelong inverse reinforcement learning. Neural Information Processing Systems, 2018. | |||
@article{Mendez2018Lifelong, title = {Lifelong inverse reinforcement learning}, author = {Jorge A. Mendez and Shashank Shivkumar and Eric Eaton}, journal = {Neural Information Processing Systems}, year = {2018}, abstract = { Methods for learning from demonstration (LfD) have shown success in acquiring behavior policies by imitating a user. However, even for a single task, LfD may require numerous demonstrations. For versatile agents that must learn many tasks via demonstration, this process would substantially burden the user if each task were learned in isolation. To address this challenge, we introduce the novel problem of lifelong learning from demonstration, which allows the agent to continually build upon knowledge learned from previously demonstrated tasks to accelerate the learning of new tasks, reducing the amount of demonstrations required. As one solution to this problem, we propose the first lifelong learning approach to inverse reinforcement learning, which learns consecutive tasks via demonstration, continually transferring knowledge between tasks to improve performance.} } | |||
Abstract:Methods for learning from demonstration (LfD) have shown success in acquiring behavior policies by imitating a user. However, even for a single task, LfD may require numerous demonstrations. For versatile agents that must learn many tasks via demonstration, this process would substantially burden the user if each task were learned in isolation. To address this challenge, we introduce the novel problem of lifelong learning from demonstration, which allows the agent to continually build upon knowledge learned from previously demonstrated tasks to accelerate the learning of new tasks, reducing the amount of demonstrations required. As one solution to this problem, we propose the first lifelong learning approach to inverse reinforcement learning, which learns consecutive tasks via demonstration, continually transferring knowledge between tasks to improve performance. | |||
Mohammad Rostami, Soheil Kolouri, Kyungnam Kim, & Eric Eaton |
Multi-agent distributed lifelong learning for collective knowledge acquisition
Download:
[ PDF ]
|
2018 | Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18) |
Mohammad Rostami, Soheil Kolouri, Kyungnam Kim, & Eric Eaton. Multi-agent distributed lifelong learning for collective knowledge acquisition. In Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18), 2018. | |||
@inproceedings{Rostami2018MultiAgent, title = {Multi-agent distributed lifelong learning for collective knowledge acquisition}, author = {Mohammad Rostami and Soheil Kolouri and Kyungnam Kim and Eric Eaton}, booktitle = {Proceedings of the Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-18)}, year = {2018}, abstract = { Lifelong machine learning methods acquire knowledge over a series of consecutive tasks, continually building upon their experience. Current lifelong learning algorithms rely upon a single learning agent that has centralized access to all data. In this paper, we extend the idea of lifelong learning from a single agent to a network of multiple agents that collectively learn a series of tasks. Each agent faces some (potentially unique) set of tasks; the key idea is that knowledge learned from these tasks may benefit other agents trying to learn different (but related) tasks. Our Collective Lifelong Learning Algorithm (CoLLA) provides an efficient way for a network of agents to share their learned knowledge in a distributed and decentralized manner, while eliminating the need to share locally observed data. We provide theoretical guarantees for robust performance of the algorithm and empirically demonstrate that CoLLA outperforms existing approaches for distributed multi-task learning on a variety of datasets.} } | |||
Abstract:Lifelong machine learning methods acquire knowledge over a series of consecutive tasks, continually building upon their experience. Current lifelong learning algorithms rely upon a single learning agent that has centralized access to all data. In this paper, we extend the idea of lifelong learning from a single agent to a network of multiple agents that collectively learn a series of tasks. Each agent faces some (potentially unique) set of tasks; the key idea is that knowledge learned from these tasks may benefit other agents trying to learn different (but related) tasks. Our Collective Lifelong Learning Algorithm (CoLLA) provides an efficient way for a network of agents to share their learned knowledge in a distributed and decentralized manner, while eliminating the need to share locally observed data. We provide theoretical guarantees for robust performance of the algorithm and empirically demonstrate that CoLLA outperforms existing approaches for distributed multi-task learning on a variety of datasets. | |||
Decebal Constantin Mocanu, Haitham Bou Ammar, Luis Puig, Eric Eaton, & Antonio Liotta |
Estimating 3D trajectories from 2D projections via disjunctive factored four-way conditional restricted Boltzmann machines
Download:
[ Website ]
|
2017 | Pattern Recognition |
Decebal Constantin Mocanu, Haitham Bou Ammar, Luis Puig, Eric Eaton, & Antonio Liotta. Estimating 3D trajectories from 2D projections via disjunctive factored four-way conditional restricted Boltzmann machines. Pattern Recognition, 69:325–335, September 2017. | |||
@article{Mocanu2017Estimating, title = {Estimating 3D trajectories from 2D projections via disjunctive factored four-way conditional restricted Boltzmann machines}, author = {Decebal Constantin Mocanu and Haitham Bou Ammar and Luis Puig and Eric Eaton and Antonio Liotta}, journal = {Pattern Recognition}, volume = {69}, pages = {325--335}, month = {September}, year = {2017}, abstract = { Estimation, recognition, and near-future prediction of 3D trajectories based on their two dimensional projections available from one camera source is an exceptionally difficult problem due to uncertainty in the trajectories and environment, high dimensionality of the specific trajectory states, lack of enough labeled data and so on. In this article, we propose a solution to solve this problem based on a novel deep learning model dubbed disjunctive factored four-way conditional restricted Boltzmann machine (DFFW-CRBM). Our method improves state-of-the-art deep learning techniques for high dimensional time-series modeling by introducing a novel tensor factorization capable of driving forth order Boltzmann machines to considerably lower energy levels, at no computational costs. DFFW-CRBMs are capable of accurately estimating, recognizing, and performing near-future prediction of three-dimensional trajectories from their 2D projections while requiring limited amount of labeled data. We evaluate our method on both simulated and real-world data, showing its effectiveness in predicting and classifying complex ball trajectories and human activities. } } | |||
Abstract:Estimation, recognition, and near-future prediction of 3D trajectories based on their two dimensional projections available from one camera source is an exceptionally difficult problem due to uncertainty in the trajectories and environment, high dimensionality of the specific trajectory states, lack of enough labeled data and so on. In this article, we propose a solution to solve this problem based on a novel deep learning model dubbed disjunctive factored four-way conditional restricted Boltzmann machine (DFFW-CRBM). Our method improves state-of-the-art deep learning techniques for high dimensional time-series modeling by introducing a novel tensor factorization capable of driving forth order Boltzmann machines to considerably lower energy levels, at no computational costs. DFFW-CRBMs are capable of accurately estimating, recognizing, and performing near-future prediction of three-dimensional trajectories from their 2D projections while requiring limited amount of labeled data. We evaluate our method on both simulated and real-world data, showing its effectiveness in predicting and classifying complex ball trajectories and human activities. | |||
Christopher Clingerman & Eric Eaton |
Lifelong machine learning with Gaussian processes
Download:
[ PDF ]
|
2017 | European Conference on Machine Learning & Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD-17) |
Christopher Clingerman & Eric Eaton. Lifelong machine learning with Gaussian processes. In Proceedings of the European Conference on Machine Learning & Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD-17), 2017. | |||
@inproceedings{Clingerman2017Lifelong, title = {Lifelong machine learning with Gaussian processes}, author = {Christopher Clingerman and Eric Eaton}, booktitle = {Proceedings of the European Conference on Machine Learning \& Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD-17)}, year = {2017}, abstract = { Recent developments in lifelong machine learning have demonstrated that it is possible to learn multiple tasks consecutively, transferring knowledge between those tasks to accelerate learning and improve performance. However, these methods are limited to using linear parametric base learners, substantially restricting the predictive power of the resulting models. We present a lifelong learning algorithm that can support nonparametric models, focusing on Gaussian processes. To enable efficient online transfer between Gaussian process models, our approach assumes a factorized formulation of the covariance functions, and incrementally learns a shared sparse basis for the models’ parameterizations. We show that this lifelong learning approach is highly computationally efficient, and outperforms existing methods on a variety of data sets. } } | |||
Abstract:Recent developments in lifelong machine learning have demonstrated that it is possible to learn multiple tasks consecutively, transferring knowledge between those tasks to accelerate learning and improve performance. However, these methods are limited to using linear parametric base learners, substantially restricting the predictive power of the resulting models. We present a lifelong learning algorithm that can support nonparametric models, focusing on Gaussian processes. To enable efficient online transfer between Gaussian process models, our approach assumes a factorized formulation of the covariance functions, and incrementally learns a shared sparse basis for the models’ parameterizations. We show that this lifelong learning approach is highly computationally efficient, and outperforms existing methods on a variety of data sets. | |||
David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor | Lifelong Learning for Disturbance Rejection on Mobile Robots | 2016 | International Conference on Intelligent Robots and Systems (IROS-16) |
David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor. Lifelong Learning for Disturbance Rejection on Mobile Robots. In Proceedings of the International Conference on Intelligent Robots and Systems (IROS-16), IEEE/RSJ, October 2016. | |||
@INPROCEEDINGS{Isele2016Lifelong, author = {David Isele and Jose Marcio Luna and Eric Eaton and Gabriel V. {de la Cruz} and James Irwin and Brandon Kallaher and Matthew E. Taylor}, year = {2016}, title = {Lifelong Learning for Disturbance Rejection on Mobile Robots}, booktitle = {Proceedings of the International Conference on Intelligent Robots and Systems (IROS-16)}, month = {October}, publisher = {IEEE/RSJ}, abstract = { No two robots are exactly the same---even for a given model of robot, different units will require slightly different controllers. Furthermore, because robots change and degrade over time, a controller will need to change over time to remain optimal. This paper leverages lifelong learning in order to learn controllers for different robots. In particular, we show that by learning a set of control policies over robots with different (unknown) motion models, we can quickly adapt to changes in the robot, or learn a controller for a new robot with a unique set of disturbances. Furthermore, the approach is completely model-free, allowing us to apply this method to robots that have not, or cannot, be fully modeled.} } | |||
Abstract:No two robots are exactly the same---even for a given model of robot, different units will require slightly different controllers. Furthermore, because robots change and degrade over time, a controller will need to change over time to remain optimal. This paper leverages lifelong learning in order to learn controllers for different robots. In particular, we show that by learning a set of control policies over robots with different (unknown) motion models, we can quickly adapt to changes in the robot, or learn a controller for a new robot with a unique set of disturbances. Furthermore, the approach is completely model-free, allowing us to apply this method to robots that have not, or cannot, be fully modeled. | |||
David Isele, Mohammad Rostami, & Eric Eaton |
Using task features for zero-shot knowledge transfer in lifelong learning Awarded sole IJCAI-16 Distinguished Student Paper |
2016 | International Joint Conference on Artificial Intelligence (IJCAI-16) |
David Isele, Mohammad Rostami, & Eric Eaton. Using task features for zero-shot knowledge transfer in lifelong learning. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-16), July 2016. | |||
@INPROCEEDINGS{Isele2016Using, author = {David Isele and Mohammad Rostami and Eric Eaton}, year = {2016}, title = {Using task features for zero-shot knowledge transfer in lifelong learning}, booktitle = {Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-16)}, month = {July}, abstract = { Knowledge transfer between tasks can improve the performance of learned models, but requires an accurate estimate of the inter-task relationships to identify the relevant knowledge to transfer. These inter-task relationships are typically estimated based on training data for each task, which is inefficient in lifelong learning settings where the goal is to learn each consecutive task rapidly from as little data as possible. To reduce this burden, we develop a lifelong reinforcement learning method based on coupled dictionary learning that incorporates high-level task descriptors to model the inter-task relationships. We show that using task descriptors improves the performance of the learned task policies, providing both theoretical justification for the benefit and empirical demonstration of the improvement across a variety of dynamical control problems. Given only the descriptor for a new task, the lifelong learner is also able to accurately predict the task policy through zero-shot learning using the coupled dictionary, eliminating the need to pause to gather training data before addressing the task.} } | |||
Abstract:Knowledge transfer between tasks can improve the performance of learned models, but requires an accurate estimate of the inter-task relationships to identify the relevant knowledge to transfer. These inter-task relationships are typically estimated based on training data for each task, which is inefficient in lifelong learning settings where the goal is to learn each consecutive task rapidly from as little data as possible. To reduce this burden, we develop a lifelong reinforcement learning method based on coupled dictionary learning that incorporates high-level task descriptors to model the inter-task relationships. We show that using task descriptors improves the performance of the learned task policies, providing both theoretical justification for the benefit and empirical demonstration of the improvement across a variety of dynamical control problems. Given only the descriptor for a new task, the lifelong learner is also able to accurately predict the task policy through zero-shot learning using the coupled dictionary, eliminating the need to pause to gather training data before addressing the task. | |||
David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor |
Work in Progress: Lifelong Learning for Disturbance Rejection on Mobile Robots Superseded by the IROS-16 paper Lifelong Learning for Disturbance Rejection on Mobile Robots.
Download:
[ PDF ] [ Slides ]
|
2016 | AAMAS'16 Workshop on Adaptive Learning Agents |
David Isele, Jose Marcio Luna, Eric Eaton, Gabriel V. de la Cruz, James Irwin, Brandon Kallaher, & Matthew E. Taylor. Work in Progress: Lifelong Learning for Disturbance Rejection on Mobile Robots. In Proceedings of the AAMAS'16 Workshop on Adaptive Learning Agents, May 2016. | |||
@INPROCEEDINGS{Isele2016Work, author = {David Isele and Jose Marcio Luna and Eric Eaton and Gabriel V. {de la Cruz} and James Irwin and Brandon Kallaher and Matthew E. Taylor}, year = {2016}, title = {Work in Progress: Lifelong Learning for Disturbance Rejection on Mobile Robots}, booktitle = {Proceedings of the AAMAS'16 Workshop on Adaptive Learning Agents}, month = {May}, abstract = { No two robots are exactly the same -- even for a given model of robot, different units will require slightly different controllers. Furthermore, because robots change and degrade over time, a controller will need to change over time to remain optimal. This paper leverages lifelong learning in order to learn controllers for different robots. In particular, we show that by learning a set of control policies over robots with different (unknown) motion models, we can quickly adapt to changes in the robot, or learn a controller for a new robot with a unique set of disturbances. Further, the approach is completely model-free, allowing us to apply this method to robots that have not, or cannot, be fully modeled. These preliminary results are an initial step towards learning robust fault-tolerant control for arbitrary robots.} } | |||
Abstract:No two robots are exactly the same -- even for a given model of robot, different units will require slightly different controllers. Furthermore, because robots change and degrade over time, a controller will need to change over time to remain optimal. This paper leverages lifelong learning in order to learn controllers for different robots. In particular, we show that by learning a set of control policies over robots with different (unknown) motion models, we can quickly adapt to changes in the robot, or learn a controller for a new robot with a unique set of disturbances. Further, the approach is completely model-free, allowing us to apply this method to robots that have not, or cannot, be fully modeled. These preliminary results are an initial step towards learning robust fault-tolerant control for arbitrary robots. | |||
Haitham Bou Ammar, Rasul Tutunov, & Eric Eaton | Safe policy search for lifelong reinforcement learning with sublinear regret | 2015 | International Conference on Machine Learning (ICML-15) |
Haitham Bou Ammar, Rasul Tutunov, & Eric Eaton. Safe policy search for lifelong reinforcement learning with sublinear regret. In Proceedings of the 32nd International Conference on Machine Learning (ICML-15), July 2015. | |||
@INPROCEEDINGS{BouAmmar2015Safe, author = {Haitham {Bou Ammar} and Rasul Tutunov and Eric Eaton}, year = {2015}, title = {Safe policy search for lifelong reinforcement learning with sublinear regret}, booktitle = {Proceedings of the 32nd International Conference on Machine Learning (ICML-15)}, month = {July}, abstract = { Lifelong reinforcement learning provides a promising framework for developing versatile agents that can accumulate knowledge over a lifetime of experience and rapidly learn new tasks by building upon prior knowledge. However, current lifelong learning methods exhibit non-vanishing regret as the amount of experience increases, and include limitations that can lead to suboptimal or unsafe control policies. To address these issues, we develop a lifelong policy gradient learner that operates in an adversarial setting to learn multiple tasks online while enforcing safety constraints on the learned policies. We demonstrate, for the first time, sublinear regret for lifelong policy search, and validate our algorithm on several benchmark dynamical systems and an application to quadrotor control.} } | |||
Abstract:Lifelong reinforcement learning provides a promising framework for developing versatile agents that can accumulate knowledge over a lifetime of experience and rapidly learn new tasks by building upon prior knowledge. However, current lifelong learning methods exhibit non-vanishing regret as the amount of experience increases, and include limitations that can lead to suboptimal or unsafe control policies. To address these issues, we develop a lifelong policy gradient learner that operates in an adversarial setting to learn multiple tasks online while enforcing safety constraints on the learned policies. We demonstrate, for the first time, sublinear regret for lifelong policy search, and validate our algorithm on several benchmark dynamical systems and an application to quadrotor control. | |||
Haitham Bou Ammar, Eric Eaton, Jose Marcio Luna, & Paul Ruvolo |
Autonomous cross-domain knowledge transfer in lifelong policy gradient reinforcement learning Finalist for IJCAI-15 Distinguished Paper award
Download:
[ PDF ]
|
2015 | International Joint Conference on Artificial Intelligence (IJCAI-15) |
Haitham Bou Ammar, Eric Eaton, Jose Marcio Luna, & Paul Ruvolo. Autonomous cross-domain knowledge transfer in lifelong policy gradient reinforcement learning. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-15), July 2015. | |||
@INPROCEEDINGS{BouAmmar2015Autonomous, author = {Haitham {Bou Ammar} and Eric Eaton and Jose Marcio Luna and Paul Ruvolo}, year = {2015}, title = {Autonomous cross-domain knowledge transfer in lifelong policy gradient reinforcement learning}, booktitle = {Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI-15)}, month = {July}, abstract = { Online multi-task learning is an important capability for lifelong learning agents, enabling them to acquire models for diverse tasks over time and rapidly learn new tasks by building upon prior experience. However, recent progress toward lifelong reinforcement learning (RL) has been limited to learning from within a single task domain. For truly versatile lifelong learning, the agent must be able to autonomously transfer knowledge between different task domains. A few methods for cross-domain transfer have been developed, but these methods are computationally inefficient for scenarios where the agent must learn tasks consecutively. In this paper, we develop the first cross-domain lifelong RL framework. Our approach efficiently optimizes a shared repository of transferable knowledge and learns projection matrices that specialize that knowledge to different task domains. We provide rigorous theoretical guarantees on the stability of this approach, and empirically evaluate its performance on diverse dynamical systems. Our results show that the proposed method can learn effectively from interleaved task domains and rapidly acquire high performance in new domains.} } | |||
Abstract:Online multi-task learning is an important capability for lifelong learning agents, enabling them to acquire models for diverse tasks over time and rapidly learn new tasks by building upon prior experience. However, recent progress toward lifelong reinforcement learning (RL) has been limited to learning from within a single task domain. For truly versatile lifelong learning, the agent must be able to autonomously transfer knowledge between different task domains. A few methods for cross-domain transfer have been developed, but these methods are computationally inefficient for scenarios where the agent must learn tasks consecutively. In this paper, we develop the first cross-domain lifelong RL framework. Our approach efficiently optimizes a shared repository of transferable knowledge and learns projection matrices that specialize that knowledge to different task domains. We provide rigorous theoretical guarantees on the stability of this approach, and empirically evaluate its performance on diverse dynamical systems. Our results show that the proposed method can learn effectively from interleaved task domains and rapidly acquire high performance in new domains. | |||
Paul Ruvolo & Eric Eaton | Online Multi-Task Learning via Sparse Dictionary Optimization | 2014 | AAAI Conference on Artificial Intelligence (AAAI-14) |
Paul Ruvolo & Eric Eaton. Online Multi-Task Learning via Sparse Dictionary Optimization. In Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14), July 2014. | |||
@INPROCEEDINGS{Ruvolo2014Online, author = {Paul Ruvolo and Eric Eaton}, title = {Online Multi-Task Learning via Sparse Dictionary Optimization}, booktitle = {Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14)}, year = {2014}, month = {July}, abstract = { This paper develops an efficient online algorithm for learning multiple consecutive tasks based on the K-SVD algorithm for sparse dictionary optimization. We first derive a batch multi-task learning method that builds upon K-SVD, and then extend the batch algorithm to train models online in a lifelong learning setting. The resulting method has lower computational complexity than other current lifelong learning algorithms while maintaining nearly identical model performance. Additionally, the proposed method offers an alternate formulation for lifelong learning that supports both task and feature similarity matrices.} } | |||
Abstract:This paper develops an efficient online algorithm for learning multiple consecutive tasks based on the K-SVD algorithm for sparse dictionary optimization. We first derive a batch multi-task learning method that builds upon K-SVD, and then extend the batch algorithm to train models online in a lifelong learning setting. The resulting method has lower computational complexity than other current lifelong learning algorithms while maintaining nearly identical model performance. Additionally, the proposed method offers an alternate formulation for lifelong learning that supports both task and feature similarity matrices. | |||
Vishnu Purushothaman Sreenivasan, Haitham Bou Ammar, & Eric Eaton |
Online Multi-Task Gradient Temporal-Difference Learning
Download:
[ PDF ]
|
2014 | AAAI Conference on Artificial Intelligence (AAAI-14) |
Vishnu Purushothaman Sreenivasan, Haitham Bou Ammar, & Eric Eaton. Online Multi-Task Gradient Temporal-Difference Learning. In Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14), July 2014. [Student Abstract] | |||
@INPROCEEDINGS{VishnuPS2014Online, author = {Vishnu Purushothaman Sreenivasan and Haitham Bou Ammar and Eric Eaton}, title = {Online Multi-Task Gradient Temporal-Difference Learning}, note = {[Student Abstract]}, booktitle = {Proceedings of the 28th AAAI Conference on Artificial Intelligence (AAAI-14)}, year = {2014}, month = {July}, abstract = { We develop an online multi-task formulation of model-based gradient temporal-difference (GTD) reinforcement learning. Our approach enables an autonomous RL agent to accumulate knowledge over its lifetime and efficiently share this knowledge between tasks to accelerate learning. Rather than learning a policy for an RL task tabula rasa, as in standard GTD, our approach rapidly learns a high performance policy by building upon the agent's previously learned knowledge. Our preliminary results on controlling different mountain car tasks demonstrates that GTD-ELLA significantly improves learning over standard GTD(0).} } | |||
Abstract:We develop an online multi-task formulation of model-based gradient temporal-difference (GTD) reinforcement learning. Our approach enables an autonomous RL agent to accumulate knowledge over its lifetime and efficiently share this knowledge between tasks to accelerate learning. Rather than learning a policy for an RL task tabula rasa, as in standard GTD, our approach rapidly learns a high performance policy by building upon the agent's previously learned knowledge. Our preliminary results on controlling different mountain car tasks demonstrates that GTD-ELLA significantly improves learning over standard GTD(0). | |||
Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor |
Online Multi-Task Learning for Policy Gradient Methods
Download:
[ PDF ]
|
2014 | International Conference on Machine Learning (ICML-14) |
Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor. Online Multi-Task Learning for Policy Gradient Methods. In Proceedings of the 31st International Conference on Machine Learning (ICML-14), June 2014. | |||
@INPROCEEDINGS{BouAmmar2014Online, author = {Haitham Bou Ammar and Eric Eaton and Paul Ruvolo and Matthew E. Taylor}, title = {Online Multi-Task Learning for Policy Gradient Methods}, booktitle = {Proceedings of the 31st International Conference on Machine Learning (ICML-14)}, year = {2014}, month = {June}, abstract = { Policy gradient algorithms have shown considerable recent success in solving high-dimensional sequential decision-making (SDM) tasks, particularly in robotics. However, these methods often require extensive experience in a domain to achieve high performance. To make agents more sample-efficient, we developed a multi-task policy gradient method to learn SDM tasks consecutively, transferring knowledge between tasks to accelerate learning. Our approach provides robust theoretical guarantees and we show empirically that it dramatically accelerates learning on a variety of dynamical systems, including an application to quadcopter control.} } | |||
Abstract:Policy gradient algorithms have shown considerable recent success in solving high-dimensional sequential decision-making (SDM) tasks, particularly in robotics. However, these methods often require extensive experience in a domain to achieve high performance. To make agents more sample-efficient, we developed a multi-task policy gradient method to learn SDM tasks consecutively, transferring knowledge between tasks to accelerate learning. Our approach provides robust theoretical guarantees and we show empirically that it dramatically accelerates learning on a variety of dynamical systems, including an application to quadcopter control. | |||
Paul Ruvolo & Eric Eaton | Active Task Selection for Lifelong Machine Learning | 2013 | AAAI Conference on Artificial Intelligence (AAAI-13) |
Paul Ruvolo & Eric Eaton. Active Task Selection for Lifelong Machine Learning. In Proceedings of the 27th AAAI Conference on Artificial Intelligence (AAAI-13), July 2013. | |||
@INPROCEEDINGS{Ruvolo2013Active, author = {Paul Ruvolo and Eric Eaton}, title = {Active Task Selection for Lifelong Machine Learning}, booktitle = {Proceedings of the 27th AAAI Conference on Artificial Intelligence (AAAI-13)}, year = {2013}, month = {July}, abstract = { In a lifelong learning framework, an agent acquires knowledge incrementally over consecutive learning tasks, continually building upon its experience. Recent lifelong learning algorithms have achieved nearly identical performance to batch multi-task learning methods while reducing learning time by three orders of magnitude. In this paper, we further improve the scalability of lifelong learning by developing curriculum selection methods that enable an agent to actively select the next task to learn in order to maximize performance on future learning tasks. We demonstrate that active task selection is highly reliable and effective, allowing an agent to learn high performance models using up to 50% fewer tasks than when the agent has no control over the task order. We also explore a variant of transfer learning in the lifelong learning setting in which the agent can focus knowledge acquisition toward a particular target task.} } | |||
Abstract:In a lifelong learning framework, an agent acquires knowledge incrementally over consecutive learning tasks, continually building upon its experience. Recent lifelong learning algorithms have achieved nearly identical performance to batch multi-task learning methods while reducing learning time by three orders of magnitude. In this paper, we further improve the scalability of lifelong learning by developing curriculum selection methods that enable an agent to actively select the next task to learn in order to maximize performance on future learning tasks. We demonstrate that active task selection is highly reliable and effective, allowing an agent to learn high performance models using up to 50% fewer tasks than when the agent has no control over the task order. We also explore a variant of transfer learning in the lifelong learning setting in which the agent can focus knowledge acquisition toward a particular target task. | |||
Paul Ruvolo & Eric Eaton | ELLA: An Efficient Lifelong Learning Algorithm | 2013 | International Conference on Machine Learning (ICML-13) |
Paul Ruvolo & Eric Eaton. ELLA: An Efficient Lifelong Learning Algorithm. In Proceedings of the 30th International Conference on Machine Learning (ICML-13), June 2013. | |||
@INPROCEEDINGS{Ruvolo2013ELLA, author = {Paul Ruvolo and Eric Eaton}, title = {ELLA: An Efficient Lifelong Learning Algorithm}, booktitle = {Proceedings of the 30th International Conference on Machine Learning (ICML-13)}, year = {2013}, month = {June}, abstract = { The problem of learning multiple consecutive tasks, known as \emph{lifelong learning}, is of great importance to the creation of intelligent, general-purpose, and flexible machines. In this paper, we develop a method for online multi-task learning in the lifelong learning setting. The proposed Efficient Lifelong Learning Algorithm (ELLA) maintains a sparsely shared basis for all task models, transfers knowledge from the basis to learn each new task, and refines the basis over time to maximize performance across all tasks. We show that ELLA has strong connections to both online dictionary learning for sparse coding and state-of-the-art batch multi-task learning methods, and provide robust theoretical performance guarantees. We show empirically that ELLA yields nearly identical performance to batch multi-task learning while learning tasks sequentially in three orders of magnitude (over 1,000x) less time.} } | |||
Abstract:The problem of learning multiple consecutive tasks, known as <i>lifelong learning</i>, is of great importance to the creation of intelligent, general-purpose, and flexible machines. In this paper, we develop a method for online multi-task learning in the lifelong learning setting. The proposed Efficient Lifelong Learning Algorithm (ELLA) maintains a sparsely shared basis for all task models, transfers knowledge from the basis to learn each new task, and refines the basis over time to maximize performance across all tasks. We show that ELLA has strong connections to both online dictionary learning for sparse coding and state-of-the-art batch multi-task learning methods, and provide robust theoretical performance guarantees. We show empirically that ELLA yields nearly identical performance to batch multi-task learning while learning tasks sequentially in three orders of magnitude (over 1,000x) less time. | |||
Paul Ruvolo & Eric Eaton |
Online Multi-Task Learning based on K-SVD Superseded by the AAAI-14 paper Online Multi-Task Learning via Sparse Dictionary Optimization.
Download:
[ PDF ]
|
2013 | ICML 2013 Workshop on Theoretically Grounded Transfer Learning |
Paul Ruvolo & Eric Eaton. Online Multi-Task Learning based on K-SVD. In Proceedings of the ICML 2013 Workshop on Theoretically Grounded Transfer Learning, June 2013. | |||
@INPROCEEDINGS{Ruvolo2013Online, author = {Paul Ruvolo and Eric Eaton}, title = {Online Multi-Task Learning based on K-SVD}, booktitle = {Proceedings of the ICML 2013 Workshop on Theoretically Grounded Transfer Learning}, year = {2013}, month = {June}, abstract = { This paper develops an efficient online algorithm based on K-SVD for learning multiple consecutive tasks. We first derive a batch multi-task learning method that builds upon the K-SVD algorithm, and then extend the batch algorithm to train models online in a lifelong learning setting. The resulting method has lower computational complexity than other current lifelong learning algorithms while maintaining nearly identical performance. Additionally, the proposed method offers an alternate formulation for lifelong learning that supports both task and feature similarity matrices.} } | |||
Abstract:This paper develops an efficient online algorithm based on K-SVD for learning multiple consecutive tasks. We first derive a batch multi-task learning method that builds upon the K-SVD algorithm, and then extend the batch algorithm to train models online in a lifelong learning setting. The resulting method has lower computational complexity than other current lifelong learning algorithms while maintaining nearly identical performance. Additionally, the proposed method offers an alternate formulation for lifelong learning that supports both task and feature similarity matrices. | |||
Eric Eaton (chair) |
Lifelong Machine Learning: Proceedings of the 2013 AAAI Spring Symposium
Download:
[ Website ]
|
2013 | AAAI Press |
Eric Eaton (chair). Lifelong Machine Learning: Proceedings of the 2013 AAAI Spring Symposium, AAAI Technical Report SS-13-05, AAAI Press, May 2013. | |||
@book{Eaton2013AAAISSS, author = {Eric Eaton (chair)}, title = {Lifelong Machine Learning: Proceedings of the 2013 AAAI Spring Symposium}, series = {AAAI Technical Report SS-13-05} month = {May}, year = {2013}, publisher = {AAAI Press}, isbn = {ISBN 978-1-57735-602-8} url = {http://www.aaai.org/Press/Reports/Symposia/Spring/ss-13-05.php} } | |||
Paul Ruvolo & Eric Eaton |
Scalable Lifelong Learning with Active Task Selection Superseded by the AAAI-13 paper Active Task Selection for Lifelong Machine Learning.
Download:
[ PDF ]
|
2013 | AAAI 2013 Spring Symposium on Lifelong Machine Learning |
Paul Ruvolo & Eric Eaton. Scalable Lifelong Learning with Active Task Selection. In Proceedings of the AAAI 2013 Spring Symposium on Lifelong Machine Learning, March 25-27 2013. | |||
@INPROCEEDINGS{Ruvolo2013Scalable, author = {Paul Ruvolo and Eric Eaton}, title = {Scalable Lifelong Learning with Active Task Selection}, booktitle = {Proceedings of the AAAI 2013 Spring Symposium on Lifelong Machine Learning}, year = {2013}, location = {Stanford, CA}, month = {March 25-27}, abstract = { The recently developed Efficient Lifelong Learning Algorithm (ELLA) acquires knowledge incrementally over a sequence of tasks, learning a repository of latent model components that are sparsely shared between models. ELLA shows strong performance in comparison to other multi-task learning algorithms, achieving nearly identical performance to batch multi-task learning methods while learning tasks sequentially in three orders of magnitude (over 1,000x) less time. In this paper, we evaluate several curriculum selection methods that allow ELLA to actively select the next task for learning in order to maximize performance on future learning tasks. Through experiments with three real and one synthetic data set, we demonstrate that active curriculum selection allows an agent to learn up to 50% more efficiently than when the agent has no control over the task order.} } | |||
Abstract:The recently developed Efficient Lifelong Learning Algorithm (ELLA) acquires knowledge incrementally over a sequence of tasks, learning a repository of latent model components that are sparsely shared between models. ELLA shows strong performance in comparison to other multi-task learning algorithms, achieving nearly identical performance to batch multi-task learning methods while learning tasks sequentially in three orders of magnitude (over 1,000x) less time. In this paper, we evaluate several curriculum selection methods that allow ELLA to actively select the next task for learning in order to maximize performance on future learning tasks. Through experiments with three real and one synthetic data set, we demonstrate that active curriculum selection allows an agent to learn up to 50% more efficiently than when the agent has no control over the task order. | |||
Transfer Learning
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim |
SAR image classification using few-shot cross-domain transfer learning
Download:
[ PDF ]
|
2019 | IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops |
Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim. SAR image classification using few-shot cross-domain transfer learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2019. | |||
@inproceedings{Rostami2019SAR, author = {Rostami, Mohammad and Kolouri, Soheil and Eaton, Eric and Kim, Kyungnam}, title = {{SAR} image classification using few-shot cross-domain transfer learning}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2019}, abstract = { Data-driven classification algorithms based on deep convolutional neural networks (CNNs) have reached human-level performance for many tasks within Electro-Optical (EO) computer vision.Despite being the prevailing visual sensory data, EO imaging is not effective in applications such as environmental monitoring at extended periods, where data collection at occluded weather is necessary.Synthetic Aperture Radar (SAR) is an effective imaging tool to circumvent these limitations and collect visual sensory information continually. However, replicating the success of deep learning on SAR domains is not straightforward. This is mainly because training deep networks requires huge labeled datasets anddata labeling is a lot more challenging in SAR domains. We develop an algorithm to transfer knowledge from EO domains to SAR domains to eliminate the need for huge labeled data points in the SAR domains. Our idea is to learn a shared domain-invariant embedding for cross-domain knowledge transfer such that the embedding is discriminative for two related EO and SAR tasks, while the latent data distributions for both domains remain similar. As a result, a classifier learned using mostly EO data can generalize well on the related task for the EO domain.} } | |||
Abstract:Data-driven classification algorithms based on deep convolutional neural networks (CNNs) have reached human-level performance for many tasks within Electro-Optical (EO) computer vision.Despite being the prevailing visual sensory data, EO imaging is not effective in applications such as environmental monitoring at extended periods, where data collection at occluded weather is necessary.Synthetic Aperture Radar (SAR) is an effective imaging tool to circumvent these limitations and collect visual sensory information continually. However, replicating the success of deep learning on SAR domains is not straightforward. This is mainly because training deep networks requires huge labeled datasets anddata labeling is a lot more challenging in SAR domains. We develop an algorithm to transfer knowledge from EO domains to SAR domains to eliminate the need for huge labeled data points in the SAR domains. Our idea is to learn a shared domain-invariant embedding for cross-domain knowledge transfer such that the embedding is discriminative for two related EO and SAR tasks, while the latent data distributions for both domains remain similar. As a result, a classifier learned using mostly EO data can generalize well on the related task for the EO domain. | |||
Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim |
Deep transfer learning for few-shot SAR image classification
Download:
[ PDF ] [ Website ]
|
2019 | Remote Sensing |
Mohammad Rostami, Soheil Kolouri, Eric Eaton, & Kyungnam Kim. Deep transfer learning for few-shot SAR image classification. Remote Sensing, 11:1374, 2019. | |||
@article{Rostami2019Deep, title = {Deep transfer learning for few-shot {SAR} image classification}, author = {Mohammad Rostami and Soheil Kolouri and Eric Eaton and Kyungnam Kim}, journal = {Remote Sensing}, volume = {11}, pages = {1374}, year = {2019}, abstract = { The emergence of Deep Neural Networks (DNNs) has lead to high-performance supervised learning algorithms for the Electro-Optical (EO) domain classification and detection problems. This success is because generating huge labeled datasets has become possible using modern crowdsourcing labeling platforms such as Amazon’s Mechanical Turk that recruit ordinary people to label data. Unlike the EO domain, labeling the Synthetic Aperture Radar (SAR) domain data can be much more challenging, and for various reasons, using crowdsourcing platforms is not feasible for labeling the SAR domain data. As a result, training deep networks using supervised learning is more challenging in the SAR domain. In the paper, we present a new framework to train a deep neural network for classifying Synthetic Aperture Radar (SAR) images by eliminating the need for a huge labeled dataset. Our idea is based on transferring knowledge from a related EO domain problem, where labeled data are easy to obtain. We transfer knowledge from the EO domain through learning a shared invariant cross-domain embedding space that is also discriminative for classification. To this end, we train two deep encoders that are coupled through their last year to map data points from the EO and the SAR domains to the shared embedding space such that the distance between the distributions of the two domains is minimized in the latent embedding space. We use the Sliced Wasserstein Distance (SWD) to measure and minimize the distance between these two distributions and use a limited number of SAR label data points to match the distributions class-conditionally. As a result of this training procedure, a classifier trained from the embedding space to the label space using mostly the EO data would generalize well on the SAR domain. We provide a theoretical analysis to demonstrate why our approach is effective and validate our algorithm on the problem of ship classification in the SAR domain by comparing against several other competing learning approaches.} } | |||
Abstract:The emergence of Deep Neural Networks (DNNs) has lead to high-performance supervised learning algorithms for the Electro-Optical (EO) domain classification and detection problems. This success is because generating huge labeled datasets has become possible using modern crowdsourcing labeling platforms such as Amazon’s Mechanical Turk that recruit ordinary people to label data. Unlike the EO domain, labeling the Synthetic Aperture Radar (SAR) domain data can be much more challenging, and for various reasons, using crowdsourcing platforms is not feasible for labeling the SAR domain data. As a result, training deep networks using supervised learning is more challenging in the SAR domain. In the paper, we present a new framework to train a deep neural network for classifying Synthetic Aperture Radar (SAR) images by eliminating the need for a huge labeled dataset. Our idea is based on transferring knowledge from a related EO domain problem, where labeled data are easy to obtain. We transfer knowledge from the EO domain through learning a shared invariant cross-domain embedding space that is also discriminative for classification. To this end, we train two deep encoders that are coupled through their last year to map data points from the EO and the SAR domains to the shared embedding space such that the distance between the distributions of the two domains is minimized in the latent embedding space. We use the Sliced Wasserstein Distance (SWD) to measure and minimize the distance between these two distributions and use a limited number of SAR label data points to match the distributions class-conditionally. As a result of this training procedure, a classifier trained from the embedding space to the label space using mostly the EO data would generalize well on the SAR domain. We provide a theoretical analysis to demonstrate why our approach is effective and validate our algorithm on the problem of ship classification in the SAR domain by comparing against several other competing learning approaches. | |||
Boyu Wang, Jorge Mendez, Mingbo Cai, & Eric Eaton | Transfer learning via minimizing the performance gap between domains | 2019 | Advances in Neural Information Processing Systems |
Boyu Wang, Jorge Mendez, Mingbo Cai, & Eric Eaton. Transfer learning via minimizing the performance gap between domains. In Advances in Neural Information Processing Systems, pp. 10645–10655, 2019. | |||
@incollection{Wang2019Transfer, title = {Transfer learning via minimizing the performance gap between domains}, author = {Wang, Boyu and Mendez, Jorge and Cai, Mingbo and Eaton, Eric}, booktitle = {Advances in Neural Information Processing Systems}, volume = {32}, pages = {10645--10655}, year = {2019}, abstract = { We propose a new principle for transfer learning, based on a straightforward intuition: if two domains are similar to each other, the model trained on one domain should also perform well on the other domain, and vice versa. To formalize this intuition, we define the performance gap as a measure of the discrepancy between the source and target domains. We derive generalization bounds for the instance weighting approach to transfer learning, showing that the performance gap can be viewed as an algorithm-dependent regularizer, which controls the model complexity. Our theoretical analysis provides new insight into transfer learning and motivates a set of general, principled rules for designing new instance weighting schemes for transfer learning. These rules lead to gapBoost, a novel and principled boosting approach for transfer learning. Our experimental evaluation on benchmark data sets shows that gapBoost significantly outperforms previous boosting-based transfer learning algorithms. } } | |||
Abstract:We propose a new principle for transfer learning, based on a straightforward intuition: if two domains are similar to each other, the model trained on one domain should also perform well on the other domain, and vice versa. To formalize this intuition, we define the performance gap as a measure of the discrepancy between the source and target domains. We derive generalization bounds for the instance weighting approach to transfer learning, showing that the performance gap can be viewed as an algorithm-dependent regularizer, which controls the model complexity. Our theoretical analysis provides new insight into transfer learning and motivates a set of general, principled rules for designing new instance weighting schemes for transfer learning. These rules lead to gapBoost, a novel and principled boosting approach for transfer learning. Our experimental evaluation on benchmark data sets shows that gapBoost significantly outperforms previous boosting-based transfer learning algorithms. | |||
Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor | Unsupervised cross-domain transfer in policy gradient reinforcement learning via manifold alignment | 2015 | AAAI Conference on Artificial Intelligence (AAAI-15) |
Haitham Bou Ammar, Eric Eaton, Paul Ruvolo, & Matthew E. Taylor. Unsupervised cross-domain transfer in policy gradient reinforcement learning via manifold alignment. In Proceedings of the 29th AAAI Conference on Artificial Intelligence (AAAI-15), January 2015. Acceptance rate: 27% | |||
@inproceedings{BouAmmar2015Unsupervised, author={Haitham {Bou Ammar} and Eric Eaton and Paul Ruvolo and Matthew E. Taylor}, title={Unsupervised cross-domain transfer in policy gradient reinforcement learning via manifold alignment}, booktitle={Proceedings of the 29th AAAI Conference on Artificial Intelligence (AAAI-15)}, month={January}, year={2015}, note={Acceptance rate: 27%}, abstract={ The success of applying policy gradient reinforcement learning (RL) to difficult control tasks hinges crucially on the ability to determine a sensible initialization for the policy. Transfer learning methods tackle this problem by reusing knowledge gleaned from solving other related tasks. In the case of multiple task domains, these algorithms require an inter-task mapping to facilitate knowledge transfer across domains. However, there are currently no general methods to learn an inter-task mapping without requiring either background knowledge that is not typically present in RL settings, or an expensive analysis of an exponential number of inter-task mappings in the size of the state and action spaces. This paper introduces an autonomous framework that uses unsupervised manifold alignment to learn intertask mappings and effectively transfer samples between different task domains. Empirical results on diverse dynamical systems, including an application to quadrotor control, demonstrate its effectiveness for cross-domain transfer in the context of policy gradient RL.}, } | |||
Abstract:The success of applying policy gradient reinforcement learning (RL) to difficult control tasks hinges crucially on the ability to determine a sensible initialization for the policy. Transfer learning methods tackle this problem by reusing knowledge gleaned from solving other related tasks. In the case of multiple task domains, these algorithms require an inter-task mapping to facilitate knowledge transfer across domains. However, there are currently no general methods to learn an inter-task mapping without requiring either background knowledge that is not typically present in RL settings, or an expensive analysis of an exponential number of inter-task mappings in the size of the state and action spaces. This paper introduces an autonomous framework that uses unsupervised manifold alignment to learn intertask mappings and effectively transfer samples between different task domains. Empirical results on diverse dynamical systems, including an application to quadrotor control, demonstrate its effectiveness for cross-domain transfer in the context of policy gradient RL. | |||
Haitham Bou Ammar, Eric Eaton, Matthew E. Taylor, Decebal Mocanu, Kurt Driessens, Gerhard Weiss, & Karl Tuyls | An automated measure of MDP similarity for transfer in reinforcement learning | 2014 | AAAI'14 Workshop on Machine Learning for Interactive Systems |
Haitham Bou Ammar, Eric Eaton, Matthew E. Taylor, Decebal Mocanu, Kurt Driessens, Gerhard Weiss, & Karl Tuyls. An automated measure of MDP similarity for transfer in reinforcement learning. In Proceedings of the AAAI'14 Workshop on Machine Learning for Interactive Systems, July 2014. | |||
@INPROCEEDINGS{BouAmmar2014Automated, author = {Haitham Bou Ammar and Eric Eaton and Matthew E. Taylor and Decebal Mocanu and Kurt Driessens and Gerhard Weiss and Karl Tuyls}, year = {2014}, title = {An automated measure of {MDP} similarity for transfer in reinforcement learning}, booktitle = {Proceedings of the AAAI'14 Workshop on Machine Learning for Interactive Systems}, month = {July}, abstract = { Transfer learning can improve the reinforcement learning of a new task by allowing the agent to reuse knowledge acquired from other source tasks. Despite their success, transfer learning methods rely on having relevant source tasks; transfer from inappropriate tasks can inhibit performance on the new task. For fully autonomous transfer, it is critical to have a method for automatically choosing relevant source tasks, which requires a similarity measure between Markov Decision Processes (MDPs). This issue has received little attention, and is therefore still a largely open problem. This paper presents a data-driven automated similarity measure for MDPs. This novel measure is a significant step toward autonomous reinforcement learning transfer, allowing agents to: (1) characterize when transfer will be useful and, (2) automatically select tasks to use for transfer. The proposed measure is based on the reconstruction error of a restricted Boltzmann machine that attempts to model the behavioral dynamics of the two MDPs being compared. Empirical results illustrate that this measure is correlated with the performance of transfer and therefore can be used to identify similar source tasks for transfer learning.} } | |||
Abstract:Transfer learning can improve the reinforcement learning of a new task by allowing the agent to reuse knowledge acquired from other source tasks. Despite their success, transfer learning methods rely on having relevant source tasks; transfer from inappropriate tasks can inhibit performance on the new task. For fully autonomous transfer, it is critical to have a method for automatically choosing relevant source tasks, which requires a similarity measure between Markov Decision Processes (MDPs). This issue has received little attention, and is therefore still a largely open problem. This paper presents a data-driven automated similarity measure for MDPs. This novel measure is a significant step toward autonomous reinforcement learning transfer, allowing agents to: (1) characterize when transfer will be useful and, (2) automatically select tasks to use for transfer. The proposed measure is based on the reconstruction error of a restricted Boltzmann machine that attempts to model the behavioral dynamics of the two MDPs being compared. Empirical results illustrate that this measure is correlated with the performance of transfer and therefore can be used to identify similar source tasks for transfer learning. | |||
Diane Oyen, Eric Eaton, & Terran Lane |
Inferring tasks for improved network structure discovery
Download:
[ PDF ]
|
2012 | Snowbird Learning Workshop |
Diane Oyen, Eric Eaton, & Terran Lane. Inferring tasks for improved network structure discovery. In Working Notes of the Snowbird Learning Workshop, April 3--6 2012. | |||
@INPROCEEDINGS{Oyen2012Inferring, author = {Diane Oyen and Eric Eaton and Terran Lane}, title = {Inferring tasks for improved network structure discovery}, booktitle = {Working Notes of the Snowbird Learning Workshop}, month = {April 3--6}, location = {Snowbird, Utah}, year = {2012}, } | |||
Eric Eaton & Marie desJardins |
Selective Transfer Between Learning Tasks Using Task-Based Boosting
Download:
[ PDF ] [ Supplementary Materials ]
|
2011 | AAAI Conference on Artificial Intelligence (AAAI-11) |
Eric Eaton & Marie desJardins. Selective Transfer Between Learning Tasks Using Task-Based Boosting. In Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11), pp. 337–342, AAAI Press, August 7--11 2011. | |||
@INPROCEEDINGS{Eaton2011Selective, author = {Eric Eaton and Marie desJardins}, title = {Selective Transfer Between Learning Tasks Using Task-Based Boosting}, booktitle = {Proceedings of the 25th AAAI Conference on Artificial Intelligence (AAAI-11)}, month = {August 7--11}, location = {San Francisco, CA}, publisher = {AAAI Press}, pages = {337--342}, year = {2011}, abstract = { The success of transfer learning on a target task is highly dependent on the selected source data. Instance transfer methods reuse data from the source tasks to augment the training data for the target task. If poorly chosen, this source data may inhibit learning, resulting in negative transfer. The current most widely used algorithm for instance transfer, TrAdaBoost, performs poorly when given irrelevant source data. We present a novel task-based boosting technique for instance transfer that selectively chooses the source knowledge to transfer to the target task. Our approach performs boosting at both the instance level and the task level, assigning higher weight to those source tasks that show positive transferability to the target task, and adjusting the weights of individual instances within each source task via AdaBoost. We show that this combination of task- and instance-level boosting significantly improves transfer performance over existing instance transfer algorithms when given a mix of relevant and irrelevant source data, especially for small amounts of data on the target task.} } | |||
Abstract:The success of transfer learning on a target task is highly dependent on the selected source data. Instance transfer methods reuse data from the source tasks to augment the training data for the target task. If poorly chosen, this source data may inhibit learning, resulting in negative transfer. The current most widely used algorithm for instance transfer, TrAdaBoost, performs poorly when given irrelevant source data. We present a novel task-based boosting technique for instance transfer that selectively chooses the source knowledge to transfer to the target task. Our approach performs boosting at both the instance level and the task level, assigning higher weight to those source tasks that show positive transferability to the target task, and adjusting the weights of individual instances within each source task via AdaBoost. We show that this combination of task- and instance-level boosting significantly improves transfer performance over existing instance transfer algorithms when given a mix of relevant and irrelevant source data, especially for small amounts of data on the target task. | |||
Eric Eaton & Terran Lane |
The Importance of Selective Knowledge Transfer for Lifelong Learning
Download:
[ PDF ]
|
2011 | AAAI-11 Workshop on Lifelong Learning from Sensorimotor Data |
Eric Eaton & Terran Lane. The Importance of Selective Knowledge Transfer for Lifelong Learning. In AAAI-11 Workshop on Lifelong Learning from Sensorimotor Data, AAAI Press, August 7 2011. | |||
@INPROCEEDINGS{Eaton2011Importance, author = {Eric Eaton and Terran Lane}, title = {The Importance of Selective Knowledge Transfer for Lifelong Learning}, booktitle = {AAAI-11 Workshop on Lifelong Learning from Sensorimotor Data}, month = {August 7}, location = {San Francisco, CA}, publisher = {AAAI Press}, year = {2011}, abstract = { As knowledge transfer research progresses from single transfer to lifelong learning scenarios, it becomes increasingly important to properly select the source knowledge that would best transfer to the target task. In this position paper, we describe our previous work on selective knowledge transfer and relate it to problems in lifelong learning. We also briefly discuss our ongoing work to develop lifelong learning methods capable of continual transfer between tasks and the incorporation of guidance from an expert human user.} } | |||
Abstract:As knowledge transfer research progresses from single transfer to lifelong learning scenarios, it becomes increasingly important to properly select the source knowledge that would best transfer to the target task. In this position paper, we describe our previous work on selective knowledge transfer and relate it to problems in lifelong learning. We also briefly discuss our ongoing work to develop lifelong learning methods capable of continual transfer between tasks and the incorporation of guidance from an expert human user. | |||
Eric Eaton & Marie desJardins |
Set-Based Boosting for Instance-level Transfer Superseded by the AAAI-11 paper Selective Transfer Between Learning Tasks Using Task-Based Boosting.
Download:
[ PDF ]
|
2009 | International Conference on Data Mining Workshop on Transfer Mining |
Eric Eaton & Marie desJardins. Set-Based Boosting for Instance-level Transfer. In Proceedings of the International Conference on Data Mining Workshop on Transfer Mining, pp. 422–428, IEEE Press, December 2009. | |||
@INPROCEEDINGS{Eaton2009SetBased, author = {Eric Eaton and Marie desJardins}, title = {Set-Based Boosting for Instance-level Transfer}, booktitle = {Proceedings of the International Conference on Data Mining Workshop on Transfer Mining}, location = {Miami, FL}, publisher = {IEEE Press}, pages = {422--428}, month = {December}, year = {2009}, abstract = { The success of transfer to improve learning on a target task is highly dependent on the selected source data. Instance-based transfer methods reuse data from the source tasks to augment the training data for the target task. If poorly chosen, this source data may inhibit learning, resulting in negative transfer. The current best performing algorithm for instance-based transfer, TrAdaBoost, performs poorly when given irrelevant source data. We present a novel set-based boosting technique for instance-based transfer. The proposed algorithm, TransferBoost, boosts both individual instances and collective sets of instances from each source task. In effect, TransferBoost boosts each source task, assigning higher weight to those source tasks which show positive transferability to the target task, and then adjusts the weights of the instances within each source task via AdaBoost. The results demonstrate that TransferBoost significantly improves transfer performance over existing instance-based algorithms when given a mix of relevant and irrelevant source data.} } | |||
Abstract:The success of transfer to improve learning on a target task is highly dependent on the selected source data. Instance-based transfer methods reuse data from the source tasks to augment the training data for the target task. If poorly chosen, this source data may inhibit learning, resulting in negative transfer. The current best performing algorithm for instance-based transfer, TrAdaBoost, performs poorly when given irrelevant source data. We present a novel set-based boosting technique for instance-based transfer. The proposed algorithm, TransferBoost, boosts both individual instances and collective sets of instances from each source task. In effect, TransferBoost boosts each source task, assigning higher weight to those source tasks which show positive transferability to the target task, and then adjusts the weights of the instances within each source task via AdaBoost. The results demonstrate that TransferBoost significantly improves transfer performance over existing instance-based algorithms when given a mix of relevant and irrelevant source data. | |||
Eric Eaton | Selective Knowledge Transfer for Machine Learning | 2009 | Ph.D. Thesis, University of Maryland Baltimore County |
Eric Eaton. Selective Knowledge Transfer for Machine Learning. Ph.D. Thesis, University of Maryland Baltimore County, 2009. | |||
@PHDTHESIS{Eaton2009Selective, author = {Eric Eaton}, title = {Selective Knowledge Transfer for Machine Learning}, school = {University of Maryland Baltimore County}, year = {2009}, abstract = { Knowledge transfer from previously learned tasks to a new task is a fundamental component of human learning. Recent work has shown that knowledge transfer can also improve machine learning, enabling more rapid learning or higher levels of performance. Transfer allows learning algorithms to reuse knowledge from a set of previously learned source tasks to improve learning on new target tasks. Proper selection of the source knowledge to transfer to a given target task is critical to the success of knowledge transfer. Poorly chosen source knowledge may reduce the effectiveness of transfer, or hinder learning through a phenomenon known as negative transfer. This dissertation proposes several methods for source knowledge selection that are based on the transferability between learning tasks. Transferability is introduced as the change in performance on a target task between learning with and without transfer. These methods show that transferability can be used to select source knowledge for two major types of transfer: instance-based transfer, which reuses individual data instances from the source tasks, and model-based transfer, which transfers components of previously learned source models. For selective instance-based transfer, the proposed TransferBoost algorithm uses a novel form of set-based boosting to determine the individual source instances to transfer in learning the target task. TransferBoost reweights instances from each source task based on their collective transferability to the target task, and then performs regular boosting to adjust individual instance weights. For model-based transfer, the learning tasks are organized into a directed network based on their transfer relationships to each other. Tasks that are close in this network have high transferability, and tasks that are far apart have low transferability. Model-based transfer is equivalent to learning a labeling function on this network. This dissertation proposes the novel Spectral Graph Labeling algorithm that constrains the smoothness of the learned function using the graph's Laplacian eigenvalues. This method is then applied to the task transferability network to learn a transfer function that automatically determines the model parameter values to transfer to a target task. Experiments validate the success of these methods for selective knowledge transfer, demonstrating significantly improved performance over existing methods.} } | |||
Abstract:Knowledge transfer from previously learned tasks to a new task is a fundamental component of human learning. Recent work has shown that knowledge transfer can also improve machine learning, enabling more rapid learning or higher levels of performance. Transfer allows learning algorithms to reuse knowledge from a set of previously learned source tasks to improve learning on new target tasks. Proper selection of the source knowledge to transfer to a given target task is critical to the success of knowledge transfer. Poorly chosen source knowledge may reduce the effectiveness of transfer, or hinder learning through a phenomenon known as negative transfer. This dissertation proposes several methods for source knowledge selection that are based on the transferability between learning tasks. Transferability is introduced as the change in performance on a target task between learning with and without transfer. These methods show that transferability can be used to select source knowledge for two major types of transfer: instance-based transfer, which reuses individual data instances from the source tasks, and model-based transfer, which transfers components of previously learned source models. For selective instance-based transfer, the proposed TransferBoost algorithm uses a novel form of set-based boosting to determine the individual source instances to transfer in learning the target task. TransferBoost reweights instances from each source task based on their collective transferability to the target task, and then performs regular boosting to adjust individual instance weights. For model-based transfer, the learning tasks are organized into a directed network based on their transfer relationships to each other. Tasks that are close in this network have high transferability, and tasks that are far apart have low transferability. Model-based transfer is equivalent to learning a labeling function on this network. This dissertation proposes the novel Spectral Graph Labeling algorithm that constrains the smoothness of the learned function using the graph's Laplacian eigenvalues. This method is then applied to the task transferability network to learn a transfer function that automatically determines the model parameter values to transfer to a target task. Experiments validate the success of these methods for selective knowledge transfer, demonstrating significantly improved performance over existing methods. | |||
Eric Eaton, Marie desJardins, & Terran Lane |
Using functions on a model graph for inductive transfer Superseded by the ECML-08 paper Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer.
Download:
[ PDF ]
|
2008 | Northeast Student Colloquium on Artificial Intelligence (NESCAI-08) |
Eric Eaton, Marie desJardins, & Terran Lane. Using functions on a model graph for inductive transfer. In Proceedings of the Northeast Student Colloquium on Artificial Intelligence (NESCAI-08), Ithaca, NY, May 2--4 2008. | |||
@inproceedings{Eaton2008Using, author = {Eric Eaton and Marie desJardins and Terran Lane}, title = {Using functions on a model graph for inductive transfer}, booktitle = {Proceedings of the Northeast Student Colloquium on Artificial Intelligence (NESCAI-08)}, year = {2008}, month = {May 2--4}, address = {Ithaca, NY}, abstract = { In this paper, we propose a novel graph-based method for knowledge transfer. We embed a set of learned background models in a graph that captures the transferability between the models. We then learn a function on this graph that automatically determines the parameters to transfer to each learning task. Transfer to a new problem proceeds by mapping the problem into the graph, then using the function to determine the parameters to transfer in learning the new model. This method is analogous to inductive transfer along a manifold that captures the transfer relationships between the tasks.} } | |||
Abstract:In this paper, we propose a novel graph-based method for knowledge transfer. We embed a set of learned background models in a graph that captures the transferability between the models. We then learn a function on this graph that automatically determines the parameters to transfer to each learning task. Transfer to a new problem proceeds by mapping the problem into the graph, then using the function to determine the parameters to transfer in learning the new model. This method is analogous to inductive transfer along a manifold that captures the transfer relationships between the tasks. | |||
Eric Eaton, Marie desJardins, & Terran Lane |
Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer
Download:
[ PDF ]
|
2008 | European Conference on Machine Learning (ECML) |
Eric Eaton, Marie desJardins, & Terran Lane. Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer. In Proceedings of the 19th European Conference on Machine Learning (ECML), pp. 317–332, Springer-Verlag, Berlin, Heidelberg, 2008. Acceptance rate: 20% | |||
@INPROCEEDINGS{Eaton2008Modeling, author = {Eric Eaton and Marie desJardins and Terran Lane}, title = {Modeling Transfer Relationships Between Learning Tasks for Improved Inductive Transfer}, booktitle = ecml08, year = {2008}, pages = {317--332}, publisher = {Springer-Verlag}, location = {Antwerp, Belgium}, address = {Berlin, Heidelberg}, note = {Acceptance rate: 20%}, abstract = { In this paper, we propose a novel graph-based method for knowledge transfer. We model the transfer relationships between source tasks by embedding the set of learned source models in a graph using transferability as the metric. Transfer to a new problem proceeds by mapping the problem into the graph, then learning a function on this graph that automatically determines the parameters to transfer to the new learning task. This method is analogous to inductive transfer along a manifold that captures the transfer relationships between the tasks. We demonstrate improved transfer performance using this method against existing approaches in several real-world domains.} } | |||
Abstract:In this paper, we propose a novel graph-based method for knowledge transfer. We model the transfer relationships between source tasks by embedding the set of learned source models in a graph using transferability as the metric. Transfer to a new problem proceeds by mapping the problem into the graph, then learning a function on this graph that automatically determines the parameters to transfer to the new learning task. This method is analogous to inductive transfer along a manifold that captures the transfer relationships between the tasks. We demonstrate improved transfer performance using this method against existing approaches in several real-world domains. | |||
Eric Eaton, Marie desJardins, & John Stevenson |
Using multiresolution learning for transfer in image classification
Download:
[ PDF ] [ Poster ]
|
2007 | National Conference on Artificial Intelligence (AAAI) |
Eric Eaton, Marie desJardins, & John Stevenson. Using multiresolution learning for transfer in image classification. In Proceedings of the 22nd National Conference on Artificial Intelligence (AAAI), AAAI Press, Vancouver, British Columbia, Canada, 2007. [Student Abstract] | |||
@INPROCEEDINGS{Eaton2007Using, author = {Eric Eaton and Marie desJardins and John Stevenson}, year = {2007}, title = {Using multiresolution learning for transfer in image classification}, booktitle = aaai07, address = {Vancouver, British Columbia, Canada}, publisher = {AAAI Press}, note = {[Student Abstract]}, abstract = { Our work explores the transfer of knowledge at multiple levels of abstraction to improve learning. By exploiting the similarities between objects at various levels of detail, multiresolution learning can facilitate transfer between image classification tasks. We extract features from images at multiple levels of resolution, then use these features to create models at different resolutions. Upon receiving a new task, the closest-matching stored model can be generalized (adapted to the appropriate resolution) and transferred to the new task.} } | |||
Abstract:Our work explores the transfer of knowledge at multiple levels of abstraction to improve learning. By exploiting the similarities between objects at various levels of detail, multiresolution learning can facilitate transfer between image classification tasks. We extract features from images at multiple levels of resolution, then use these features to create models at different resolutions. Upon receiving a new task, the closest-matching stored model can be generalized (adapted to the appropriate resolution) and transferred to the new task. | |||
Eric Eaton |
Multi-Resolution Learning for Knowledge Transfer
Download:
[ PDF ]
|
2006 | National Conference on Artificial Intelligence (AAAI) |
Eric Eaton. Multi-Resolution Learning for Knowledge Transfer. In Proceedings of the 21st National Conference on Artificial Intelligence (AAAI), AAAI Press, Boston, MA, July 16--20 2006. [Doctoral Consortium] | |||
@INPROCEEDINGS{Eaton2006MultiResolution, author = {Eric Eaton}, title = {Multi-Resolution Learning for Knowledge Transfer}, booktitle = aaai06, year = {2006}, address = {Boston, MA}, month = {July 16--20}, publisher = {AAAI Press}, note = {[Doctoral Consortium]}, keywords = {knowledge transfer, multiresolution learning, aaai doctoral consortium}, abstract = { Related objects may look similar at low-resolutions; differences begin to emerge naturally as the resolution is increased. By learning across multiple resolutions of input, knowledge can be transfered between related objects. My dissertation develops this idea and applies it to the problem of multitask transfer learning.} } | |||
Abstract:Related objects may look similar at low-resolutions; differences begin to emerge naturally as the resolution is increased. By learning across multiple resolutions of input, knowledge can be transfered between related objects. My dissertation develops this idea and applies it to the problem of multitask transfer learning. | |||
Eric Eaton & Marie desJardins |
Knowledge Transfer with a Multiresolution Ensemble of Classifiers
Download:
[ PDF ]
|
2006 | ICML-06 Workshop on Structural Knowledge Transfer for Machine Learning |
Eric Eaton & Marie desJardins. Knowledge Transfer with a Multiresolution Ensemble of Classifiers. In Proceedings of the ICML-06 Workshop on Structural Knowledge Transfer for Machine Learning, Pittsburgh, PA, June 29 2006. | |||
@INPROCEEDINGS{Eaton2006Knowledge, author = {Eric Eaton and Marie desJardins}, title = {Knowledge Transfer with a Multiresolution Ensemble of Classifiers}, booktitle = {Proceedings of the ICML-06 Workshop on Structural Knowledge Transfer for Machine Learning}, year = {2006}, address = {Pittsburgh, PA}, month = {June 29}, keywords = {knowledge transfer, multiresolution learning}, abstract = { We demonstrate transfer via an ensemble of classifiers, where each member focuses on one resolution of data. Lower-resolution ensemble members are shared between tasks, providing a medium for knowledge transfer.} } | |||
Abstract:We demonstrate transfer via an ensemble of classifiers, where each member focuses on one resolution of data. Lower-resolution ensemble members are shared between tasks, providing a medium for knowledge transfer. | |||
Interactive and Interpretable Learning
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Efstathios D. Gennatas, Jerome H. Friedman, Lyle H. Ungar, Romain Pirracchio, Eric Eaton, Lara G. Reichmann, Yannet Interian, Jose Marcio Luna, Charles B. Simone, Andrew Auerbach, Elier Delgado, Mark J. van der Laan, Timothy D. Solberg, & Gilmer Valdes | Expert-augmented machine learning | 2020 | Proceedings of the National Academy of Sciences 117(9) |
Efstathios D. Gennatas, Jerome H. Friedman, Lyle H. Ungar, Romain Pirracchio, Eric Eaton, Lara G. Reichmann, Yannet Interian, Jose Marcio Luna, Charles B. Simone, Andrew Auerbach, Elier Delgado, Mark J. van der Laan, Timothy D. Solberg, & Gilmer Valdes. Expert-augmented machine learning. Proceedings of the National Academy of Sciences, 117(9):4571–4577, National Academy of Sciences, 2020. | |||
@article {Gennatas2020ExpertAugmented, author = {Gennatas, Efstathios D. and Friedman, Jerome H. and Ungar, Lyle H. and Pirracchio, Romain and Eaton, Eric and Reichmann, Lara G. and Interian, Yannet and Luna, Jose Marcio and Simone, Charles B. and Auerbach, Andrew and Delgado, Elier and van der Laan, Mark J. and Solberg, Timothy D. and Valdes, Gilmer}, title = {Expert-augmented machine learning}, volume = {117}, number = {9}, pages = {4571--4577}, year = {2020}, publisher = {National Academy of Sciences}, abstract = { Machine learning is increasingly used across fields to derive insights from data, which further our understanding of the world and help us anticipate the future. The performance of predictive modeling is dependent on the amount and quality of available data. In practice, we rely on human experts to perform certain tasks and on machine learning for others. However, the optimal learning strategy may involve combining the complementary strengths of humans and machines. We present expert-augmented machine learning, an automated way to automatically extract problem-specific human expert knowledge and integrate it with machine learning to build robust, dependable, and data-efficient predictive models.Machine learning is proving invaluable across disciplines. However, its success is often limited by the quality and quantity of available data, while its adoption is limited by the level of trust afforded by given models. Human vs. machine performance is commonly compared empirically to decide whether a certain task should be performed by a computer or an expert. In reality, the optimal learning strategy may involve combining the complementary strengths of humans and machines. Here, we present expert-augmented machine learning (EAML), an automated method that guides the extraction of expert knowledge and its integration into machine-learned models. We used a large dataset of intensive-care patient data to derive 126 decision rules that predict hospital mortality. Using an online platform, we asked 15 clinicians to assess the relative risk of the subpopulation defined by each rule compared to the total sample. We compared the clinician-assessed risk to the empirical risk and found that, while clinicians agreed with the data in most cases, there were notable exceptions where they overestimated or underestimated the true risk. Studying the rules with greatest disagreement, we identified problems with the training data, including one miscoded variable and one hidden confounder. Filtering the rules based on the extent of disagreement between clinician-assessed risk and empirical risk, we improved performance on out-of-sample data and were able to train with less data. EAML provides a platform for automated creation of problem-specific priors, which help build robust and dependable machine-learning models in critical applications. }, journal = {Proceedings of the National Academy of Sciences}, } | |||
Abstract:Machine learning is increasingly used across fields to derive insights from data, which further our understanding of the world and help us anticipate the future. The performance of predictive modeling is dependent on the amount and quality of available data. In practice, we rely on human experts to perform certain tasks and on machine learning for others. However, the optimal learning strategy may involve combining the complementary strengths of humans and machines. We present expert-augmented machine learning, an automated way to automatically extract problem-specific human expert knowledge and integrate it with machine learning to build robust, dependable, and data-efficient predictive models.Machine learning is proving invaluable across disciplines. However, its success is often limited by the quality and quantity of available data, while its adoption is limited by the level of trust afforded by given models. Human vs. machine performance is commonly compared empirically to decide whether a certain task should be performed by a computer or an expert. In reality, the optimal learning strategy may involve combining the complementary strengths of humans and machines. Here, we present expert-augmented machine learning (EAML), an automated method that guides the extraction of expert knowledge and its integration into machine-learned models. We used a large dataset of intensive-care patient data to derive 126 decision rules that predict hospital mortality. Using an online platform, we asked 15 clinicians to assess the relative risk of the subpopulation defined by each rule compared to the total sample. We compared the clinician-assessed risk to the empirical risk and found that, while clinicians agreed with the data in most cases, there were notable exceptions where they overestimated or underestimated the true risk. Studying the rules with greatest disagreement, we identified problems with the training data, including one miscoded variable and one hidden confounder. Filtering the rules based on the extent of disagreement between clinician-assessed risk and empirical risk, we improved performance on out-of-sample data and were able to train with less data. EAML provides a platform for automated creation of problem-specific priors, which help build robust and dependable machine-learning models in critical applications. | |||
Jose Marcio Luna, Efstathios D. Gennatas, Lyle H. Ungar, Eric Eaton, Eric S. Diffenderfer, Shane T. Jensen, Charles B. Simone, Jerome H. Friedman, Timothy D. Solberg, & Gilmer Valdes | Building more accurate decision trees with the additive tree | 2019 | Proceedings of the National Academy of Sciences 116(40) |
Jose Marcio Luna, Efstathios D. Gennatas, Lyle H. Ungar, Eric Eaton, Eric S. Diffenderfer, Shane T. Jensen, Charles B. Simone, Jerome H. Friedman, Timothy D. Solberg, & Gilmer Valdes. Building more accurate decision trees with the additive tree. Proceedings of the National Academy of Sciences, 116(40):19887–19893, National Academy of Sciences, 2019. | |||
@article{Luna2019Building, author = {Luna, Jose Marcio and Gennatas, Efstathios D. and Ungar, Lyle H. and Eaton, Eric and Diffenderfer, Eric S. and Jensen, Shane T. and Simone, Charles B. and Friedman, Jerome H. and Solberg, Timothy D. and Valdes, Gilmer}, title = {Building more accurate decision trees with the additive tree}, volume = {116}, number = {40}, pages = {19887--19893}, year = {2019}, publisher = {National Academy of Sciences}, journal = {Proceedings of the National Academy of Sciences}, abstract = { As machine learning applications expand to high-stakes areas such as criminal justice, finance, and medicine, legitimate concerns emerge about high-impact effects of individual mispredictions on people{\textquoteright}s lives. As a result, there has been increasing interest in understanding general machine learning models to overcome possible serious risks. Current decision trees, such as Classification and Regression Trees (CART), have played a predominant role in fields such as medicine, due to their simplicity and intuitive interpretation. However, such trees suffer from intrinsic limitations in predictive power. We developed the additive tree, a theoretical approach to generate a more accurate and interpretable decision tree, which reveals connections between CART and gradient boosting. The additive tree exhibits superior predictive performance to CART, as validated on 83 classification tasks.The expansion of machine learning to high-stakes application domains such as medicine, finance, and criminal justice, where making informed decisions requires clear understanding of the model, has increased the interest in interpretable machine learning. The widely used Classification and Regression Trees (CART) have played a major role in health sciences, due to their simple and intuitive explanation of predictions. Ensemble methods like gradient boosting can improve the accuracy of decision trees, but at the expense of the interpretability of the generated model. Additive models, such as those produced by gradient boosting, and full interaction models, such as CART, have been investigated largely in isolation. We show that these models exist along a spectrum, revealing previously unseen connections between these approaches. This paper introduces a rigorous formalization for the additive tree, an empirically validated learning technique for creating a single decision tree, and shows that this method can produce models equivalent to CART or gradient boosted stumps at the extremes by varying a single parameter. Although the additive tree is designed primarily to provide both the model interpretability and predictive performance needed for high-stakes applications like medicine, it also can produce decision trees represented by hybrid models between CART and boosted stumps that can outperform either of these approaches.}, } | |||
Abstract:As machine learning applications expand to high-stakes areas such as criminal justice, finance, and medicine, legitimate concerns emerge about high-impact effects of individual mispredictions on people\textquoterights lives. As a result, there has been increasing interest in understanding general machine learning models to overcome possible serious risks. Current decision trees, such as Classification and Regression Trees (CART), have played a predominant role in fields such as medicine, due to their simplicity and intuitive interpretation. However, such trees suffer from intrinsic limitations in predictive power. We developed the additive tree, a theoretical approach to generate a more accurate and interpretable decision tree, which reveals connections between CART and gradient boosting. The additive tree exhibits superior predictive performance to CART, as validated on 83 classification tasks.The expansion of machine learning to high-stakes application domains such as medicine, finance, and criminal justice, where making informed decisions requires clear understanding of the model, has increased the interest in interpretable machine learning. The widely used Classification and Regression Trees (CART) have played a major role in health sciences, due to their simple and intuitive explanation of predictions. Ensemble methods like gradient boosting can improve the accuracy of decision trees, but at the expense of the interpretability of the generated model. Additive models, such as those produced by gradient boosting, and full interaction models, such as CART, have been investigated largely in isolation. We show that these models exist along a spectrum, revealing previously unseen connections between these approaches. This paper introduces a rigorous formalization for the additive tree, an empirically validated learning technique for creating a single decision tree, and shows that this method can produce models equivalent to CART or gradient boosted stumps at the extremes by varying a single parameter. Although the additive tree is designed primarily to provide both the model interpretability and predictive performance needed for high-stakes applications like medicine, it also can produce decision trees represented by hybrid models between CART and boosted stumps that can outperform either of these approaches. | |||
Kiri Wagstaff, Marie desJardins, & Eric Eaton |
Modeling and learning user preferences over sets
Download:
[ PDF ]
|
2010 | Journal of Experimental & Theoretical Artificial Intelligence 22(3) |
Kiri Wagstaff, Marie desJardins, & Eric Eaton. Modeling and learning user preferences over sets. Journal of Experimental & Theoretical Artificial Intelligence, 22(3):237–268, September 2010. | |||
@ARTICLE{Wagstaff2010Modeling, author = {Kiri Wagstaff and Marie desJardins and Eric Eaton}, year = {2010}, title = {Modeling and learning user preferences over sets}, journal = {Journal of Experimental & Theoretical Artificial Intelligence}, volume = {22}, number = {3}, pages = {237--268}, month = {September}, abstract = { Although there has been significant research on modeling and learning user preferences for various types of objects, there has been relatively little work on the problem of representing and learning preferences over <i>sets</i> of objects. We introduce a representation language, DD-PREF, that balances preferences for particular objects with preferences about the properties of the set. Specifically, we focus on the <i>depth</i> of objects (i.e. preferences for specific attribute values over others) and on the <i>diversity</i> of sets (i.e. preferences for broad vs. narrow distributions of attribute values). The DD-PREF framework is general and can incorporate additional object- and set-based preferences. We describe a greedy algorithm, DD-Select, for selecting satisfying sets from a collection of new objects, given a preference in this language. We show how preferences represented in DD-PREF can be learned from training data. Experimental results are given for three domains: a blocks world domain with several different task-based preferences, a real-world music playlist collection, and rover image data gathered in desert training exercises. } } | |||
Abstract:Although there has been significant research on modeling and learning user preferences for various types of objects, there has been relatively little work on the problem of representing and learning preferences over <i>sets</i> of objects. We introduce a representation language, DD-PREF, that balances preferences for particular objects with preferences about the properties of the set. Specifically, we focus on the <i>depth</i> of objects (i.e. preferences for specific attribute values over others) and on the <i>diversity</i> of sets (i.e. preferences for broad vs. narrow distributions of attribute values). The DD-PREF framework is general and can incorporate additional object- and set-based preferences. We describe a greedy algorithm, DD-Select, for selecting satisfying sets from a collection of new objects, given a preference in this language. We show how preferences represented in DD-PREF can be learned from training data. Experimental results are given for three domains: a blocks world domain with several different task-based preferences, a real-world music playlist collection, and rover image data gathered in desert training exercises. | |||
Eric Eaton, Gary Holness, & Daniel McFarlane |
Interactive Learning using Manifold Geometry
Download:
[ PDF ] [ Slides ]
|
2010 | AAAI Conference on Artificial Intelligence (AAAI-10) |
Eric Eaton, Gary Holness, & Daniel McFarlane. Interactive Learning using Manifold Geometry. In Proceedings of the 24th AAAI Conference on Artificial Intelligence (AAAI-10), pp. 437–443, AAAI Press, July 11--15 2010. | |||
@INPROCEEDINGS{Eaton2010Interactive, author = {Eric Eaton and Gary Holness and Daniel McFarlane}, title = {Interactive Learning using Manifold Geometry}, booktitle = {Proceedings of the 24th AAAI Conference on Artificial Intelligence (AAAI-10)}, month = {July 11--15}, location = {Atlanta, GA}, publisher = {AAAI Press}, pages = {437--443}, year = {2010}, abstract = { We present an interactive learning method that enables a user to iteratively refine a regression model. The user examines the output of the model, visualized as the vertical axis of a 2D scatterplot, and provides corrections by repositioning individual data instances to the correct output level. Each repositioned data instance acts as a control point for altering the learned model, using the geometry underlying the data. We capture the underlying structure of the data as a manifold, on which we compute a set of basis functions as the foundation for learning. Our results show that manifold-based interactive learning improves performance monotonically with each correction, outperforming alternative approaches.} } | |||
Abstract:We present an interactive learning method that enables a user to iteratively refine a regression model. The user examines the output of the model, visualized as the vertical axis of a 2D scatterplot, and provides corrections by repositioning individual data instances to the correct output level. Each repositioned data instance acts as a control point for altering the learned model, using the geometry underlying the data. We capture the underlying structure of the data as a manifold, on which we compute a set of basis functions as the foundation for learning. Our results show that manifold-based interactive learning improves performance monotonically with each correction, outperforming alternative approaches. | |||
Eric Eaton, Gary Holness, & Daniel McFarlane | Interactive Learning using Manifold Geometry Superseded by the AAAI-10 conference paper Interactive Learning using Manifold Geometry. |
2009 | AAAI Fall Symposium on Manifold Learning and Its Applications (AAAI Technical Report FS-09-04) |
Eric Eaton, Gary Holness, & Daniel McFarlane. Interactive Learning using Manifold Geometry. In Proceedings of the AAAI Fall Symposium on Manifold Learning and Its Applications (AAAI Technical Report FS-09-04), pp. 10–17, AAAI Press, November 5--7 2009. | |||
@INPROCEEDINGS{Eaton2009Interactive, author = {Eric Eaton and Gary Holness and Daniel McFarlane}, title = {Interactive Learning using Manifold Geometry}, booktitle = {Proceedings of the AAAI Fall Symposium on Manifold Learning and Its Applications (AAAI Technical Report FS-09-04)}, month = {November 5--7}, location = {Arlington, VA}, publisher = {AAAI Press}, pages = {10--17}, year = {2009}, } | |||
Marie desJardins, Eric Eaton, & Kiri Wagstaff |
Learning user preferences for sets of objects Awarded recognition as a NASA Tech Brief in 2008 |
2006 | International Conference on Machine Learning (ICML) |
Marie desJardins, Eric Eaton, & Kiri Wagstaff. Learning user preferences for sets of objects. In Proceedings of the 23rd International Conference on Machine Learning (ICML), Pittsburgh, PA, June 25--29 2006. Acceptance rate: 20% | |||
@INPROCEEDINGS{desJardins2006Learning, author = {Marie desJardins and Eric Eaton and Kiri Wagstaff}, title = {Learning user preferences for sets of objects}, booktitle = icml06, year = {2006}, month = {June 25--29}, address = {Pittsburgh, PA}, note = {Acceptance rate: 20%}, abstract = { Most work on preference learning has focused on pairwise preferences or rankings over individual items. In this paper, we present a method for learning preferences over sets of items. Our learning method takes as input a collection of positive examples--that is, one or more sets that have been identified by a user as desirable. Kernel density estimation is used to estimate the value function for individual items, and the desired set diversity is estimated from the average set diversity observed in the collection. Since this is a new learning problem, we introduce a new evaluation methodology and evaluate the learning method on two data collections: synthetic blocks-world data and a new real-world music data collection that we have gathered.} } | |||
Abstract:Most work on preference learning has focused on pairwise preferences or rankings over individual items. In this paper, we present a method for learning preferences over sets of items. Our learning method takes as input a collection of positive examples--that is, one or more sets that have been identified by a user as desirable. Kernel density estimation is used to estimate the value function for individual items, and the desired set diversity is estimated from the average set diversity observed in the collection. Since this is a new learning problem, we introduce a new evaluation methodology and evaluate the learning method on two data collections: synthetic blocks-world data and a new real-world music data collection that we have gathered. | |||
Eric Eaton, Marie desJardins, & Kiri Wagstaff |
DDPref Software: Learning preferences for sets of objects
Download:
[ Code ]
|
2006 | Available online at: http://maple.cs.umbc.edu/ ericeaton/software/DDPref.zip |
Eric Eaton, Marie desJardins, & Kiri Wagstaff. DDPref Software: Learning preferences for sets of objects. Available online at: http://maple.cs.umbc.edu/ ericeaton/software/DDPref.zip, 2006. | |||
@misc{Eaton2006DDPrefSoftware, author = {Eric Eaton and Marie desJardins and Kiri Wagstaff}, year = {2006}, title = {DDPref Software: Learning preferences for sets of objects}, howpublished = {Available online at: http://maple.cs.umbc.edu/~ericeaton/software/DDPref.zip}, } | |||
Marie desJardins, Eric Eaton, & Kiri Wagstaff |
A context-sensitive and user-centric approach to developing personal assistants
Download:
[ PDF ]
|
2005 | AAAI Spring Symposium on Persistent Assistants |
Marie desJardins, Eric Eaton, & Kiri Wagstaff. A context-sensitive and user-centric approach to developing personal assistants. In Proceedings of the AAAI Spring Symposium on Persistent Assistants, pp. 98–100, Stanford, CA, March 21--23 2005. | |||
@INPROCEEDINGS{desJardins2005ContextSensitive, author = {Marie desJardins and Eric Eaton and Kiri Wagstaff}, title = {A context-sensitive and user-centric approach to developing personal assistants}, booktitle = {Proceedings of the AAAI Spring Symposium on Persistent Assistants}, year = {2005}, month = {March 21--23}, address = {Stanford, CA}, pages = {98--100}, abstract = { Several ongoing projects in the MAPLE (Multi-Agent Planning and LEarning) lab at UMBC and the Machine Learning Systems Group at JPL focus on problems that we view as central to the development of persistent agents. This position paper describes our current research in this area, focusing on four topics in particular: effective use of observational and active learning, utilizing repeated behavioral contexts, clustering with annotated constraints, and learning user preferences.} } | |||
Abstract:Several ongoing projects in the MAPLE (Multi-Agent Planning and LEarning) lab at UMBC and the Machine Learning Systems Group at JPL focus on problems that we view as central to the development of persistent agents. This position paper describes our current research in this area, focusing on four topics in particular: effective use of observational and active learning, utilizing repeated behavioral contexts, clustering with annotated constraints, and learning user preferences. | |||
Constrained Clustering
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Eric Eaton, Marie desJardins, & Sara Jacob |
Multi-view constrained clustering with an incomplete mapping between views
Download:
[ PDF ]
|
2014 | Knowledge and Information Systems 38(1) |
Eric Eaton, Marie desJardins, & Sara Jacob. Multi-view constrained clustering with an incomplete mapping between views. Knowledge and Information Systems, 38(1):231–257, January 2014. | |||
@ARTICLE{Eaton2012MultiView, author = {Eric Eaton and Marie desJardins and Sara Jacob}, title = {Multi-view constrained clustering with an incomplete mapping between views}, journal = {Knowledge and Information Systems}, volume = {38}, number = {1}, pages = {231--257}, month = {January}, year = {2014}, notes = {Published online November 2012, <a href="http://link.springer.com/content/pdf/10.1007%2Fs10115-012-0577-7">doi 10.1007/s10115-012-0577-7</a>.}, abstract = { Multi-view learning algorithms typically assume a complete bipartite mapping between the different views in order to exchange information during the learning process. However, many applications provide only a partial mapping between the views, creating a challenge for current methods. To address this problem, we propose a multi-view algorithm based on constrained clustering that can operate with an incomplete mapping. Given a set of pairwise constraints in each view, our approach propagates these constraints using a local similarity measure to those instances that can be mapped to the other views, allowing the propagated constraints to be transferred across views via the partial mapping. It uses co-EM to iteratively estimate the propagation within each view based on the current clustering model, transfer the constraints across views, and then update the clustering model. By alternating the learning process between views, this approach produces a unified clustering model that is consistent with all views. We show that this approach significantly improves clustering performance over several other methods for transferring constraints and allows multi-view clustering to be reliably applied when given a limited mapping between the views. Our evaluation reveals that the propagated constraints have high precision with respect to the true clusters in the data, explaining their benefit to clustering performance in both single- and multi-view learning scenarios.} } | |||
Abstract:Multi-view learning algorithms typically assume a complete bipartite mapping between the different views in order to exchange information during the learning process. However, many applications provide only a partial mapping between the views, creating a challenge for current methods. To address this problem, we propose a multi-view algorithm based on constrained clustering that can operate with an incomplete mapping. Given a set of pairwise constraints in each view, our approach propagates these constraints using a local similarity measure to those instances that can be mapped to the other views, allowing the propagated constraints to be transferred across views via the partial mapping. It uses co-EM to iteratively estimate the propagation within each view based on the current clustering model, transfer the constraints across views, and then update the clustering model. By alternating the learning process between views, this approach produces a unified clustering model that is consistent with all views. We show that this approach significantly improves clustering performance over several other methods for transferring constraints and allows multi-view clustering to be reliably applied when given a limited mapping between the views. Our evaluation reveals that the propagated constraints have high precision with respect to the true clusters in the data, explaining their benefit to clustering performance in both single- and multi-view learning scenarios. | |||
Eric Eaton, Marie desJardins, & Sara Jacob |
Multi-View Clustering with Constraint Propagation for Learning with an Incomplete Mapping Between Views
Download:
[ PDF ] [ Slides ]
|
2010 | Conference on Information and Knowledge Management (CIKM'10) |
Eric Eaton, Marie desJardins, & Sara Jacob. Multi-View Clustering with Constraint Propagation for Learning with an Incomplete Mapping Between Views. In Proceedings of the Conference on Information and Knowledge Management (CIKM'10), pp. 389–398, ACM Press, October 26--30 2010. | |||
@INPROCEEDINGS{Eaton2010MultiView, author = {Eric Eaton and Marie desJardins and Sara Jacob}, title = {Multi-View Clustering with Constraint Propagation for Learning with an Incomplete Mapping Between Views}, booktitle = {Proceedings of the Conference on Information and Knowledge Management (CIKM'10)}, month = {October 26--30}, location = {Toronto, Ontario, Canada}, publisher = {ACM Press}, pages = {389--398}, year = {2010}, abstract = { Multi-view learning algorithms typically assume a complete bipartite mapping between the different views in order to exchange information during the learning process. However, many applications provide only a partial mapping between the views, creating a challenge for current methods. To address this problem, we propose a multi-view algorithm based on constrained clustering that can operate with an incomplete mapping. Given a set of pairwise constraints in each view, our approach propagates these constraints using a local similarity measure to those instances that can be mapped to the other views, allowing the propagated constraints to be transferred across views via the partial mapping. It uses co-EM to iteratively estimate the propagation within each view based on the current clustering model, transfer the constraints across views, and update the clustering model, thereby learning a unified model for all views. We show that this approach significantly improves clustering performance over several other methods for transferring constraints and allows multi-view clustering to be reliably applied when given a limited mapping between the views. } } | |||
Abstract:Multi-view learning algorithms typically assume a complete bipartite mapping between the different views in order to exchange information during the learning process. However, many applications provide only a partial mapping between the views, creating a challenge for current methods. To address this problem, we propose a multi-view algorithm based on constrained clustering that can operate with an incomplete mapping. Given a set of pairwise constraints in each view, our approach propagates these constraints using a local similarity measure to those instances that can be mapped to the other views, allowing the propagated constraints to be transferred across views via the partial mapping. It uses co-EM to iteratively estimate the propagation within each view based on the current clustering model, transfer the constraints across views, and update the clustering model, thereby learning a unified model for all views. We show that this approach significantly improves clustering performance over several other methods for transferring constraints and allows multi-view clustering to be reliably applied when given a limited mapping between the views. | |||
Eric Eaton |
Clustering with Propagated Constraints
Download:
[ PDF ]
|
2005 | Master's Thesis, University of Maryland Baltimore County |
Eric Eaton. Clustering with Propagated Constraints. Master's Thesis, University of Maryland Baltimore County, 2005. | |||
@MASTERSTHESIS{Eaton2005MastersThesis, author = {Eric Eaton}, title = {Clustering with Propagated Constraints}, school = {University of Maryland Baltimore County}, year = {2005}, abstract = { Background knowledge in the form of constraints can dramatically improve the quality of generated clustering models. In constrained clustering, these constraints typically specify the relative cluster membership of pairs of points. They are tedious to specify and expensive from a user perspective, yet are very useful in large quantities. Existing constrained clustering methods perform well when given large quantities of constraints, but do not focus on performing well when given very small quantities. This thesis focuses on providing a high-quality clustering with small quantities of constraints. It proposes a method for propagating pairwise constraints to nearby instances using a Gaussian function. This method takes a few easily specified constraints, and propagates them to nearby pairs of points to constrain the local neighborhood. Clustering with these propagated constraints can yield superior performance with fewer constraints than clustering with only the original user-specified constraints. The experiments compare the performance of clustering with propagated constraints to that of established constrained clustering algorithms on several real-world data sets.} } | |||
Abstract:Background knowledge in the form of constraints can dramatically improve the quality of generated clustering models. In constrained clustering, these constraints typically specify the relative cluster membership of pairs of points. They are tedious to specify and expensive from a user perspective, yet are very useful in large quantities. Existing constrained clustering methods perform well when given large quantities of constraints, but do not focus on performing well when given very small quantities. This thesis focuses on providing a high-quality clustering with small quantities of constraints. It proposes a method for propagating pairwise constraints to nearby instances using a Gaussian function. This method takes a few easily specified constraints, and propagates them to nearby pairs of points to constrain the local neighborhood. Clustering with these propagated constraints can yield superior performance with fewer constraints than clustering with only the original user-specified constraints. The experiments compare the performance of clustering with propagated constraints to that of established constrained clustering algorithms on several real-world data sets. | |||
Relational Network Analysis
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Rasul Tutunov, Haitham Bou Ammar, Ali Jadbabaie, & Eric Eaton |
On the degree distribution of Pólya urn graph processes
Download:
[ PDF ]
|
2014 | arXiv:1410.8515 Preprint |
Rasul Tutunov, Haitham Bou Ammar, Ali Jadbabaie, & Eric Eaton. On the degree distribution of Pólya urn graph processes. arXiv:1410.8515 Preprint, October 2014. | |||
@ARTICLE{Tutunov2014DegreeDistribution, author = {Rasul Tutunov and Haitham {Bou Ammar} and Ali Jadbabaie and Eric Eaton}, year = {2014}, title = {On the degree distribution of P\'{o}lya urn graph processes}, journal = {arXiv:1410.8515 Preprint}, month = {October}, abstract = { This paper presents a tighter bound on the degree distribution of arbitrary Polya urn graph processes, proving that the proportion of vertices with degree d obeys a power-law distribution P(d) proportional to d^(-gamma) for d <= n^(1/6 - epsilon) for any epsilon > 0, where n represents the number of vertices in the network. Previous work by Bollobas et al. formalized the well-known preferential attachment model of Barabasi and Albert, and showed that the power-law distribution held for d <= n^(1/15) with gamma = 3. Our revised bound represents a significant improvement over existing models of degree distribution in scale-free networks, where its tightness is restricted by the Azuma-Hoeffding concentration inequality for martingales. We achieve this tighter bound through a careful analysis of the first set of vertices in the network generation process, and show that the newly acquired is at the edge of exhausting Bollobas model in the sense that the degree expectation breaks down for other powers.} } | |||
Abstract:This paper presents a tighter bound on the degree distribution of arbitrary Polya urn graph processes, proving that the proportion of vertices with degree d obeys a power-law distribution P(d) proportional to d^(-gamma) for d <= n^(1/6 - epsilon) for any epsilon > 0, where n represents the number of vertices in the network. Previous work by Bollobas et al. formalized the well-known preferential attachment model of Barabasi and Albert, and showed that the power-law distribution held for d <= n^(1/15) with gamma = 3. Our revised bound represents a significant improvement over existing models of degree distribution in scale-free networks, where its tightness is restricted by the Azuma-Hoeffding concentration inequality for martingales. We achieve this tighter bound through a careful analysis of the first set of vertices in the network generation process, and show that the newly acquired is at the edge of exhausting Bollobas model in the sense that the degree expectation breaks down for other powers. | |||
Eric Eaton & Rachael Mansbach | A spin-glass model for semi-supervised community detection | 2012 | AAAI Conference on Artificial Intelligence (AAAI-12) |
Eric Eaton & Rachael Mansbach. A spin-glass model for semi-supervised community detection. In Proceedings of the 26th AAAI Conference on Artificial Intelligence (AAAI-12), pp. 900–906, AAAI Press, July 22--26 2012. | |||
@INPROCEEDINGS{Eaton2012SpinGlass, author = {Eric Eaton and Rachael Mansbach}, title = {A spin-glass model for semi-supervised community detection}, booktitle = {Proceedings of the 26th AAAI Conference on Artificial Intelligence (AAAI-12)}, month = {July 22--26}, location = {Toronto, Canada}, publisher = {AAAI Press}, pages = {900--906}, year = {2012}, abstract = { Current modularity-based community detection methods show decreased performance as relational networks become increasingly noisy. These methods also yield a large number of diverse community structures as solutions, which is problematic for applications that impose constraints on the acceptable solutions or in cases where the user is focused on specific communities of interest. To address both of these problems, we develop a semi-supervised spin-glass model that enables current community detection methods to incorporate background knowledge in the forms of individual labels and pairwise constraints. Unlike current methods, our approach shows robust performance in the presence of noise in the relational network, and the ability to guide the discovery process toward specific community structures. We evaluate our algorithm on several benchmark networks and a new political sentiment network representing cooperative events between nations that was mined from news articles over six years.} } | |||
Abstract:Current modularity-based community detection methods show decreased performance as relational networks become increasingly noisy. These methods also yield a large number of diverse community structures as solutions, which is problematic for applications that impose constraints on the acceptable solutions or in cases where the user is focused on specific communities of interest. To address both of these problems, we develop a semi-supervised spin-glass model that enables current community detection methods to incorporate background knowledge in the forms of individual labels and pairwise constraints. Unlike current methods, our approach shows robust performance in the presence of noise in the relational network, and the ability to guide the discovery process toward specific community structures. We evaluate our algorithm on several benchmark networks and a new political sentiment network representing cooperative events between nations that was mined from news articles over six years. | |||
Computational Sustainability
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Pengyuan Shen, William Braham, Yun Kyu Yi, & Eric Eaton |
Rapid multi-objective optimization with multi-year future weather condition and decision-making support for building retrofit
Download:
[ PDF ] [ Website ]
|
2019 | Energy |
Pengyuan Shen, William Braham, Yun Kyu Yi, & Eric Eaton. Rapid multi-objective optimization with multi-year future weather condition and decision-making support for building retrofit. Energy, 172:892–912, Elsevier Limited, April 2019. | |||
@article{Shen2019Rapid, title = "Rapid multi-objective optimization with multi-year future weather condition and decision-making support for building retrofit", keywords = "Building retrofit, Climate change, Heuristic method, Hierarchical clustering, Optimization, Pareto fronts", author = "Pengyuan Shen and William Braham and Yi, {Yun Kyu} and Eric Eaton", year = "2019", month = {April}, volume = "172", pages = "892--912", journal = "Energy", publisher = "Elsevier Limited", abstract = { A method of fast multi-objective optimization and decision-making support for building retrofit planning is developed, and lifecycle cost analysis method taking into account of future climate condition is used in evaluating the retrofit performance. In order to resolve the optimization problem in a fast manner with recourse to non-dominate sorting differential evolution algorithm, the simplified hourly dynamic simulation modeling tool SimBldPy is used as the simulator for objective function evaluation. Moreover, the generated non-dominated solutions are treated and rendered by a layered scheme using agglomerative hierarchical clustering technique to make it more intuitive and sense making during the decision-making process as well as to be better presented. The suggested optimization method is implemented to the retrofit planning of a campus building in UPenn with various energy conservation measures (ECM) and costs, and more than one thousand Pareto fronts are obtained and being analyzed according to the proposed decision-making framework. Twenty ECM combinations are eventually selected from all generated Pareto fronts. It is manifested that the developed decision-making support scheme shows robustness in dealing with retrofit optimization problem and is able to provide support for brainstorming and enumerating various possibilities during the decision-making process.}, } | |||
Abstract:A method of fast multi-objective optimization and decision-making support for building retrofit planning is developed, and lifecycle cost analysis method taking into account of future climate condition is used in evaluating the retrofit performance. In order to resolve the optimization problem in a fast manner with recourse to non-dominate sorting differential evolution algorithm, the simplified hourly dynamic simulation modeling tool SimBldPy is used as the simulator for objective function evaluation. Moreover, the generated non-dominated solutions are treated and rendered by a layered scheme using agglomerative hierarchical clustering technique to make it more intuitive and sense making during the decision-making process as well as to be better presented. The suggested optimization method is implemented to the retrofit planning of a campus building in UPenn with various energy conservation measures (ECM) and costs, and more than one thousand Pareto fronts are obtained and being analyzed according to the proposed decision-making framework. Twenty ECM combinations are eventually selected from all generated Pareto fronts. It is manifested that the developed decision-making support scheme shows robustness in dealing with retrofit optimization problem and is able to provide support for brainstorming and enumerating various possibilities during the decision-making process. | |||
Eric Eaton, Carla Gomes, & Brian Williams, editors |
Special Issue of AI Magazine on Computational Sustainability
Download:
[ Website ]
|
2014 | AAAI Press |
Eric Eaton, Carla Gomes, & Brian Williams, editors. Special Issue of AI Magazine on Computational Sustainability, AI Magazine 35(2-3) AAAI Press, June 2014. | |||
@book{Eaton2014AIMagazine, editor = {Eric Eaton and Carla Gomes and Brian Williams}, title = {Special Issue of AI Magazine on Computational Sustainability}, series = {AI Magazine}, volume = {35}, number = {2-3}, month = {June}, year = {2014}, publisher = {AAAI Press}, url = {http://www.aaai.org/ojs/index.php/aimagazine/issue/view/206} } | |||
Eric Eaton, Carla Gomes, & Brian Williams |
Computational Sustainability
Download:
[ PDF ]
|
2014 | AI Magazine 35(2) |
Eric Eaton, Carla Gomes, & Brian Williams. Computational Sustainability. AI Magazine, 35(2):3–7, June 2014. | |||
@article{Eaton2014CompSustainability, author = {Eric Eaton and Carla Gomes and Brian Williams}, title = {Computational Sustainability}, journal = {AI Magazine}, volume = {35}, number = {2}, pages = {3--7}, year = {2014}, month = {June}, abstract = { Computational sustainability problems, which exist in dynamic environments with high amounts of uncertainty, provide a variety of unique challenges to artificial intelligence research and the opportunity for significant impact upon our collective future. This editorial provides an overview of artificial intelligence for computational sustainability, and introduces this special issue of AI Magazine.} } | |||
Abstract:Computational sustainability problems, which exist in dynamic environments with high amounts of uncertainty, provide a variety of unique challenges to artificial intelligence research and the opportunity for significant impact upon our collective future. This editorial provides an overview of artificial intelligence for computational sustainability, and introduces this special issue of AI Magazine. | |||
Medicine
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Efstathios D. Gennatas, Jerome H. Friedman, Lyle H. Ungar, Romain Pirracchio, Eric Eaton, Lara G. Reichmann, Yannet Interian, Jose Marcio Luna, Charles B. Simone, Andrew Auerbach, Elier Delgado, Mark J. van der Laan, Timothy D. Solberg, & Gilmer Valdes | Expert-augmented machine learning | 2020 | Proceedings of the National Academy of Sciences 117(9) |
Efstathios D. Gennatas, Jerome H. Friedman, Lyle H. Ungar, Romain Pirracchio, Eric Eaton, Lara G. Reichmann, Yannet Interian, Jose Marcio Luna, Charles B. Simone, Andrew Auerbach, Elier Delgado, Mark J. van der Laan, Timothy D. Solberg, & Gilmer Valdes. Expert-augmented machine learning. Proceedings of the National Academy of Sciences, 117(9):4571–4577, National Academy of Sciences, 2020. | |||
@article {Gennatas2020ExpertAugmented, author = {Gennatas, Efstathios D. and Friedman, Jerome H. and Ungar, Lyle H. and Pirracchio, Romain and Eaton, Eric and Reichmann, Lara G. and Interian, Yannet and Luna, Jose Marcio and Simone, Charles B. and Auerbach, Andrew and Delgado, Elier and van der Laan, Mark J. and Solberg, Timothy D. and Valdes, Gilmer}, title = {Expert-augmented machine learning}, volume = {117}, number = {9}, pages = {4571--4577}, year = {2020}, publisher = {National Academy of Sciences}, abstract = { Machine learning is increasingly used across fields to derive insights from data, which further our understanding of the world and help us anticipate the future. The performance of predictive modeling is dependent on the amount and quality of available data. In practice, we rely on human experts to perform certain tasks and on machine learning for others. However, the optimal learning strategy may involve combining the complementary strengths of humans and machines. We present expert-augmented machine learning, an automated way to automatically extract problem-specific human expert knowledge and integrate it with machine learning to build robust, dependable, and data-efficient predictive models.Machine learning is proving invaluable across disciplines. However, its success is often limited by the quality and quantity of available data, while its adoption is limited by the level of trust afforded by given models. Human vs. machine performance is commonly compared empirically to decide whether a certain task should be performed by a computer or an expert. In reality, the optimal learning strategy may involve combining the complementary strengths of humans and machines. Here, we present expert-augmented machine learning (EAML), an automated method that guides the extraction of expert knowledge and its integration into machine-learned models. We used a large dataset of intensive-care patient data to derive 126 decision rules that predict hospital mortality. Using an online platform, we asked 15 clinicians to assess the relative risk of the subpopulation defined by each rule compared to the total sample. We compared the clinician-assessed risk to the empirical risk and found that, while clinicians agreed with the data in most cases, there were notable exceptions where they overestimated or underestimated the true risk. Studying the rules with greatest disagreement, we identified problems with the training data, including one miscoded variable and one hidden confounder. Filtering the rules based on the extent of disagreement between clinician-assessed risk and empirical risk, we improved performance on out-of-sample data and were able to train with less data. EAML provides a platform for automated creation of problem-specific priors, which help build robust and dependable machine-learning models in critical applications. }, journal = {Proceedings of the National Academy of Sciences}, } | |||
Abstract:Machine learning is increasingly used across fields to derive insights from data, which further our understanding of the world and help us anticipate the future. The performance of predictive modeling is dependent on the amount and quality of available data. In practice, we rely on human experts to perform certain tasks and on machine learning for others. However, the optimal learning strategy may involve combining the complementary strengths of humans and machines. We present expert-augmented machine learning, an automated way to automatically extract problem-specific human expert knowledge and integrate it with machine learning to build robust, dependable, and data-efficient predictive models.Machine learning is proving invaluable across disciplines. However, its success is often limited by the quality and quantity of available data, while its adoption is limited by the level of trust afforded by given models. Human vs. machine performance is commonly compared empirically to decide whether a certain task should be performed by a computer or an expert. In reality, the optimal learning strategy may involve combining the complementary strengths of humans and machines. Here, we present expert-augmented machine learning (EAML), an automated method that guides the extraction of expert knowledge and its integration into machine-learned models. We used a large dataset of intensive-care patient data to derive 126 decision rules that predict hospital mortality. Using an online platform, we asked 15 clinicians to assess the relative risk of the subpopulation defined by each rule compared to the total sample. We compared the clinician-assessed risk to the empirical risk and found that, while clinicians agreed with the data in most cases, there were notable exceptions where they overestimated or underestimated the true risk. Studying the rules with greatest disagreement, we identified problems with the training data, including one miscoded variable and one hidden confounder. Filtering the rules based on the extent of disagreement between clinician-assessed risk and empirical risk, we improved performance on out-of-sample data and were able to train with less data. EAML provides a platform for automated creation of problem-specific priors, which help build robust and dependable machine-learning models in critical applications. | |||
Jose Marcio Luna, Efstathios D. Gennatas, Lyle H. Ungar, Eric Eaton, Eric S. Diffenderfer, Shane T. Jensen, Charles B. Simone, Jerome H. Friedman, Timothy D. Solberg, & Gilmer Valdes | Building more accurate decision trees with the additive tree | 2019 | Proceedings of the National Academy of Sciences 116(40) |
Jose Marcio Luna, Efstathios D. Gennatas, Lyle H. Ungar, Eric Eaton, Eric S. Diffenderfer, Shane T. Jensen, Charles B. Simone, Jerome H. Friedman, Timothy D. Solberg, & Gilmer Valdes. Building more accurate decision trees with the additive tree. Proceedings of the National Academy of Sciences, 116(40):19887–19893, National Academy of Sciences, 2019. | |||
@article{Luna2019Building, author = {Luna, Jose Marcio and Gennatas, Efstathios D. and Ungar, Lyle H. and Eaton, Eric and Diffenderfer, Eric S. and Jensen, Shane T. and Simone, Charles B. and Friedman, Jerome H. and Solberg, Timothy D. and Valdes, Gilmer}, title = {Building more accurate decision trees with the additive tree}, volume = {116}, number = {40}, pages = {19887--19893}, year = {2019}, publisher = {National Academy of Sciences}, journal = {Proceedings of the National Academy of Sciences}, abstract = { As machine learning applications expand to high-stakes areas such as criminal justice, finance, and medicine, legitimate concerns emerge about high-impact effects of individual mispredictions on people{\textquoteright}s lives. As a result, there has been increasing interest in understanding general machine learning models to overcome possible serious risks. Current decision trees, such as Classification and Regression Trees (CART), have played a predominant role in fields such as medicine, due to their simplicity and intuitive interpretation. However, such trees suffer from intrinsic limitations in predictive power. We developed the additive tree, a theoretical approach to generate a more accurate and interpretable decision tree, which reveals connections between CART and gradient boosting. The additive tree exhibits superior predictive performance to CART, as validated on 83 classification tasks.The expansion of machine learning to high-stakes application domains such as medicine, finance, and criminal justice, where making informed decisions requires clear understanding of the model, has increased the interest in interpretable machine learning. The widely used Classification and Regression Trees (CART) have played a major role in health sciences, due to their simple and intuitive explanation of predictions. Ensemble methods like gradient boosting can improve the accuracy of decision trees, but at the expense of the interpretability of the generated model. Additive models, such as those produced by gradient boosting, and full interaction models, such as CART, have been investigated largely in isolation. We show that these models exist along a spectrum, revealing previously unseen connections between these approaches. This paper introduces a rigorous formalization for the additive tree, an empirically validated learning technique for creating a single decision tree, and shows that this method can produce models equivalent to CART or gradient boosted stumps at the extremes by varying a single parameter. Although the additive tree is designed primarily to provide both the model interpretability and predictive performance needed for high-stakes applications like medicine, it also can produce decision trees represented by hybrid models between CART and boosted stumps that can outperform either of these approaches.}, } | |||
Abstract:As machine learning applications expand to high-stakes areas such as criminal justice, finance, and medicine, legitimate concerns emerge about high-impact effects of individual mispredictions on people\textquoterights lives. As a result, there has been increasing interest in understanding general machine learning models to overcome possible serious risks. Current decision trees, such as Classification and Regression Trees (CART), have played a predominant role in fields such as medicine, due to their simplicity and intuitive interpretation. However, such trees suffer from intrinsic limitations in predictive power. We developed the additive tree, a theoretical approach to generate a more accurate and interpretable decision tree, which reveals connections between CART and gradient boosting. The additive tree exhibits superior predictive performance to CART, as validated on 83 classification tasks.The expansion of machine learning to high-stakes application domains such as medicine, finance, and criminal justice, where making informed decisions requires clear understanding of the model, has increased the interest in interpretable machine learning. The widely used Classification and Regression Trees (CART) have played a major role in health sciences, due to their simple and intuitive explanation of predictions. Ensemble methods like gradient boosting can improve the accuracy of decision trees, but at the expense of the interpretability of the generated model. Additive models, such as those produced by gradient boosting, and full interaction models, such as CART, have been investigated largely in isolation. We show that these models exist along a spectrum, revealing previously unseen connections between these approaches. This paper introduces a rigorous formalization for the additive tree, an empirically validated learning technique for creating a single decision tree, and shows that this method can produce models equivalent to CART or gradient boosted stumps at the extremes by varying a single parameter. Although the additive tree is designed primarily to provide both the model interpretability and predictive performance needed for high-stakes applications like medicine, it also can produce decision trees represented by hybrid models between CART and boosted stumps that can outperform either of these approaches. | |||
Julia E. Reid & Eric Eaton |
Artificial intelligence for pediatric ophthalmology
Download:
[ PDF ] [ Website ]
|
2019 | Current Opinion in Ophthalmology 30(5) |
Julia E. Reid & Eric Eaton. Artificial intelligence for pediatric ophthalmology. Current Opinion in Ophthalmology, 30(5):337–346, 2019. | |||
@article{Reid2019Artificial, title = {Artificial intelligence for pediatric ophthalmology}, author = {Reid, Julia E. and Eaton, Eric}, journal = {Current Opinion in Ophthalmology}, volume = {30}, number = {5}, pages = {337-346}, year = {2019}, abstract = { PURPOSE OF REVIEW Despite the impressive results of recent artificial intelligence applications to general ophthalmology, comparatively less progress has been made toward solving problems in pediatric ophthalmology using similar techniques. This article discusses the unique needs of pediatric patients and how artificial intelligence techniques can address these challenges, surveys recent applications to pediatric ophthalmology, and discusses future directions. RECENT FINDINGS The most significant advances involve the automated detection of retinopathy of prematurity, yielding results that rival experts. Machine learning has also been applied to the classification of pediatric cataracts, prediction of postoperative complications following cataract surgery, detection of strabismus and refractive error, prediction of future high myopia, and diagnosis of reading disability. In addition, machine learning techniques have been used for the study of visual development, vessel segmentation in pediatric fundus images, and ophthalmic image synthesis. SUMMARY Artificial intelligence applications could significantly benefit clinical care by optimizing disease detection and grading, broadening access to care, furthering scientific discovery, and improving clinical efficiency. These methods need to match or surpass physician performance in clinical trials before deployment with patients. Owing to the widespread use of closed-access data sets and software implementations, it is difficult to directly compare the performance of these approaches, and reproducibility is poor. Open-access data sets and software could alleviate these issues and encourage further applications to pediatric ophthalmology. }, } | |||
Abstract:PURPOSE OF REVIEW Despite the impressive results of recent artificial intelligence applications to general ophthalmology, comparatively less progress has been made toward solving problems in pediatric ophthalmology using similar techniques. This article discusses the unique needs of pediatric patients and how artificial intelligence techniques can address these challenges, surveys recent applications to pediatric ophthalmology, and discusses future directions. RECENT FINDINGS The most significant advances involve the automated detection of retinopathy of prematurity, yielding results that rival experts. Machine learning has also been applied to the classification of pediatric cataracts, prediction of postoperative complications following cataract surgery, detection of strabismus and refractive error, prediction of future high myopia, and diagnosis of reading disability. In addition, machine learning techniques have been used for the study of visual development, vessel segmentation in pediatric fundus images, and ophthalmic image synthesis. SUMMARY Artificial intelligence applications could significantly benefit clinical care by optimizing disease detection and grading, broadening access to care, furthering scientific discovery, and improving clinical efficiency. These methods need to match or surpass physician performance in clinical trials before deployment with patients. Owing to the widespread use of closed-access data sets and software implementations, it is difficult to directly compare the performance of these approaches, and reproducibility is poor. Open-access data sets and software could alleviate these issues and encourage further applications to pediatric ophthalmology. | |||
Education
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Eric Eaton |
A lightweight approach to academic research group management using online tools: Spend more time on research and less on management
Download:
[ PDF ] [ Website ]
|
2019 | Educational Advances in Artificial Intelligene (EAAI) Symposium |
Eric Eaton. A lightweight approach to academic research group management using online tools: Spend more time on research and less on management. In Proceedings of the Educational Advances in Artificial Intelligene (EAAI) Symposium, pp. 9644–9647, 2019. | |||
@inproceedings{Eaton2019Lightweight, title = {A lightweight approach to academic research group management using online tools: Spend more time on research and less on management}, author = {Eric Eaton}, booktitle = {Proceedings of the Educational Advances in Artificial Intelligene (EAAI) Symposium}, pages = {9644--9647}, year = {2019}, abstract = { After years of taking a trial-and-error approach to managing a moderate-size academic research group, I settled on using a set of online tools and protocols that seem effective, require relatively little effort to use and maintain, and are inexpensive. This paper discusses this approach to communication, project management, document and code management, and logistics. It is my hope that other researchers, especially new faculty and research scientists, might find this set of tools and protocols useful when determining how to manage their own research group. This paper is targeted toward research groups based in mathematics and engineering, although faculty in other disciplines may find inspiration in some of these ideas. }, } | |||
Abstract:After years of taking a trial-and-error approach to managing a moderate-size academic research group, I settled on using a set of online tools and protocols that seem effective, require relatively little effort to use and maintain, and are inexpensive. This paper discusses this approach to communication, project management, document and code management, and logistics. It is my hope that other researchers, especially new faculty and research scientists, might find this set of tools and protocols useful when determining how to manage their own research group. This paper is targeted toward research groups based in mathematics and engineering, although faculty in other disciplines may find inspiration in some of these ideas. | |||
Eric Eaton |
Teaching integrated AI through interdisciplinary project-driven courses
Download:
[ PDF ] [ Website ]
|
2017 | AI Magazine 38(2) |
Eric Eaton. Teaching integrated AI through interdisciplinary project-driven courses. AI Magazine, 38(2):13–21, 2017. | |||
@article{Eaton2017Teaching, title = {Teaching integrated {AI} through interdisciplinary project-driven courses}, author = {Eric Eaton}, journal = {AI Magazine}, volume = {38}, number = {2}, pages = {13--21}, year = {2017}, abstract = { Different subfields of AI (such as vision, learning, reasoning, planning, and others) are often studied in isolation, both in individual courses and in the research literature. This promulgates the idea that these different AI capabilities can easily be integrated later, whereas, in practice, developing integrated AI systems remains an open challenge for both research and industry. Interdisciplinary project-driven courses can fill this gap in AI education, providing challenging problems that require the integration of multiple AI methods. This article explores teaching integrated AI through two project-driven courses: a capstone-style graduate course in advanced robotics, and an undergraduate course on computational sustainability and assistive computing. In addition to studying the integration of AI techniques, these courses provide students with practical applications experience and exposure to social issues of AI and computing. My hope is that other instructors find these courses as useful examples for constructing their own project-driven courses to teach integrated AI. } } | |||
Abstract:Different subfields of AI (such as vision, learning, reasoning, planning, and others) are often studied in isolation, both in individual courses and in the research literature. This promulgates the idea that these different AI capabilities can easily be integrated later, whereas, in practice, developing integrated AI systems remains an open challenge for both research and industry. Interdisciplinary project-driven courses can fill this gap in AI education, providing challenging problems that require the integration of multiple AI methods. This article explores teaching integrated AI through two project-driven courses: a capstone-style graduate course in advanced robotics, and an undergraduate course on computational sustainability and assistive computing. In addition to studying the integration of AI techniques, these courses provide students with practical applications experience and exposure to social issues of AI and computing. My hope is that other instructors find these courses as useful examples for constructing their own project-driven courses to teach integrated AI. | |||
Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes |
Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text
Download:
[ PDF ] [ Poster ]
|
2012 | AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-12) |
Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes. Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text. In Proceedings of the 3rd AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-12), AAAI Press, July 23--24 2012. | |||
@INPROCEEDINGS{Fisher2012IncorporatingEAAI, author = {Douglas Fisher and Bistra Dilkina and Eric Eaton and Carla Gomes}, title = {Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text}, booktitle = {Proceedings of the 3rd AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-12)}, month = {July 23--24}, location = {Toronto, Canada}, publisher = {AAAI Press}, year = {2012}, abstract = { We introduce a laboratory text on environmental and societal sustainability applications that can be a supplemental resource for any undergraduate AI course. The lab text, entitled Artificial Intelligence for Computational Sustainability: A Lab Companion, is brand new and incomplete; freely available through Wikibooks; and open to community additions of projects, assignments, and explanatory material on AI for sustainability. The project adds to existing educational efforts of the computational sustainability community, encouraging the flow of knowledge from research to education and public outreach. Besides summarizing the laboratory book, this paper touches on its implications for integration of research and education, for communicating science to the public, and other broader impacts.} } | |||
Abstract:We introduce a laboratory text on environmental and societal sustainability applications that can be a supplemental resource for any undergraduate AI course. The lab text, entitled Artificial Intelligence for Computational Sustainability: A Lab Companion, is brand new and incomplete; freely available through Wikibooks; and open to community additions of projects, assignments, and explanatory material on AI for sustainability. The project adds to existing educational efforts of the computational sustainability community, encouraging the flow of knowledge from research to education and public outreach. Besides summarizing the laboratory book, this paper touches on its implications for integration of research and education, for communicating science to the public, and other broader impacts. | |||
Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes |
Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text [Oral Presentation]
Download:
[ PDF ]
|
2012 | International Conference on Computational Sustainability (CompSust'12) |
Douglas Fisher, Bistra Dilkina, Eric Eaton, & Carla Gomes. Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text [Oral Presentation]. In Proceedings of the 3rd International Conference on Computational Sustainability (CompSust'12), July 5--6 2012. | |||
@INPROCEEDINGS{Fisher2012IncorporatingCompSust, author = {Douglas Fisher and Bistra Dilkina and Eric Eaton and Carla Gomes}, title = {Incorporating computational sustainability into AI education through a freely-available, collectively-composed supplementary lab text [Oral Presentation]}, booktitle = {Proceedings of the 3rd International Conference on Computational Sustainability (CompSust'12)}, month = {July 5--6}, location = {Copenhagen, Denmark}, year = {2012}, } | |||
Eric Eaton |
Gridworld Search and Rescue: A Project Framework for a Course in Artificial Intelligence
Download:
[ PDF ] [ Website ]
|
2008 | AAAI-08 AI Education Colloquium |
Eric Eaton. Gridworld Search and Rescue: A Project Framework for a Course in Artificial Intelligence. In Proceedings of the AAAI-08 AI Education Colloquium, Chicago, IL, July 13 2008. | |||
@INPROCEEDINGS{Eaton2008Gridworld, author = {Eric Eaton}, title = {Gridworld Search and Rescue: A Project Framework for a Course in Artificial Intelligence}, booktitle = {Proceedings of the AAAI-08 AI Education Colloquium}, year = {2008}, month = {July 13}, address = {Chicago, IL}, keywords = {education, project framework, search and rescue, gridworld}, abstract = { This paper describes the Gridworld Search and Rescue simulator: freely available educational software that allows students to develop an intelligent agent for a search and rescue application in a partially observable gridworld. It permits students to focus on high-level AI issues for solving the problem rather than low-level robotic navigation. The complexity of the search and rescue problem supports a wide variety of solutions and AI techniques, including search, logical reasoning, planning, and machine learning, while the high-level GSAR simulator makes the complex problem manageable. The simulator represents a 2D disaster-stricken building for multiple rescue agents to explore and rescue autonomous injured victims. It was successfully used as the semester project for CMSC 471 (Artificial Intelligence) in Fall 2007 at UMBC.} } | |||
Abstract:This paper describes the Gridworld Search and Rescue simulator: freely available educational software that allows students to develop an intelligent agent for a search and rescue application in a partially observable gridworld. It permits students to focus on high-level AI issues for solving the problem rather than low-level robotic navigation. The complexity of the search and rescue problem supports a wide variety of solutions and AI techniques, including search, logical reasoning, planning, and machine learning, while the high-level GSAR simulator makes the complex problem manageable. The simulator represents a 2D disaster-stricken building for multiple rescue agents to explore and rescue autonomous injured victims. It was successfully used as the semester project for CMSC 471 (Artificial Intelligence) in Fall 2007 at UMBC. | |||
Eric Eaton |
Gridworld Search and Rescue Software
Download:
[ Website ]
|
2008 | Available online at: http://maple.cs.umbc.edu/ ericeaton/searchandrescue/ |
Eric Eaton. Gridworld Search and Rescue Software. Available online at: http://maple.cs.umbc.edu/ ericeaton/searchandrescue/, 2008. | |||
@misc{Eaton2008GridworldSoftware, author = {Eric Eaton}, year = {2008}, title = {Gridworld Search and Rescue Software}, howpublished = {Available online at: http://maple.cs.umbc.edu/~ericeaton/searchandrescue/}, } | |||
Other
AUTHORS | TITLE | YEAR | VENUE / PUBLISHER |
---|---|---|---|
Gilmer Valdes, Jose Marcio Luna, Eric Eaton, Charles B. Simone II, Lyle H. Ungar, & Timothy D. Solberg | MediBoost: a Patient Stratification Tool for Interpretable Decision Making in the Era of Precision Medicine | 2016 | Scientific Reports |
Gilmer Valdes, Jose Marcio Luna, Eric Eaton, Charles B. Simone II, Lyle H. Ungar, & Timothy D. Solberg. MediBoost: a Patient Stratification Tool for Interpretable Decision Making in the Era of Precision Medicine. Scientific Reports, 6:37854, November 2016. | |||
@article{Valdes2016MediBoost, title = {MediBoost: a Patient Stratification Tool for Interpretable Decision Making in the Era of Precision Medicine}, author = {Gilmer Valdes and Jose Marcio Luna and Eric Eaton and Charles B. {Simone II} and Lyle H. Ungar and Timothy D. Solberg}, journal = {Scientific Reports}, volume = {6}, pages = {37854}, year = {2016}, month = {November}, abstract = { Machine learning algorithms that are both interpretable and accurate are essential in applications such as medicine where errors can have a dire consequence. Unfortunately, there is currently a tradeoff between accuracy and interpretability among state-of-the-art methods. Decision trees are interpretable and are therefore used extensively throughout medicine for stratifying patients. Current decision tree algorithms, however, are consistently outperformed in accuracy by other, less-interpretable machine learning models, such as ensemble methods. We present MediBoost, a novel framework for constructing decision trees that retain interpretability while having accuracy similar to ensemble methods, and compare MediBoost’s performance to that of conventional decision trees and ensemble methods on 13 medical classification problems. MediBoost significantly outperformed current decision tree algorithms in 11 out of 13 problems, giving accuracy comparable to ensemble methods. The resulting trees are of the same type as decision trees used throughout clinical practice but have the advantage of improved accuracy. Our algorithm thus gives the best of both worlds: it grows a single, highly interpretable tree that has the high accuracy of ensemble methods.} } | |||
Abstract:Machine learning algorithms that are both interpretable and accurate are essential in applications such as medicine where errors can have a dire consequence. Unfortunately, there is currently a tradeoff between accuracy and interpretability among state-of-the-art methods. Decision trees are interpretable and are therefore used extensively throughout medicine for stratifying patients. Current decision tree algorithms, however, are consistently outperformed in accuracy by other, less-interpretable machine learning models, such as ensemble methods. We present MediBoost, a novel framework for constructing decision trees that retain interpretability while having accuracy similar to ensemble methods, and compare MediBoost’s performance to that of conventional decision trees and ensemble methods on 13 medical classification problems. MediBoost significantly outperformed current decision tree algorithms in 11 out of 13 problems, giving accuracy comparable to ensemble methods. The resulting trees are of the same type as decision trees used throughout clinical practice but have the advantage of improved accuracy. Our algorithm thus gives the best of both worlds: it grows a single, highly interpretable tree that has the high accuracy of ensemble methods. | |||
Eric Eaton, Caio Mucchiani, Mayumi Mohan, David Isele, Jose Marcio Luna, & Christopher Clingerman | Design of a low-cost platform for autonomous mobile service robots | 2016 | IJCAI-16 Workshop on Autonomous Mobile Service Robots |
Eric Eaton, Caio Mucchiani, Mayumi Mohan, David Isele, Jose Marcio Luna, & Christopher Clingerman. Design of a low-cost platform for autonomous mobile service robots. In IJCAI-16 Workshop on Autonomous Mobile Service Robots, July 2016. | |||
@INPROCEEDINGS{Eaton2016Design, author = {Eric Eaton and Caio Mucchiani and Mayumi Mohan and David Isele and Jose Marcio Luna and Christopher Clingerman}, year = {2016}, title = {Design of a low-cost platform for autonomous mobile service robots}, booktitle = {IJCAI-16 Workshop on Autonomous Mobile Service Robots}, month = {July}, abstract = { Most current autonomous mobile service robots are either expensive commercial platforms or custom manufactured for research environments, limiting their availability. We present the design for a low-cost service robot based on the widely used TurtleBot 2 platform, with the goal of making service robots affordable and accessible to the research, educational, and hobbyist communities. Our design uses a set of simple and inexpensive modifications to transform the TurtleBot 2 into a 4.5ft (1.37m) tall tour-guide or telepresence-style robot, capable of performing a wide variety of indoor service tasks. The resulting platform provides a shoulder-height touchscreen and 3D camera for interaction, an optional low-cost arm for manipulation, enhanced onboard computation, autonomous charging, and up to 6 hours of runtime. The resulting platform can support many of the tasks performed by significantly more expensive service robots. For compatibility with existing software packages, the service robot runs the Robot Operating System (ROS). } } | |||
Abstract:Most current autonomous mobile service robots are either expensive commercial platforms or custom manufactured for research environments, limiting their availability. We present the design for a low-cost service robot based on the widely used TurtleBot 2 platform, with the goal of making service robots affordable and accessible to the research, educational, and hobbyist communities. Our design uses a set of simple and inexpensive modifications to transform the TurtleBot 2 into a 4.5ft (1.37m) tall tour-guide or telepresence-style robot, capable of performing a wide variety of indoor service tasks. The resulting platform provides a shoulder-height touchscreen and 3D camera for interaction, an optional low-cost arm for manipulation, enhanced onboard computation, autonomous charging, and up to 6 hours of runtime. The resulting platform can support many of the tasks performed by significantly more expensive service robots. For compatibility with existing software packages, the service robot runs the Robot Operating System (ROS). | |||
Unspecified
This material is presented to ensure timely dissemination of scholarly work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.
Generated by bib2html.pl (written by Patrick Riley ) on Tue Mar 16, 2021 18:23:17
Website Last Updated: 03/16/21
Copyright © 2002-2013 Eric Eaton.
Copyright restrictions apply to all files under the
website's directory. All rights reserved.