Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
sdmia_schedule [2015/10/18 09:39] matthijs |
sdmia_schedule [2015/11/13 12:08] (current) matthijs |
||
---|---|---|---|
Line 21: | Line 21: | ||
|03:30pm| 04:00pm | //Coffee break// || | |03:30pm| 04:00pm | //Coffee break// || | ||
|04:00pm| 04:45pm |Mykel Kochenderfer, Stanford |[[sdmia_invited_speakers|Decision Theoretic Planning for Air Traffic Applications]]| | |04:00pm| 04:45pm |Mykel Kochenderfer, Stanford |[[sdmia_invited_speakers|Decision Theoretic Planning for Air Traffic Applications]]| | ||
- | |04:45pm| 05:10pm |M. Hausknecht and P. Stone |Deep Recurrent Q-Learning for Partially Observable MDPs| | + | |04:45pm| 05:10pm |B. Lacerda, D. Parker and N. Hawes |Nested Value Iteration for Partially Satisfiable Co-Safe LTL Specifications (Extended Abstract) | |
|05:10pm| 05:35pm |K. Wray and S. Zilberstein |A Parallel Point-Based POMDP Algorithm Leveraging GPUs | | |05:10pm| 05:35pm |K. Wray and S. Zilberstein |A Parallel Point-Based POMDP Algorithm Leveraging GPUs | | ||
|06:00pm| 07:00pm | //Reception// || | |06:00pm| 07:00pm | //Reception// || | ||
Line 34: | Line 34: | ||
|11:00am| 11:25am |E. Durfee and S. Singh |Commitment Semantics for Sequential Decision Making Under Reward Uncertainty | | |11:00am| 11:25am |E. Durfee and S. Singh |Commitment Semantics for Sequential Decision Making Under Reward Uncertainty | | ||
|11:25am| 11:50am |A. Iwasaki, T. Sekiguchi, S. Yamamoto and M. Yokoo |How is cooperation/collusion sustained in repeated multimarket contact with observation errors? | | |11:25am| 11:50am |A. Iwasaki, T. Sekiguchi, S. Yamamoto and M. Yokoo |How is cooperation/collusion sustained in repeated multimarket contact with observation errors? | | ||
- | |11:50am| 12:10pm |S. Mcgregor, H. Buckingham, R. Houtman, C. Montgomery, R. Metoyer and T. Dietterich |Facilitating Testing and Debugging of Markov Decision Processes with Interactive Visualization | | + | |11:50am| 12:10pm |S. Mcgregor, H. Buckingham, R. Houtman, C. Montgomery, R. Metoyer and T. Dietterich |MDPvis: An Interactive Visualization for Testing Markov Decision Processes | |
|12:10pm| 12:30pm |F. Oliehoek, M. T. J. Spaan, P. Robbel and J. Messias |The MADP Toolbox: An Open-Source Library for Planning and Learning in (Multi-)Agent Systems | | |12:10pm| 12:30pm |F. Oliehoek, M. T. J. Spaan, P. Robbel and J. Messias |The MADP Toolbox: An Open-Source Library for Planning and Learning in (Multi-)Agent Systems | | ||
|12:30pm| 02:00pm | //Lunch// || | |12:30pm| 02:00pm | //Lunch// || | ||
Line 43: | Line 43: | ||
|04:00pm| 04:45pm |Emma Brunskill, CMU |[[sdmia_invited_speakers|Quickly Learning to Make Good Decisions]] | | |04:00pm| 04:45pm |Emma Brunskill, CMU |[[sdmia_invited_speakers|Quickly Learning to Make Good Decisions]] | | ||
|04:45pm| 05:10pm |M. Allen |Complexity of Self-Preserving, Team-Based Competition in Partially Observable Stochastic Games | | |04:45pm| 05:10pm |M. Allen |Complexity of Self-Preserving, Team-Based Competition in Partially Observable Stochastic Games | | ||
- | |05:10pm| 05:35pm |B. Lacerda, D. Parker and N. Hawes |Nested Value Iteration for Partially Satisfiable Co-Safe LTL Specifications (Extended Abstract) | | + | |05:10pm| 05:35pm |M. Hausknecht and P. Stone |Deep Recurrent Q-Learning for Partially Observable MDPs | |
|06:00pm| 07:30pm | //Plenary session// || | |06:00pm| 07:30pm | //Plenary session// || | ||
Line 49: | Line 49: | ||
^ ^ ^ Authors ^ Title ^ | ^ ^ ^ Authors ^ Title ^ | ||
- | |09:00am| 09:45am |Alan Fern, Oregon State |[[sdmia_invited_speakers|TBA]] | | + | |09:00am| 09:45am |Alan Fern, Oregon State |[[sdmia_invited_speakers|Kinder and Gentler Teaching Modes for Human-Assisted Policy Learning]] | |
|09:45am| 10:00am | application discussion || | |09:45am| 10:00am | application discussion || | ||
|10:00am| 10:25am |A. Reyes, P. H. Ibarguengoytia, I. Romero, D. Pech and M. Borunda |Open questions for building optimal operation policies for dam management using Factored Markov Decision Processes | | |10:00am| 10:25am |A. Reyes, P. H. Ibarguengoytia, I. Romero, D. Pech and M. Borunda |Open questions for building optimal operation policies for dam management using Factored Markov Decision Processes | |