Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
sdmia_schedule [2015/10/18 09:39]
matthijs
sdmia_schedule [2015/11/13 12:08] (current)
matthijs
Line 21: Line 21:
 |03:​30pm| 04:​00pm | ​ //Coffee break// || |03:​30pm| 04:​00pm | ​ //Coffee break// ||
 |04:​00pm| 04:​45pm |Mykel Kochenderfer,​ Stanford |[[sdmia_invited_speakers|Decision Theoretic Planning for Air Traffic Applications]]| |04:​00pm| 04:​45pm |Mykel Kochenderfer,​ Stanford |[[sdmia_invited_speakers|Decision Theoretic Planning for Air Traffic Applications]]|
-|04:​45pm| 05:​10pm |MHausknecht ​and PStone |Deep Recurrent Q-Learning ​for Partially ​Observable MDPs|+|04:​45pm| 05:​10pm |BLacerda, D. Parker ​and NHawes |Nested Value Iteration ​for Partially ​Satisfiable Co-Safe LTL Specifications (Extended Abstract) ​|
 |05:​10pm| 05:​35pm |K. Wray and S. Zilberstein |A Parallel Point-Based POMDP Algorithm Leveraging GPUs | |05:​10pm| 05:​35pm |K. Wray and S. Zilberstein |A Parallel Point-Based POMDP Algorithm Leveraging GPUs |
 |06:​00pm| 07:​00pm | ​ //​Reception//​ || |06:​00pm| 07:​00pm | ​ //​Reception//​ ||
Line 34: Line 34:
 |11:​00am| 11:​25am |E. Durfee and S. Singh |Commitment Semantics for Sequential Decision Making Under Reward Uncertainty | |11:​00am| 11:​25am |E. Durfee and S. Singh |Commitment Semantics for Sequential Decision Making Under Reward Uncertainty |
 |11:​25am| 11:​50am |A. Iwasaki, T. Sekiguchi, S. Yamamoto and M. Yokoo |How is cooperation/​collusion sustained in repeated multimarket contact with observation errors? | |11:​25am| 11:​50am |A. Iwasaki, T. Sekiguchi, S. Yamamoto and M. Yokoo |How is cooperation/​collusion sustained in repeated multimarket contact with observation errors? |
-|11:​50am| 12:​10pm |S. Mcgregor, H. Buckingham, R. Houtman, C. Montgomery, R. Metoyer and T. Dietterich |Facilitating ​Testing ​and Debugging of Markov Decision Processes ​with Interactive Visualization ​|+|11:​50am| 12:​10pm |S. Mcgregor, H. Buckingham, R. Houtman, C. Montgomery, R. Metoyer and T. Dietterich |MDPvis: An Interactive Visualization for Testing Markov Decision Processes |
 |12:​10pm| 12:​30pm |F. Oliehoek, M. T. J. Spaan, P. Robbel and J. Messias |The MADP Toolbox: An Open-Source Library for Planning and Learning in (Multi-)Agent Systems | |12:​10pm| 12:​30pm |F. Oliehoek, M. T. J. Spaan, P. Robbel and J. Messias |The MADP Toolbox: An Open-Source Library for Planning and Learning in (Multi-)Agent Systems |
 |12:​30pm| 02:​00pm | ​ //​Lunch// ​ ||  |12:​30pm| 02:​00pm | ​ //​Lunch// ​ ||
Line 43: Line 43:
 |04:​00pm| 04:​45pm |Emma Brunskill, CMU |[[sdmia_invited_speakers|Quickly Learning to Make Good Decisions]] | |04:​00pm| 04:​45pm |Emma Brunskill, CMU |[[sdmia_invited_speakers|Quickly Learning to Make Good Decisions]] |
 |04:​45pm| 05:​10pm |M. Allen |Complexity of Self-Preserving,​ Team-Based Competition in Partially Observable Stochastic Games | |04:​45pm| 05:​10pm |M. Allen |Complexity of Self-Preserving,​ Team-Based Competition in Partially Observable Stochastic Games |
-|05:​10pm| 05:​35pm |BLacerda, D. Parker ​and NHawes |Nested Value Iteration ​for Partially ​Satisfiable Co-Safe LTL Specifications (Extended Abstract) ​|+|05:​10pm| 05:​35pm |MHausknecht ​and PStone |Deep Recurrent Q-Learning ​for Partially ​Observable MDPs |
 |06:​00pm| 07:​30pm | ​ //Plenary session// ​ ||  |06:​00pm| 07:​30pm | ​ //Plenary session// ​ ||
  
Line 49: Line 49:
  
 ^ ^ ^ Authors ^ Title ^ ^ ^ ^ Authors ^ Title ^
-|09:​00am| 09:​45am |Alan Fern, Oregon State |[[sdmia_invited_speakers|TBA]] |+|09:​00am| 09:​45am |Alan Fern, Oregon State |[[sdmia_invited_speakers|Kinder and Gentler Teaching Modes for Human-Assisted Policy Learning]] |
 |09:​45am| 10:​00am | ​ application discussion || |09:​45am| 10:​00am | ​ application discussion ||
 |10:​00am| 10:​25am |A. Reyes, P. H. Ibarguengoytia,​ I. Romero, D. Pech and M. Borunda |Open questions for building optimal operation policies for dam management using Factored Markov Decision Processes ​ | |10:​00am| 10:​25am |A. Reyes, P. H. Ibarguengoytia,​ I. Romero, D. Pech and M. Borunda |Open questions for building optimal operation policies for dam management using Factored Markov Decision Processes ​ |
Recent changes RSS feed Creative Commons License Donate Minima Template by Wikidesign Driven by DokuWiki