ML Track
The Machine Learning Track is to award the team that conducts and reports the most novel machine learning study using our simulation environment. To win this track, one should:
- rank in the top 50 of Round 1
- submit a paper describing your machine learning approach in NeurIPS format
Our review board led by Xue Bin (Jason) Peng, xbpeng@berkeley.edu) will select the paper that presents the most novel machine learning study using our simulation environment. The topics of interest include but are not limited to:
- reinforcement learning (e.g. novel RL algorithms or improvements to existing techniques [Kakade, 2001; Silver et al., 2014; Haarnoja et al., 2018])
- imitation learning (e.g. methods for learning from demonstrations [Peng et al., 2018; Ho and Ermon, 2016])
- reward functions (e.g. designing reward function that encourage natural behaviours [Wang et al., 2009; Yu et al., 2018])
- control structures (e.g. task-specific control structures for motion control [Wang et al., 2018; Lee et al., 2019])
The prizes for the winner are:
- 2x NVIDIA GPU
- Travel grant (up to $1,000 to NeurIPS 2019 or Stanford)
- Paper acceptance into the NeurIPS Deep RL workshop
References
- Haarnoja, T., Zhou, A., Abbeel, P., & Levine, S. (2018). Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor. Proceedings of the 35th International Conference on Machine Learning (ICML 2018).
- Ho, J., & Ermon, S. (2016). Generative Adversarial Imitation Learning. Advances in Neural Information Processing Systems (NIPS 2016).
- Kakade, S. (2001). A Natural Policy Gradient. Proceedings of the 14th International Conference on Neural Information Processing Systems (NIPS 2001).
- Lee, S., Lee, K., Park, M., and Lee, J. (2019). Scalable Muscle-actuated Human Simulation and Control. ACM Transactions on Graphics (Proc. SIGGRAPH 2019).
- Peng, X. B., Abbeel, P., Levine, S., & van de Panne, M. (2018). Deepmimic: Example-guided deep reinforcement learning of physics-based character skills. ACM Transactions on Graphics (Proc. SIGGRAPH 2018).
- Silver, D., Lever, G., Heess, N., Degris, T., Weirstra, D., & Riedmiller, M. (2014). Deterministic Policy Gradient Algorithms. Proceedings of the 31st International Conference on Machine Learning (ICML 2014).
- Yu, W., Turk, G., & Liu, C.K. (2018). Learning Symmetric and Low-Energy Locomotion. ACM Transactions on Graphics (Proc. SIGGRAPH 2018).
- Wang, J., Fleet, D., & Hertzmann, A. (2009). Optimizing Walking Controllers. ACM Transactions on Graphics (Proc. SIGGRAPH Asia 2009).
- Wang, T., Liao, R., Ba, J., & Fidler, S. (2018). NerveNet: Learning Structured Policy with Graph Neural Networks. International Conference on Learning Representations (ICLR 2019).