fruit_08.jpg

 

Jaeyoung Lee | Curriculum Vitae | Resume

Research Fellow (2024.01~Present)

at Centre for AI Fundamentals (Computer Science)

University of Manchester, Manchester, UK.

Research Associate (2021.02~2022.03) / Postdoctoral Fellow (2018.01~2021.01)

at Waterloo Intelligent Software Engineering (WISE) Lab. (Electrical and Computer Engineering)

University of Waterloo, Waterloo, ON, Canada.

Postdoctoral Fellow (2015.9~2017.12)

at Reinforcement Learning and Artificial Intelligence (RLAI) Lab. (Computing Science)

University of Alberta, Edmonton, AB, Canada.

 

Education

Ph.D. in Electrical & Electronics Engineering, Yonsei University, Seoul, South Korea, 2015

Area of Specialization: Reinforcement Learning and Optimal Control [ Ph. D. dissertation ]

B.E. in Information and Control Engineering, Kwangwoon University, Seoul, Korea, 2006

Major: Information & Control Engineering.   Minor: Electronic Engineering

 

Area of Specialization: Reinforcement Learning

Current Research Interests: User-Modelling and AI-Assistant

1) Inverse (constrained) reinforcement learning with computational rationality

2) Model-based Bayesian reinforcement learning

3) Experimental-design-based exploration and exploitation trade-off

Other Research Interests on Decision-Making and Optimal Control

1) Safe/deep reinforcement and imitation learning, e.g. for autonomous driving

2) Reinforcement learning and optimal control in continuous domain

3) Control, game and multi-agent theory, e.g. for autonomous driving

 

Selected Publications
(* equally contributed)

Topic 1. Constrained Reinforcement Learning for Safety-critical Systems

Lee, J.*, Sedwards, S.* & Czarnecki, K. (2023).

Uniformly Constrained Reinforcement Learning. [ Springer ]

Journal of Autonomous Agents and Multi-Agent Systems, 38(1), 51 pages.

Lee, J.*, Sedwards, S.* and Czarnecki, K. (2021).

Recursive Constraints to Prevent Instability in Constrained Reinforcement Learning. [ arXiv | video | slides ]

In: Proc. 1st Multi-Objective Decision Making Workshop (MODeM 2021). Online at http://modem2021.cs.nuigalway.ie. (cited: 1)

Topic 2. Distillation and Imitation of Deep Q-Network by Decision Tree, for Formal Verification

Abdelzad, V.*, Lee, J.*, Sedwards, S.*, Soltani S.* and Czarnecki, K. (2021).

Non-divergent Imitation for Verification of Complex Learned Controllers. [ IEEEXplorer | video | slides ]

In: 2021 International Joint Conference on Neural Networks (IJCNN). Shenzhen, China (virtual). (cited: 1)

Jhunjhunwala, A., Lee, J., Sedwards S., Abdelzad V. and Czarnecki K. (2020).

Improved Policy Extraction via Online Q-value Distillation. [ IEEEXplorer ]

In: 2020 International Joint Conference on Neural Networks (IJCNN, in 2020 IEEE WCCI). Glasgow, U.K. (cited: 2)

Topic 3. Deep Reinforcement Learning (for Autonomous Driving)

Lee, J.*, Balakrishnan, A.*, Gaurav, A.*, Czarnecki, K. and Sedwards, S.* (2019).

WiseMove: a Framework to Investigate Safe Deep Reinforcement Learning for Autonomous Driving.
[ Springer | arXiv | Git | slides ]

In: Parker D., Wolf V. (eds) Quantitative Evaluation of Systems. QEST 2019. Lecture Notes in Computer Science, vol. 11785.
(cited: 20)

Balakrishnan, A., Lee, J., Gaurav A., Czarnecki, K. and Sedwards, S. (2021).

Transfer Reinforcement Learning for Autonomous Driving: from WiseMove to WiseSim. [ ACM | Git ]

In: ACM Transactions on Modeling and Computer Simulation (TOMAC), 31(3), article no. 15, 26 pages. (cited: 1)

Lee, S., Lee, J. and Hasuo I. (2020 & 2021).

Predictive PER: Balancing Priority and Diversity towards Stable Deep Reinforcement Learning.
[ IEEEXplorer | deep RL workshop | arXiv | video with slides ]

In: (1) 2021 International Joint Conference on Neural Networks (IJCNN). Shenzhen, China (virtual). (cited: 1)
     (2) Deep Reinforcement Learning Workshop in 2020 NeurIPS. (cited: 3)

Topic 4. Reinforcement Learning and Optimal Control in Continuous Domain

Lee, J. and Sutton, R.S. (2021).

Policy Iterations for Reinforcement Learning Problems in Continuous Time and Space: Fundamental Theory and Methods. [ Elsevier | arXiv | Git ]

Automatica, 126, 109421, 15 pages. (cited: 38)

Lee, J. and Sutton, R.S. (2017).

Policy Iteration for Discounted Reinforcement Learning Problems in Continuous Time and Space.
[ extended abstracts book | preprint | poster | slides ]

In: 2017 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM). Ann Arbor, MI, USA. (cited: 1)

Lee, J.Y., Park, J.B. and Choi, Y.H. (2012).

Integral Q-learning and Explorized Policy Iteration for Adaptive Optimal Control of Continuous-time Linear Systems. [ Elsevier | preprint | Git ]

Automatica, 48(11), pp. 2850~2859. (cited: 177)

Lee, J.Y., Park, J.B. and Choi, Y.H. (2014).

Integral Reinforcement Learning for a Class of Nonlinear Systems with Invariant Explorations.
[ IEEEXplorer | preprint | Git ]

IEEE Transactions on Neural Networks and Learning Systems, 26(5), pp. 916~932. (cited: 109)

Lee, J.Y., Park, J.B. and Choi, Y.H. (2014).

On Integral Generalized Policy Iteration for Continuous-time Linear Quadratic Regulations.
[ IEEEXplorer | preprint ]

Automatica, 50(2), pp. 475~489. (cited: 35)

Topic 5. Multi-agent Consensus Techniques; Applications to Vehicles' Formation Control

Lee, J.Y., Choi, Y.H. and Park, J.B. (2014).

Inverse Optimal Design of the Distributed Consensus Protocol for Formation Control of Multiple Mobile Robots.
[ IEEEXploer | preprint ]

In: Proc. 53rd IEEE Conference on Decision and Control (CDC), pp. 2222~2227. Los Angeles, CA, USA. (cited: 4)

Lee, G.U., Lee, J.Y., Park, J.B. and Choi, Y.H. (2018).

On Stability and Inverse Optimality for a Class of Multi-agent Linear Consensus Protocols. [ Springer ]

International Journal of Control, Automation and Systems (IJCAS), 16(3), pp. 1194~1206. (cited: 6)

 

Other Materials: Studies and Self-writings on Measure Theory and Integrations

Study Materials on Folland, G.B. (2013). Real Analysis: Modern Techniques and their Applications. John Wiley & Sons

and more... Lee, J. (2022). Real Analysis, Probability, and Random Processes with Measure Theory (in progress) [PDF]

 

Jaeyoung Lee @ Christabel Pankhurst Building, Dover St, Manchester, United Kingdom M13 9PS. Mobile: +44 07918 718261.