基本信息
views: 8
![](https://originalfileserver.aminer.cn/sys/aminer/icon/show-trajectory.png)
Bio
Research
My primary research interest is reinforcement learning (RL), particularly enabling RL agents to efficiently adapt to unseen tasks (meta/multi-task RL) by learning "nice" task representations (resp. good coverage policies) in the presence (resp. absence) of training rewards.
As the dynamics of control tasks are commonly governed by physical laws, I embarked upon the quest of developing RL agents that explicitly model dynamics with equations as I believe that it can enable prior knowledge and/or inductive bias injection, sample efficiency gains, better domain randomization (à-la-Sim2Real) and risk-control thanks to interpretability... Practically, this involves using symbolic regression (SR), the search of analytic expressions composed of mathematical operators, e.g. cos, exp, constants and variables. Due to the lack of SR algorithms that infer accurate expressions in reasonable time, I have worked on developing transformer-based models, trained on synthetically-generated datasets, that search with order of magnitudes less time.
My primary research interest is reinforcement learning (RL), particularly enabling RL agents to efficiently adapt to unseen tasks (meta/multi-task RL) by learning "nice" task representations (resp. good coverage policies) in the presence (resp. absence) of training rewards.
As the dynamics of control tasks are commonly governed by physical laws, I embarked upon the quest of developing RL agents that explicitly model dynamics with equations as I believe that it can enable prior knowledge and/or inductive bias injection, sample efficiency gains, better domain randomization (à-la-Sim2Real) and risk-control thanks to interpretability... Practically, this involves using symbolic regression (SR), the search of analytic expressions composed of mathematical operators, e.g. cos, exp, constants and variables. Due to the lack of SR algorithms that infer accurate expressions in reasonable time, I have worked on developing transformer-based models, trained on synthetically-generated datasets, that search with order of magnitudes less time.
Research Interests
Papers共 15 篇Author StatisticsCo-AuthorSimilar Experts
By YearBy Citation主题筛选期刊级别筛选合作者筛选合作机构筛选
时间
引用量
主题
期刊级别
合作者
合作机构
F. O. de Franca,M. Virgolin, M. Kommenda,M. S. Majumder,M. Cranmer,G. Espada,L. Ingelse,A. Fonseca, M. Landajuela,B. Petersen,R. Glatt, N. Mundhenk,
CoRR (2023)
ICMLpp.15655-15668, (2023)
arXiv (Cornell University) (2022)
arXiv (Cornell University) (2022)
CoRR (2021)
Cited3Views0EIBibtex
3
0
user-5d4bc4a8530c70a9b361c870(2021)
Cited0Views0Bibtex
0
0
Load More
Author Statistics
Co-Author
Co-Institution
D-Core
- 合作者
- 学生
- 导师
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn