File Download

There are no files associated with this item.

Supplementary

Conference Paper: Text2Reward: Automated Dense Reward Function Generation for Reinforcement Learning

TitleText2Reward: Automated Dense Reward Function Generation for Reinforcement Learning
Authors
Issue Date10-May-2024
Abstract

We introduce Lemur and Lemur-Chat, openly accessible language models optimizedfor both natural language and coding capabilities to serve as the backboneof versatile language agents. The evolution from language chat models tofunctional language agents demands that models not only master human interaction,reasoning, and planning but also ensure grounding in the relevant environments.This calls for a harmonious blend of language and coding capabilitiesin the models. Lemur and Lemur-Chat are proposed to address this necessity,demonstrating balanced proficiencies in both domains, unlike existingopen-source models that tend to specialize in either. Through meticulous pretrainingusing a code-intensive corpus and instruction fine-tuning on text and codedata, our models achieve state-of-the-art averaged performance across diversetext and coding benchmarks. Comprehensive experiments demonstrate Lemur’ssuperiority over existing open-source models and its proficiency across variousagent tasks involving human communication, tool usage, and interaction underfully- and partially- observable environments. The harmonization between naturaland programming languages enables Lemur-Chat to significantly narrow thegap with proprietary models on agent abilities, providing key insights into developingadvanced open-source agents adept at reasoning, planning, and operatingseamlessly across environments.


Persistent Identifierhttp://hdl.handle.net/10722/354461

 

DC FieldValueLanguage
dc.contributor.authorXie, Tianbao-
dc.contributor.authorZhao, Siheng-
dc.contributor.authorWu, Chen Henry-
dc.contributor.authorLiu, Yitao-
dc.contributor.authorLuo, Qian-
dc.contributor.authorZhong, Victor-
dc.contributor.authorYang, Yanchao-
dc.contributor.authorYu, Tao-
dc.date.accessioned2025-02-08T00:51:33Z-
dc.date.available2025-02-08T00:51:33Z-
dc.date.issued2024-05-10-
dc.identifier.urihttp://hdl.handle.net/10722/354461-
dc.description.abstract<p>We introduce Lemur and Lemur-Chat, openly accessible language models optimizedfor both natural language and coding capabilities to serve as the backboneof versatile language agents. The evolution from language chat models tofunctional language agents demands that models not only master human interaction,reasoning, and planning but also ensure grounding in the relevant environments.This calls for a harmonious blend of language and coding capabilitiesin the models. Lemur and Lemur-Chat are proposed to address this necessity,demonstrating balanced proficiencies in both domains, unlike existingopen-source models that tend to specialize in either. Through meticulous pretrainingusing a code-intensive corpus and instruction fine-tuning on text and codedata, our models achieve state-of-the-art averaged performance across diversetext and coding benchmarks. Comprehensive experiments demonstrate Lemur’ssuperiority over existing open-source models and its proficiency across variousagent tasks involving human communication, tool usage, and interaction underfully- and partially- observable environments. The harmonization between naturaland programming languages enables Lemur-Chat to significantly narrow thegap with proprietary models on agent abilities, providing key insights into developingadvanced open-source agents adept at reasoning, planning, and operatingseamlessly across environments.<br></p>-
dc.languageeng-
dc.relation.ispartofInternational Conference on Learning Representations (ICLR), 2024 (07/05/2024-11/05/2024, Vienna, Austria)-
dc.titleText2Reward: Automated Dense Reward Function Generation for Reinforcement Learning-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats