Authors :
S R Abhiram; Suhas L; Tejas S; Tejaswini K. P.
Volume/Issue :
Volume 9 - 2024, Issue 11 - November
Google Scholar :
https://tinyurl.com/4ftamdvu
Scribd :
https://tinyurl.com/5dskea44
DOI :
https://doi.org/10.5281/zenodo.14524942
Abstract :
This survey examines advancements in
augmenting language models (LMs) with enhanced
reasoning abilities and tool-usage capabilities. Reasoning in
this context involves breaking down complex tasks into
simpler subtasks, while tool use refers to engaging with
external modules, such as a code interpreter. LMs can
apply these capabilities independently or together through
heuristics or through learning from example
demonstrations. By utilizing various, often non-parametric
external modules, these enhanced LMs expand their ability
to process context, shifting beyond traditional language
modeling. This type of model is referred to as an
Augmented Language Model (ALM). The standard
missing token objective enables ALMs to develop reasoning
skills, utilize tools, and even perform actions, while still
handling typical language tasks—and in some cases,
outperforming standard LMs in benchmark tests. This
survey concludes that ALMs could potentially overcome
significant limitations found in traditional LMs, including
issues with interpretability, consistency, and scalability.
Keywords :
Reasoning, Tool Use, Non-Parametric Module, Missing Token Prediction), Heuristics, Demonstrations, Interpretability, Consistency, Scalability.
References :
- Deborah Cohen, Moonkyung Ryu, Yinlam Chow, Orgad Keller, Ido Greenberg, Avinatan Hassidim, Michael Fink, Yossi Matias, Idan Szpektor, Craig Boutilier, et al. Dynamic planning in open-ended dialogue using reinforcement learning. arXiv preprint
- Antonia Creswell and Murray Shanahan. Faithful reasoning using large language models.
- Satanjeev Banerjee and Alon Lavie. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pages 65–72.
- Antonia Creswell, Murray Shanahan, and Irina Higgins. Selection-inference: Exploiting large language models for interpretable logical reasoning
- Ishita Dasgupta, Andrew K Lampinen, Stephanie CY Chan, Antonia Creswell, Dharshan Kumaran, James L McClelland, and Felix Hill. Language models show human-like content effects on reasoning.
- Ishita Dasgupta, Christine Kaeser-Chen, Kenneth Marino, Arun Ahuja, Sheila Babayan, Felix Hill, and Rob Fergus. Collaborating with language models for embodied reasoning, 2023.
- Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the North American Chapter of the Association for Computational Linguistics (NAACL), 2019.
- Pierre L Dognin, Inkit Padhi, Igor Melnyk, and Payel Das. Regen: Reinforcement learning for text and knowledge base generation using pretrained language models. Conference on Empirical Methods in Natural Language Processing (EMNLP), 2021.
- Chris Donahue, Mina Lee, and Percy Liang. Enabling language models to fill in the blanks. In Proceedings of the Annual Meeting of the Association for Computational Linguistics(ACL), 2020.
- Iddo Drori, Sarah Zhang, Reece Shuttleworth, Leonard Tang, Albert Lu, Elizabeth Ke, Kevin Liu, Linda Chen, Sunny Tran, Newman Cheng, et al. A neural network solves, explains, and generates university math problems by program synthesis and few-shot learning at human level. Proceedings of the National Academy of Sciences, 119(32), 2022.
- Andrew Drozdov, Nathanael Schärli, Ekin Akyürek, Nathan Scales, Xinying Song, Xinyun Chen, Olivier Bousquet, and Denny Zhou. Compositional semantic parsing with large language models.
- Dheeru Dua, Shivanshu Gupta, Sameer Singh, and Matt Gardner. Successive prompting for decomposing complex questions. Conference on Empirical Methods in Natural Language Processing (EMNLP), 2022.
- Luyu Gao, Aman Madaan, Shuyan Zhou, Uri Alon, Pengfei Liu, Yiming Yang, Jamie Callan, and Graham Neubig. Pal: Program-aided language models, 2022.
- lge Akkaya, Marcin Andrychowicz, Maciek Chociej, Mateusz Litwin, Bob McGrew, Arthur Petron, Alex Paino, Matthias Plappert, Glenn Powell, Raphael Ribas, et al. Solving rubik’s cube with a robot hand.
- Jean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, Arthur Mensch, Katie Millican, Malcolm Reynolds, et al. Flamingo: a visual language model for few-shot learning. Advances in Neural Information Processing Systems (NeurIPS), 2022.
- Daniel Andor, Luheng He, Kenton Lee, and Emily Pitler. Giving BERT a calculator: Finding operations and arguments with reading comprehension. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019.
- Akari Asai, Xinyan Yu, Jungo Kasai, and Hannaneh Hajishirzi. One question answering model for many languages with cross-lingual dense passage retrieval. Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Akari Asai, Timo Schick, Patrick Lewis, Xilun Chen, Gautier Izacard, Sebastian Riedel, Hannaneh Hajishirzi, and Wen-tau Yih. Task-aware retrieval with instructions
- Lalit R. Bahl, Frederick Jelinek, and Robert L. Mercer. A maximum likelihood approach to continuous speech recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, PAMI-5(2):179–190, 1983.
- Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, et al. Constitutional ai: Harmlessness from ai feedback. arXiv preprint arXiv:2212.08073, 2022.
This survey examines advancements in
augmenting language models (LMs) with enhanced
reasoning abilities and tool-usage capabilities. Reasoning in
this context involves breaking down complex tasks into
simpler subtasks, while tool use refers to engaging with
external modules, such as a code interpreter. LMs can
apply these capabilities independently or together through
heuristics or through learning from example
demonstrations. By utilizing various, often non-parametric
external modules, these enhanced LMs expand their ability
to process context, shifting beyond traditional language
modeling. This type of model is referred to as an
Augmented Language Model (ALM). The standard
missing token objective enables ALMs to develop reasoning
skills, utilize tools, and even perform actions, while still
handling typical language tasks—and in some cases,
outperforming standard LMs in benchmark tests. This
survey concludes that ALMs could potentially overcome
significant limitations found in traditional LMs, including
issues with interpretability, consistency, and scalability.
Keywords :
Reasoning, Tool Use, Non-Parametric Module, Missing Token Prediction), Heuristics, Demonstrations, Interpretability, Consistency, Scalability.