Glenn Matlin
Mika Okamoto
Huzaifa Pardawala
Yang Yang
Sudheer Chava
June 18, 2025
Publication
A comprehensive evaluation framework for studying language models against reasoning-reinforced LMs across 20+ core NLP tasks in finance.
Published
June 18, 2025
Authors
Glenn Matlin, Mika Okamoto, Huzaifa Pardawala, Yang Yang, Sudheer Chava
Venue
Findings of the Association for Computational Linguistics: ACL 2025
This is the first research paper to comprehensively study language models against reasoning-reinforced LMs, with an empirical study of 23 foundation LMs over 20 core NLP tasks in finance. FLaME provides a grounded evaluation framework that reveals significant differences in financial reasoning, with leading models achieving accuracy levels near 80%.
Continue exploring
Return to the publication archive or step back to the broader research agenda.