IDEAS home Printed from https://ideas.repec.org/p/chf/rpseri/rp23112.html
   My bibliography  Save this paper

When Does Aggregating Multiple Skills with Multi-Task Learning Work? A Case Study in Financial NLP

Author

Listed:
  • Jingwei Ni

    (ETH Zurich)

  • Zhijing Jin

    (ETH Zurich)

  • Qian Wang

    (University of Zurich)

  • Mrinmaya Sachan

    (ETH Zürich)

  • Markus Leippold

    (University of Zurich; Swiss Finance Institute)

Abstract

Multi-task learning (MTL) aims at achieving a better model by leveraging data and knowledge from multiple tasks. However, MTL does not always work – sometimes negative transfer occurs between tasks, especially when aggregating loosely related skills, leaving it an open question when MTL works. Previous studies show that MTL performance can be improved by algorithmic tricks. However, what tasks and skills should be included is less well explored. In this work, we conduct a case study in Financial NLP where multiple datasets exist for skills relevant to the domain, such as numeric reasoning and sentiment analysis. Due to the task difficulty and data scarcity in the Financial NLP domain, we explore when aggregating such diverse skills from multiple datasets with MTL can work. Our findings suggest that the key to MTL success lies in skill diversity, relatedness between tasks, and choice of aggregation size and shared capacity. Specifically, MTL works well when tasks are diverse but related, and when the size of the task aggregation and the shared capacity of the model are balanced to avoid overwhelming certain tasks.

Suggested Citation

  • Jingwei Ni & Zhijing Jin & Qian Wang & Mrinmaya Sachan & Markus Leippold, 2023. "When Does Aggregating Multiple Skills with Multi-Task Learning Work? A Case Study in Financial NLP," Swiss Finance Institute Research Paper Series 23-112, Swiss Finance Institute.
  • Handle: RePEc:chf:rpseri:rp23112
    as

    Download full text from publisher

    File URL: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4456653
    Download Restriction: no
    ---><---

    More about this item

    Keywords

    Multi-Task Learning; Sentiment Analysis; Financial Datasets; FinBERT;
    All these keywords.

    NEP fields

    This paper has been announced in the following NEP Reports:

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:chf:rpseri:rp23112. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Ridima Mittal (email available below). General contact details of provider: https://edirc.repec.org/data/fameech.html .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.