Skip to content
/ fingeit Public

FinGEITje 7B is a large open Dutch financial language model with 7 billion parameters, optimized for understanding and generating Dutch financial text. Built on Mistral 7B, it provides advanced proficiency in Dutch and offers comprehensive insights into financial topics.

License

Notifications You must be signed in to change notification settings

snoels/fingeit

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

🐐 FinGEITje 7B: a large open Dutch Financial language model

📄 README

GEITje is a large open Dutch financial language model with 7 billion parameters, based on Mistral 7B. It has been (further) trained on Dutch financial texts. As a result, it has learned better Dutch and has more knowledge about Dutch financial topics.

DALL·E 3: "Create a logo for a Dutch large language model's Github readme. Incorporate a hyper realistic cute baby goat painting on a Dutch landscape with a few finance skyscrapers. The cute baby goat wears a business suit and has a financial background."

Getting Started

  1. Install Dependencies: Run poetry install. FinGEITje uses Poetry as a dependency manager. Running this command will create a virtual environment and install the necessary Python packages.

  2. Download the Original Dataset: Run data_downloader to download the original dataset.

  3. Translate the Dataset: Run translation_service to translate the original dataset.

  4. Post-Process the Translated Dataset: Run post_process to post-process the translated dataset.

  5. Format the Translation: Run translation_formatter to format the translation into the original dataset format.

  6. Training Configuration: The training configuration can be found in config_qlora. This is a recipe as described in the Alignment Handbook, and we used the Alignment Handbook to spawn the whole training pipeline.

  7. Evaluation Package: The evaluation package can be found in evaluation. To evaluate the model, a set of metrics are defined per task. The tasks can be grouped per dataset so every dataset is evaluated separately.

Notebooks

Paper

This repository is based on the following paper:

A Dutch Financial Large Language Model
Link to the paper

Citation

If you use FinGEITje in your work, please cite:

@inproceedings{10.1145/3677052.3698628,
author = {Noels, Sander and De Blaere, Jorne and De Bie, Tijl},
title = {A Dutch Financial Large Language Model},
year = {2024},
isbn = {9798400710810},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3677052.3698628},
doi = {10.1145/3677052.3698628},
abstract = {This paper presents FinGEITje, the first Dutch financial Large Language Model (LLM) specifically designed and optimized for various financial tasks. Together with the model, we release a specialized Dutch financial instruction tuning dataset with over 140,000 samples, constructed employing an automated translation and data processing method. The open-source data construction method is provided, facilitating the creation of financial instruction datasets in different languages. To evaluate model performance, the study introduces the first Dutch financial evaluation benchmark, along with an automated evaluation method that utilizes an LLM as an independent evaluator, reducing manual intervention in performance evaluation. The experimental results highlight the superior performance of FinGEITje across five critical Dutch and English financial tasks.},
booktitle = {Proceedings of the 5th ACM International Conference on AI in Finance},
pages = {283–291},
numpages = {9},
keywords = {Financial Large Language Model, Instruction Tuning., Natural Language Processing},
location = {Brooklyn, NY, USA},
series = {ICAIF '24}
}

Contributing

Contributions are welcome! If you have ideas, suggestions, or find issues, please open an issue or submit a pull request. We appreciate your help in improving FinGEITje.

Acknowledgements

We would like to thank:

  • Rijgersberg (GitHub) for creating one of the first Dutch foundation models called GEITje: a Dutch large open language model with 7 billion parameters, based on Mistral 7B. It has been further trained on 10 billion tokens of Dutch text. The model can be found here: Rijgersberg/GEITje-7B.

  • Bram Vanroy (GitHub) for creating one of the first Dutch open source chat models, GEITje-7B-ultra, and for open-sourcing its training, translation (dutch-instruction-datasets), and evaluation details.

  • Silverfin for their collaboration in this research. Silverfin, a Belgian scale-up focused on building an accountancy cloud service, provided valuable insights and resources that were instrumental in the development of FinGEITje. More about their work can be found at Silverfin.

We also extend our gratitude to the contributors of the Alignment Handbook for providing valuable resources that aided in the development of FinGEITje.

License

This project is licensed under the Apache License 2.0.

Contact

For any inquiries or questions, please contact Sander Noels.

About

FinGEITje 7B is a large open Dutch financial language model with 7 billion parameters, optimized for understanding and generating Dutch financial text. Built on Mistral 7B, it provides advanced proficiency in Dutch and offers comprehensive insights into financial topics.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published