FinGPT, an open-source project for financial large language models, aims to democratize access to advanced financial modeling tools. It emphasizes cost-effective model adaptation and fine-tuning.
Recent issues and pull requests suggest a focus on improving usability and training methodologies. Key issues include the non-functional Hugging Face demo (#188) and challenges with GPU memory management. These indicate ongoing user engagement and areas needing improvement.
Yuncong Liu (ycsama0703):
FinGPT_ Training with LoRA and Meta-Llama-3-8B.ipynb
.Bruce Yanghy (ByFinTech): No recent activity besides collaboration with Yuncong Liu.
Other team members have been inactive, indicating a concentrated effort by Yuncong Liu on specific tasks related to model training.
int8
to kbit
training reflects adaptation to new insights.Timespan | Opened | Closed | Comments | Labeled | Milestones |
---|---|---|---|---|---|
7 Days | 0 | 0 | 0 | 0 | 0 |
30 Days | 0 | 0 | 0 | 0 | 0 |
90 Days | 4 | 0 | 2 | 3 | 1 |
1 Year | 52 | 17 | 66 | 42 | 1 |
All Time | 105 | 36 | - | - | - |
Like all software activity quantification, these numbers are imperfect but sometimes useful. Comments, Labels, and Milestones refer to those issues opened in the timespan in question.
Developer | Avatar | Branches | PRs | Commits | Files | Changes |
---|---|---|---|---|---|---|
Yuncong Liu | 1 | 1/1/0 | 5 | 3 | 3083 | |
RAVI GAUTAM (RGIIST) | 0 | 0/0/1 | 0 | 0 | 0 | |
ByFinTech | 0 | 0/0/0 | 0 | 0 | 0 |
PRs: created by that dev and opened/merged/closed-unmerged during the period
The FinGPT project currently has 69 open issues, indicating ongoing user engagement and potential areas for improvement. Notably, several issues are related to critical functionalities, such as the Hugging Face demo not working (#188) and various error messages encountered during model training and inference.
A recurring theme in the issues is the struggle with model compatibility and resource management, particularly concerning GPU memory limitations and software dependencies. This suggests that users may be facing challenges in effectively utilizing the models due to environmental constraints or outdated documentation.
Issue #188: The Hugging Face demo does not work.
Issue #187: Inquiry about the training dataset for FinGPT_Sentiment_Analysis_v1.
Issue #186: Error encountered while trying to run the forecaster.
Issue #185: Request for guidance on how to call the model locally.
Issue #179: Zero rows after running prepare_data.ipynb.
Overall, these insights reflect both the strengths and challenges faced by the FinGPT project as it continues to evolve in response to user needs and technological advancements.
The provided data includes a series of pull requests (PRs) from the AI4Finance-Foundation/FinGPT repository, detailing contributions to the FinGPT project, which focuses on developing financial large language models. The PRs cover various aspects such as code updates, documentation improvements, and feature enhancements.
PR #192: A minor update to the rag.py
file, correcting a typo in comments and ensuring consistent formatting. This PR is straightforward and improves code readability without affecting functionality.
PR #184: Fixes an installation issue with the bitsandbytes
package in a Jupyter notebook, ensuring compatibility with Google Colab. This PR addresses a practical issue faced by users and is crucial for maintaining usability across different environments.
PR #173: Updates the README file to correct a Python version badge and fix grammatical errors. This PR enhances documentation accuracy and clarity.
PR #172: Modifies train_lora.py
to replace int8
training with kbit
training in the PEFT import. This change reflects an update in training methodology and could impact model performance or training efficiency.
PR #167: Allows benchmarking on CPU by updating benchmark scripts to dynamically move data to the device where the model is located. This PR broadens accessibility for testing on machines without GPUs.
The analysis of these pull requests reveals several key themes:
Continuous Improvement: The open PRs demonstrate ongoing efforts to improve code quality, usability, and documentation. For instance, PR #192 focuses on code readability, while PR #184 addresses installation issues that could hinder user experience.
Adaptation to New Methodologies: Changes like those in PR #172 indicate an adaptation to evolving methodologies in model training. The shift from int8
to kbit
training suggests a response to new research or practical insights that could enhance model performance or training efficiency.
Enhanced Accessibility: The ability to benchmark on CPU (PR #167) reflects a commitment to making the tools accessible to a wider audience, including those without access to high-end hardware.
User-Centric Enhancements: The closed PR (#194) that introduces new features for training tutorials highlights a focus on enhancing user experience through better guidance and tools.
Documentation and Clarity: Regular updates to documentation (e.g., PR #173) are crucial for maintaining clarity as the project evolves. Accurate documentation helps users understand new features or changes in methodology.
Overall, these pull requests illustrate a proactive approach to development within the FinGPT project, emphasizing quality improvement, methodological adaptation, user accessibility, and clear documentation.
Bruce Yanghy (ByFinTech)
Yuncong Liu (ycsama0703)
FinGPT_ Training with LoRA and Meta-Llama-3-8B.ipynb
.FinGPT_ Training with LoRA and Llama3-7B.ipynb
and Copy_of_FinGPT_Training_with_LoRA_and_ChatGLM2–6B.ipynb
).Likun Lin (llk010502)
Daniel (Neng) Wang (Noir97)
Boyu Zhang (boyuZh)
Eleni Verteouri
Gason Bai
William Gazeley
Kalyani Mhala
Alfonso Amayuelas
Surav Shrestha
Peter Schofield
Tianyu Zhou (raphaelzhou1)
Shivam Singh
Kris248
mac
RGIIST
The development team shows a concentrated effort from Yuncong Liu, while other members have been inactive recently. This may indicate a focused sprint on specific features related to model training, particularly with LoRA and Meta-Llama models, while other areas remain static for now.