LitGPT, a framework for large language models, has seen active development with a focus on improving compatibility across platforms and optimizing performance. The project supports pretraining, finetuning, and deployment of LLMs, emphasizing ease of use and scalability.
Recent issues and pull requests indicate ongoing efforts to address compatibility challenges, particularly with CUDA errors and multi-GPU setups. The team is also working on documentation improvements for custom datasets and advanced features like LoRA.
Sebastian Raschka (rasbt)
Motsepe-Jr (challenger)
Jirka Borovec (Borda)
apaz-cli (apaz)
batched_generate_fn()
(14 days ago).Thomas Viehmann (t-vi)
Sander Land (sanderland)
Andrei-Aksionov
The team is actively collaborating, with Sebastian Raschka leading major contributions and others focusing on specific areas like testing and documentation.
Timespan | Opened | Closed | Comments | Labeled | Milestones |
---|---|---|---|---|---|
7 Days | 6 | 2 | 14 | 0 | 1 |
30 Days | 27 | 13 | 72 | 0 | 1 |
90 Days | 78 | 46 | 178 | 2 | 1 |
1 Year | 388 | 199 | 1119 | 147 | 2 |
All Time | 746 | 538 | - | - | - |
Like all software activity quantification, these numbers are imperfect but sometimes useful. Comments, Labels, and Milestones refer to those issues opened in the timespan in question.
Developer | Avatar | Branches | PRs | Commits | Files | Changes |
---|---|---|---|---|---|---|
Sebastian Raschka | 3 | 14/14/0 | 22 | 29 | 2100 | |
apaz | 2 | 3/3/1 | 7 | 9 | 942 | |
Thomas Viehmann | 2 | 3/2/0 | 14 | 7 | 196 | |
Sander Land | 2 | 2/2/0 | 3 | 4 | 114 | |
None (Andrei-Aksionov) | 1 | 1/0/1 | 2 | 4 | 69 | |
Jirka Borovec | 1 | 1/1/0 | 1 | 1 | 10 | |
challenger | 1 | 1/1/0 | 1 | 2 | 6 |
PRs: created by that dev and opened/merged/closed-unmerged during the period
The Lightning-AI/litgpt repository currently has 208 open issues, with recent activity indicating a mix of bug reports, feature requests, and user inquiries. Notably, issues related to CUDA errors and model loading problems have been prevalent, suggesting potential challenges with GPU compatibility and model configuration.
Several issues highlight common themes such as difficulties in multi-GPU training, the need for better documentation on custom datasets, and inconsistencies in model performance across different setups. The presence of multiple unresolved queries about quantization and LoRA further indicates that users are actively seeking clarity on advanced features.
Issue #1733: cuda error when serve with workers_per_device > 1
Issue #1729: use initial_checkpoint_dir for continue-pretraining but can't load model correctly
Issue #1727: Question about tie_embeddings
Issue #1723: Is Support for the DeepSeek v2.5 model on the roadmap?
Issue #1717: Cannot attend to 9904, block size is only 4096
Issue #1715: llm.generate issue on CPU machines
Issue #1714: llm.generate function does not work on Mac (MPS) devices anymore
Issue #1711: Manual convert_to_litgpt for Phi-3.5-mini-instruct downloaded weights from HF
The complexity of integrating various features like quantization and LoRA into existing workflows appears to be a significant pain point for users, highlighting the need for improved guidance and examples in the documentation.
The analysis of the pull requests (PRs) for the Lightning-AI/litgpt project reveals a vibrant and active development environment. The project has seen significant contributions in terms of new features, bug fixes, and enhancements aimed at improving performance, usability, and compatibility with various hardware setups. The PRs indicate a strong focus on expanding model support, optimizing training and inference processes, and enhancing the overall user experience.
PR #1725: bump macos to m1
PR #1538: Do not wrap LoRA layers with FSDP
PR #1421: WIP: TensorParallel with new strategy
PR #1354: Add resume for adapter_v2, enable continued finetuning for adapter
PR #1350: Add LongLora for both full and lora fine-tuning
PR #1331: example for full finetuning with python code done!
PR #1232: Correct an apparent logger output directory bug
PR #1179: Improved Lora finetuning script
PR #1057: [WIP] Simplified preparation of pretraining datasets
PR #1013: Drop interleave placement in QKV matrix
PR #1728: Add Chainlit Studio
PR #1726: Simplify MPS support
PR #1724: Enable MPS support for LitGPT
index_copy_
.Additional PRs focused on version bumps, bug fixes, and minor enhancements that reflect ongoing maintenance and improvement efforts within the project.
The pull requests indicate a strong focus on enhancing compatibility across different hardware platforms (e.g., macOS M1 support), optimizing memory usage (e.g., not wrapping LoRA layers with FSDP), and expanding functionality (e.g., adding LongLora support). The discussions within these PRs highlight collaborative efforts among contributors to address complex challenges such as memory management during model training and inference, ensuring consistent behavior across different environments, and providing clearer guidance to users through improved examples and documentation.
The presence of multiple open PRs related to fine-tuning techniques (e.g., LoRA, LongLora) suggests an active interest in advancing model training methodologies within the community. Additionally, the quick turnaround time for merging PRs that fix bugs or improve usability reflects a commitment to maintaining high-quality standards in the project's development lifecycle.
Overall, the analysis of these pull requests showcases a dynamic development environment where contributors are actively working towards enhancing the capabilities, performance, and user experience of LitGPT while also addressing technical challenges associated with large language models.
Sebastian Raschka (rasbt)
Motsepe-Jr (challenger)
Jirka Borovec (Borda)
apaz-cli (apaz)
batched_generate_fn()
(14 days ago).Thomas Viehmann (t-vi)
Sander Land (sanderland)
Andrei-Aksionov
The development team is actively engaged in enhancing the LitGPT framework through collaborative efforts, focusing on both feature development and stability improvements. The leadership of Sebastian Raschka is evident in the volume of contributions, while other team members support specific areas such as testing and documentation. Overall, the team's recent activities reflect a strong commitment to advancing the project effectively.