Training Dynamics of a 1.7B LLaMa Model: A Data-Efficient Approach
Published in IJCNN 2025, 2025
This paper is about our study on training a 1.7B LLaMa model
Recommended citation: Miles Q. Li, Benjamin CM Fung, and Shih-Chia Huang. Training Dynamics of a 1.7B LLaMa Model: A Data-Efficient Approach. In Proceedings of the International Conference on Joint Conference on Neural Networks (IJCNN), pages 1-10, Roma, Italy: IEEE, June 2025. https://arxiv.org/abs/2412.13335