The Research on Algorithms, Algorithm Optimization, and Applications of Large-Scale AI Models
PDF

Keywords

algorithm optimization; large-scale model; distributed training; efficient fine-tuning of parameters; memory optimization

Abstract

In recent years, artificial intelligence systems, such as Large Language Models (LLMs) and multimodal large models, have developed rapidly. With the exponential increase in model parameters, computational complexity, algorithm designs, and engineering implementations face incomparable challenges. How to train models with parameters in the range of hundred billion to trillion with restricted computing capacity and storage resources, efficient fine-tuning, deployment, further latency reduction, and decreased memory footprint have become the most fundamental problems in modern studies. Benefiting from the systematic analysis on respective reviews, different papers, and studies in these years, based on above analysis, in this article, there would be analysis on three different aspects, separately: Firstly, at algorithm foundations, there’s investigation on evolution in optimizers, network architectures, and training patterns. Secondly, in algorithm levels, there’s summary on core areas like memory, resource, efficient parameter fine-tuning, model compression, and other directions. Thirdly, in levels concerning applications, problems, challenges, there’s observation on applications, their respective feasibilities, limitations in modern models with optimized algorithms. By comparison on different studies on these areas, there’s illustration on trends in algorithm optimizations for large-scale AI models, identification on key problems in modern studies, and illustration on different possibilities in new studies. This article looks for providing a systematic literature analysis for researchers, engineers, participants in studies concerning algorithms in large models, optimizations, in AI studies, for necessary references in their studies.

https://doi.org/10.63808/decs.v1i3.262
PDF

References

[1] Bai, G., Chai, Z., Ling, C., Wang, S., Lu, J., Zhang, N., Shi, T., Yu, Z., Zhu, M., Zhang, Y., Yang, C., Cheng, Y., & Zhao, L. (2023). Beyond efficiency: A systematic survey of resource-efficient large language models. arXiv. https://doi.org/10.48550/arXiv.2401.00625

[2] Bottou, L. (2018). Optimization methods for large-scale machine learning. SIAM Review, 60(2), 223–311.

[3] He, X., Xue, F., Ren, X., & You, Y. (2021). Large-scale deep learning optimizations: A comprehensive survey. arXiv. https://doi.org/10.48550/arXiv.2111.00856

[4] Liu, J., Tang, P., Wang, W., Ren, Y., Hou, X., Guo, M., & Li, C. (2024). A survey on inference optimization techniques for mixture of experts models. arXiv. https://doi.org/10.48550/arXiv.2412.14219

[5] Mei, T., Zi, Y., Cheng, X., Gao, Z., Wang, Q., & Yang, H. (2024). Efficiency optimization of large-scale language models based on deep learning in natural language processing tasks. arXiv. https://doi.org/10.48550/arXiv.2405.11704

[6] Shen, L., Sun, Y., Yu, Z., Ding, L., Tian, X., & Tao, D. (2023). On efficient training of large-scale deep learning models: A literature review. arXiv. https://doi.org/10.48550/arXiv.2304.03589

[7] Tang, Z., Shi, S., Wang, W., Li, B., & Chu, X. (2020). Communication-efficient distributed deep learning: A comprehensive survey. arXiv. https://doi.org/10.48550/arXiv.2003.06307

[8] Tian, X., Li, Y., Zhang, Q., & Zhou, W. (2025). A survey on memory-efficient large-scale model training. arXiv. https://doi.org/10.48550/arXiv.2501.11847

[9] Tu, X., He, Z., Huang, Y., Zhang, Z.-H., Yang, M., & Zhao, J. (2024). An overview of large AI models and their applications. Visual Intelligence, 2, Article 34.

[10] Zeng, F., Li, X., Wang, Y., Chen, H., & Zhao, Q. (2025). Distributed training of large language models: A survey. Information Systems. Advance online publication.

Creative Commons License

This work is licensed under a Creative Commons Attribution 4.0 International License.

Copyright (c) 2025 Xiang Zhang, Hazirah Bee Yusof Ali, Qi Xi