Alexander Müller (2025) “Optimizing Large Language Model Inference: Strategies for Latency Reduction, Energy Efficiency, and Cybersecurity Applications”, Academic Reseach Library for International Journal of Computer Science & Information System, 10(11), pp. 93–97. Available at: https://colomboscipub.com/index.php/arlijcsis/article/view/58 (Accessed: 17 January 2026).