[1]
2025. Optimizing LLM Inference: Metrics that Matter for Real Time Applications. Journal of Artificial Intelligence & Cloud Computing. 4, 1 (Jan. 2025), 1–4. DOI:https://doi.org/10.47363/JAICC/2025(4)446.