[1]
2024. Optimizing LLM Deployments through Inference Backends. Journal of Artificial Intelligence & Cloud Computing. 3, 4 (Jul. 2024), 1–4. DOI:https://doi.org/10.47363/JAICC/2024(3)E128.