- Understanding how to host LLMs using edge compute and integrate them into generative AI systems
- Exploring the role of reinforcement learning and fine-tuning to customize models for industry-wide and agentic AI use cases
- Reviewing existing LLM hosting platforms—open-source and commercial—with emphasis on low-latency needs and forward-looking hosting strategies
- Evaluating experience of varying hosting models as well future and proposed solutions
Technical paper presentation.