-
Deploy LLMs with opeanai compatible API which can be used as drop in replacement for actual openai api for which access is paid and you have little control over it
-
Learn to develop Langchain based application which will use these APIs
-
Compare different LLMs for the same langchain agents, the quality of response and other KPIs (TBD)
-
Understand the cost of LLM+application
- FastChat: provides easy to deploy interface for most populat LLMs (Vicuna, maybe LLaMA), supports PERF
- Find other easy methods ... TBD
- Vicuna (because it is safest with FastChat, but not very good quality with 7b model, especially the Oneshot_REACT agents fail to follow REACT framework)
- LLaMA-2
- Dolly-2
- Falcon