Setting Up Local Language Models for Your App
Setting up local language models for your app can significantly enhance its capabilities, enabling it to understand and generate text in multiple languages without relying on external APIs. By integrating local language models, you can improve privacy, reduce latency, and ensure continuous functionality even in offline environments. Here's a comprehensive guide on how to set up local language models for your application:
Steps:
For cloud version LLM change:
Step 1
Visit the chat screen and you will be to see the default LLM selected.
Step 2
Click on it and you will get a drop down of various LLM's available to choose.
Step 3
Choose the LLM of your choice.
Video Demo

For Open source llm change:
Step 1
For open source version please edit LLM_NAME
, MODEL_NAME
and others in the .env file. Refer to ⚙️ App Configuration for more information.
Step 2
Visit ☁️ Cloud Providers for the updated list of online models. Make sure you have the right API_KEY and correct LLM_NAME. For self-hosted please visit 🖥️ Local Inference.