Chat with the model

To run with a locally run model, make sure that iLab model serve is running in a seperate terminal. If you are unsure on how to do this, please visit the Intro to serve and chat section of this document.

If you go to Playground > Chat by using the side navigation bar, you can interact with the merlinite and granite models.

UI No Model Response

If you are running the ui within a dev environment, the model won't reply because a granite/merinite model endpoint hasn't been given. In this case, we will create a new custom model endpoint, using our locally hosted quantised model.

To add a custom model endpoint, go to Playground > Custom Model Endpoints and press the Add Endpoint button on the right side.

You will have 3 fields to fill out

  • The URL, where your customised model is hosted, if hosting locally, the URL would be http://127.0.0.1:8000/

  • The Model Name, merlinite-7b-lab-Q4_K_M.gguf

  • API Key, you may put any text in here; in this case I've usedrandomCharacters. If you are setting up an API key, please provide the key in this section.

UI Custom Model Endpoint

Go back to the playground chat, select newly added model and chat.

UI Model Response

Next Steps