LLM and Generative AI

Prompts and Requests Library

If you have integrated a Custom LLM, continue to add prompts. If not, skip this article and see Co-Pilot and Dynamic Conversations features. Effective prompts play a crucial role in enhancing response accuracy when interacting with LLMs. The new Prompts Library module empowers bot designers by allowing them to create…

LLM Integration

To use LLM and Generative AI features, you must configure the integration with a pre-built or custom LLM. Pre-built LLM Integration The XO Platform offers seamless integration with leading AI services like Azure OpenAI, OpenAI, and Anthropic. Utilizing pre-configured prompts and APIs, you can effortlessly tap into the core capabilities…

Co-Pilot Features

The Co-Pilot features add design-time capabilities to accelerate your bot development process using tailored LLM features. By default, all the features are disabled. To enable the feature, select the model, prompt (for custom model only), and then toggle the status to enable it. You can select another supported model for…

Dynamic Conversations Features

The Dynamic Conversations features boost your virtual assistant’s performance with LLM-powered runtime features designed to streamline development and reduce time and effort. By default, all the features are disabled. To enable the feature, select the model and toggle the status to enable it. You can select another supported model for…

Introduction to LLM and Generative AI

The Kore.ai XO Platform helps enhance your bot development process and enrich end-user conversational experiences by integrating pre-built (OpenAI, Azure OpenAI, Anthropic) or custom models in the backend. In addition to the out-of-box integration with pre-built models, the Platform supports the bring-your-own (BYO) model framework to integrate with externally hosted…
Menu