Developers need to provide sample utterances for each intent (task) the bot needs to identify to train the machine learning model. The platform ML engine will build a model that will try to map a user utterance to one of the bot intents.
Kore.ai’s Bots Platform allows fully unsupervised machine learning to constantly expand the language capabilities of your chatbot – without human intervention. Unlike unsupervised models in which chatbots learn from any input – good or bad – the Kore.ai Bots Platform enables chatbots to automatically increase their vocabulary only when the chatbot successfully recognizes the intent and extracts the entities of a human’s request to complete a task.
However, we recommend keeping Supervised learning enabled to monitor the bot performance and manually tune where required. Using the bots platform, developers can evaluate all interaction logs, easily change NL settings for failed scenarios, and use the learnings to retrain the bot for better conversations.
Multiple Intent Model
(introduced in v8.1)
Training of “similar intents” with different purposes is usually difficult as the training given for an intent can add noise or conflict with the training given to the other intent. This is more evident in cases where the intents have a contextually different meaning or purpose.
Consider the following case, here when the user is in the Place Order task, any query pertaining to returns policy or delivery options should be answered in the placed order context. But the query from the generic Return a product FAQ would be triggered.
Enabling the Multiple Intent Models from the Advanced NLP Configurations (see here for how) allows you to have a dedicated ML model only for the primary intents and separate ML Models for each of the dialogs with their associated sub-intents so that the intent detection of sub-intents gets preferential treatment.
All the primary intents of the bot will be part of the Bot Level Intent Model. Each of the Dialog tasks will have its own ML Model consisting of all the sub-intents added to it. The Thresholds and Configurations can be individually configured for each of the models.
Adding Machine Learning Utterances
- Open the bot for which you want to add sample user utterances.
- From the left pane, click Natural Language -> Training.
- Select the Machine Learning Utterances tab.
- A list of all Intents would be provided, you can use the filter option to restrict the display items to Dialog, Sub-dialog, or Sub-intents.
- Select the Intent for which you want to add the utterances, the user utterance page would open
- Here enter the utterances.
Note: Utterances added should be unique, but in the case of multiple intent models, the same utterance can be used across different models.
The negation of trained intents will be ignored by the platform.
For example, consider a Banking Bot with trained utterance – Funds Transfer. Then a user utterance “My account is debited even without doing funds transfer” will not trigger the “funds transfer” task.
Named Entity Recognition
Apart from the intent, you can train your Bot to recognize the entities, if present, in the user utterance. For example, if the user says “Book Flight from Hyderabad to Mumbai” apart from recognizing the intent as “Book Flight” the source and destination of the flight should also be recognized. This can be achieved by marking the entities in the user utterance during training.
The platform will also try to identify and mark the entities, you have the option to accept or discard these suggestions. The platform will identify the entities based upon:
- System entities;
- Static List of items either enumerated or lookup;
- NER trained entities (from above).
For each of the entities thus marked, the confidence scores identified by the ML engine are displayed. This was introduced in v8.0 of the platform and is available only when Conditional Random Field is selected as the NER model.
Further, if you have enabled Entity Placeholders the platform will replace the entity values in the training utterance with entity name placeholders for training the ML model. Using actual entity values as well as multiple additions of an utterance with just a change in the entity value will have an adverse impact on the ML training model. The name of entities also starts contributing highly to the intent detection model.
Training your Bot
After you add user utterances, you should train the Kore.ai interpreter to recognize the utterances and the associated user intent. When you have untrained utterances in your bot, the following message is displayed:
“You have untrained utterances in your machine learning model. Unless the engine is trained the bot will not identify intents and entities based on the un-trained utterances. Click on ‘Train’ button to update the bot with all your utterances.”
In the User Says section, click Train. A status bar is displayed to show progress for utterance training. When complete, the Utterances trained successfully message is displayed. The user utterances are added to the Machine Learning Database. You can further configure the ML engine (post rel 8.0), identify the dummy intents when a user utterance contains the words that are not used in the bot’s training i.e. bot vocabulary, refer here for more details.
By default, machine learning is automatically trained for any defined user utterances whenever a task is:
- changed from a status of In-Progress to Configured.
- updated with a new
- task name or intent name,
- entity name or parameter name,
- entity type,
- bot name.
- suspended by the Bots Admin.
- deleted by the Bots Admin.
In Bot Builder when auto-train is in progress, a warning message that untrained user utterances cannot be identified is displayed if you try to test the bot before auto-train is complete.
You can set the Auto Train option as follows:
- Open the bot for which you want to modify the settings.
- Hover over the side navigation panel and then click Natural Language -> Training.
- Select the Advanced Settings tab.
- Enable or Disable the Auto Training option as per your requirements.
Negative patterns can be used to eliminate intents detected by the Fundamental Meaning or Machine Learning models. Refer here to know more.
Threshold & Configurations
To train and improve the performance Threshold and Configurations can be specified for all three NLP engines – FM, KG, and ML. You can access these settings under Natural Language > Training > Thresholds & Configurations.
NOTE: If your Bot is multilingual, you can set the Thresholds differently for different languages. If not set, the Default Settings will be used for all languages.
The settings for the ML engine is discussed in detail in the following sections.
The Bots Platform ver 6.3 upgraded its Machine Learning (ML) model to v3. This includes a host of improvements and also allows developers to fine-tune the model using parameters to suit business requirements. The developers can change parameters like stopword usage, synonym usage, thresholds, and n-grams, as well as opt between Deep Neural Network or Conditional Random Field-based algorithm for the Named-Entity Recognition (NER) model.
In v8.0 of the platform, provision has been enabled to use the v5 of the ML intent model and externalize several hyperparameters. This can be achieved through the Advanced NLP Configuration, refer here for details.
In v8.1 of the platform, the Multiple Intent Models feature has been introduced. When the ‘multiple intents model’ option is enabled, the ML Engine maintains multiple intent models for the bot as follows:
- Bot level Intent Model containing all the Primary Intents of the bot which includes Primary Dialog Intents, and Alert Task Intents.
- Dialog Intent Models – one for every primary dialog intent and sub-dialog intent which includes the Sub-intent nodes added to the dialog definition, Sub-intents scoped as part of the Group nodes and Interruption exceptions added to the dialog definition.
You can configure the Thresholds and Configurations separately for each of the intent models. This includes
- All the configurations under Thresholds and Configurations – ML Engine as discussed in the below section;
- All the ML Engine configurations under the Advanced NLP Configurations discussed in detail here.
Configuring the Machine Learning Parameters
The Bots Platform provides language-wise defaults for the following parameters related to the ML performance of your bot. You can customize them to suit your particular needs.
Points to note in ML configurations:
- The following is the list of all possible configurations and these are available for both single and multiple intent models.
- When the multiple intent model is enabled, you can configure the individual models by selecting the configure link against the model.
- While there is only one Bot level intent model, you can add multiple dialog intent models and configure each as per your requirements.
- Advanced ML Configurations can be applied from here or from the Advanced NLP Configurations section refer here for details.
You can choose the Neural Network that you would like to use to train the intent models. This setting has been moved to Machine Learning from Advanced NLP Configurations in v8.1.
You can choose between the following types. Based on the selection additional configurations can be done from the Advanced NLP Configurations section, refer here for details.
- MLP-BOW – The bag-of-words model is a simplifying representation used in natural language processing and information retrieval. In this model, a text is represented as the bag of its words, disregarding grammar and even word order but keeping multiplicity.
- MLP-WordEmbeddings – Word embedding is the collective name for a set of language modeling and feature learning techniques in natural language processing where words or phrases from the vocabulary are mapped to vectors of real numbers.
- LSTM (Long Short-Term Memory) is an artificial recurrent neural network (RNN) architecture used in the field of deep learning. LSTM has feedback connections and hence has the ability to capture long-term dependencies for texts of any length and is well suited for longer texts.
- CNN (convolutional neural networks) is a class of deep neural networks in deep learning most commonly applied to analyzing visual imagery. It makes use of the word order for a specific region size and has achieved remarkable results on various text classification tasks.
- Transformers use a Universal Sentence encoder in the vectorization stage of the Training pipeline. The output of the sentence encoder is fed to a Multi-Layer perceptron network for training. SentenceEncoder has an inbuilt capability of understanding the semantic similarity between sentences taking into account the synonyms and various usage patterns of the same sentence.
The Universal Sentence Encoder encodes text into high-dimensional vectors that can be used for text classification, semantic similarity, clustering, and other natural language tasks. The model is trained and optimized for greater-than-word length text, such as sentences, phrases, or short paragraphs. It is trained on a variety of data sources and a variety of tasks with the aim of dynamically accommodating a wide variety of natural language understanding tasks. The input is the variable-length English text and the output is a 512-dimensional vector.
Define the minimum ML score to qualify an intent as a probable match. Learn more about ML Scoring.
ML Definitive Score
Configure the threshold score for definite matches, which can be set to a value between 80-100%.
This setting is Disabled by default. Enable this option if you would like to consider intent synonyms in building the ML model.
Choose the NER model to be used for entity detection.
This setting is Disabled by default. Enable this option if you would like to remove the stop words in the training utterances in building the ML model. Not valid when Network Type is set to Transformer.
Using this option (introduced in ver8.0) you can associate the ML intent model with the preferred algorithm. Not valid when Network Type is set to MLP WordEmbeddings, LSTM, CNN, and Transformer.
The options being:
- n-gram – this is the default setting and can be used to define the contiguous sequence of words to be used from training sentences to train the model.
For example, if Generate sales forecast report is the user utterance and if n-gram is set to 2, then Generate sales, Sales forecast, and Forecast report are used in training the model. If n-gram is set to 3, then Generate sales forecast, and Sales forecast report will be used in training the model.
You can set the n-gram using the n-gram Sequence Length – The minimum n-gram limit is 1 by default. You can set the maximum limit up to 4.
- skip-gram – when the corpus is very limited or when the training sentences, in general, contain fewer words then skip-gram would be a better option. For this you need to define:
- Sequence Length – the length for skip-gram sequences, with a minimum of 2 and a maximum of 4
- Maximum Skip Distance – the maximum words to skip to form the grams, with a minimum of 1 and a maximum of 3.
Enable to replace entity values present in the training utterances with the corresponding entity placeholders in the training model. Not valid when Network Type is set to Transformer.
Upgrading the ML Model
All new bots that are created use the new ML model by default. Developers can upgrade the ML model for old bots or downgrade the model for the bots using the new model.
If you are using a previous model of ML in the bots platform, you can upgrade it as follows:
- Open the bot for which you want to upgrade the ML model and go to Natural Language > Advanced Settings.
- Expand Machine Learning. Under ML Upgrade section, click the Upgrade Now button. It opens a confirmation window.
- Click Upgrade and Train. You can see new customizable options under the Machine Learning section.
You can also downgrade the ML model for new or upgraded bots from here by clicking Switch to older version. However, note that the older version of the ML model will be deprecated soon. So, we strongly recommend staying on the latest version to receive continued support and future enhancements.
Note: If a bot is exported using the older model (V2) and imported as a new bot, it continues to be in V2 model until you upgrade it.
Exporting and Importing Machine Learning Utterances
You can import and export ML utterances of a bot into another in CSV and JSON formats. You can choose between ‘In-Development’ or ‘Published’ tasks to export, whereas importing utterances always replace the latest copy of the task in the bot.
How to Export or Import ML Utterances
- On the bot’s menu, click Natural Language > Machine Learning Utterances.
- The ‘In-Development’ version of the bot’s ML utterances open by default. If you want to see the utterances in the ‘Published’ version, toggle on the top right side of the window to Published.
Note: The export of ML utterances vary based on this selection as explained in the Versioning and Behavior of the Exported UUtterances section below.
- Click the options icon and select an option:
- Click Import and upload a CSV or JSON file with the utterances to import. Read Versioning and Behavior of the Imported Utterances section below for more information.-OR-
- Click Export and select CSV or JSON formats to export the utterances. Read Versioning and Behavior or the Exported Utterances section below for more information.
Versioning and Behavior of Imported Utterances
- The imported utterances in CSV/JSON entirely replace the utterances present in the latest copy of the tasks.
- If the task is in the Configured status, the utterances in the task get entirely replaced with the new utterances for the task present in the imported file.
- If the task is in Upgrade in Progress status, the utterances related to the task get entirely replaced with the task utterances present in the imported file. The utterances in the Published copy of the task aren’t affected.
- If the task is in the Published status, an Upgrade in Progress copy of the task gets created by default and the new utterances present in the imported file will be added to the upgraded copy. The utterances in the Published copy of the task aren’t affected.
Versioning and Behavior of Exported Utterances
- When you export a bot’s utterances, all the utterances related to every task type – alert, action, information, dialog – get exported.
- When you export an In Development copy of the bot, the utterances of all tasks in the latest available copy get exported.
- When you export a Published copy of the bot, all the utterances in the published state get exported.
- In the case of multi-language bots, the export of utterances only happens for the selected bot language.
- Export of utterances to JSON include NER tagging present in the tasks, whereas CSV export doesn’t include them.
ML Training Recommendations
- Give a balanced training for all the intents that the bot needs to detect, add approximately the same number of sample utterances. A skewed model may result in skewed results.
- Provide at least 8-10 sample utterances against each intent. The model with just 1-2 utterances will not yield any machine learning benefits. Ensure that the utterances are varied and you do not provide variations that use the same words in a different order.
- Avoid training common phrases that could be applied to every intent, for example, “I want to”. Ensure that the utterances are varied for larger variety and learning.
- After every change, train the model and check the model. Ensure that all the dots in the ML model are diagonal (in the True-positive and True-negative) quadrant and you do not have scattered utterances in other quadrants. Train the model until you achieve this.
- Regularly train the bot with new utterances.
- Regularly review the failed or abandoned utterances and add them to utterance list against a valid task or intent.
NLP Intent Detection Training Recommendations
- If there are a good number of sample utterances, try training the bot using Machine Learning approach first, before trying to train the fundamental meaning model.
- Define bot synonyms to build a domain dictionary such as pwd for a password; SB for a savings bank account.
- After every change to the model training, run the batch testing modules. Test suites are a means to perform regression testing of your bot’s ML model.
NLP Entity Detection Training Recommendations
The best approach to train entities is based on the type of entity as explained below:
- Entity type like List of Items (enumerated, lookup), City, Date, Country do not need any training unless the same entity type is used multiple types in the same task. If the same entity type is used in a bot task, use either of the training models to find the entity within the user utterances.
- When the entity type is String or Description, the recommended approach is to use Entity patterns and synonyms.
- For all other entity types, both NER and Patterns can be used in combination.
Entity Training Recommendations
- Use NER training where possible – NER coverage is higher than patterns.
- NER approach best suits detecting an entity where information is provided as unformatted data. For entities like Date and Time, the platform has been trained with a large set of data.
- NER is a neural network-based model and will need to be trained with at least 8-10 samples to work effectively.
You might want to read on ML Model, refer here.