I am building a bot with Rasa.ai.When training the bot with Rasa NLU, we use a training data file where the text, intent, entity etc. are specified. For example for a simple restaurant chatbot, the training file data.json
may contain
{
"text": "central indian restaurant",
"intent": "restaurant_search",
"entities": [
{
"start": 0,
"end": 7,
"value": "central",
"entity": "location"
},
{
"start": 8,
"end": 14,
"value": "indian",
"entity": "cuisine"
}
]
}
We use this to train the model. But we need to create this training file manually (or through a GUI).
Is there any tool where I can feed sentences and it can automatically create intent and entity?
Sample Input: Is there any central Indian restaurant?
Sample Output: The above data.json
EDIT:
To better explain this question - suppose I have a huge set of customer service call log. My understanding is with Rasa (or other similar framework) - a human being need to go through the call log and understand all possible intents, entity combination that happened in the past and create a file like data.json
such as above before training the model. This seems like a really unscalable problem. Is there a way to generate that data.json
file from those GB size call logs without involving a human being? Am I missing something here?