I am new to testing AI chatbots and had a few questions.
The chatbots I am testing (text and voice) use DialogFlow ES as the NLP agent. Each agent has training phrases (each phrase has entities and synonyms). In order to test:
- Do I test using the training phrases or create new phrases not part of that list in Dialoglow?
- I read something about a 80/20 rule for training and test data. Does that apply to testing Dialoglow powered bots? If so, can i split the current training phrases into training (80%) + test (20%) data?
- If I already have a separate testing dataset, what happens if I find an issue, i.e. a particular phrase did not return the correct intent/response? Does that test case become part of training data to cause the agent to “learn” or should test and training phrases never be mixed?