简体   繁体   English

NLC或R&R的再训练方法

[英]Re-training Approach for NLC or R&R

The ground truth we know is used to re-train the NLC or R&R. 我们知道的基本事实用于重新训练NLC或R&R。

The ground truth is a question level training data. 基本事实是问题级别的训练数据。

eg 例如

"How hot is it today?,temperature" “今天天气热吗?温度”

The question "how hot is it today?" 问题“今天有多热?” is therefore classified to "temperature" class. 因此被归为“温度”等级。

Once the application is up, real user questions will be received. 应用程序启动后,将收到真实的用户问题。 Some are the same (ie the question from the real users are the same to the question in the ground truth), some are similar terms, some are new questions. 有些是相同的(即,来自真实用户的问题与基本事实中的问题是相同的),有些是相似的术语,有些是新问题。 Assume the application has a feedback loop to know whether or not the class (for NLC) or answer (for R&R) are relevant. 假设应用程序有一个反馈循环,以了解该类(对于NLC)或答案(对于R&R)是否相关。

About the new questions, the approach seems to just add the them to the ground truth, which is then used to re-train the NLC/R&R?
For the questions with similar terms, do we just add them like the new questions, or do we just ignore them, given that similar terms can also be scored well even similar terms are not used to train the classifier?
In the case of the same questions, there seems nothing to do on the ground truth for NLC, however, to the R&R, are we just increase or decrease 1 for the relevance label in the ground truth?

The main question here is, in short, about what the re-training approach is for NLC & R&R... 简而言之,这里的主要问题是关于NLC&R&R的再培训方法是什么...

Once your application has gone live, you should periodically review your feedback log for opportunities for improvement. 应用程序上线后,您应该定期查看反馈日志,以寻求改进的机会。 For NLC, if there are texts being incorrectly classified, then you can add those texts to the training set and retrain in order to improve your classifier. 对于NLC,如果有一些文本分类不正确,则可以将这些文本添加到训练集中并重新训练,以改善分类器。

It is not necessary to capture every imaginable variation of a class, as long as your classifier is returning acceptable responses. 只要您的分类器返回可接受的响应,就不必捕获类的每个可能的变体。

You could use the additional examples of classes from your log to assemble a test set of texts that do not feature in your training set. 您可以使用日志中的其他类示例来组装测试集中没有的文本测试集。 Running this test set when you make changes will enable you to determine whether or not a change has inadvertently caused a regression. 进行更改时运行此测试集将使您能够确定更改是否无意间导致了回归。 You can run this test either by calling the classifier using a REST client, or via the Beta Natural Language Classifier toolkit. 您可以通过使用REST客户端调用分类器或通过Beta自然语言分类器工具包来运行此测试。

A solid retraining approach should be getting feedback from live users. 可靠的再培训方法应该是从实时用户那里获得反馈。 Your testing and validation of any retrained NLC (or R&R for that matter) should be guided by some of the principles that James Ravenscroft has outlined here ( https://brainsteam.co.uk/2016/03/29/cognitive-quality-assurance-an-introduction/ ). 您对任何经过再培训的NLC(或有关R&R)的测试和验证应遵循James Ravenscroft在此处概述的一些原则( https://brainsteam.co.uk/2016/03/29/cognitive-quality-保证简介/ )。

The answer by @davidgeorgeuk is correct, but fails to extend the thought to the conclusion that you are looking for. @davidgeorgeuk 的回答是正确的,但未能将想法扩展到您正在寻找的结论。 I would have a monthly set of activities where I would go through application logs where REAL users are indicating that your not classifying things correctly, and also incorporate any new classes to your classifier. 我将每月进行一系列活动,其中将浏览应用程序日志,其中REAL用户指示您未正确分类事物,并且还将任何新类合并到分类器中。 I would retrain a second instance of NLC with the new data, and go through the test scenarios outlined above. 我将使用新数据重新训练NLC的第二个实例,并进行上面概述的测试方案。

Once you are satisfied that you have IMPROVED your model, I would then switch my code to point at the new NLC instance, and the old NLC instance would be your "backup" instance, and the one that you would use for this exercise the next month. 当您对模型进行了改进后,我将切换代码以指向新的NLC实例,而旧的NLC实例将成为您的“备份”实例,下一个将用于本练习的实例月。 It's just applying a simple DevOps approach to managing your NLC instances. 它只是使用一种简单的DevOps方法来管理您的NLC实例。 You could extend this to a development, QA, production scenario if you wanted. 如果需要,可以将其扩展到开发,质量检查,生产方案。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 Watson R&R服务能否指定特定用户是否可以访问受限答案? - Can Watson R&R service specify whether specific user can access to restricted answer? 如何查看我在其中一个R&R群集中消耗了多少空间? - How can I check how much space I have consumed in one of my R&R clusters? Watson NLC培训数据和用户问题中的逗号 - Watson NLC training data and commas in user questions IBM Watson NLC - 使用超过 20,000 个文本示例进行培训? - IBM Watson NLC - Training with more than 20,000 text examples? 使用 IBM Watson NLU 和 IBM Watson NLC 服务与相同训练集组装的分类器产生不同的结果 - Classifiers assembled with identical training sets using IBM Watson NLU and IBM Watson NLC services yield different results 为什么在Watson NLC中对训练(1024个字符)和生产(2048个字符)有不同的大小限制? - Why are there different size limitations in Watson NLC for training (1024 chars) and for production (2048 chars)? 如何使用Watson NLC Toolkit管理多个培训集 - How do I manage multiple training sets using the Watson NLC Toolkit 如何在HTTP请求中发送多部分表单数据(用于Watson NLC培训)? - How do I send multipart form data in an HTTP request (for Watson NLC training)? 在bluemix上创建NLC分类器 - creating NLC classifier on bluemix 从 Watson NLC 转换为 NLU - Convert from Watson NLC to NLU
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM