Alexa AI team discusses NeurIPS workshop best paper award

海外精选
海外精选的内容汇集了全球优质的亚马逊云科技相关技术内容。同时,内容中提到的“AWS” 是 “Amazon Web Services” 的缩写,在此网站不作为商标展示。
0
0
{"value":"![image.png](https://dev-media.amazoncloud.cn/e209f177a337435b92b0af83092e9dc8_image.png)\n\nFrom top to bottom: Di Jin, applied scientist; Shuyang Gao, applied scientist; Seokhwan Kim, principal applied scientist; Yang Liu, principal applied scientist; and Dilek Hakkani-Tür, senior principal scientist, won a best paper award for \"[Towards Zero- and Few-shot Knowledge-seeking Turn Detection in Task-oriented Dialogue Systems](https://www.amazon.science/publications/towards-zero-and-few-shot-knowledge-seeking-turn-detection-in-task-oriented-dialogue-systems)\".\n\nA team of Alexa AI researchers in December won the best paper award at the NeurIPS 2021 workshop on Efficient Natural Language and Speech Processing (ENLSP).\n\n[Di Jin](https://www.linkedin.com/in/jindi11/?locale=en_US), an applied scientist, won the award along with [Shuyang Gao](https://www.linkedin.com/in/biubiugao/), applied scientist; [Seokhwan Kim](https://www.linkedin.com/in/seokhwan-kim/), principal applied scientist; [Yang Liu](https://www.linkedin.com/in/yang-liu-8555143/), principal applied scientist; and [Dilek Hakkani-Tür](https://www.linkedin.com/in/dilek-hakkani-tur-9517543/), senior principal scientist, for their paper, “Towards Zero- and Few-shot Knowledge-seeking Turn Detection in Task-oriented Dialogue Systems”.\n\nPresently, the authors note, task-oriented dialogue systems frequently “rely on pre-defined APIs to complete tasks and filter out any other requests beyond the APIs as out-of-domain cases.” The paper focuses on how to more efficiently process out-of-domain customer requests “by incorporating external domain knowledge from the web or any other sources.”\n\nThe issue, Jin explained, results primarily from the gap between training data and actual user requests.\n\n“It's very hard to guarantee that all user queries or user input text are in the exact same distribution as the training data,” he said. “Our APIs are based on common user queries, so we needed to enhance the model to detect the out-of-domain data and route those user queries elsewhere to be addressed.”\n\nThe authors designed a model that more efficiently identifies and routes out-of-domain requests, and named their model REDE because it utilizes adaptive REpresentation learning and DEnsity estimation.\n\n“The most typical way to handle this kind of issue is to train a binary classifier, e.g., a large-scale pre-trained language model like BERT,” Kim explained. “To achieve this, we need positive and negative instances and we can build a machine learning model to decide whether a given input can be addressed by the API or requires external knowledge. But because open domain conversational AI systems allow customers to ask anything, it is difficult to collect a sufficient number of out-of-domain instances to train a classifier.\n\n“What we proposed is not to train the classifier based on some training dataset, but instead to adapt the existing representation,” Jin said. “We transform that representation so that the new representation has enough distinctive power between the two classes, the seen — the instances the current API can field — and the potentially unseen, or out-of-domain ones.”\n\nThe paper notes the REDE model outperformed binary classifiers, for both standard and out-of-domain requests, for both low-resource and high-resource settings. And in zero-shot and few-shot scenarios, the REDE model gains an even larger performance margin relative to traditional binary classifiers like BERT.\n\n“The key takeaway is that this kind of simple transformation of the representation works very well and efficiently,” Kim said. “That will help us to develop even more robust conversational models with much smaller datasets and smaller models.”\n\nABOUT THE AUTHOR\n\n#### **Staff writer**","render":"<p><img src=\"https://dev-media.amazoncloud.cn/e209f177a337435b92b0af83092e9dc8_image.png\" alt=\"image.png\" /></p>\n<p>From top to bottom: Di Jin, applied scientist; Shuyang Gao, applied scientist; Seokhwan Kim, principal applied scientist; Yang Liu, principal applied scientist; and Dilek Hakkani-Tür, senior principal scientist, won a best paper award for “<a href=\"https://www.amazon.science/publications/towards-zero-and-few-shot-knowledge-seeking-turn-detection-in-task-oriented-dialogue-systems\" target=\"_blank\">Towards Zero- and Few-shot Knowledge-seeking Turn Detection in Task-oriented Dialogue Systems</a>”.</p>\n<p>A team of Alexa AI researchers in December won the best paper award at the NeurIPS 2021 workshop on Efficient Natural Language and Speech Processing (ENLSP).</p>\n<p><a href=\"https://www.linkedin.com/in/jindi11/?locale=en_US\" target=\"_blank\">Di Jin</a>, an applied scientist, won the award along with <a href=\"https://www.linkedin.com/in/biubiugao/\" target=\"_blank\">Shuyang Gao</a>, applied scientist; <a href=\"https://www.linkedin.com/in/seokhwan-kim/\" target=\"_blank\">Seokhwan Kim</a>, principal applied scientist; <a href=\"https://www.linkedin.com/in/yang-liu-8555143/\" target=\"_blank\">Yang Liu</a>, principal applied scientist; and <a href=\"https://www.linkedin.com/in/dilek-hakkani-tur-9517543/\" target=\"_blank\">Dilek Hakkani-Tür</a>, senior principal scientist, for their paper, “Towards Zero- and Few-shot Knowledge-seeking Turn Detection in Task-oriented Dialogue Systems”.</p>\n<p>Presently, the authors note, task-oriented dialogue systems frequently “rely on pre-defined APIs to complete tasks and filter out any other requests beyond the APIs as out-of-domain cases.” The paper focuses on how to more efficiently process out-of-domain customer requests “by incorporating external domain knowledge from the web or any other sources.”</p>\n<p>The issue, Jin explained, results primarily from the gap between training data and actual user requests.</p>\n<p>“It’s very hard to guarantee that all user queries or user input text are in the exact same distribution as the training data,” he said. “Our APIs are based on common user queries, so we needed to enhance the model to detect the out-of-domain data and route those user queries elsewhere to be addressed.”</p>\n<p>The authors designed a model that more efficiently identifies and routes out-of-domain requests, and named their model REDE because it utilizes adaptive REpresentation learning and DEnsity estimation.</p>\n<p>“The most typical way to handle this kind of issue is to train a binary classifier, e.g., a large-scale pre-trained language model like BERT,” Kim explained. “To achieve this, we need positive and negative instances and we can build a machine learning model to decide whether a given input can be addressed by the API or requires external knowledge. But because open domain conversational AI systems allow customers to ask anything, it is difficult to collect a sufficient number of out-of-domain instances to train a classifier.</p>\n<p>“What we proposed is not to train the classifier based on some training dataset, but instead to adapt the existing representation,” Jin said. “We transform that representation so that the new representation has enough distinctive power between the two classes, the seen — the instances the current API can field — and the potentially unseen, or out-of-domain ones.”</p>\n<p>The paper notes the REDE model outperformed binary classifiers, for both standard and out-of-domain requests, for both low-resource and high-resource settings. And in zero-shot and few-shot scenarios, the REDE model gains an even larger performance margin relative to traditional binary classifiers like BERT.</p>\n<p>“The key takeaway is that this kind of simple transformation of the representation works very well and efficiently,” Kim said. “That will help us to develop even more robust conversational models with much smaller datasets and smaller models.”</p>\n<p>ABOUT THE AUTHOR</p>\n<h4><a id=\"Staff_writer_26\"></a><strong>Staff writer</strong></h4>\n"}
目录
亚马逊云科技解决方案 基于行业客户应用场景及技术领域的解决方案
联系亚马逊云科技专家
亚马逊云科技解决方案
基于行业客户应用场景及技术领域的解决方案
联系专家
0
目录
关闭
contact-us