New Alexa features: Interactive teaching by customers

海外精选
海外精选的内容汇集了全球优质的亚马逊云科技相关技术内容。同时,内容中提到的“AWS” 是 “Amazon Web Services” 的缩写,在此网站不作为商标展示。
0
0
{"value":"*Today in Seattle, Dave Limp, Amazon’s senior vice president for devices, unveiled the latest lineup of products and services from his organization. During the presentation, Rohit Prasad, Amazon vice president and Alexa head scientist, described three new advances from the Alexa science team. One of those is interactive teaching by customers.*\n\nRead Alexa head scientist Rohit Prasad's overview of today's Alexa-related announcements [on Amazon's Day One blog](https://blog.aboutamazon.com/devices/ai-advances-make-alexa-more-natural-conversational-and-useful).\n\nLast year, we launched a [self-learning](https://www.amazon.science/blog/how-we-taught-alexa-to-correct-her-own-defects) feature that enables Alexa to automatically correct interpretation errors, based on cues such as customers’ rephrasing of requests or interruptions of Alexa’s responses. Millions of customers today enjoy the benefit of this capability. But what if the customer says something that Alexa doesn’t know how to interpret?\n\nTo allow customers to directly help Alexa learn the correct interpretation, we have given Alexa the ability to engage in live interactive teaching sessions with a customer, learn new concepts on the fly, generalize those concepts to new contexts, and associate them with the customer’s account.\n\nFor instance, if a customer says, “Alexa, set the living room light to study mode”, Alexa might now respond, “I don't know what study mode is. Can you teach me?” Alexa extracts a definition from the customer’s answer, and when the customer later makes the same request — or a similar request — Alexa responds with the learned action.\n\n![image.png](https://dev-media.amazoncloud.cn/566add27ef7b46aea5d01ba2defe7e09_image.png)\n\nUnlike Alexa [Routines](https://www.amazon.com/gp/help/customer/display.html?nodeId=202200080), where customers use the Alexa app to associate actions with verbal triggers (such as turning off the lights when the customer says “good night”), interactive teaching lets Alexa engage in a conversation to ask about unknown or unresolved concepts, in order to complete tasks that would fail otherwise.\n\nInteractive teaching allows Alexa to learn two different types of concepts. One is **entity concepts**: in the example above, “study mode” is a new entity that Alexa must learn. The other type is **declarative concepts**. With declarative concepts, Alexa learns how to interpret instructions that are disguised as declarations, such as “Alexa, it’s too dark in this room.”\n\n![image.png](https://dev-media.amazoncloud.cn/20a2a2012f904de3b2edc2dfd0d97504_image.png)\n\nInteractive teaching is a [conversational-AI](https://www.amazon.science/research-areas/conversational-ai-natural-language-processing) solution that uses the predictions of multiple deep-learning models to determine its next output during a teaching session. Those models have four chief functions:\n- **understanding-gap detection**, or automatically identifying the part of an utterance that Alexa doesn’t understand;\n- **concept interpretation**, or eliciting and extracting the definition of a concept from interactions with the customer;\n- **dialogue management**, or keeping conversations about new concepts on track; and\n- **declarative reasoning**, or evaluating the actions available to Alexa (e.g., controlling smart-home appliances) for the best matches to a declarative instruction (e.g., “It’s dark in here”).\n\nAlexa’s natural-language-understanding models classify customer utterances by domain — broad functional areas such as music or weather — and intent — the action the customer wants performed, such as playing music.\n\nThey also identify the slots and slot-values in the utterance, or the specific entities and entity types the intent should operate upon. For instance, in the utterance “Alexa, play ‘Blinding Lights’ by the Weeknd”, “Blinding Lights” is the value of the slot Song_Name, and “the Weeknd” is the value of the slot “Artist_Name”.\n\nWhen the probabilities of the top-ranked slots are low, the **understanding-gap-detection** model recognizes an opportunity to learn new slot concepts (such as “study mode” in the utterance “set the living room light to study mode”). The model is also trained to reject utterances such as “set the lights to, umm, never mind”.\n\nOnce the customer engages in a teaching session, the **concept interpretation** model elicits and extracts the interpretation of the new concept from the customer’s free-form speech.\n\nFor example, the customer could respond to the question “What do you mean by ‘study mode’?” by saying, “Well, you know, I usually study at night by setting the light to 50% brightness”.\n\nThe concept interpretation model would extract the phrase “50% brightness” from that utterance and store it as the definition of “study mode”.\n\nThe **dialogue management** model checks whether a customer’s answer to a question is within the scope of the question or not. For example, when Alexa asks, “What do you mean by ‘study mode’?”, the customer might reply, “Set it to a good brightness level for reading”. The model would recognize that that answer doesn’t provide a suitable concept definition.\n\nAfter every failed attempt to elicit a definition, the dialogue manager reduces the complexity of the follow-up question. For example, if the concept extraction model fails to extract a definition of “study mode” after one round of questioning, the dialogue manager might ask the more direct question “Can you provide me a value for brightness or color?”\n\nFinally, the **declarative-reasoning** model combines machine learning and machine reasoning to predict actions that correspond to customers’ declarative utterances. The model also helps verify that the chosen action is semantically appropriate in the context of the declarative utterance before deciding to store it for future re-use.\n\nAfter a successful teaching session, the previously learned concepts can be reused in relevant contexts. For instance, when a customer has taught Alexa that in the living room, “study mode” means setting the lights to 50%, Alexa knows to apply the same concept in the office, as well.\n\nSimilarly, if the customer has taught Alexa to respond to a declarative utterance such as “It’s dark in here” by turning on a light, Alexa knows that the subsequent utterance “I can’t see anything here” should trigger the same action.\n\nIn addition to automatically generalizing taught concepts, the teachable-AI capability will allow the customer to explicitly instruct Alexa to forget either the most recently learned or all learned concepts.\n\nAt launch, interactive teaching by customers will be available for Alexa smart-home devices, and it will expand to other features over time. This is an exciting step forward not just for Alexa but for AI services’ being explicitly taught by end users.\n\n**More coverage of Alexa announcements**\n- [Speaking-style adaptation](https://www.amazon.science/blog/new-text-to-speech-generator-and-rephraser-move-alexa-toward-concept-to-speech)\n- [Natural turn-taking](https://www.amazon.science/blog/change-to-alexa-wake-word-process-adds-natural-turn-taking)\n- [The science behind Echo Show 10](https://www.amazon.science/blog/the-science-behind-echo-show-10)\n\nABOUT THE AUTHOR\n\n#### **[Govind Thattai](https://www.amazon.science/author/govind-thattai)**\n\nGovind Thattai is a senior applied scientist in the Alexa AI organization.\n\n#### **[Gokhan Tur](https://www.amazon.science/author/gokhan-tur)**\n\nGokhan Tur is a senior principal scientist in the Alexa AI organization.\n\n#### **[Prem Natarajan](https://www.amazon.science/author/prem-natarajan)**\n\nPrem Natarjan is the Alexa AI vice president of natural understanding","render":"<p><em>Today in Seattle, Dave Limp, Amazon’s senior vice president for devices, unveiled the latest lineup of products and services from his organization. During the presentation, Rohit Prasad, Amazon vice president and Alexa head scientist, described three new advances from the Alexa science team. One of those is interactive teaching by customers.</em></p>\n<p>Read Alexa head scientist Rohit Prasad’s overview of today’s Alexa-related announcements <a href=\"https://blog.aboutamazon.com/devices/ai-advances-make-alexa-more-natural-conversational-and-useful\" target=\"_blank\">on Amazon’s Day One blog</a>.</p>\n<p>Last year, we launched a <a href=\"https://www.amazon.science/blog/how-we-taught-alexa-to-correct-her-own-defects\" target=\"_blank\">self-learning</a> feature that enables Alexa to automatically correct interpretation errors, based on cues such as customers’ rephrasing of requests or interruptions of Alexa’s responses. Millions of customers today enjoy the benefit of this capability. But what if the customer says something that Alexa doesn’t know how to interpret?</p>\n<p>To allow customers to directly help Alexa learn the correct interpretation, we have given Alexa the ability to engage in live interactive teaching sessions with a customer, learn new concepts on the fly, generalize those concepts to new contexts, and associate them with the customer’s account.</p>\n<p>For instance, if a customer says, “Alexa, set the living room light to study mode”, Alexa might now respond, “I don’t know what study mode is. Can you teach me?” Alexa extracts a definition from the customer’s answer, and when the customer later makes the same request — or a similar request — Alexa responds with the learned action.</p>\n<p><img src=\"https://dev-media.amazoncloud.cn/566add27ef7b46aea5d01ba2defe7e09_image.png\" alt=\"image.png\" /></p>\n<p>Unlike Alexa <a href=\"https://www.amazon.com/gp/help/customer/display.html?nodeId=202200080\" target=\"_blank\">Routines</a>, where customers use the Alexa app to associate actions with verbal triggers (such as turning off the lights when the customer says “good night”), interactive teaching lets Alexa engage in a conversation to ask about unknown or unresolved concepts, in order to complete tasks that would fail otherwise.</p>\n<p>Interactive teaching allows Alexa to learn two different types of concepts. One is <strong>entity concepts</strong>: in the example above, “study mode” is a new entity that Alexa must learn. The other type is <strong>declarative concepts</strong>. With declarative concepts, Alexa learns how to interpret instructions that are disguised as declarations, such as “Alexa, it’s too dark in this room.”</p>\n<p><img src=\"https://dev-media.amazoncloud.cn/20a2a2012f904de3b2edc2dfd0d97504_image.png\" alt=\"image.png\" /></p>\n<p>Interactive teaching is a <a href=\"https://www.amazon.science/research-areas/conversational-ai-natural-language-processing\" target=\"_blank\">conversational-AI</a> solution that uses the predictions of multiple deep-learning models to determine its next output during a teaching session. Those models have four chief functions:</p>\n<ul>\n<li><strong>understanding-gap detection</strong>, or automatically identifying the part of an utterance that Alexa doesn’t understand;</li>\n<li><strong>concept interpretation</strong>, or eliciting and extracting the definition of a concept from interactions with the customer;</li>\n<li><strong>dialogue management</strong>, or keeping conversations about new concepts on track; and</li>\n<li><strong>declarative reasoning</strong>, or evaluating the actions available to Alexa (e.g., controlling smart-home appliances) for the best matches to a declarative instruction (e.g., “It’s dark in here”).</li>\n</ul>\n<p>Alexa’s natural-language-understanding models classify customer utterances by domain — broad functional areas such as music or weather — and intent — the action the customer wants performed, such as playing music.</p>\n<p>They also identify the slots and slot-values in the utterance, or the specific entities and entity types the intent should operate upon. For instance, in the utterance “Alexa, play ‘Blinding Lights’ by the Weeknd”, “Blinding Lights” is the value of the slot Song_Name, and “the Weeknd” is the value of the slot “Artist_Name”.</p>\n<p>When the probabilities of the top-ranked slots are low, the <strong>understanding-gap-detection</strong> model recognizes an opportunity to learn new slot concepts (such as “study mode” in the utterance “set the living room light to study mode”). The model is also trained to reject utterances such as “set the lights to, umm, never mind”.</p>\n<p>Once the customer engages in a teaching session, the <strong>concept interpretation</strong> model elicits and extracts the interpretation of the new concept from the customer’s free-form speech.</p>\n<p>For example, the customer could respond to the question “What do you mean by ‘study mode’?” by saying, “Well, you know, I usually study at night by setting the light to 50% brightness”.</p>\n<p>The concept interpretation model would extract the phrase “50% brightness” from that utterance and store it as the definition of “study mode”.</p>\n<p>The <strong>dialogue management</strong> model checks whether a customer’s answer to a question is within the scope of the question or not. For example, when Alexa asks, “What do you mean by ‘study mode’?”, the customer might reply, “Set it to a good brightness level for reading”. The model would recognize that that answer doesn’t provide a suitable concept definition.</p>\n<p>After every failed attempt to elicit a definition, the dialogue manager reduces the complexity of the follow-up question. For example, if the concept extraction model fails to extract a definition of “study mode” after one round of questioning, the dialogue manager might ask the more direct question “Can you provide me a value for brightness or color?”</p>\n<p>Finally, the <strong>declarative-reasoning</strong> model combines machine learning and machine reasoning to predict actions that correspond to customers’ declarative utterances. The model also helps verify that the chosen action is semantically appropriate in the context of the declarative utterance before deciding to store it for future re-use.</p>\n<p>After a successful teaching session, the previously learned concepts can be reused in relevant contexts. For instance, when a customer has taught Alexa that in the living room, “study mode” means setting the lights to 50%, Alexa knows to apply the same concept in the office, as well.</p>\n<p>Similarly, if the customer has taught Alexa to respond to a declarative utterance such as “It’s dark in here” by turning on a light, Alexa knows that the subsequent utterance “I can’t see anything here” should trigger the same action.</p>\n<p>In addition to automatically generalizing taught concepts, the teachable-AI capability will allow the customer to explicitly instruct Alexa to forget either the most recently learned or all learned concepts.</p>\n<p>At launch, interactive teaching by customers will be available for Alexa smart-home devices, and it will expand to other features over time. This is an exciting step forward not just for Alexa but for AI services’ being explicitly taught by end users.</p>\n<p><strong>More coverage of Alexa announcements</strong></p>\n<ul>\n<li><a href=\"https://www.amazon.science/blog/new-text-to-speech-generator-and-rephraser-move-alexa-toward-concept-to-speech\" target=\"_blank\">Speaking-style adaptation</a></li>\n<li><a href=\"https://www.amazon.science/blog/change-to-alexa-wake-word-process-adds-natural-turn-taking\" target=\"_blank\">Natural turn-taking</a></li>\n<li><a href=\"https://www.amazon.science/blog/the-science-behind-echo-show-10\" target=\"_blank\">The science behind Echo Show 10</a></li>\n</ul>\n<p>ABOUT THE AUTHOR</p>\n<h4><a id=\"Govind_Thattaihttpswwwamazonscienceauthorgovindthattai_57\"></a><strong><a href=\"https://www.amazon.science/author/govind-thattai\" target=\"_blank\">Govind Thattai</a></strong></h4>\n<p>Govind Thattai is a senior applied scientist in the Alexa AI organization.</p>\n<h4><a id=\"Gokhan_Turhttpswwwamazonscienceauthorgokhantur_61\"></a><strong><a href=\"https://www.amazon.science/author/gokhan-tur\" target=\"_blank\">Gokhan Tur</a></strong></h4>\n<p>Gokhan Tur is a senior principal scientist in the Alexa AI organization.</p>\n<h4><a id=\"Prem_Natarajanhttpswwwamazonscienceauthorpremnatarajan_65\"></a><strong><a href=\"https://www.amazon.science/author/prem-natarajan\" target=\"_blank\">Prem Natarajan</a></strong></h4>\n<p>Prem Natarjan is the Alexa AI vice president of natural understanding</p>\n"}
目录
亚马逊云科技解决方案 基于行业客户应用场景及技术领域的解决方案
联系亚马逊云科技专家
亚马逊云科技解决方案
基于行业客户应用场景及技术领域的解决方案
联系专家
0
目录
关闭