Amazon has amplified its drive into the artificial intelligence (AI) sector, sparking debates regarding the extent of customer data utilized in shaping its extensive large language model (LLM).
A declaration last week indicated the forthcoming integration of generative AI technology into Alexa and Amazon’s intelligent home devices.
However, a subsequent interview aired on Bloomberg TV has ignited apprehensions regarding the nature of the data being employed to fine-tune Alexa’s AI capabilities.
A report by NBC News on Tuesday implied that employing user dialogues as training data might be a newfound or disclosed practice — a portrayal that was challenged by an Amazon representative in a conversation with Decrypt.
The dispute was initiated following a Q&A session on Bloomberg TV featuring David Limp, Amazon’s Senior Vice President of Devices and Services.
Limp purportedly stated that “users would be volunteering their voice data and conversations for Amazon’s LLM training purposes” upon consenting to utilize a more “personalized” variant of Alexa.
A spokesperson from Amazon countered, informing Decrypt that Limp’s remarks were misconstrued in the NBC report.
The scenario unfolds as AI advances continually blur the boundaries of technological capabilities and ethical data use, positioning Amazon in the spotlight amidst discussions on customer data usage in AI model development.
“[Amazon] has always believed that training Alexa with real-world requests is essential to delivering an experience to customers that’s accurate and personalized and constantly getting better,” the spokesperson said.
“But in tandem, we give customers control over whether their Alexa voice recordings are used to improve the service, and we always honor our customer preferences when we train our models.”
Amazon highlighted that an upcoming AI-enhanced Alexa update will introduce more natural voice interactions, including real-time news and smart home functions.