Previous Blogs

July 19, 2022
Qualcomm Accelerates Wearables with W5 Platforms

July 12, 2022
New Research Highlights Opportunities and Challenges for Private 5G

June 29, 2022
Arm Aims to Make Mobile Graphics “Immortal-is”

June 14, 2022
Cisco Brings Simplicity and Observability to Networks, Collaboration and Cloud Apps

May 24, 2022
Microsoft Unveils Foundation for AI-Powered Client/Cloud Hybrid Loop

May 18, 2022
Citrix to Integrate with Microsoft Windows 365

May 3, 2022
Dell Expands APEX, Adds Analytics and Data Recovery

April 27, 2022
Arm Simplifies and Modernizes IoT Development with Virtual Hardware

April 21, 2022
Amazon’s Launch of Buy with Prime Highlights Growth of Logistics Business

March 30, 2022
Intel Spices Up PC Market with Arc GPU Launch

March 22, 2022
Nvidia GTC Announcements Confirm it’s a Connected, Multi-Chip World

March 15, 2022
Lenovo and AMD Announcements Highlight Spring PC Refresh

March 8, 2022
The Future of Semiconductors is UCIe

March 2, 2022
Qualcomm Demos Future of Connectivity with WiFi 7 and X70 5G Chips

February 24, 2022
5G Edge Computing Challenges Remain

February 9, 2022
Samsung Raises the Bar with Ultra Versions of S22 and Tab S8

January 20, 2022
US 5G Market Just Got Much More Interesting

January 4, 2022
Qualcomm Extends Automotive Offerings with Snapdragon Ride Vision, Digital Chassis

2021 Blogs

2020 Blogs

2019 Blogs

2018 Blogs

2017 Blogs

2016 Blogs

2015 Blogs

2014 Blogs

2013 Blogs

TECHnalysis Research Blog

July 20, 2022
Amazon Extends Alexa’s Reach with New Tools

By Bob O'Donnell

Ever since Amazon first introduced the Alexa voice assistant via its Echo smart speakers, I’ve longed for the day when I could just talk to devices and have them do what I wanted them to. Unfortunately, we’re not there just yet, but thanks to some of the many new software development tools Amazon announced at its Alexa Live 2022 conference, we’re definitely getting closer.

A big part of the problem is that getting machines to understand natural language interactions is a lot harder than it first appeared. Many of us learned this to some degree in the early days of voice assistants when what seemed like very reasonable information requests often ended up being answered with frustratingly nonsensical responses. It turns out human beings are much better at understanding the subtle nuances (or very obvious differences) between what someone meant versus what they actually said.

One of the issues that became apparent is that the structure and syntax of spoken language that we all understand intuitively often needs to be broken down into many different sub-components before they can be “understood” by machines. Practically speaking, that means the evolution of machine intelligence has been slower than many hoped because of the need to figure out the incremental steps necessary to really make sense of a given request. Even today, some of the most sophisticated natural language AI models are running into walls when it comes to doing any kind of simple reasoning that requires the kind of independent thinking that a young child can do.

On top of this, when it comes to smart home-focused devices—which is where voice-assistant powered machines continue to make their mark—there has been a frustrating wealth of incompatible standards that have made it physically challenging to get devices to work together. Thankfully, the new Matter standard—which Amazon, Apple, Google and many others are planning to support—goes a long way towards solving this challenge. As a result, the very real problem of getting multiple devices from different vendors or even different smart home ecosystems to seamlessly work together may soon be little more than a distant memory.

With all this context in mind, the many different software developer focused announcements that Amazon made at Alexa Live 2022 make even more sense. First, the company debuted the Connect Kit SDK (Software Development Kit) for Matter. This extends a range of Amazon developed connection services to any Matter-capable device that supports it. This means that companies building smart home devices can leverage the work Amazon has done for critical features like cloud connectivity, over-the-air software updates, activity logging, metrics and more. Ultimately, the goal is to get a baseline of functionality that will encourage users to purchase and install multiple Matter-capable smart home products.

Of course, once devices are connected, they still need to communicate with each other in intelligent ways to provide more functionality. To address this, Amazon also unveiled the Alexa Ambient Home Dev Kit, which combines services and software APIs (application programming interfaces) that allow multiple devices to work together easily and silently in the background. Amazon and others call this “ambient computing”, because it’s meant to provide a mesh of essentially invisible computing services. The first version of this dev kit includes Home State APIs to do things like simultaneously put all your smart home devices into different modes (such as Sleep, Dinner Time, Home, etc.). Safety and Security APIs automatically send alarms from connected sensors, such as smoke alarms, to other connected devices and applications to ensure the alarms are noticed/heard. API for Credentials makes user setup across multiple devices easier by sharing Thread network credentials (a key part of the Matter standard) so that users don’t have to do it more than once.

Speaking of easier setup, Amazon also announced plans to let its “Frustration-Free Setup” features be used by non-Amazon branded devices purchased in other retail stores. The company plans to leverage the Matter standard to enable this, emphasizing once again how important Matter is going to be for future devices.

For those working with voice-based interfaces, Amazon is also working to enable some of the first real capabilities for an industry development called the Voice Interoperability Initiative, or VII. First announced in 2019, VII is designed to let multiple voice assistants work together in a seamless way to provide more complex interactions. As an example, Amazon said it is working with Skullcandy and Native Voice to allow use of Alexa along with the “Hey Skullcandy” assistants and commands at the same time. For example, you can use “Hey Skullcandy” to enable voice-based control of headphone settings and media playback, but also ask Alexa for the latest news headlines and have them play back over the Skullcandy headphones.

Amazon also debuted the Alexa Voice Service (AVS) SDK 3.0 which combines Alexa voice-based capabilities with the previously separate set Alexa Smart Screen SDK for generating smart screen-based responses. Using this would allow companies to potentially do things like have a voice-based interface with visual confirmations on screen or to create multi-modal interfaces that leverage both at the same time.

Finally, Amazon also unveiled a host of new Skills, Skill Development, Skill Promotion, and Skill education tools designed to help developers who want to create Skill “apps” for the Alexa ecosystem across a wide range of different devices, including TVs, PCs, tablets, smart displays, cars, and more. All told, it looks to be a comprehensive range of capabilities that should make a tangible difference for those who want to leverage the installed base of roughly 300 million Alexa-capable devices.

Unfortunately, browsing through multi-level screen-based menus, pushing numerous combinations of buttons, and trying to figure out the mindset of the engineers who designed the user interfaces is still the reality of many gadgets today. I, for one, look forward to the ability to do something like plug a new device in, tell it to connect my other devices, have it speak to me through some connected speaker to tell me that it did so (or if it didn’t, what needs to be done to fix that), answer questions about what it can and can’t do and how I can control it, and finally, keep me up-to-date verbally about any problems that may arise or new capabilities it acquires.

As these new tools and capabilities start to get deployed, the potential for significantly easier, voice-based control of a multitude of digital devices is getting tantalizingly closer.

Here’s a link to the original column:

Bob O’Donnell is the president and chief analyst of TECHnalysis Research, LLC a market research firm that provides strategic consulting and market research services to the technology industry and professional financial community. You can follow him on Twitter @bobodtech.