Previous Blogs

August 15, 2017
The Myth of General Purpose Wearables

August 8, 2017
IoT Connections Made Easy

August 1, 2017
Smarter Computing

July 25, 2017
The Value of Limits

July 18, 2017
Tech in the Heartland

June 27, 2017
Business Realities vs. Tech Dreams

June 20, 2017
The Power of Hidden Tech

June 13, 2017
Computing Evolves from Outside In to Inside Out

June 6, 2017
The Overlooked Surprises of Apple’s WWDC Keynote

May 30, 2017
Are AR and VR Only for Special Occasions?

May 23, 2017
The Digital Car

May 16, 2017
Digital Assistants Drive New Meta-Platform Battle

May 9, 2017
Getting Smart on Smart Speakers

May 5, 2017
Intel Opens High-Tech "Garage"

May 2, 2017
The Hidden Value of Analog

April 28, 2017
Google’s Waymo Starts Driving Passengers

April 25, 2017
The Robotic Future

April 21, 2017
Sony Debuts New Pro Camera

April 18, 2017
Should Apple Build a Car?

April 14, 2017
PC Market Outlook Improving

April 11, 2017
Little Data Analytics

April 7, 2017
Facebook Debuts Free Version of Workplace Collaboration Tool

April 4, 2017
Samsung Building a Platform Without an OS

March 31, 2017
Microsoft Announces Windows 10 Creators Update Release Date

March 28, 2017
Augmented Reality Finally Delivers on 3D Promise

March 24, 2017
Intel Creates AI Organization

March 21, 2017
Chip Magic

March 17, 2017
Microsoft Unveils Teams Chat App

March 14, 2017
Computing on the Edge

March 7, 2017
Cars Need Digital Safety Standards Too

February 28, 2017
The Messy Path to 5G

February 24, 2017
AMD Launches Ryzen CPU

February 21, 2017
Rethinking Wearable Computing

February 17, 2017
Samsung Heir Arrest Unlikely to Impact Sales

February 14, 2017
Modern Workplaces Still More Vision Than Reality

February 10, 2017
Lenovo Develops Energy-Efficient Soldering Technology

February 7, 2017
The Missing Map from Silicon Valley to Main Street

January 31, 2017
The Network vs. The Computer

January 27, 2017
Facebook Adds Support For FIDO Security Keys

January 24, 2017
Voice Drives New Software Paradigm

January 20, 2017
Tesla Cleared of Fault in NHTSA Crash Probe

January 17, 2017
Inside the Mind of a Hacker

January 13, 2017
PC Shipments Stumble but Turnaround is Closer

January 10, 2017
Takeaways from CES 2017

January 3, 2017
Top 10 Tech Predictions for 2017

2016 Blogs

2015 Blogs

2014 Blogs

2013 Blogs

TECHnalysis Research Blog

August 22, 2017
The Evolution of Smart Speakers

By Bob O'Donnell

For a relatively nascent product category, smart speakers like Amazon Echo and Google Home are already seeing a huge influx of attention from both consumers and potential competitors eager to enter the market. Apple has announced the HomePod and numerous other vendors have either unveiled or are heavily rumored to be working on versions of their own.

Harman Kardon (in conjunction with Microsoft), GE Lighting and Lenovo have announced products in the US, while Alibaba, Xiaomi and, among others, have said they will be bringing products out in China. In addition, Facebook is rumored to be building a screen-equipped smart speaker called Gizmo.

One obvious question after hearing about all the new entrants is, how can they all survive? The short answer, of course, is they won’t. Nevertheless, expect to see a lot of jockeying, marketing and positioning over the next year or two because it’s still very early days in the world of AI-powered and personal assistant-driven smart speakers.

Yes, Amazon has built an impressive and commanding presence with the Echo line, but there are many limitations to Echos and all current smart speakers that frustrate existing users. Thankfully, technology improvements are coming that will enable competitors to differentiate themselves from others in ways which reduce the frustration and increase the satisfaction that consumers have with smart speakers.

Part of the work involves the overall architecture of the devices and how they interact with cloud-based services. For example, one of the critical capabilities that many users want is the ability to accurately recognize different individuals that speak to the device, so that responses can be customized for different members of a household. To achieve this as quickly and accurately as possible, it doesn’t make sense to try and send the audio signal to the cloud and then wait for the response. Even with superfast network connections, the inevitable delays make interactions with the device feel somewhat awkward.

The same problem exists when you try to move beyond the simple single query requests that most people are making to their smart speakers today. (Alexa, play music by horn bands or Alexa, what is the capital of Iceland?) In order to have naturally flowing, multi-question or multi-statement conversations, the delays (or latency) have to be dramatically reduced.

The obvious answer to the problem is to do more of the recognition and response work locally on the device and not rely on a cloud-based network connection to do so. In fact, this is a great example of the larger trend of edge computing, where we are seeing devices or applications that use to rely solely on big data centers in the cloud start to do more of the computational work on their own.

That’s part of the reason you’re starting to see companies like Qualcomm and Intel, among others, develop chips that are designed to enable more powerful local computing work on devices like smart speakers. The ability to learn and then recognize different individuals, for example, is something that the DSP (digital signal processor) component of new chips from these vendors can do.

Another technological challenge facing current generation products is recognition accuracy. Everyone who has used a smart speaker or digital assistant on other device has had the experience of not being understood. Sometimes that’s due to how the question or command is phrased, but it’s often due to background noises, accents, intonation or other factors that essentially end up providing an imperfect audio signal to the cloud-based recognition engine. Again, more local audio signal processing can often improve the audio signal to be sent, thereby enhancing overall recognition.

Going further, most of the AI-based learning algorithms used to recognize and accurately respond to speech will likely need to be run in very large, compute-intensive cloud data centers. However, the idea of being able to start do pattern recognition of common phrases (a form of inferencing—the second key aspect of machine learning and AI) locally with the right kind of computing engines and hardware architectures is becoming increasingly possible. It may be a long time before all that kind of work can be done within smart speakers and other edge devices, but even doing some speech recognition on the device should enable higher accuracy and longer conversations. In short, a much better user experience.

As new entrants try to differentiate their products in an increasingly crowded space, the ability to offer some key tech-based improvements is going to be essential. Clearly there’s a great deal of momentum behind the smart speaker phenomenon, but it’s going to take these kind performance improvements to move them beyond idle curiosities and into truly useful, everyday kinds of tools.

Here's a link to the column:

Bob O’Donnell is the president and chief analyst of TECHnalysis Research, LLC a market research firm that provides strategic consulting and market research services to the technology industry and professional financial community. You can follow him on Twitter @bobodtech.

Leveraging more than 10 years of award-winning, professional radio experience, TECHnalysis Research participates in regular audio podcasts in conjunction with the team at
  Research Offerings
TECHnalysis Research offers a wide range of research deliverables that you can read about here.