All posts in “virtual assistant”

Ecobee’s new voice-powered light switch moves closer to whole-home Alexa

Alexa is already everywhere in a lot of homes, thanks to the affordability and ease of installation/setup of the Echo Dot. But Alexa could become even more seamlessly integrated into your home, if you think about it. And Canadian smart home tech maker ecobee did think about it, which is how they came up with the ecobee Switch+.

Ecobee is probably most known for their connected thermostats, which are one of the strongest competitors out there for Nest. The company’s been building other products, too, however, and developing closer ties with Amazon and its Alexa virtual assistant. The Switch+ has the closest ties yet, since it includes Alexa Voice Service and far-field voice detection microphone arrays to essential put an Echo in your wall wherever you have a light switch handy.

The ecobee Switch+ is still a connected light switch that works like similar offerings from Belkin’s Wemo, too, and offers full compatibility with Alexa, HomeKit and Assistant for remote voice control. But it goes a step further with Alexa, acting not only as the connected home smart device, but also the command center, too.

The Switch+ is now available for pre-order from ecobee and select retail partners including, unsurprisingly, Amazon, in both the U.S. and Canada for a retail price of $99 U.S. or $119 Canadian. It should work with most standard light switches, although not 2-way switches where multiple switches control the same light or lights. In-store availability and shipping starts on March 26.

AI voice assistant developer Rokid raises $100M Series B extension to build its US presence

Rokid founder and CEO Mingming Zhu

Rokid, a Chinese startup that makes an AI voice assistant and smart devices, just raised a Series B extension round led by Temasek Holdings, with participation from Credit Suisse, IDG Capital and CDIB Capital. The size of the round was not released, but a source familiar with the deal told TechCrunch that it is $100 million.

The company’s previous funding was its Series B round, which was announced in November 2016. Founder and chief executive officer Mingming Zhu says Rokid raised a Series B+ instead of a C round because the company, which is based in Hangzhou, China with research centers in Beijing and San Francisco that develop its proprietary natural language processing, image processing, face recognition and robotics technology, is still in its early stages. Rokid wants to focus on gathering more resources and bringing in strategic investors like Temasek Holdings before moving on to a Series C. An investment holding company owned by the Singaporean government, Temasek Holdings counts artificial intelligence and robotics among its main investment areas and its other portfolio companies include Magic Leap.

Rokid Glass

The company’s product lineup already includes smart speakers called Rokid Pebble and Alien, which are currently sold in China. During CES, Rokid debuted its newest offering, Rokid Glass, augmented glasses created specifically for consumer use, as well as an open-source platform, called the Rokid Full Stack Open Platform. Created in partnership with Alibaba, the platform gives third-party hardware developers who use Rokid’s voice assistant access to free resources, including software blueprints and content for IoT devices. Rokid hopes that both will help build its name recognition and presence in the United States.

Reynold Wu, Rokid’s director of product management, describes the Full Stack Open Platform as a turnkey solution that not only gives developers access to Rokid’s AI technology, but also hardware solutions and services. Released with Aliyun, Alibaba’s cloud computing business, the cloud platform opened to third-party developers in China earlier this year, and will launch in the U.S. soon.

Rokid wants the platform to serve as a bridge between the two countries by giving U.S. developers an easy way to enter the Chinese market and also encouraging the development of more content for devices running Rokid’s technology, which founder and chief executive officer Mingming Zhu says is vital to attracting consumers.

“AI products are born to be global, not just for local market,” explains Zhu. “The only issue for Rokid is that we’re not ready for the U.S. market because the most important thing is content and we are not ready if there is only local content or services.”

The Pebble and Alien will be up against Google Home and Amazon Echo, which have become almost synonymous with “smart speaker” in the minds of many consumers, while Rokid Glass will inevitably be compared to Google Glass. The success of the Pebble and Alien hinge not only on how well users think Rokid’s voice assistant compares to Google Assistant and Amazon Alexa, but also the library of content and apps that the startup is able to build for its smart speakers.

While Google Glass flopped among consumers, but saw more success as an enterprise device. Rokid hopes its smart glasses, which run on its proprietary AI voice and imaging algorithms, will be able to succeed where Google Glass wasn’t because it was designed specifically for consumer applications. Early reviews from CES say the Rokid Glass is promising and praised features like face recognition, but said it still needs work to become more responsive. Once it goes on sale, the Rokid Glass will compete with smart glasses from Vuxiz, Sony and Epson. Its price hasn’t been revealed yet, but Zhu says it will be sold at a consumer-friendlier price point than its competitors (many augmented reality smart glasses from Rokid’s rivals are currently priced in the range of $600 to $1,500).

“I think we are the only product that is really consumer-centric in not only design and weight, but also energy use,” says Wu. “A lot of players design for the enterprise market first and then try consumer opportunities, but we have developed consumer products over the past three years. All of them have entered the market successfully and we have users because of that, so we have confidence in our consumer products.”

Featured Image: Rokid

Sony reboots Aibo with AI and extra kawaii

The rumors had it right: Sony is rebooting its robot dog, Aibo, announcing a new four-legged companion AI-powered bot incoming with the same brand name but more rounded corners and visible facial features for extra kawaii, including a pair of expressive, puppy-dog eyes.

Deep learning tech, fish-eye cameras and a series of other embedded sensors enable Aibo to detect and analyze sounds and images so that it can learn and respond to its environment and interact with its owner so it appears less, well, robotic.

Sony claims Aibo’s adaptive behavior includes being able to actively seek out its owners; detect words of praise; smiles; head and back scratches; petting, and more.

Thanks to the embedded cameras you can also instruct Aibo to take a photo for you — should you want a dog’s eye view of yourself/your home life.

“Aibo’s AI learns from interactions with its owners and develops a unique personality over time,” it writes. “Further, with its owners’ permission, aibo can collect data from these interactions, then connect to the cloud and access the knowledge accumulated from interactions between different owners and their aibo to become even more clever.”

So, basically, if you’re comfortable installing a roaming camera, microphone and location sensing device in your home that’s designed to watch your every action, listen to what you’re saying and upload all this intel to the cloud to yield new programming instructions so the device changes how it acts around you in a pantomime of intelligence — this could be the artificial pet for you.

Alternatively, you could make like Sebastian Thrun and become the proud owner of a real life puppy.

But if Aibo’s hairless lines and plastic smells are melting your heart, then prepare to shell out some serious yen: The Aibo robot itself is priced at 198,000 JPY (~$1,735) but you also need a subscription plan to connect to the cloud service that powers Aibo’s AI.

A basic three-year subscription plan costs 2,980 JPY ($26) per month (or ~$790 if you pay up-front for the full three years).

It also looks like Sony is limiting Aibo’s release to its home market — at least for now. It’s accepting pre-orders for Aibo from today, via its website, but won’t be shipping until January according to Reuters.

Owners of the pet bot can also access an Aibo store via the companion app to shell out yet more money to buy additional tricks. Add on hardware accessories are also on the slate, such as a toy bone with a “tentative” price of ~$26.

Sony is also offering a further optional support care plan which gives Aibo owners a 50 per cent discount on repair fees in the event of malfunction or damage, and 50 per cent on “checkups and inspections”, for ~$15 per month. (Presumably that’s if you want to pretend you’re taking your pretend pet to the vet.)

Aibo’s battery is good for two hours of activity before it’ll need a recharge on its charging mat (taking three hours to be fully juiced).

Under it’s white plastic hood, the bot is powered by a 64bit Quad-Core CPU and is packing LTE and Wi-Fi:IEEE 802.11b/g/n radios for connectivity.

On the moveable parts front, the robopup has 22 degrees of freedom. While its eyes are comprised of two OLED screens, hence enabling new Aibo to have a range of puppy dog looks.

The full list of embedded sensors inside the bot are:

2 Cameras (Front camera, SLAM camera)
4 Microphones
ToF sensor
2 PSD sensors
Pressure sensitive/capacitive type touch sensor (Back sensor)
Capacitive type touch sensor (Head sensor, jaw sensor)
6 axis detection system(3 axis gyro/3 axis acceleration)×2 (Head, Torso)
Motion sensor
Light sensor
4 Paw pads

The original Aibo went on sale in 1999, with Sony going on to ship around 150,000 of the robots before production was ceased in 2006.

Speaking at a launch event for its rebooted Aibo robopup today, Sony CEO Kazuo Hirai said: “It was a difficult decision to stop the project in 2006, but we continued development in AI and robotics.

“I asked our engineers a year and a half ago to develop Aibo because I strongly believe robots capable of building loving relationships with people help realize Sony’s mission.”

Another AI chatbot shown spouting offensive views

Yandex, Russia’s homegrown Google rival — which offers a suite of similar products in its target non-U.S. markets, from search to webmail to maps — outed another equivalent offering earlier this month: An AI assistant.

The Russian-speaking “intelligent assistant” — named Alice — utilizes machine learning and voice recognition tech to respond voice queries like asking it for a nearby restaurant recommendation. So far so standard voice assistant fare.

But it also has what Yandex dubbed “a neural network based ‘chit-chat’ engine” to allow Alice to have “free-flowing conversations about anything”. So yes, you can see where this story is going.

The company claimed this feature is “unique”, and that users find it “surprisingly delightful and different from other major voice assistants”.

Well, safe to say, Alice’s AI chit-chat feature is not uniquely proofed against controversy — and apparently quickly went off the rails, a la Microsoft’s Tay AI bot last year. (Which, after being let loose on Twitter quickly learnt — thanks to the help of hate-loving Twitter trolls — to parrot racist and sexist views.)

Screengrabs of some of the unsavory opinions that Yandex’s Alice AI has apparently been exposing can be seen here, via Facebook user Darya Chermoshanskaya. They are said to include pro-Stalin views; support for wife-beating, child abuse and suicide, to name a few.

On the Facebook post, Chermoshanskaya writes that standard words on controversial topics can trigger a content lock — whereby Alice says she does not know how to talk about that topic yet.

However switching to synonyms apparently circumvents the built in safety measure, and the AI appears to change tone and “willingly” participate in conversations — and can thus appear to be advocating for Stalin’s regime of terror, spousal abuse, and so on.

In a statement regarding its Alice AI being shown espousing violent views, Yandex told us: “We tested and filtered Alice’s responses for many months before releasing it to the public. We take the responsibility very seriously to train our assistant to be supportive and polite and to handle sensitive subjects, but this is an ongoing task and in a few cases among it’s widespread use, Alice has offended users.  We apologize to our users for any offensive responses and in the case you referenced in your email, we did so directly on Facebook where a user identified an issue.”

“We review all feedback and make necessary changes to Alice so any flagged content for inappropriate responses won’t appear again,” it added. “We are committed to constant improvement with all our products and services to provide a high-quality user experience.  We will continue to regularly monitor social and traditional media and will correct our assistant’s behavior when necessary.”

Asked how many users the AI has at this stage Yandex declined to specify — saying only that Alice has “millions” of daily interactions with users.

As ever, when it comes to AI, the ‘intelligence’ on display is derived from training data fed into learning models — and thus an AI will absorb and reflect existing biases and prejudices in the data-set.

Ergo, the resulting product may not prove quite so “delightful” as you’d hoped.

Featured Image: Thomas/Flickr UNDER A CC BY-ND 2.0 LICENSE

Alexa, don’t talk to strangers!

This story began when I was giving a TV interview about a recent incident where police in the U.S. obtained a search warrant for an Amazon Echo owned by a murder suspect in the hopes that it would help convict him — or at least shed light on the case (the case is not relevant to our story, but here’s an article about it, anyway).

The Amazon Echo, a smart helper running Alexa (I know I’m going to make some of my friends at Amazon angry by writing it, but Alexa is Amazon’s Siri), is a very cool device that can do anything from playing music to ordering things from Amazon. To activate it, you just have to say “Alexa” and then ask for whatever it is that you want: “Alexa, please wake me up tomorrow at 5 am.

The day after my interview, several people told me that every time I said “Alexa” on TV (and, as you can imagine, I said it several times during that interview), their device turned on and entered a waiting-for-your-commend mode.

It took me a few seconds to grasp the meaning of what they had just told me… and then I had one of those OMG moments. This is exactly the kind of thing I love to hear. Immediately I started planning what I would say to mess with Echo owners who were unlucky enough to be watching the next time I was on TV. Actually, I don’t even need them to watch — I just need them to leave the TV on. 

Finding the best way to mess with Echo owners

In the process of learning more about Alexa’s capabilities, I found an extensive list of commands that work with Alexa; I also consulted several Echo owners.

My initial idea was to make Echo play the scariest and noisiest heavy metal song I could find (“Alexa, please play Behemoth Slaves Shall Serve!”). Then I thought — maybe I could have Echo order something interesting from Amazon: “Alexa, please buy a long white wedding dress.”

This would definitely create some interesting situations.

Then I thought, maybe I could make Alexa run dubious internet searches that might cause the police to knock on Echo owners’ doors:

“Alexa, what do you do when your neighbor just bought a set of drums?”

“Alexa, where can I get ear plugs urgently?”

“Alexa, how many pills would kill a person?”

“Alexa, how do you hide a body?”

As I started getting more into the Echo command possibilities, I realized there is an interesting attack vector here — and it is bigger than messing with people over the TV.

Echo, personal data and much more

Many use Echo to add items to their to-do list, their schedule, their shopping list, etc. Advanced users also can install plug-ins that connect Alexa to other resources, such as their Google Drive and Gmail. As expected, Echo also can retrieve these items and read them out loud.

And because Echo has no user identification process, this actually means that anyone within talking range can get access to Echo users’ sensitive emails, documents and other personal data.

The world of IoT was destined to introduce problems of a new type. And here they are.

And it goes on. Alexa has a kit that allows it to control smart houses. There are alarm systems that you can turn on and off with voice commands and, in some cases, even unlock the front door.

The fact that Alexa doesn’t identify users means that for “her,” all us humans are the same. Actually, you don’t even need a real person to speak. In an experiment that we’ve made, we controlled Alexa by using a text to speech system, and it worked wonderfully.

Why Alexa?

Though other companies have introduced personal helpers, Alexa is unique in the sense that it is open and has many interfaces and plug-ins. Even Hyundai integrates Alexa in some of its cars, which means you can control its locks, climate, lights and horn, and even turn it on with voice commands.

The death of secure physical boundaries

But there is a much greater issue here. And it relates to the fact that placing an Echo device in a physically secured environment does not actually mean that threat actors cannot access it.

The world of IoT was destined to introduce problems of a new type. And here they are.

Until now, we have (generally) divided the world into two sections: secured areas and unsecured areas. And we had straightforward means to facilitate the separation between the two: walls, barbed-wire fences, gates, guards, cameras, alarm systems, etc.

But now, devices like Alexa force us to define a new type of physical perimeter — the “vocal” perimeter. This perimeter has different boarders that are much harder to define and measure because they are not definite. They are derived from the power of your sound source. For example, for a human with no additional voice amplification tools, the vocal perimeter might look like this:

But if you have a truck with huge speakers on it and a microphone, you might be able to drive down the street and operate all the Alexa devices around:

Vocal perimeter security solutions?

Wi-Fi networks suffer from a similar problem. Companies that wanted to implement Wi-Fi networks had to deal with the fact that they might be accessible from outside their physical office. But with Wi-Fi, we had a simple solution — encryption.

This is obviously not relevant to voice-operated devices. You cannot encrypt the conversation between the user and the device. Or can you? No, you can’t.

Nevertheless, an obvious solution to this problem is to integrate a biometric vocal identification system in Alexa. If done, Alexa could just ignore you if you are not an authorized user. But this is not an easy thing to accomplish. Even though the world of biometric vocal identification constantly innovates and improves, there are still challenges and vocal identification solutions are still struggling with a bad false positives versus false negative ratio.

And then there is Adobe’s VOCO. This amazing application can actually synthesize human voices (watch the demo video).

So there is nothing that can be done?

There is always an inherent trade-off between security and user experience. You can set up Alexa to only operate when you press a button, but that makes no sense.

Related Articles

We could implement a password request for any command, but this is stupid — you cannot say the password out loud; obviously it is not going to work:

Me: “Alexa, 23Dd%%2ew, what’s the time?

Alexa: “23Dd%%2ew is the wrong password

Me: “Arrrrr I meant capital W

Alexa: “ArrrrrImeantcapitalW is the wrong password

At least you can change Amazon’s startup word, but this will only partially mitigate the problem. (By the way, for the Trekkies among us, amazon has recently added the word “Computer” as a wake word. Yes. You’re welcome.)

Well, at least Echo doesn’t have a camera integrated into it. Yet.

Bottom line, Amazon and other vocally controlled device manufacturers will have to integrate vocal identification in the years to come. 

Additional thoughts…

Should we ban Hollywood from producing movies that have a character going by the name “Alexa”? What if Alexa the character is watching her boyfriend in trouble and in his last breath he yells “Alexa, call 911!”

Funny enough, after starting to work on this article, a TV anchor accidentally said “Alexa, order me a dollhouse.” And it did. For many people.

Call for action: If you are an entrepreneur, build a startup company with a product that solves this issue (vocal firewall?). Sell it to Amazon. Make billions.

Thanks to all Amazon Echo users who help in the process, and specifically to Uri Shamay.

Bonus links

If you are unconvinced that vocal identification is a complicated thing, you can watch this video. Or this one (warning: explicit language!).

And how can we end without XKCD: