Interesting announcements from Amazon at its AWS event this week: the company is rolling out a suite of artificial intelligence APIs for developers to plug their apps into. These tools are based on the AWS cloud (which a lot of your favorite apps and services already use) and they leverage the same AI and deep learning that has also powered Alexa, the software behind the Amazon Echo.
Here’s April Glaser, writing for Recode:
Drawing on the artificial intelligence that powers Amazon’s popular home assistant Alexa, the new tools will allow developers to build apps that have conversational interfaces, can turn text into speech and use computer vision that is capable of recognizing faces and objects.
Amazon’s latest push follows moves from Google and Microsoft, both of which have cloud computing platforms that already use artificial intelligence.
Google’s G Suite, for example, uses AI to power Smart Reply in Gmail, instant translation and smart scheduling functions in its calendar. Likewise, Microsoft recently announced it’s bringing artificial intelligence to its Office 365 service to add search within Word, provide productivity tracking and build maps from Excel with geographic data.
It’s increasingly starting to look like “AI as an SDK” will become a requirement for modern apps and services. Deep learning and AI aren’t limited to playing chess and recognizing cat videos anymore; developers are using this new kind of computing power for all kinds of features – see Plex, Spotify, and Todoist for two recent examples. I’ve also been hearing about iOS apps using Google’s Cloud Vision a lot more frequently over the past few months.
I think this trend will only accelerate as AI reshapes how software gets more and better work done for us. And I wonder if Apple is considering an expansion of their neural network APIs to match what others are doing – competition in this field is heating up quickly.