At betaworks, we are very excited about the opportunities conversational interfaces offer for building new kinds of products, services, distribution channels and user experiences. We have been investing and building in this space for quite some time now. Given our curiosity and experience in this space, we introduced Botcamp back in April this year and we were blown away by the quality and quantity of applications in this space. We received over 400 applications from early stage bot companies from all over the world. These developers are building conversational services in various verticals from enterprise, finance, and commerce to travel, analytics and news on different platforms. Some products were thin bots, some had taken a human-assisted approach and some were a result of over a year’s work on language analysis and machine learning in specific verticals.
Let us now talk about the one platform that is absent from the graph below because it wasn’t launched until after our application deadline.
Apple recently opened up its Messages, Siri and Maps products and launched the beta version of its iOS 10 SDK. Apple also introduced rich notifications and comprehensive widgets in iOS 10. It might seem that Apple hasn’t joined the bot world that has been dominated by the likes of Slack and Facebook. But, when we look at this landscape closely, it is fascinating how Apple has indeed joined the race but in a manner most of us did not expect.
Bots can be thought of as simple, contextual ways of accessing services without changing the context of where you already were and what you were doing. Bots reduce and often eliminate the need for downloading and using traditional mobile apps.
We can broadly put these bots in three categories —
- Textual or Conversational bots — These are conversational products living within chat interfaces. Messaging platforms like Slack, Facebook Messenger, Telegram, and Kik allow third-party developers to build services within their chat threads. These bots are navigated via commands and are the most common type of bots developers are pursuing in 2016. Some of these bots do have visual components. For instance, Slack now has buttons and Facebook Messenger now has quick replies, media support and persistent menu. But, both Slack and Facebook have taken a predominantly conversational approach for the most part.
- Visual bots — These are visual products living within chat interfaces. WeChat, contrary to common understanding, has web views within its chat interface. Google, on the other hand, has enabled streamable, modular experiences called Instant Apps. These “bots” are navigated via most of the visual elements that traditional mobile apps have. The focus is still on accessibility and convenience. They don’t require a user to download a traditional app but enable them to enjoy most benefits of having an app. From a user perspective, using these bots feels like using apps.
- Audio bots — These are voice-based products. These bots include Apple’s Siri, Microsoft’s Cortana, Google’s Now, Viv and Amazon’s Alexa.
Apple, with its updates in iOS 10, took an interesting approach of having visual bots and strengthening its audio bot in its ecosystem.
In the context of the way I defined bots and the categorization of bots, we could argue that the new notifications, widgets and iMessage apps in iOS 10 are Apple’s visual bots. We now also have card-like interactions in Maps and Siri that could also be treated like visual bots.
Let’s unbundle each of these “Apple bots” —
Up until now, the notifications on our iPhones offered limited context, information, and actions.
For instance, Messages app currently has quick replies. When I get a new text message while I am catching up on emails or watching a video on Youtube on my iPhone, I have two options: 1. I could read the latest message, pull down the push notification and respond 2. I could click on that notification to pull up the chat thread in my Messages app. The former offers limited context and the latter involves complete change of context.
In iOS 10, we can 3D Touch a notification to open up a popover window with greater context. After acting on those notifications, we can seamlessly go back to what we were doing without changing context or app. We can also 3D Touch on our lockscreen.
A lot of textual chatbots these days are extensions of installed apps on our phones. Such chatbots are just an easy way of accessing a mobile app via frameworks and API calls.
Similarly, rich notifications in iOS 10 can be treated like app extensions. They live in the notifications layer, increase engagement and help users in unique ways without expecting them to pull up an app. Imagine being able to dip into a group conversation on a messaging app, get the context of prior messages and respond without leaving Instagram.
We often hear that the messaging inbox is the new homescreen and that different chat threads are the new “apps” for infrequent use cases.
With iOS 10, Apple is launching iMessage apps. These apps live within the Messages app. They have their own store which is different from the App Store that we use for discovering and downloading apps. Once you install iMessage apps from the store, they reside in what Apple calls an App Drawer within the Messages app.
For the purpose of making things simpler to understand, I categorize iMessage apps into 3 types.
- Sticker apps — These apps contain packs of buyable stickers. No programming is needed to create a basic sticker app. Stickers are already an effective way of communicating with people and a driver of business on Line and WeChat. Stickers offer a lot of new opportunities for fun, story-telling, branded content, monetization, and communication.
- App extensions — These apps are extensions of existing iOS apps. The main mobile app lives in the App Store and its iMessage extension lives in the iMessages Store. This type is the closest Apple equivalent to several bots that we are seeing grow on Facebook and Slack. We can imagine ride-sharing, ticket-booking, food-delivery, and utility apps adding such app extensions very soon.
- iMessage-only apps. These are iMessage-first apps. They don’t live in the App Store. These are downloadable only in the iMessages store. While Telegram, Sack, and other messaging platforms have native chatbots that live in chat threads or channels, Apple now has what could be referred to as “micro-apps” in chat threads. There is a lot of potential for native experiences for the developers to build on the iMessages platform.
The WWDC presentation of two friends placing an order on DoorDash over iMessage introduces a notion of shared context. Participants in a chat thread can get something done collaboratively via exchange of content and state. Such real-time exchange could enable games, collaborative tools, and other services that require active engagement of parties in a chat thread.
Apple has taken widgets from the notification center and deeply integrated them within the operating system. They are accessible on the lockscreen, to the left of the homescreen, and also via 3D touch on an app’s icon. These widgets would convert our lockscreen from a place for passive consumption of notifications to homescreen-like center for actually carrying out tasks. We shall see whether the messaging inbox or the lockscreen ends up becoming the new homescreen.
The widget layer could be thought of as an additional runtime for new interactions on top of the notifications and iMessages layer. They reduce the need for opening up apps by providing more context, interactions, and accessibility to content.
With SiriKit, developers can enable users to access content and services using voice. It offers third-party integrations with a limited set of verticals —
- Messaging apps
- Ride Sharing
- Climate & Audio (This is designed for CarPlay apps)
These services are extensions of mobile apps that you already have installed on your phone. Apple doesn’t support “Siri-first” or “Siri-only” services yet.
Siri was opened up in a way that involves the practice of installing an app on our device. As a consumer, to add functionality to Siri, we must install an iOS app. Developers need to create special additions to their iOS applications that enable Siri to interact with their apps. Siri extensions cannot be built or added outside the context of an iOS app. At least for now.
Changes to Siri raise a bunch of interesting questions —
- When and whether Apple will integrate the app extensions in Siri to standalone apps on the iMessages platform? Deep integration would make for cohesive user experience across the operating system and tighten Apple’s ecosystem.
- Will there be a Siri App store the way Amazon has an Alexa skills store? Having a store for “Siri-first” apps might be a huge step towards a hands-free computing future.
- Will we see voice or visual bots go mainstream before textual bots do? It is hard to neglect the fact that a lot of people are visual learners and would enjoy the convenience of hands-free computing.
- Will Siri become the main bot that communicates with a plethora of other smaller bots? Imagine asking Siri to do something really complex like “Tell me how widely the Medium piece that John Borthwick wrote a couple years ago on Betaworks changing its logo got distributed on social networks”. Siri could then break this user query down into small parts and assign each of those parts to relevant domain-specific bots.
- Will Siri across devices have the same access to user information and be capable of same functionalities? Farhad Manjoo, at The New York Times, did a great job at explaining this issue —
Another problem is that Siri is still hopelessly tied to each Apple device. Siri on your iPhone doesn’t really know anything about Siri on your Mac or Apple TV. On each device, Siri has different capabilities: On your iPhone it can call an Uber, if you have the Uber app installed, but Siri on your Mac can’t. Siri on your Apple TV can search YouTube for clips of Stephen Curry, but Siri on your iPhone can’t.
Every change we have discussed so far is limited to either a communication channel or the digital context of how we interact with app and services. Apple, with iOS 10, has opened up its Maps product for the first time. This enables developers to create new experiences that involve users’ geo-context.
Developers can now build extensions for certain type of apps like restaurant bookings and ride-sharing for Apple Maps. The Maps extensions will work similarly to intents for Siri and will allow developers to build apps with map functionality.
While Apple allows developers to create a limited set of interactions in its Maps app, it isn’t hard to imagine what could potentially be done in the future with greater flexibility and access. Imagine getting recommendations for contextual reads within the Maps app based off of where exactly you are standing in a foreign country on your vacation.
Amongst all these changes, we are most excited about what could be built on the iMessages platform. The standalone iMessages app offers the conversational context and required flexibility to build new native experiences. It is important to note how Apple is moving towards enabling developers to build visual experiences in a conversational context the way WeChat did with in-app web views. Other dominant messaging platforms are focusing on conversational experiences in a chat-thread context.
While it is too soon to comment on which methodology will be more user-friendly and effective, we are excited about both approaches. Each approach has its own set of challenges and opportunities. Both approaches help people do a variety of things without frequently changing context or downloading new apps.
It is fascinating how two tech titans (ie. Google and Apple) that have been competing on mobile apps and operating systems have taken a similar path (ie. visual bots) in their respective styles. Apple has taken a distributed approach of having visual bots integrated within its entire operating system while Google has redefined the very concept of mobile apps by changing their status from “downloadable” to “streamable”. As always, Apple is focusing on simplicity and user experience while Google is focusing on accessibility and developer flexibility. It is equally interesting how messaging platform rivals have also taken a similar, conversational approach to the bot paradigm. Both sets of companies seem to be playing to their strengths.
Thinking of bots as textual mediums narrows down the scope of what bots mean, what they will be capable of and how they will be used. Thinking of a bot as a new kind of user experience and delivery model will enable us to be as creative as possible in engineering bot services that people would love to use rather than coming up with solutions that developers and platform companies wish people wanted to use for their own interests. Attributing the rise of bots to declining app downloads and worsening app discovery issues seems to be the wrong way of looking at this trend. When bots truly go mainstream, we will most likely face the issues of discovery, competition, monetization, quality control, privacy, security and platform controls if we do not approach this new paradigm in a fresh manner. It will largely be the responsibility of the underlying platforms companies to strike a balance between their own interests, user experience and developer growth.
In the the weeks leading up to Botcamp’s application deadline, there was an expectation amongst the developers community that Apple will probably open up its Messages platform. But, it wasn’t as easy to predict what approach they would eventually take. It is not surprising that Botcamp applicants built products for every platform other than that of Apple. However, we are now very curious about what sort of experiences could be built using the iOS 10 SDK. Now that developers are able to build on this new platform, it opens up a whole new world of opportunities. As builders and investors, we are always interested in learning about new products. If you are working on anything you think we might find interesting, reach out to me at firstname.lastname@example.org
A version of this post appeared on Medium.
The audio problem: Learn how new cloud-based API solutions are solving imperfect, frustrating audio in video conferences. Access here