Google (NASDAQ:GOOG) has announced "Android Wear", a new extension of Android to power smart watches (it also realised some teaser renders of Motorola smart watches that are due for this summer). The Wear concept is that smart watches are remote touch displays for an Android smartphone. They will show the time, accept touch and voice input, display the Google Now feed and they will display all the notifications that apps on your phone produce.
Developers have options (which will be enhanced in future) to customise how the notifications their phone apps produce behave on the watch. But they don't get native code at all - the developer isn't running code on the watch, really. The device is really an extension of the phone's Android OS itself, not an extension of your app.
In effect, the watch is a device for using Google Now and cards that apps on the phone send to it.
Now contrast this with the rumours of a new Apple (NASDAQ:AAPL) "Healthbook" app. I hate speculating upon Apple rumours, because they could come true next week, next year or never, but they provoke an interesting idea.
This would be one answer to why Apple's recent hires of "wearables experts" sound a bit like a team for a hospital device rather than a watch, measuring various, quite technical things - because Apple plans to enable such devices, not try to pack every single one into its own device. That is, the straightforward sensors should live in the phone (like the pedometer that's already in the iPhone 5S), and the complex and demanding ones should be enabled by an Apple platform, not become part of an Apple device.
Today, you can manage a bunch of heath sensors with a bunch of apps, but that seems less... obvious, to use an "Ivism". If I have a wearable sleep sensor, a pedometer in my phone and a Wi-Fi scale (without even getting into glucose metres and more specialised things), should that be three apps that I install separately and open separately? If I buy a small computer I wear on my wrist, should it run apps (especially given that with the current state of technology, it'll need to use your phone to go online anyway)? If you have multiple devices, where should the code live, and how do you shape the user flow based on what makes sense rather than on where you put the code? Does a sensor need a screen? Does a screen or a sensor need to be smart? Is the right UI something totally custom that's installed from a store, or something more standardised?
This question of where the code lives also, of course, applies to TVs and to cars as much as to wearables. With Apple TV and Chromecast and CarPlay, Apple and Google are saying that though everything is becoming a computer, actually the "smart" part should be concentrated in the smartphone or tablet - something that's easy to update, that's replaced every couple of years, and that has a rich touch interface, and everything else should be a dumb sensor, or dumb glass or both. And so, the apps should only be in one place, and whether it should be an "app" in a strictly technical sense is also up for debate.
During the "apps versus HTML" argument of a year to two ago, someone said that the issue is not what coding language you use, but how you get an icon onto the user's home screen, and whether indeed they want your icon on their home screen. The conversation more or less crystallised around the position that apps are for the head of the tail and the web is for the rest. But Android Wear is not the web or an app. Neither is Google Now, and nor is the Healthbook I just described.
Now, suppose you hesitate outside a restaurant and look at your phone, and iBeacon has already activated a Yelp review card on your phone or watch, or Google Now has put a scraped review up, or Facebook (NASDAQ:FB) tells you 10 of your friends liked it? Is that the web? Or apps? How do you do SEO for that? What's the acquisition channel? Some of that might be HTML, but you'll never see a URL.
It seems to me that the key question this year is that now that the platform war is over, and Apple and Google won, what happens on top of those platforms? How do Apple and Google, but also a bunch of other companies, drive interaction models forward? I've said quite often that on mobile, the internet is in a pre-Pagerank phase, lacking the "one good" discovery mechanism that the desktop web had, but it's also in a pre-Netscape phase, lacking one interaction model in the way that the web dominated the desktop internet for the last 20 years. Of course, that doesn't mean there'll be one, but right now, everything is wide open.
This thought, incidentally, is one of the things that prompted this article.
Disclosure: No positions