Right, so I don't think the smart phone delayed anything since the price point for the hardware is STILL not in range today.
But I think the vision has legs, just not yet. My guess is that we have now crossed the rubicon where voice driven in-home ubiquitous interfaces are here.
Because voice is omni-directional and distant, one can cover ones home (or a work/car environment) with an inexpensive device. The biggest issue is that even a tiny smartphone screen can provide 3-10x more info per second than voice response can provide.
My intuition is that *IF* google glass can be made comfortable/practical then in tandem with a nearby smart-phone for processing, one can build ever more intuitive ubiquitous interfaces.
Imagine if you tilt your head in a particular way and use a slight change in intonation in order to tell your assistant you are talking to them, then any convenient surface/wall or open space is used to provide visual and auditory feedback.
Ultimately it could be superior to direct interaction with another person... but the devil would be in the details. Today I think things would be lagging and very crufty feeling. Its just so hard to do AR well...
But what I do like about that kind of a solution is (like the smart phone) just one expensive device will cover the whole universe for you, no matter where you are or where you are looking.
p.s. I used to work at IBM research, we had a ubiquitous computing group there... they did cool stuff, but like Xerox, IBM has not really seen cash from it!