On Risk


You’re welcome. Discuss.

iPad and the Brain


The iPad promises to be a very big deal: not just because it’s the next big over-hyped thing from Apple, but because it fundamentally shifts the way that humans will interact with computing.

Let’s call this the “fourth turning” of the computing paradigm.

Calculators

Early “computers” were electro-mechanical, then electric, and then later all electronic. But the metaphor was constant: you pushed buttons to enter either values or operators, and you had to adhere to a fixed notation to obtain the desired results. This model was a “technology” in the truest sense of the word, replacing “how” a pre-existing task got done. It didn’t fundamentally change the user, it just made a hard task easier.

8-Bit Computers: Keyboards

The early days of computing were characterized by business machines (CP/M, DOS, and character-based paradigms) and by low-end “graphics and sound” computers like the Atari 800, Apple II, and Commodore 64.

The promise here was “productivity” and “fun,” offering someone a more orderly typewriting experience or the opportunity to touch the edges of the future with some games and online services. But the QWERTY keyboard (and its derivatives) date back to at least 1905. And the first typewriters were made by Remington, the arms manufacturer.

The keyboard input model enforces a verbal, semantic view of the world. The command line interface scared the hell out of so many people because they didn’t know what they might “say” to a computer, and they were often convinced they’d “mess it up.” During this era, computing was definitely still not a mainstream activity.

More of the population was older (relative to computing) and had no experience with the concepts.

The Mouse, GUI, and the Web

Since the introduction of the Macintosh, and later Windows, the metaphors of the mouse, GUI, and the web have become so pervasive we don’t even think about them anymore.

But the reality is that the mouse is a 1970’s implementation of a 1950’s idea, stolen by Apple for the Lisa from Xerox PARC. Windows is a copy of the Macintosh.

The graphical computing metaphor, combined with the web, has opened the power of the Internet to untold millions, but it’s not hard to argue that we’re all running around with Rube Goldberg-like contraptions, cobbled together from parts from 1905, 1950, and 1984 respectively. Even so, the mouse alone has probably done more to open up computing than anything else so far.

The mouse enforces certain modes of use. The mouse is an analog proxy for the movement of our hands. Most people are right handed, and the right hand is controlled by the left hemisphere of the brain, which science has long argued is responsible for logic and reason. While a good percentage of the population is left handed, the fact remains that our interactions with mice are dominated by one half of the brain. Imagine how different your driving is when you only use one hand.

While we obviously use two hands to interact with a keyboard, some cannot do that well, and it continues a semantic, verbal mode of interaction.

iPad

The iPad will offer the first significant paradigm shift since the introduction of the mouse. And let me be clear: it doesn’t matter whether hardcore geeks like it now, or think it lacks features, or agree with Apple’s App Store policies.

The iPad will open up new parts of the human brain.

By allowing a tactile experience, by allowing people to interact with the world using two hands, by promoting and enabling ubiquitous network connections, the iPad will extend the range and the reach of computing to places we haven’t yet conceived.

Seriously. The world around us is reflected by our interactions with it. We create based on what we can perceive, and we perceive what we can sense. The fact that you can use two hands with this thing and that it appears to be quick and responsive is a really big deal. It will light up whole new parts of the brain, especially the right hemisphere — potentially making our computing more artistic and visual.

Just as the mouse ushered in 25 years of a new computing paradigm, pushing computing technology out over a much larger portion of the market, the iPad marks the beginning of the next 25 years of computing.

And before you get worried about how people will type their papers and design houses and edit video without traditional “computers,” let me answer: no one knows. We’ll use whatever’s available until something better comes along.

But computing platforms are created and shaped by raw numbers and the iPad has every opportunity to reach people in numbers as-yet unimagined. That will have the effect of making traditional software seem obsolete nearly overnight.

When the Macintosh was released, it was widely derided as a “toy” by the “business computing” crowd. We see how well that turned out.

This time, expect a bright line shift: BIP and AIP (before iPad and after iPad). It’s the first time that an entirely new design has been brought to market, answering the question, “Knowing everything you know now, what would you design as the ultimate computer for people to use with the global network?”

It’s 2010, and we don’t need to be tied down to paradigms from 1950 or 1905. Everything is different now, and it’s time our tools evolved to match the potential of our brains and bodies.

How Google Will Make the Suburbs Obsolete

There’s been a lot of speculation about Google’s plans to deploy Gigabit fiberoptic Internet. Where will they deploy? What are the criteria? How many homes will they serve? Will they favor cities, or rural areas?

Your guess is as good as mine. But as a part of the global tech community and as someone who has spent a lot of time at Google and with people from Silicon Valley, these are my guesses about what they might do.

Cities Offer Higher Returns

Cities have the kind of density required to deliver a lower cost-per-home deployment. Less cable, a single point of negotiation and contact, and the ability to deploy using lateral construction from fiber conduits means lower overall costs.

Multi-family housing means more customers per square mile. Baltimore has a city-owned conduit system which can serve over 90% of the area of the city — without requiring the use of poles or negotiating with third party utilities.

Rural Areas Cannot Be Served Profitably

Telephone companies receive funds from the Universal Service Fund to subsidize service in areas that otherwise cannot be profitably served. Google is not subject to the regulatory framework (Communications Acts of 1934 and 1996) that would give it access to USF funds; in fact, it has every incentive to fight to avoid falling under such regulation.

Google is not a charity, it’s not being subsidized by the government, and it is not a monopoly. There is no special reason why Google should care about making services available in rural areas, and there is certainly no profit motive. Rural service requires fuel, vehicles, and people on the ground. Every part of this is expensive; it’s why it loses money and why it has to be subsidized by USF funds.

Google simply has no motive at all to serve rural areas. I’ll eat cat meat if Google selects a rural area for this trial. It just won’t happen.

Tech Is Opinionated

Google has opinions. In the tech world, people take a stand: Google and Apple both expressed strong opinions about how a smartphone ought to operate. Opinionated software is an emerging trend in software tools. Software designers bake their opinions into the tools they create. People who use those tools end up adopting those opinions; if they don’t, the tools become counterproductive, and they are better off using different tools.

There is every reason to believe that Google’s opinion is that the suburbs are obsolete, and that that opinion will inform their strategy for building out a fiber network. Here’s why Google likely believes the suburbs are obsolete:

  1. Suburbs rely on car culture, which consumes time; that’s time that people can’t spend on the Internet, making money for Google.
  2. Suburbs are not energy efficient, requiring lifestyles that generate more CO2 emissions. Google has said it wants to see greater energy efficiency in America.
  3. Google CEO Eric Schmidt has said he wants to see America close its innovation deficit. There’s nothing innovative about the design of the suburbs. It’s a tired model.
  4. Schmidt has supported Al Gore politically and in his efforts to combat global warming. Regardless of what you might think of Al Gore or global warming, we have a pretty good idea what Google thinks of the issue.
  5. Gigabit Fiber in cities could utterly revitalize them. We’ve been looking for ways to fix our cities for the last 50 years. The last renaissance was powered by large-scale economics; a new renaissance can be launched with large-scale communications investment.
  6. Google’s employees are young, idealistic, and believe in self-powered transportation. It’s worth pointing out that the Google Fiber project lead, Minnie Ingersoll, is an avid cyclist.

The Suburbs Are Done

I’ve said it before. So have others. But I’m not promoting that they be subject to some kind of post-apocalyptic ghettoization, either, so calm down. No one’s threatening your commute or your backyard barbecue.

But what I am saying is that at some point we need to take a stand about where we’re going to invest in our future. About where we believe we can regain competitive advantage and efficiency.

I believe our only hope to do that is with smart, well-designed urban cores, connected with world-class communications infrastructure and fast, green, and efficient people-powered transportation. And I think Google believes that too. Bet on it.

Fiber Economics

With the release of the FCC’s National Broadband Plan, Google’s announced intention to build gigabit fiber-to-the-home networks, and Verizon’s indications that they are not likely to be expanding their FIOS service to new areas, it’s a good time to review where we really stand with fiber.

The Real Reasons You Don’t Have Fiber

What are the real economics of broadband infrastructure? It’s not so simple as market opportunity, investment, and subscribers; Verizon and Comcast have different regulatory histories and see the world differently apotheke-zag.de. Google, as a potential new entrant, has completely different motivations.

Let’s take a look at the regulatory background, and then get a sense of what’s really motivating Verizon, Comcast, and Google.

Regulatory Background

We have gradually come to think of Verizon and Comcast as equals: big, for-profit telecom companies — competitors for TV, Internet, and telephone service. But they got to their current positions through very different routes. Here’s a brief (and rather incomplete) history.

In 1984, the former AT&T was busted up into seven Baby Bells: Ameritech, Bell Atlantic, BellSouth, NYNEX, Pacific Telesis, Southwestern Bell, and US West. Terminator-like, these companies have been spending the last 26 years reconstituting themselves, merging into very large firms. Bell Atlantic changed its name to Verizon in 2000 after acquiring GTE.

Telecommunications regulation in the United States has a long history and reflects theory originally applied to railroads and other public utilities. The idea was that communications was a public good and because the network had to be large and interoperable to be effective, it was best served by a natural monopoly. So, assets like public rights-of-way were made available for the monopoly to use, in exchange for an agreement to provide Universal Service, covering the entire population.

To keep the monopoly from charging unreasonable prices, regulators mandated that their services should be marketed at cost, plus a reasonable and sustaining profit margin. This means that there is no incentive for them to keep costs down; in fact, the higher their costs, the more raw dollars they make.

Verizon today operates under this kind of regulatory background, which was outlined initially in the Communications Act of 1934, and then amended by the Telecommunications Act of 1996 — which has subsequently been eroded and modified by case law and other FCC actions.

The FCC, under Bush-appointed Chairmen Michael Powell and Kevin Martin, tended towards the opinion that the best way to foster competition and innovation would be to empower a small number of well-capitalized firms and let them compete together in the marketplace.

Comcast, for its part, came together very differently. Cable TV franchises were primarily granted by local municipalities, starting in the 1950’s. Comcast acquired dozens of these small firms, each with their own regulatory agreements with cities and counties. By 2000 or so, this aggregation started to resemble the sort of “large firm” that the FCC thought could be an effective competitor to the telephone companies.

So that’s how we got here. Verizon is heir to the top-down, cost-based monopoly regulation subsidized by the Universal Service Fund, which requires that it provide telephone service even in rural areas. Comcast is the product of the roll-up of dozens (if not hundreds) of small cable TV firms. Now let’s take a look at their interests in the current landscape.

Verizon

Verizon, in many ways, is just the current-day incarnation of a big chunk of the original AT&T. It’s still the primary telephone infrastructure provider and the bulk of its physical wiring plant is copper. It operates the same switching facilities that AT&T did back in 1984. In many important ways, nothing has changed.

What about FIOS? Isn’t Verizon innovating there? Aren’t they making this investment to “make money?” It’s complicated.

Verizon made the decision to install FIOS primarily to block competition. The Telecom Act of 1996 required that telcos make copper wireline infrastructure for competitors to run alternative services. This is where alternative telcos like Cavalier, Covad, Adelphia, and many others came into the market. You’ll notice that almost all of those companies are now defunct or severely hamstrung.

This is in part because Verizon (and its peers) set out on a strategy to make a competitive business model all but impossible. FIOS was part of that strategy.

When Verizon installs FIOS, they almost always remove the copper wires that could otherwise have been used by competitors; and this has effectively shut them out.

Verizon has spent over $20 Billion to build out FIOS in its service area. Ostensibly this might look like “investment in innovation” to observers. But in fact, this spending was mostly done to block competitors and to destroy the pro-competition provisions of the Telecom Act of ’96.

It should thus come as no surprise that Verizon has recently announced that they are unlikely to expand their FIOS network further. This isn’t because they can’t get more new subscribers in new areas (like downtown Boston, which is still not served); rather, it’s because they have calculated that the costs of future expansion exceed the downside risk of lost profits caused by competitors in the areas that remain.

They have put FIOS in all the places where it was either easy to do so or where the competition was too strong to ignore. Now that the competitors are mostly defunct or severely weakened, the threat is just not there to justify expansion.

Like feudal warlords, they invested just enough in FIOS to block out competitors, rejigger the regulations, and maintain a status quo of mediocrity. And we’re supposed to think this is innovation?

Comcast

Comcast has different problems. Because all of their regulatory agreements are negotiated with individual municipalities, it’s more difficult for them to make investments across their entire footprint. This is why Comcast often rolls out new products and services in trial communities, and then rolls them out to new areas one at a time.

Comcast does have a very large television service footprint, and their acquisition of NBC and other content providers over the years, like HTS, is an attempt to establish themselves as a vertically integrated entertainment provider. They control the entire stack, from the physical cable, to the viewer, to the content itself. This means that they are protected from threats from content providers who might try to command high rates for popular content. Disney (who controls ESPN and ABC) often finds itself in battles about rates with cable providers. Acquiring NBC/Universal means one less potential threat of rate hikes for Comcast, and higher overall profits.

But Comcast’s physical plant is dominated by aging coaxial cable infrastructure. While local head-ends are fed by fiberoptic backbones, local distribution to the home is through co-axial cable, which can degrade in performance when it rains, is subject to lightning damage, and can only go so fast. Fiber-to-the-home is vastly superior, but it would cost Comcast billions to upgrade its plant. In the absence of competitive pressure (such as that which was faced by Verizon), they have no incentive to do so. Instead they are happy to push their existing plant as hard as it can go, using standards such as DOCSIS 3, and invest in fiber-to-the-home infrastructure only as necessary or convenient.

Google

Google has recently announced that they would like to spawn innovation, and potentially build out gigabit fiberoptic infrastructure in one or more communities in the United States. I helped organize Baltimore’s municipal response to Google’s Request For Information for this project.

Google’s proposing something very different from what Verizon and Comcast offer: an open-access network, over which new entrants could provide Internet or other services. This is exactly the paradigm that Verizon has fought to destroy with FIOS.

Comcast has also fought open access repeatedly; before Verizon settled on FIOS as its primary anti-competitive strategy, Verizon tried to force cable companies to become subject to the same kind of infrastructure-sharing to which it was subject under the Telecom Act of 1996.

And Comcast fought this effort mightily; in 2002, I testified before the Maryland House of Delegates in support of a bill that would force Comcast to open its network, and Comcast’s lobbyists managed to defeat it.

Also in 2002, working alongside Verizon-supplied lobbyists, I testified before the FCC with TCP/IP co-inventor Vint Cerf (now a VP at Google) arguing that cable companies should be forced to provide “open access” to their networks because it would promote competition and entrepreneurship. At that hearing, FCC Commissioner Robert Pepper made it clear that the FCC believed that Verizon and Comcast could provide all the competition we would ever need. We see how that’s turned out.

To date, there has not been any significant open access network deployment in the United States. And with the decline of competitive telco-based services, telecom innovation has now stalled entirely. It’s time for something new.

Net Neutrality

Google has a different potential problem on its radar. In the US, Comcast and Verizon control access to a large percentage of its customers. Currently, the Internet operates under a doctrine of “Net Neutrality,” which is to say that customers and Google all just pay for access to the network, and each can communicate freely with anyone else on the network.

Various telecom executives (most notably former AT&T CEO Ed Whitacre, now CEO of GM) have argued that companies like Google should no longer get free access to their customers. Folks like Whitacre believe that the natural role of a company like Verizon or Comcast is to act as a toll-gate, charging both content providers (Google) as well as customers for access to each other.

As you might imagine, Google heartily opposes this idea: it could dramatically increase their costs and would destroy the “level playing field” which has dominated the Internet from the beginning. Startups could be stifled because they might need to negotiate an agreement with broadband providers to get access to customers. This is a war that Verizon and Comcast appear ready to start, and people like News Corp’s Rupert Murdoch are fanning the flames.

Google’s Fiber Plan

Google’s announcement that it intends to build ultra-fast open access fiber networks is its declaration of war against the threatened end of net-neutrality. Further, this is a productive use of Google’s vast stockpile of cash; it’s something tangible it can do to ensure its market position.

And it’s a move that’s ideologically compatible with its mission. Google genuinely believes that the expansion of a fast, net-neutral Internet has positive effects on society, and it’s also good for its bottom-line. More people online means more ad-views which means more advertising, and more dollars for Google. There’s no downside for them; it’s an expensive proposition to be sure, but it’s less expensive than paying for access to customers in a world without net-neutrality.

By promoting itself as a good citizen, wrapped in the banner of open-access, innovation, and net-neutrality, ideologically-sympathetic regulators such as the FCC’s new Obama-appointed Chairman Julius Genachowski are likely to view Google’s approach favorably. This would allow Google to establish a vertically-integrated, long-term market position which would be hard for Verizon or Comcast to disrupt.

And the kicker? The open-access network Google’s proposing really would promote innovation and entrepreneurship. The United States is ranked 15th in broadband penetration worldwide today. This is a chance to change that.

Don’t believe that Verizon or Comcast will make these investments unless forced to do so. And while Google may also feel it has no choice but to build its own network, Google at least has a vision that goes far beyond just sustaining a mediocre status-quo; they truly believe in the level playing field that has given birth to so much innovation.

It’s time for America’s bandwidth to finally match our ambitions and our talent. Let’s go Google!