Оценить:
 Рейтинг: 0

Zucked: How Users Got Used and What We Can Do About It

Автор
Год написания книги
2019
<< 1 2 3 4 5 6 >>
На страницу:
4 из 6
Настройки чтения
Размер шрифта
Высота строк
Поля

But before that, in the fall of my senior year, I enrolled in Introduction to Music Theory, a brutal two-semester course for music majors. I was convinced that a basic knowledge of music theory would enable me to write better songs for my band. They randomly assigned me to one of a dozen sections, each with fifteen students, all taught by graduate students. The first class session was the best hour of classroom time I had ever experienced, so I told my roommate to switch from his section to mine. Apparently many others did the same thing, as forty people showed up the second day. That class was my favorite at Yale. The grad student who taught the class, Ann Kosakowski, did not teach the second semester, but early in the new semester, I ran into her as she exited the gymnasium, across the street from my dorm. She was disappointed because she had narrowly lost a squash match in the fifth game to the chair of the music department, so I volunteered to play her the next day. We played squash three days in a row, and I did not win a single point. Not one. But it didn’t matter. I had never played squash and did not care about the score. Ann was amazing. I wanted to get to know her. I invited her on a date to see the Jerry Garcia Band right after Valentine’s Day. A PhD candidate in music theory, Ann asked, “What instrument does Mr. Garcia play?” thinking perhaps it might be the cello. Ann and I are about to celebrate the thirty-ninth anniversary of that first date.

Ann and I graduated together, she a very young PhD, me an old undergraduate. She received a coveted tenure-track position at Swarthmore College, outside of Philadelphia. I could not find a job in Philadelphia, so I enrolled at the Tuck School of Business at Dartmouth, in Hanover, New Hampshire. So began a twenty-one-year interstate commute.

My first job after business school was at T. Rowe Price, in Baltimore, Maryland. It was a lot closer to Philadelphia than Hanover, but still too far to commute every day. That’s when I got hit by two game-changing pieces of good luck: my start date and my coverage group. My career began on the first day of the bull market of 1982, and they asked me to analyze technology stocks. In those days, there were no tech-only funds. T. Rowe Price was the leader in the emerging growth category of mutual funds, which meant they focused on technology more than anyone. I might not be able to make the first personal organizer, I reasoned, but I would be able to invest in it when it came along.

In investing, they say that timing is everything. By assigning me to cover tech on the first day of an epic bull market, T. Rowe Price basically put me in a position where I had a tailwind for my entire career. I can’t be certain that every good thing in my career resulted from that starting condition, but I can’t rule it out either. It was a bull market, so most stocks were going up. In the early days, I just had to produce reports that gave the portfolio managers confidence in my judgment. I did not have a standard pedigree for an analyst, so I decided to see if I could adapt the job to leverage my strengths.

I became an analyst by training, a nerd who gets paid to understand the technology industry. When my career started, most analysts focused primarily on financial statements, but I changed the formula. I have been successful due to an ability to understand products, financial statements, and trends, as well as to judge people. I think of it as real-time anthropology, the study of how humans and technology evolve and interact. I spend most of my time trying to understand the present so I can imagine what might happen in the future. From any position on the chessboard, there are only a limited number of moves. If you understand that in advance and study the possibilities, you will be better prepared to make good choices each time something happens. Despite what people tell you, the technology world does not actually change that much. It follows relatively predictable patterns. Major waves of technology last at least a decade, so the important thing is to recognize when an old cycle is ending and when a new one is starting. As my partner John Powell likes to say, sometimes you can see which body is tied to the railroad tracks before you can see who is driving the train.

The personal computer business started to take off in 1985, and I noticed two things: everyone was my age, and they convened at least monthly in a different city for a conference or trade show. I persuaded my boss to let me join the caravan. Almost immediately I had a stroke of good luck. I was at a conference in Florida when I noticed two guys unloading guitars and amps from the back of a Ford Taurus. Since all guests at the hotel were part of the conference, I asked if there was a jam session I could join. There was. It turns out that the leaders of the PC industry didn’t go out to bars. They rented instruments and played music. When I got to my first jam session, I discovered I had an indispensable skill. Thanks to many years of gigs in bands and bars, I knew a couple hundred songs from beginning to end. No one else knew more than a handful. This really mattered because the other players included the CEO of a major software company, the head of R&D from Apple, and several other industry big shots. Microsoft cofounder Paul Allen played with us from time to time, but only on songs written by Jimi Hendrix. He could shred. Suddenly, I was part of the industry’s social fabric. It is hard to imagine this happening in any other industry, but I was carving my own path.

My next key innovation related to earnings models. Traditional analysts used spreadsheets to forecast earnings, but spreadsheets tend to smooth everything. In tech, where success is binary, hot products always beat the forecast, and products that are not hot always fall short. I didn’t need to worry about earnings models. I just needed to figure out which products were going to be hot. Forecasting products was not easy, but I did not need to be perfect. As with the two guys being chased by a bear, I only needed to do it better than the other guy.

I got my first chance to manage a portfolio in late 1985. I was asked to run the technology sector of one of the firm’s flagship funds; tech represented about 40 percent of the fund. It was the largest tech portfolio in the country at the time, so it was a big promotion and an amazing opportunity. I had been watching portfolio managers for three years, but that did not really prepare me. Portfolio management is a game played with real money. Everyone makes mistakes. What differentiates great portfolio managers is their ability to recognize mistakes early and correct them. Portfolio managers learn by trial and error, with lots of errors. The key is to have more money invested in your good ideas than your bad ones.

T. Rowe launched a pure-play Science & Technology Fund, managed by two of my peers, on September 30, 1987. Nineteen days later, the stock market crashed. Every mutual fund got crushed, and Science & Tech was down 31 percent after only a month in business. While the number was terrible, it was actually better than competitors because the portfolio managers had invested only half their capital when the market collapsed. In the middle of 1988, with the viability of the fund in doubt, the firm reassigned the two managers and asked me to take over. I agreed to do so on one condition: I would run the fund my way. I told my bosses that I intended to be aggressive.

Another piece of amazing luck hit me when T. Rowe Price decided to create a growth-stage venture fund. I was already paying attention to private companies, because in those days, the competition in tech came from startups, not established companies. Over the next few years, I led three key growth-stage venture investments: Electronic Arts, Sybase, and Radius. The lead venture investor in all three companies was Kleiner Perkins Caufield & Byers, one of the leading venture capital firms in Silicon Valley. All three went public relatively quickly, making me popular both at T. Rowe Price and Kleiner Perkins. My primary contact at Kleiner Perkins was a young venture capitalist named John Doerr, whose biggest successes to that point had been Sun Microsystems, Compaq Computer, and Lotus Development. Later, John would be the lead investor in Netscape, Amazon, and Google.

My strategy with the Science & Technology Fund was to focus entirely on emerging companies in the personal computer, semiconductor, and database software industries. I ignored all the established companies, a decision that gave the fund a gigantic advantage. From its launch through the middle of 1991, a period that included the 1987 crash and a second mini-crash in the summer of 1990, the fund achieved a 17 percent per annum return, against 9 percent for the S&P 500 and 6 percent for the technology index. That was when I left T. Rowe Price with John Powell to launch Integral Capital Partners, the first institutional fund to combine public market investments with growth-stage venture capital. We created the fund in partnership with Kleiner Perkins—with John Doerr as our venture capitalist—and Morgan Stanley. Our investors were the people who know us best, the founders and executives of the leading tech companies of that era.

Integral had a charmed run. Being inside the offices of Kleiner Perkins during the nineties meant we were at ground zero for the internet revolution. I was there the day that Marc Andreessen made his presentation for the company that became Netscape, when Jeff Bezos did the same for Amazon, and when Larry Page and Sergey Brin pitched Google. I did not imagine then how big the internet would become, but it did not take long to grasp its transformational nature. The internet would democratize access to information, with benefits to all. Idealism ruled. In 1997, Martha Stewart came in with her home-decorating business, which, thanks to an investment by Kleiner Perkins, soon went public as an internet stock, which seemed insane to me. I was convinced that a mania had begun for dot-coms, embodied in the Pets.com sock puppet and the slapping of a little “e” on the front of a company’s name or a “.com” at the end. I knew that when the bubble burst, there would be a crash that would kill Integral if we did not do something radical.

I took my concerns to our other partner, Morgan Stanley, and they gave me some money to figure out the Next Big Thing in tech investing, a fund that could survive a bear market. It took two years, but Integral launched Silver Lake Partners, the first private equity fund focused on technology. Our investors shared our concerns and committed one billion dollars to the new fund.

Silver Lake planned to invest in mature technology companies. Once a tech company matured in those days, it became vulnerable to competition from startups. Mature companies tend to focus on the needs of their existing customers, which often blinds them to new business opportunities or new technologies. In addition, as growth slows, so too does the opportunity for employees to benefit from stock options, which startups exploit to recruit the best and brightest from established companies. My vision for Silver Lake was to reenergize mature companies by recapitalizing them to enable investment in new opportunities, while also replicating the stock compensation opportunities of a startup. The first Silver Lake fund had extraordinary results, thanks to three investments: Seagate Technology, Datek, and Gartner Group.

During the Silver Lake years, I got a call from the business manager of the Grateful Dead, asking for help. The band’s leader, Jerry Garcia, had died a few years before, leaving the band with no tour to support a staff of roughly sixty people. Luckily, one of the band’s roadies had created a website and sold merchandise directly to fans. The site had become a huge success, and by the time I showed up, it was generating almost as much profit as the band had made in its touring days. Unfortunately, the technology was out of date, but there was an opportunity to upgrade the site, federate it to other bands, and prosper as never before. One of the bands that showed an interest was U2. They found me through a friend of Bono’s at the Department of the Treasury, a woman named Sheryl Sandberg. I met Bono and the Edge at Morgan Stanley’s offices in Los Angeles on the morning after the band had won a Grammy for the song “Beautiful Day.” I could not have named a U2 song, but I was blown away by the intelligence and business sophistication of the two Irishmen. They invited me to Dublin to meet their management. I made two trips during the spring of 2001.

On my way home from that second trip, I suffered a stroke. I didn’t realize it at the time, and I tried to soldier on. Shortly thereafter, after some more disturbing symptoms, I found myself at the Mayo Clinic, where I learned that I had in fact suffered two ischemic strokes, in addition to something called a transient ischemic attack in my brain stem. It was a miracle I had survived the strokes and suffered no permanent impairment.

The diagnosis came as a huge shock. I had a reasonably good diet, a vigorous exercise regime, and a good metabolism, yet I had had two strokes. It turned out that I had a birth defect in my heart, a “patent foramen ovale,” basically the mother of all heart murmurs. I had two choices: I could take large doses of blood thinner and live a quiet life, or I could have open-heart surgery and eliminate the risk forever. I chose surgery.

I had successful surgery in early July 2001, but my recovery was very slow. It took me nearly a year to recover fully. During that time, Apple shipped the first iPod. I thought it was a sign of good things to come and reached out to Steve Jobs to see if he would be interested in recapitalizing Apple. At the time, Apple’s share price was about twelve dollars per share, which, thanks to stock splits, is equivalent to a bit more than one dollar per share today. The company had more than twelve dollars in cash per share, which meant investors were attributing zero value to Apple’s business. Most of the management options had been issued at forty dollars per share, so they were effectively worthless. If Silver Lake did a recapitalization, we could reset the options and align interests between management and shareholders. Apple had lost most of its market share in PCs, but thanks to the iPod and iMac computers, Apple had an opportunity to reinvent itself in the consumer market. The risk/reward of investing struck me as especially favorable. We had several conversations before Steve told me he had a better idea. He wanted me to buy up to 18 percent of Apple shares in the public market and take a board seat.

After a detailed analysis, I proposed an investment to my partners in the early fall of 2002, but they rejected it out of hand. The decision would cost Silver Lake’s investors the opportunity to earn more than one hundred billion dollars in profits.

In early 2003, Bono called up with an opportunity. He wanted to buy Universal Music Group, the world’s largest music label. It was a complicated transaction and took many months of analysis. A team of us did the work and presented it to my other three partners in Silver Lake in September. They agreed to do the deal with Bono, but they stipulated one condition: I would not be part of the deal team. They explained their intention for Silver Lake to go forward as a trio, rather than as a quartet. There had been signals along the way, but I had missed them. I had partnered with deal guys—people who use power when they have it to gain advantages where they can get them—and had not protected myself.

I have never believed in staying where I’m not wanted, so I quit. If I had been motivated by money, I would have hung in there, as there was no way they could force me out. I had conceived the fund, incubated it, brought in the first billion dollars of assets, and played a decisive role on the three most successful investments. But I’m not wired to fight over money. I just quit and walked out. I happened to be in New York and called Bono. He asked me to come to his apartment. When I got there, he said, “Screw them. We’ll start our own fund.” Elevation Partners was born.

In the long term, my departure from Silver Lake worked out for everyone. The second Silver Lake fund got off to a rocky start, as my cofounders struggled with stock picking, but they figured it out and built the firm into an institution that has delivered good investment returns to its investors.

2

Silicon Valley Before Facebook

I think technology really increased human ability. But technology cannot produce compassion. —DALAI LAMA

The technology industry that gave birth to Facebook in 2004 bore little resemblance to the one that had existed only half a dozen years earlier. Before Facebook, startups populated by people just out of college were uncommon, and few succeeded. For the fifty years before 2000, Silicon Valley operated in a world of tight engineering constraints. Engineers never had enough processing power, memory, storage, or bandwidth to do what customers wanted, so they had to make trade-offs. Engineering and software programming in that era rewarded skill and experience. The best engineers and programmers were artists. Just as Facebook came along, however, processing power, memory, storage, and bandwidth went from being engineering limits to turbochargers of growth. The technology industry changed dramatically in less than a decade, but in ways few people recognized. What happened with Facebook and the other internet platforms could not have happened in prior generations of technology. The path the tech industry took from its founding to that change helps to explain both Facebook’s success and how it could do so much damage before the world woke up.

The history of Silicon Valley can be summed in two “laws.” Moore’s Law, coined by a cofounder of Intel, stated that the number of transistors on an integrated circuit doubles every year. It was later revised to a more useful formulation: the performance of an integrated circuit doubles every eighteen to twenty-four months. Metcalfe’s Law, named for a founder of 3Com, said that the value of any network would increase as the square of the number of nodes. Bigger networks are geometrically more valuable than small ones. Moore’s Law and Metcalfe’s Law reinforced each other. As the price of computers fell, the benefits of connecting them rose. It took fifty years, but we eventually connected every computer. The result was the internet we know today, a global network that connects billions of devices and made Facebook and all other internet platforms possible.

Beginning in the fifties, the technology industry went through several eras. During the Cold War, the most important customer was the government. Mainframe computers, giant machines that were housed in special air-conditioned rooms, supervised by a priesthood of technicians in white lab coats, enabled unprecedented automation of computation. The technicians communicated with mainframes via punch cards connected by the most primitive of networks. In comparison to today’s technology, mainframes could not do much, but they automated large-scale data processing, replacing human calculators and bookkeepers with machines. Any customer who wanted to use a computer in that era had to accept a product designed to meet the needs of government, which invested billions to solve complex problems like moon trajectories for NASA and missile targeting for the Department of Defense. IBM was the dominant player in the mainframe era and made all the components for the machines it sold, as well as most of the software. That business model was called vertical integration. The era of government lasted about thirty years. Data networks as we think of them today did not yet exist. Even so, brilliant people imagined a world where small computers optimized for productivity would be connected on powerful networks. In the sixties, J. C. R. Licklider conceived the network that would become the internet, and he persuaded the government to finance its development. At the same time, Douglas Engelbart invented the field of human-computer interaction, which led to him to create the first computer mouse and to conceive the first graphical interface. It would take nearly two decades before Moore’s Law and Metcalfe’s Law could deliver enough performance to enable their vision of personal computing and an additional decade before the internet took off.

Beginning in the seventies, the focus of the tech industry began to shift toward the needs of business. The era began with a concept called time sharing, which enabled many users to share the use of a single computer, reducing the cost to everyone. Time sharing gave rise to minicomputers, which were smaller than mainframes but still staggeringly expensive by today’s standards. Data networking began but was very slow and generally revolved around a single minicomputer. Punch cards gave way to terminals, keyboards attached to the primitive network, eliminating the need for a priesthood of technicians in white lab coats. Digital Equipment, Data General, Prime, and Wang led in minicomputers, which were useful for accounting and business applications but were far too complicated and costly for personal use. Although they were a big step forward relative to mainframes, even minicomputers barely scratched the surface of customer needs. Like IBM, the minicomputer vendors were vertically integrated, making most of the components for their products. Some minicomputers—Wang word processors, for example—addressed productivity applications that would be replaced by PCs. Other applications survived longer, but in the end, the minicomputer business would be subsumed by personal computer technology, if not by PCs themselves. Main frames have survived to the present day, thanks in large part to giant, custom applications like accounting systems, which were created for the government and corporations and are cheaper to maintain on old systems than to re-create on new ones. (Massive server farms based on PC technology now attract any new application that needs mainframe-class processing; it is a much cheaper solution because you can use commodity hardware instead of proprietary mainframes.)

ARPANET, the predecessor to today’s internet, began as a Department of Defense research project in 1969 under the leadership of Bob Taylor, a computer scientist who continued to influence the design of systems and networks until the late nineties. Douglas Engelbart’s lab was one of the first nodes on ARPANET. The goal was to create a nationwide network to protect the country’s command and control infrastructure in the event of a nuclear attack.

The first application of computer technology to the consumer market came in 1972, when Al Alcorn created the game Pong as a training exercise for his boss at Atari, Nolan Bushnell. Bushnell’s impact on Silicon Valley went far beyond the games produced by Atari. He introduced the hippie culture to tech. White shirts with pocket protectors gave way to jeans and T-shirts. Nine to five went away in favor of the crazy, but flexible hours that prevail even today.

In the late seventies, microprocessors made by Motorola, Intel, and others were relatively cheap and had enough performance to allow Altair, Apple, and others to make the first personal computers. PCs like the Apple II took advantage of the growing supply of inexpensive components, produced by a wide range of independent vendors, to deliver products that captured the imagination first of hobbyists, then of consumers and some businesses. In 1979, Dan Bricklin and Bob Frankston introduced VisiCalc, the first spreadsheet for personal computers. It is hard to overstate the significance of VisiCalc. It was an engineering marvel. A work of art. Spreadsheets on Apple IIs transformed the productivity of bankers, accountants, and financial analysts.

Unlike the vertical integration of mainframes and minicomputers, which limited product improvement to the rate of change of the slowest evolving part in the system, the horizontal integration of PCs allowed innovation at the pace of the most rapidly improving parts in the system. Because there were multiple, competing vendors for each component, systems could evolve far more rapidly than equivalent products subject to vertical integration. The downside was that PCs assembled this way lacked the tight integration of mainframes and minicomputers. This created a downstream cost in terms of training and maintenance, but that was not reflected in the purchase price and did not trouble customers. Even IBM took notice.

When IBM decided to enter the PC market, it abandoned vertical integration and partnered with a range of third-party vendors, including Microsoft for the operating system and Intel for the microprocessor. The first IBM PC shipped in 1981, signaling a fundamental change in the tech industry that only became obvious a couple of years later, when Microsoft’s and Intel’s other customers started to compete with IBM. Eventually, Compaq, Hewlett-Packard, Dell, and others left IBM in the dust. In the long run, though, most of the profits in the PC industry went to Microsoft and Intel, whose control of the brains and heart of the device and willingness to cooperate forced the rest of the industry into a commodity business.

ARPANET had evolved to become a backbone for regional networks of universities and the military. PCs continued the trend of smaller, cheaper computers, but it took nearly a decade after the introduction of the Apple II before technology emerged to leverage the potential of clusters of PCs. Local area networks (LANs) got their start in the late eighties as a way to share expensive laser printers. Once installed, LANs attracted developers, leading to new applications, such as electronic mail. Business productivity and engineering applications created incentives to interconnect LANs within buildings and then tie them all together over proprietary wide area networks (WANs) and then the internet. The benefits of connectivity overwhelmed the frustration of incredibly slow networks, setting the stage for steady improvement. It also created a virtuous cycle, as PC technology could be used to design and build better components, increasing the performance of new PCs that could be used to design and build even better components.

Consumers who wanted a PC in the eighties and early nineties had to buy one created to meet the needs of business. For consumers, PCs were relatively expensive and hard to use, but millions bought and learned to operate them. They put up with character-mode interfaces until Macintosh and then Windows finally delivered graphical interfaces that did not, well, totally suck. In the early nineties, consumer-centric PCs optimized for video games came to market.

The virtuous cycle of Moore’s Law for computers and Metcalfe’s Law for networks reached a new level in the late eighties, but the open internet did not take off right away. It required enhancements. The English researcher Tim Berners-Lee delivered the goods when he invented the World Wide Web in 1989 and the first web browser in 1991, but even those innovations were not enough to push the internet into the mainstream. That happened when a computer science student by the name of Marc Andreessen created the Mosaic browser in 1993. Within a year, startups like Yahoo and Amazon had come along, followed in 1995 by eBay, and the web that we now know had come to life.

By the mid-nineties, the wireless network evolved to a point that enabled widespread adoption of cell phones and alphanumeric pagers. The big applications were phone calls and email, then text messaging. The consumer era had begun. The business era had lasted nearly twenty years—from 1975 to 1995—but no business complained when it ended. Technology aimed at consumers was cheaper and somewhat easier to use, exactly what businesses preferred. It also rewarded a dimension that had not mattered to business: style. It took a few years for any vendor to get the formula right.

The World Wide Web in the mid-nineties was a beautiful thing. Idealism and utopian dreams pervaded the industry. The prevailing view was that the internet and World Wide Web would make the world more democratic, more fair, and more free. One of the web’s best features was an architecture that inherently delivered net neutrality: every site was equal. In that first generation, everything on the web revolved around pages, every one of which had the same privileges and opportunities. Unfortunately, the pioneers of the internet made omissions that would later haunt us all. The one that mattered most was the choice not to require real identity. They never imagined that anonymity would lead to problems as the web grew.

Time would expose the naïveté of the utopian view of the internet, but at the time, most participants bought into that dream. Journalist Jenna Wortham described it this way: “The web’s earliest architects and pioneers fought for their vision of freedom on the Internet at a time when it was still small forums for conversation and text-based gaming. They thought the web could be adequately governed by its users without their need to empower anyone to police it.” They ignored early signs of trouble, such as toxic interchanges on message boards and in comments sections, which they interpreted as growing pains, because the potential for good appeared to be unlimited. No company had to pay the cost of creating the internet, which in theory enabled anyone to have a website. But most people needed tools for building websites, applications servers and the like. Into the breach stepped the “open source” community, a distributed network of programmers who collaborated on projects that created the infrastructure of the internet. Andreessen came out of that community. Open source had great advantages, most notably that its products delivered excellent functionality, evolved rapidly, and were free. Unfortunately, there was one serious problem with the web and open source products: the tools were not convenient or easy to use. The volunteers of the open source community had one motivation: to build the open web. Their focus was on performance and functionality, not convenience or ease of use. That worked well for the infrastructure at the heart of the internet, but not so much for consumer-facing applications.

The World Wide Web took off in 1994, driven by the Mosaic/Netscape browser and sites like Amazon, Yahoo, and eBay. Businesses embraced the web, recognizing its potential as a better way to communicate with other businesses and consumers. This change made the World Wide Web geometrically more valuable, just as Metcalfe’s Law predicted. The web dominated culture in the late nineties, enabling a stock market bubble and ensuring near-universal adoption. The dot-com crash that began in early 2000 left deep scars, but the web continued to grow. In this second phase of the web, Google emerged as the most important player, organizing and displaying what appeared to be all the world’s information. Apple broke the code on tech style—their products were a personal statement—and rode the consumer wave to a second life. Products like the iMac and iPod, and later the iPhone and iPad, restored Apple to its former glory and then some. At this writing, Apple is the most valuable company in the world. (Fortunately, Apple is also the industry leader in protecting user privacy, but I will get to that later.)

In the early years of the new millennium, a game changing model challenged the page-centric architecture of the World Wide Web. Called Web 2.0, the new architecture revolved around people. The pioneers of Web 2.0 included people like Mark Pincus, who later founded Zynga; Reid Hoffman, the founder of LinkedIn; and Sean Parker, who had cofounded the music file sharing company Napster. After Napster, Parker launched a startup called Plaxo, which put address books in the cloud. It grew by spamming every name in every address book to generate new users, an idea that would be copied widely by social media platforms that launched thereafter. In the same period, Google had a brilliant insight: it saw a way to take control of a huge slice of the open internet. No one owned open source tools, so there was no financial incentive to make them attractive for consumers. They were designed by engineers, for engineers, which could be frustrating to non-engineers.

Google saw an opportunity to exploit the frustration of consumers and some business users. Google made a list of the most important things people did on the web, including searches, browsing, and email. In those days, most users were forced to employ a mix of open source and proprietary tools from a range of vendors. Most of the products did not work together particularly well, creating a friction Google could exploit. Beginning with Gmail in 2004, Google created or acquired compelling products in maps, photos, videos, and productivity applications. Everything was free, so there were no barriers to customer adoption. Everything worked together. Every app gathered data that Google could exploit. Customers loved the Google apps. Collectively, the Google family of apps replaced a huge portion of the open World Wide Web. It was as though Google had unilaterally put a fence around half of a public park and then started commercializing it.

The steady march of technology in the half century prior to 2000 produced so much value—and so many delightful surprises—that the industry and customers began to take positive outcomes for granted. Technology optimism was not equivalent to the law of gravity, but engineers, entrepreneurs, and investors believed that everything they did made the world a better place. Most participants bought into some form of the internet utopia. What we did not realize at the time was that the limits imposed by not having enough processing power, memory, storage, and network bandwidth had acted as a governor, limiting the damage from mistakes to a relatively small number of customers. Because the industry had done so much good in the past, we all believed that everything it would create in the future would also be good. It was not a crazy assumption, but it was a lazy one that would breed hubris.

When Zuck launched Facebook in early 2004, the tech industry had begun to emerge from the downturn caused by the dot-com meltdown. Web 2.0 was in its early stages, with no clear winners. For Silicon Valley, it was a time of transformation, with major change taking place in four arenas: startups, philosophy, economics, and culture. Collectively, these changes triggered unprecedented growth and wealth creation. Once the gravy train started, no one wanted to get off. When fortunes can be made overnight, few people pause to ask questions or consider side effects.

The first big Silicon Valley change related to the economics of startups. Hurdles that had long plagued new companies evaporated. Engineers could build world-class products quickly, thanks to the trove of complementary software components, like the Apache server and the Mozilla browser, from the open source community. With open source stacks as a foundation, engineers could focus all their effort on the valuable functionality of their app, rather than building infrastructure from the ground up. This saved time and money. In parallel, a new concept emerged—the cloud—and the industry embraced the notion of centralization of shared resources. The cloud is like Uber for data—customers don’t need to own their own data center or storage if a service provides it seamlessly from the cloud. Today’s leader in cloud services, Amazon Web Services (AWS), leveraged Amazon.com’s retail business to create a massive cloud infrastructure that it offered on a turnkey basis to startups and corporate customers. By enabling companies to outsource their hardware and network infrastructure, paying a monthly fee instead of the purchase price of an entire system, services like AWS lowered the cost of creating new businesses and shortened the time to market. Startups could mix and match free open source applications to create their software infrastructure. Updates were made once, in the cloud, and then downloaded by users, eliminating what had previously been a very costly and time-consuming process of upgrading individual PCs and servers. This freed startups to focus on their real value added, the application that sat on top of the stack. Netflix, Box, Dropbox, Slack, and many other businesses were built on this model.

Thus began the “lean startup” model. Without the huge expense and operational burden of creating a full tech infrastructure, new companies did not have to aim for perfection when they launched a new product, which had been Silicon Valley’s primary model to that point. For a fraction of the cost, they could create a minimum viable product (MVP), launch it, and see what happened. The lean startup model could work anywhere, but it worked best with cloud software, which could be updated as often as necessary. The first major industry created with the new model was social media, the Web 2.0 startups that were building networks of people rather than pages. Every day after launch, founders would study the data and tweak the product in response to customer feedback. In the lean startup philosophy, the product is never finished. It can always be improved. No matter how rapidly a startup grew, AWS could handle the load, as it demonstrated in supporting the phenomenal growth of Netflix. What in earlier generations would have required an army of experienced engineers could now be accomplished by relatively inexperienced engineers with an email to AWS. Infrastructure that used to require a huge capital investment could now be leased on a monthly basis. If the product did not take off, the cost of failure was negligible, particularly in comparison to the years before 2000. If the product found a market, the founders had alternatives. They could raise venture capital on favorable terms, hire a bigger team, improve the product, and spend to acquire more users. Or they could do what the founders of Instagram and WhatsApp would eventually do: sell out for billions with only a handful of employees.

Facebook’s motto—“Move fast and break things”—embodies the lean startup philosophy. Forget strategy. Pull together a few friends, make a product you like, and try it in the market. Make mistakes, fix them, repeat. For venture investors, the lean startup model was a godsend. It allowed venture capitalists to identify losers and kill them before they burned through much cash. Winners were so valuable that a fund needed only one to provide a great return.

When hardware and networks act as limiters, software must be elegant. Engineers sacrifice frills to maximize performance. The no-frills design of Google’s search bar made a huge difference in the early days, providing a competitive advantage relative to Excite, Altavista, and Yahoo. A decade earlier, Microsoft’s early versions of Windows failed in part because hardware in that era could not handle the processing demands imposed by the design. By 2004, every PC had processing power to spare. Wired networks could handle video. Facebook’s design outperformed MySpace in almost every dimension, providing a relative advantage, but the company did not face the fundamental challenges that had prevailed even a decade earlier. Engineers had enough processing power, storage, and network bandwidth to change the world, at least on PCs. Programming still rewarded genius and creativity, but an entrepreneur like Zuck did not need a team of experienced engineers with systems expertise to execute a business plan. For a founder in his early twenties, this was a lucky break. Zuck could build a team of people his own age and mold them. Unlike Google, Facebook was reluctant to hire people with experience. Inexperience went from being a barrier to being an advantage, as it kept labor costs low and made it possible for a young man in his twenties to be an effective CEO. The people in Zuck’s inner circle bought into his vision without reservation, and they conveyed that vision to the rank-and-file engineers. On its own terms, Facebook’s human resources strategy worked exceptionally well. The company exceeded its goals year after year, creating massive wealth for its shareholders, but especially for Zuck. The success of Facebook’s strategy had a profound impact on the human resources culture of Silicon Valley startups.

In the early days of Silicon Valley, software engineers generally came from the computer science and electrical engineering programs at MIT, Caltech, and Carnegie Mellon. By the late seventies, Berkeley and Stanford had joined the top tier. They were followed in the mid-nineties by the University of Illinois at Urbana-Champaign, the alma mater of Marc Andreessen, and other universities with strong computer science programs. After 2000, programmers were coming from just about every university in America, including Harvard.
<< 1 2 3 4 5 6 >>
На страницу:
4 из 6