The rise of machine intelligence at Codecon 2016, by Steven Sinofsky of a16z
Geege Schuman stashed this in Machine Learning
Stashed in: Awesome, Amazon, @sherylsandberg, Turing, @a16z, Jeff Bezos, Ford, Accelerating Returns, @sundarpichai, GPU, Machine Learning, IBM, Artificial Intelligence, Machines Writing Code, Deep Learning
Facebook doing machine intelligence:
Perhaps the most frequent use of AI any of us experience is through Facebook on our mobile phones, and together onstage, Facebook COO Sheryl Sandberg and CTO Mike Schroepfer outlined numerous efforts in AI across the site, advertisers and, of course, in virtual reality. The depth of both product features and applied research going on in AI at Facebook is not as widely appreciated, I believe. For example, a few weeks ago, Facebook contributed a dozen papers to the major AI conference ICLR. As mentioned below, the role each of us play as people (not "user," according to both speakers) is a huge contributor to Facebook’s ability to deliver AI-based features like photo tagging that we find so valuable.
AI is already here. Get used to it.
Some years are remembered as launch years (like the iPod, or even Windows or Office releases!), and others are marked more by debates over disruption (such as net neutrality or music distribution). Rarely do we get to experience a year when the breadth of speakers collectively express both optimism and execution plans for a technology shift as we saw this year. By any name, machine learning, artificial intelligence, deep simulation — or, as IBM’s CEO insisted, cognitive computing — were front and center for nearly every speaker.
Over the course of the previous 12 conferences, there have been themes that everyone might have touched on, but I don’t recall a case where there was such uniform aggressiveness at staking a claim on the technology. Across the board, speakers went to great lengths to talk about how their customers are going to benefit from the use of intelligence technologies (let’s call this "AI") in products and services.
Why is this not simply a waypoint on the way through the hype cycle? The short answer is because we, as customers, are already "using" intelligence every day on our smart phones. It is worth pausing to acknowledge that somehow over the course of the past year, AI has gone from passing reference through implementation to daily use.
Today’s use of AI is not hype, but reality.
Down the road, the interesting question will not be which companies used AI, but which companies made the most of AI in novel ways.
The most fascinating aspect of listening to the speakers respond to questions about how AI will or does play a role in their respective enterprises is how the technology spans devices, strategies and business models. While each of us might be familiar with a specific example, looking across the speakers paints a picture of an incredibly rapid and deep technology diffusion.
Amazon CEO Jeff Bezos spoke at great length about the role of AI in the company's breakout product, Echo. In many ways, Echo has come to symbolize the true potential of multiple technologies across voice commands, agents and machine learning, all packaged up in super-simple consumer devices. In addition, Bezos outlined how Alexa — the technology underpinnings of Echo — is both a customizable platform and an embeddable technology. Developers can build new "skills" for Alexa and contribute "learning" to offer new capabilities (and as Mossberg noted, owners receive weekly email detailing the latest skills added). Makers can embed Alexa technology in their own devices — one example mentioned was an alarm clock — and turn a mundane technology into another AI-enabled endpoint.
Ford CEO Mark Fields worked to convince the Code crowd that while Ford’s intrinsic strength in AI or cloud might not be there, they are actively partnering (with Pivotal, for example) and bringing in-house the skills needed to make sure Ford vehicles participate in this technology wave. While Fields did not articulate a deep vision of AI, he used examples around maintenance and navigation to illustrate the company's commitment.
In an approach similar to Ford, Cisco CEO Chuck Robbins talked about committing his company's thousands of software engineers to AI, and incorporating it into Ford's existing product lines. He spoke of using AI to bring better management and understanding to modern networks that will contain millions of end points of all sorts.
While spending a good deal of time differentiating his company from Amazon, eBay CEO Devin Wenig talked in detail about eBay's amazing work to improve the shopping experience, and to eliminate fraud by using AI. Wenig characterized the use of AI as the way the company will deliver a highly personalized eBay — one that is curated and relevant. Fraud, he claimed (in a follow-up interview by Lauren Goode, from Recode's Vox partner The Verge), has become a "meaningless" number through the use of technology.
IBM CEO Ginni Rometty traced IBM’s longstanding efforts in AI, and outlined what amounts to a "bet the company" investment in the technology. She discussed scenarios from health care to education, from business IT to third-party developers, and from cloud to on-premises as ways IBM is working to contribute to and support the use of AI. Most interesting were the efforts IBM is making to provide open source or freely available solutions, which Rometty said ran counter to IBM’s history, but were essential to see the company moving forward. While many of us think of IBM’s Big Blue or "Jeopardy" building up to Watson, IBM’s history with AI goes to the very start of the field, where the famed Watson Labs worked to pioneer the earliest ideas in translation, language processing, speech and handwriting recognition, and more. Even with so many challenges, one has to be impressed by Rometty's outlook on the technology and the depth of IBM’s engagement.
Fresh off the heels of the I/O Conference, Google CEO Sundar Pichai took us through the history of AI at Google. There is clearly no other company that has the depth of efforts and the broad use of AI in products over time. Without a doubt, AI has always defined Google, and it is only in the past year or so that this has become broadly understood. Photos, inbox, search, advertising, Assistant, self-driving cars — and the list continues —were all examples used by Pichai to illustrate Google’s ongoing commitment to AI. If AI is itself a platform, Google is most certainly the most invested and best positioned to be the leader in such a platform.
That AI is so mainstream today that dozens of CEOs can articulate company execution plans using the technology is directly attributable to four important changes in the technology landscape.
I think it's worth reflecting on how these came together to make it possible to see such a rapid move from lab to deployed feature in such a high-tech endeavor.
Raw compute power for models: While we love to talk about Moore’s law as enabling so much, when it comes to AI, it is Moore’s law applied to parallel architectures, not Intel’s scalar ones. The application of even more transistors to graphics processing units (GPU) has been a key enabler of AI technology. Cloud architecture plays an incredibly important part in this because companies do not need to build out their own GPU data centers in order to tap into the power of AI training models, but can tap into on-demand scale.
Massive data capacity for training: Everyone has come to learn that more data is the only way to train AI models, and you can never have too much data (but it is easy to have too little). It is only recently that cloud architectures have become "routine" and "economical" at both retaining and accessing the quantities of data required for training. Facebook provides the easiest-to-visualize view of this as it trains recognition of people on the more than 300 million photos a day (or half a petabyte). IBM’s examples using radiology offer another view into just how important the evolution of storage is to AI.
Incredible availability of labeled data: As important as data is to training, data without "labels" isn’t very helpful. This is where our role using technology plays an important part, as well as the openness of the internet. It isn’t just that we upload photos to Facebook, but that we tag people we know, and in doing so, train the image-recognition engine. It isn’t enough for eBay to say it wants to offer a customized store, but that we are signed in and purchasing items to inform the customization engine. This is such an advance over the way we used to think of click streams or guessing if someone is a return visitor.
In addition to all of this, sensors in our phones offer motion and location data, enhancing everything we are doing. Clearly and obviously, there are privacy and security questions with all of this, but at the same time, never before have there been such personal benefits to each of us as we use services. The availability of data goes beyond that which I personally generate (and label), and includes data sets and APIs that are now available simply because of openness and cloud-based solutions (for example, economic and demographic data from governments) that can be incorporated as part of training models.
Open implementations of technology underpinnings: The most fascinating aspect of the rise of AI solutions has been that so much of the core technology has been developed in the open (often by the research arms of companies) or is at least contributed in an open way relatively early in the evolution of the technology. Google’s TensorFlow, Facebook’s Torch, IBM’s SystemML and UC Berkeley’s CAFFE, along with technologies for data (such as Spark) are all openly available platform elements. This is most certainly following the same pattern as HTML/HTTP, which means the economics will be elsewhere in the system (in the data, training and models, of course).
Soon we won't program computers. We will train them.
http://pandawhale.com/post/72195/soon-we-wont-program-computers-well-train-them-like-dogs
1:43 PM Jun 07 2016