To put this in perspective, 5000 people at $200k per year (which is conservative if you include benefits, etc.) is 1 billion dollars in comp per year.
So OpenAI is spending a billion dollars over the next several years.
Microsoft is spending a billion dollars per year.
Google, etc. do the math.
There's literally billions of dollars now being spent on moving deep learning forward. Pretty amazing when I think back to 2011 and there were machine learning conferences where literally no one I spoke to had heard of deep learning.
People worrying about a second AI winter are like the people that have been worrying about an "internet bubble" since 2004. It's fine to be worried, there will be bubbles, but this time it's different and there are many reasons for that. There is no "internet industry" anymore; it's become more segmented and just plain bigger. Similarly, there will be no "AI industry"; it will branch out, and there are more potential applications of "understanding data and automating decisions" than there were in the 80s.
> something that they are branding AI - I might wait and see how much of it is on real AI - what ever that is.
Yeah, I wouldn't get hung up on that, they're calling it AI because that's easier to explain to reporters than deep learning. The difference is deep learning techniques are already being used in released products and companies are looking to do more of that, so there is a very real definition and set of goals associated with what these groups are doing, it's not just, "hey everyone, let's make an ai!"
I think it's important to understand that is really happening here. Microsoft is merging Microsoft Research with their application oriented AI work. See the press release from Microsoft (http://mnc.ms/2ddF5dC).
MSR is already a pretty big group at Microsoft and does all sorts of work that is completely unrelated to AI. Now, I am confident the teams within MSR will have to start thinking about AI, and how their work relates to AI, but this initiative is much more of an interdisciplinary effort than a 5000 person deep learning team.
Well, companies are certainly spending a lot of money on something that they are branding AI - I might wait and see how much of it is on real AI - what ever that is.
While most people in the industry were busy debating the usefulness and definition of "cloud", Amazon and others were busy building it. To me this is similar to that situation. There is real progress happening with practical benefit in AI, machine learning, deep learning, and Microsoft is committed to being at the forefront. Personally I'm not waiting. I'm teaching myself data modeling and other techniques that'll let me be part of the coming revolution.
I think 200k as an average is good, since some will be making a whole lot more, and there will be the 100K per year positions too.
My equivalent to your dearth of people in the know in 2011, was studying AL (Artificial Life) in 1990s, and nobody heard of it. It included the study of ANNs, GAs, GP and AI in general (which I prefer to call CI - Computational Intelligence nowadays). The book that started it all for me [0].
There were no immediate applications aside from expert systems here and there, and the fuzzy logic appliance controllers coming out of Japan. However, now, with self-driving cars, recommender systems, image recognition (face-matching surveillance post 9/11) have big pay-offs or budgets to spend on it.
VR is having its second renaissance. I thought the first time it would have taken off even with the clunky glasses and headsets, since gaming was already such a huge money industry.
And now with modeling becoming prominent again, AL paradigms are being modified, created and repurposed for all sorts of cool things. I play with NetLogo since it is a fun environment for that [1].
My only regret is that I was in at the beginning, but left it to pursue other things, and so I am not at the level of practice to get one of those 200K jobs. I still kept studying it all these years though, and I have coded my own bits and pieces, but mainly for conceptual pieces, art and music, not practical applications. I play with Darknet [2] now, since C was my second language after Assembly (6502 and then x86_32), and it is great fun, and fast. A very understandable and manageable platform.
I am hoping it all leads someday to keeping me alive longer to enjoy studying some more, because as I get older that's really what I enjoy most aside from family!
Well its not 5000 new people I assume ~70% would be restructuring (E.g. as the article states from Bing and Cortana). Even then at 1500 new employees, its a huge undertaking and vote of confidence into importance of AI.
Except for severance and hiring bonuses, what's the difference between restructuring someone's job, or firing them and hiring another person? In the end, you're still committing $1B/year to AI jobs.
- Salary cost = gross salary / 1800 (or whatever # hours/yr your company uses)
- Benefits = proportional cost of vacation, 401K, etc
- G&A = General expenses and admin (office, equipment, free food, etc)
- Overhead = amount needed to cover everyone else who isn't billable (HR, Finance, IT, assistants, President, etc)
If you do the math, you'll see $200-300 is a standard rate in high cost markets (NYC, SFO) for someone with a six-digit figure. More junior positions are ~$150, while very senior roles (Partner, Director, etc) would be in the $300-500 range.
Because your boss has no billable hours. Nor his boss, nor the boss above him. The front desk guy, the lady in accounting, the payroll people, facilities, most of the IT folk.
You need one of each these people for every N developers (different value of N for different classes of service). Therefore the cost of having you includes the 1/N cost of each of these people.
If you are client-facing consultant, your billable rate is paying:
- Your salary
- Your benefits
- Your expenses (travel, per diem, equipment, software, anything else)
- And then... Salary, benefits, expenses of all your NON-client-facing teammates: Salespeople, marketing, management, client relationship, administrative, HR, etc.
So as a consultant, you are the product, and as such have to be sold at a price that nets enough profit to pay for rest of the organization.
Normally when they say "you cost $150/hour", what they mean is, "you, and all the support framework behind you, needs to be billed at $150/hour to pay off".
Hope that helps. If you're in-house developer and not client-facing, it's harder to understand and justify cost figures, but typically would be nowhere near that rate...
If you look at MS's financial statements you can (almost) do the math. Divide R&D spending by R&D personnel and you end up at $300k, which can serve as an upper bound. That supposedly includes vendor budget (not included in #personnel), of which MS employs quite a few, and overhead.
If you do the same with Google you end up at $500k, which seems insane.
Also, am I misunderstanding, or is the "Deep" part of Deep Learning not a very quantifiable thing? I.e. there's no rule that says "Your neural net isn't 'deep' unless it has X amount of layers or N amount of nodes per layer" or anything. It just specifies using large neural networks for examining a huge amount of data and feature set. Right?
So really, any AI-based learning is going to be "deep" from now on, simply because we've reached the point where we can handle large neural nets and complex datasets, right?
>... there will be no "AI industry"; it will branch out, and there are more potential applications of "understanding data and automating decisions" than there were in the 80s.
IBM was the first company with a massive investment in AI (during this cycle) - they are now trying to push Watson as a product. Did it pay off so far? I am not quite sure.
Can you help us understand what you think this new division will be producing? What are a few deep learning-enabled products or projects that are in production and generating revenues?
Microsoft historically pays slightly better than Google and other valley companies. And, the cost of an employee is usually significantly higher than the salary alone. Something like 30%-60% higher, when factoring in benefits, cost of providing offices, stock options, etc. So, $200k is pretty conservative in many tech hubs, not just the valley, particularly when you're talking about people involved in research (who are more likely to have obtained graduate degrees).
Where are you getting this information from? I got an offer from Microsoft and it definitely wasn't even close to what people who get offers from FB or Google get.
My information may be out of date. My friends who've chosen between Google and Microsoft got better offers from MS; though often ended up going to Google anyway, because they just liked it better. But, it was all several years ago.
Microsoft pays well, but this is the first time I've heard them paying better than google. When I was at MS, there were always complaints about losing potential hires to Google over much higher salaries.
Comparing MS salaries in Redmond to Google salaries in Mountain View is rather apples and oranges. Cost of living is different (esp. once you start shopping for houses), state taxes are drastically different etc.
Now, Google also has an office in Kirkland. That would be an interesting place to compare. But it's also not particularly large.
I don't appreciate the "cost of living" salary differential excuse, and let me tell you why... Ask yourself, does Microsoft charge me less for Windows or Office if I live in, say, Kansas? No, they don't.
Why would that matter in the slightest when you're actually shopping for a job? Either the pay is more (in terms of quality of life it translates to while residing where you need to live to commute), or not.
Relatively little of your money is spent on those kinds of things though. If you're like most people, most of your money goes to taxes, a place to live, and food. Those are location-dependent.
I feel like, as a solo app developer, I'm being left out of this 'revolution'.
It seems like deep learning only makes sense if you have enough data to feed the algorithm with. The kind of data that only big companies can produce or harvest.
Sure you can produce some video and audio data and you can spider the web a little bit, but that doesn't even come close to the resources that these big corporations have and the 'depth' of learning that they can achieve.
So I'm not even trying.
Or should I ?
Is there any place for solo/indy developers in this field ?
I've just started tinkering with deep learning for a solo project; I recognized pretty quickly that there are many areas where I can't "play with the big boys", but I also realized there's a lot of low-hanging fruit that is accessible to me exactly because the big guys are open sourcing so much.
So, I can build products that aren't big enough to interest Google, but include a bunch of tech developed by Google (and others) and leverages their APIs to provide a unique service that is feasible for me to build, and will be useful to a wide variety of people. I can offer it for very little money (one person's side project), and hopefully have some fun learning about deep learning.
I'm so new to it that I'm not even thinking about advancing the state of the art or doing novel work. But, in a couple of years, who knows. Just tinkering with things in a new industry often provides pathways to cool stuff because so many doors are opening all the time. This is like being involved in the Internet in the early-to-mid 90s. You probably won't become the next Google, but the odds of finding a highly profitable smaller niche seems pretty high.
Also, there's going to be a ton of acquisitions in the AI/deep learning space over the next decade. Every company that even does a little tech will "need" an AI story to keep their investors happy. Your tiny thing could be one of those acqui-hires, or maybe not.
Then again, if you have an interest in other stuff, and really don't feel excited about it...probably not worth forcing yourself to get into it. Life is short, you should do stuff that's fun, even if you have to ring the cash register now and then.
I'll be posting a Show HN post in a few weeks with one of those ideas.
But, there's a bunch of ideas I've brainstormed around using things like sentiment analysis and other kinds of very simple-to-use AI concepts for automating tedious stuff. Things like automatically triaging support requests based on how angry the customer sounds, or based on keywords and an analysis of earlier requests; off-the-shelf NLP algorithms can do this today (and Google uses it that way for their own support tools, but doesn't make it widely available in that form, though Inbox has some of that kind of tech working in it). All you need is training data and some familiarity with Python.
My brainstorming exercise goes something like this: Append "with spooky powers" to a bunch of common things until one of them seems cool and useful to me. So, "forum notifications bot with spooky powers", "IRC bot with spooky powers", "twitter bot with spooky powers", "customer relationship management with spooky powers", "analytics with spooky powers", "server monitoring with spooky powers", "log analysis with spooky powers", etc. And I try to think of what I would use such a thing for, if it existed. Then, I sit down and see if I can make it real. The Yahoo NSFW image detection announcement reminded me of ideas I had and tinkered with a decade ago when I worked on a content filtering system for schools...the difference is that now we have the horsepower, the data sets, and the algorithms to actually make it work (but, I haven't worked in that field in a decade and never really liked being a purveyor of censorship tools, even if only for children, anyway).
Anyway, the possibilities are kinda huge and wide open. Many of these ideas will fizzle out, even the ones that look promising, but as with the Internet a lot of millionaires are going to be made by people saying, "It's like X, but with AI." just as people used to say, "It's like X, but on the Internet."
That's a really neat way to brainstorm. Let me add one more thing.
The big companies have an advantage in hardware and research. But they dont care about niche applications of their tech, because prizes worth less than $1B don't matter at their scale. That's where I try to focus on.
The key challenge is data. Too many AI startups get stuck in the "give us your data and we'll do some awesome stuff." That almost never works. [This](http://mattturck.com/2016/09/29/building-an-ai-startup/) talk does a really good job explaining why. The trick is figuring out how to get enough initial data to deliver value upfront.
Good slide deck, and I agree with most of it (and what I don't agree with is probably my own ignorance of the field).
I suspect there will be very few "pure AI" startups, and a ton of regular old tech startups that figure out how AI fits into their business faster than their competitors or figure out how to use it for a business advantage or to deliver a service that couldn't exist in that way before AI. With the early "like X but on the Internet" startups, the ones that succeeded in the biggest way (Amazon, for example) were the ones that built a great X that leveraged the internet to make it an order of magnitude better X.
So, Amazon was the best book store because they got everything right about being a regular bookstore (good prices, good service, efficient sales channel, solid relationships with publishers) and had damned near every book and could serve customers everywhere; a thing that is only possible on the Internet.
So, the best "X except with AI" company will be a great X company, and then AI will allow them to do some kind of force multiplier to push them to the top of the heap. That means we need to look for opportunities that currently require a lot of resources (say, people, or vehicles, or ) and can have AI added to it to make it produce 10x value given the same inputs. Even 2x value could be a big enough difference to beat your competitors at market, but the real out-of-the-park success stories probably need an order-of-magnitude boost from AI, even if it starts out slower because AI is still clumsy and most of the small companies are starting with tiny data sets (relatively speaking).
Anyway, mostly I think it's cool to play with. I think I see some ways to provide value and make some money with it, but it'll be as much an experiment as a business plan in the short term.
I don't know much about it, but there are public datasets you can train on and contests you can enter.
But this seems more like doing research in hopes of coming up with something new that big companies will be interested in. Or, if that doesn't happen, learning enough so that they can hire you as a researcher or consultant.
You can do some cool stuff with the APIs and datasets the big companies provide.
Most are typically priced per API call with generous free tiers. I expect these will get cheaper. I'm working on a search engine for lectures (https://www.findlectures.com) - for ~30k API requests I've been able to do everything free.
A lot of interesting large data sets are hosted in the "cloud" too for you to use for research, so you can get at them that way.
Depending on what you're trying to do you have some options, but you should probably outsource.
Option 1: If you're working on image recognition or anything similar, it's easy to get a ton of data. There are many corpora of image data available, probably on the order of hundreds of terabytes, with tags or at least some structured data. If you're going to "spider the web" you should use Common Crawl instead, and can access all of Blekko's data for the cost of data transfer via AWS. Same for text data, use the Google N-gram corpus.
Knowing that all of that exists, I would still recommend Option 2: outsourcing your AI needs unless you're a researcher or have a decent budget for AI development. Go search "ai api" and pick one that matches what you want to do. Match your skill and risk tolerance; nobody got fired for using IBM, but you may get a better experience from a startup with the possibility of flaming out in a year. You'll get the leverage of whatever company is spending those dollars on your behalf, and you'll be able to concentrate on the user experience instead of the science-ey part.
Option 3 is "if you can't beat 'em, join 'em". Go work for MS / GOOG / FB / IBM building something, and get access to those resources for yourself. Then at some point in the future you'll know their API interface, and you can go back to option #2 with much better data.
There is a lot of design and UX and market exploration to do around learning systems. The way that a user will consume the outputs of a learning system are far from well defined right now. Larger companies have an advantage in many ways (like Google Now can exploit google's reach into android home screens) but consider a tool that helps you garden, find a restaurant or dress yourself based on your preferences - what would that UI look like? Google doesn't have an answer to that, yet.
If someone proves that deep learning is applicable to most software categories (and I honestly don't think it is, at least not for a very long time), then someone will sell a solution you can plug into your app.
We already have a disturbing quantity and variety of user metric apps for the web. Realtime feedback just requires someone to bring engineering discipline to bear on the space and produce a functional and efficient version. Instead of a bunch of code monkey asshats making my beautiful, fast web app soul crushingly slow because my bosses said yes to one more tracking addon.
The obvious focus is deep learning, but those models need data, trained and deployed. There is a lot of work on those edges that require some AI knowledge. Also don't forget there are lots of AI techniques that are just as important now. So focus on an area that interests you, learn some of those AI methods and go! Every developer in the future will have to understand some of these methods at a basic level.
It's hard to just have a product centered around AI, but you can add AI to one of the components in your product. E.g. drawing a box around the face in a photo using face recognition technique is fairly approachable. It just means you need to call a library. A lots of the AI are well understood and encapsulated.
Prisma is an example of small team doing something novel with AI. Over time they are going to have a huge amount of data to further improve their product.
A friend of mine was an early investor in DeepMind. For like a year and a half, because that's how long it took Google to buy them out for somewhere around $400mm.
At the time, I thought that sounded like an amazing exit (OK, it still sounds like an amazing exit), and wasn't clear how Google could get that value out in a reasonable timeframe.
I was so, so, wrong. The amount of value hidden and public in that acquisition is astounding. Whoever put it together deserves a massive bonus, ideally in Alphabet stock.
MS putting $1bn a year in on AI is a catch-up game. They may do very well at it, but make no mistake -- we are only seeing the public side of the value Google is generating. I don't imagine we'll ever see blogposts about how they're tuning adwords using AI, for instance. But you can bet the same sort of gains they are seeing with translations, audio generation, game playing they are seeing in the ad space.
Remember that this is, in large part, a restructuring of existing orgs and investments. I wouldn't interpret this as the starting gun for AI/deep-learning investments at Microsoft. There's been heavy R&D investment for years in this space making its way into products like Cortana, Bing, PowerBI, Skype, Cognitive Services, etc. Similarly, there's a non-public side to the results at Microsoft I'm sure. So I don't think catchup is a fair analysis.
Here's what I imagine - Google will move from a PUSH model where advertisers specify demographics to sell to, to a model where Google tells the advertisers or product owners that they have 200,000 people they could target their ad to and get a 50% sell-through rate guaranteed, which they then take a much higher cut.
Or frankly they probably already do this, and are just increasing that sell-through rate via AI.
" and wasn't clear how Google could get that value out in a reasonable timeframe."
Tell us, how the valuation can be justified in terms of dollars.
Which 'AI' products are useful in the portfolio today, that makes products useful to you and I, that are derived from DeepMind?
The fact is - 'AI' is really short-hand for Multi-layer Neural nets - and they are applying those things in some very specific areas such as voice recognition and image recognition.
I think there will be many more places where we can do this - but I think it's going to be a very long time before we get to 'true AI'.
Admittedly, that's pretty 'cool' (pun intended) but not $400M kill. I think we still have a long way to go to see AI reap direct rewards for consumers. Payoffs have been minimal/incremental for now.
Just the PR value from Alphago is probably worth more than $400mm. Traditionally google has significantly delayed announcements of very interesting tech for years; essentially waiting until they had replaced an entire technology before announcing the old one.
It seems kind of unlikely that Google is doing that in this case, (if so, the rev rate on new AI tech is insane, and, well, I'm waiting for our AI overlords politely), but I would be very surprised if they were pushing out news about their most bleeding-edge tech.
"Just the PR value from Alphago is probably worth more than $400mm. "
No - it's definitely not.
Go and walk down the street in any part of the world and ask how many people know about this acquisition? Nobody will.
Even the vast majority of technical people will have never heard of this acquisition.
There is no 'PR' at all, really, from this acquisition.
I don't think any acquisition in history was worth it's 'PR' value.
Unless there is mass market news on it, in which households are getting to know about it, then possibly, but even then, I can't think of a single acquisition that was worth it.
Zuckerberg bought some tech that enables you to wear 'masks' on your face while on video - when he made the announcement, it was fun (he was wearing and Iron Man mask) - and it got picked up globally. If they paid less than $2M for that company, maybe it was worth it for the 'PR value'. But even then it's shaky.
I'm not sure if that is the right metric for research groups. If someone said "the startup has a promising cure for cancer in trials" would you ask if it had produced revenue yet?
"the startup has a promising cure for cancer in trials"
A) First, I would ask what they likelihood they have in making through trials, what is the real 'cure rate', what is the operational cost of the therapy, what kind of cancer it cures (obscure?).
B) They are not curing cancer. They have some intangible AI technology that doesn't necessarily or may not ever do anything.
As far as 'cutting the cooling bills on infrastructure' - there's no reason to think that that was an issue they were looking at solving, and used the tech as an 'example case' - but that some other, normal technology couldn't have solved th e problem just as well.
B) To confirm the ridiculous hype around this etheric technology, somebody, here on HN is equating this 'black box' to 'curing cancer'. Please.
There was an internet marketing manager on a podcast I listened to once that said the number of people using adblockers on the internet is essentially negligible, it's in the millions but a small percentage of all users (<5% if I recall).
This[1] is from a year ago, before ad-blocking became widely available on mobile, and they're still looking at hundreds of millions of users using adblockers.
I don't know how reputable PageFair is, but they estimate more than 20% of smartphone users are adblocking now[2].
That sounds like a very high number. At least anecdotally, I'm not aware of anyone in my circle of influence who uses a mobile ad blocker. My suspicion would be that less people use mobile ad blockers than on a browser.
Reading this carefully, the 5000-person group is "AI and Research" not just AI. (5000 people would be a lot to have working on AI.) This group includes Bing, Cortana and the current research group, so there are a lot of people not working directly on AI. That said, it is a significant change in focus, making AI a priority.
Correct. The team still includes everyone who works on Bing and essentially the Non Office part of what Qi used to run. Its less 5000 person AI Group as 5000 people running various businesses whose collective mantra is now AI.
We seem to be at an inflection point with AI -- companies across the board are investing billions of dollars into AI R&D and I expect that we'll start to see some really amazing products and services coming out of this in the coming decades.
This is a real concern. But as I see it, the difference is that the last time around, everyone was sold on the potential applications -- which turned out to be way harder than anyone realized.
This time around, the technology is seeing real applications today, so the valuations are more grounded in reality. So while people are definitely investing based on the future potential, the worst case scenario -- that we're going to hit a wall next week where no further progress can be made in machine learning research -- wouldn't be as devastating as the last AI winter.
At this point there's a lot of work to do, and money to be made, applying the current state of the art even if no further progress can be made.
The next AI winter will come when we maxed out on the current technology, i.e. when model training will become too expensive. If all that is left to do is to run training on top of all of Youtube videos, one might have to wait for 10 years until that becomes feasible.
It probably will be something more like the Kinect: a few products come out that aren't useful at all.
It took decades before people figured out Aluminum was actually useful for things for example. It was a chemical curiosity for the later-half of the 1800s (hmm, this is a cheap metal that is found everywhere. But its weaker than steel, what should we use it for?)
Just because you discover something useful doesn't mean you figure out what to do with it.
Yup. The reason why it was not used for a long time is that it was hard to make. Aluminum is a very reactive metal; when combined with oxygen or other stuff, it sits at the bottom of a deep pit of energy. It takes a lot of effort to get it out of there. The electrolytic process is basically a brute-force approach: throw enough energy at anything, and it will start moving eventually.
Napoleon III had his fancy-dinner utensils made from aluminum, for those occasions when gold did not seem lavish enough. And then cheap manufacturing was invented, and the rest is history.
I know IBM and Watson often get a bad rap here. But IBM made the same exact move almost 3 years ago with the creation of the Watson unit. There is way too much PR around Watson, but IBM should be credited for having called the current round of AI investments way before anybody else.
Disclosure: I was part of the initial IBM Watson team, left recently.
I love the fact that it took Ballmer to leave for them to really get serious about cutting-edge tech again. Bill Gates are probably wishing he never met Ballmer at Harvard.
This was Satya sending the strong message that no-one is untouchable and everyone must get inline or they will be next. It's actually a common strategy that requires the victims to be their best people. Obviously you can't to this too many times. Politics as usual.
Can you elaborate on how that works? If the victims are their best people who are doing solid, profitable work, laying them off seems like it would create a state of uncertainty which, in this climate, might just cause people to jump ship because competitive jobs are just around the corner.
Or was there some issue with the group where they had to sacrifice the good with the bad due to lack of overall profitability?
It's about internal power, even at the expense of the companies wellbeing. These layoffs were in the context of a larger set of layoffs that Satya undertook as his first big move at MS. Satya was getting across the board pushback from lots of different warring orgs at MS. He even had to bring the Big G out of retirement to help him out. So by taking out the top dogs he becomes the top dog. I don't even blame Satya for doing it - it's human nature that required it. If he didn't do it he would have eventually been pushed out of MS for not getting anything done. It's why drug cartels are so violent, it's why our election choices are Clinton and Trump. At least he's less evil than Sinofsky and Ballmer who were both cartoonishly horrible.
Good people have been leaving MS for a very long time. But big company corporate politics are really toxic in the US. MS is actually one of the better ones. Intel, Amazon, Netflix are much worse. Facebook and Google are on the decline. Yahoo.... yeah
As for this AI thing, my bet is that it is mostly fluff. Most people will ignore it until the next re-org comes along. There was a similar thing with Big Data at MSR a number of years ago and look what happened there.
You think you are better than the ML group? No? See I've just fired all of them. Don't think for a second I won't fire your ass on a whim. Now get to work.
So, Microsoft is going to put 5000 people
on applications of artificial intelligence
(AI). Likely they will also include
machine learning (ML).
IMHO, there is some value there.
But, IMHO they would be better off just
drawing all they can, including AI/ML but
much more from the QA section of research
libraries. There they will find oceans of
material, where in comparison AI/ML look
like farm ponds, in pure and applied math
as math but also operations research,
statistics, optimization, control theory,
applied probability, stochastic processes,
mathematical finance, mathematical parts
of high end electronic engineering, signal
processing, experimental design,
quantitative methods in business, and much
more.
Despite the general positive spin around it ("we did it as a learning project"), most people would agree that Tay was both a technical and a PR failure.
But the pattern does repeat: Microsoft releases an AI which fails. Tesla's autopilot cannot "see" white object on white background. Apparently, Google also had a crash which is recently being claimed as human error. My guess is that this list is not going to stop here.
Suppose I ask you to build me a teleporting machine. You try, and like the movie Spaceballs, my torso and up comes out aligned wrong. This is now declared part of the iterative learning process, except that the cost borne by the corporations for the failure is quite minuscule compared to the cost borne by the affected party (risk asymmetry).
So while people talk about the huge advancements in AI, shouldn't we be quite skeptical especially at this point? Since none of us have seen the alternate parallel universes, and considering
a) the resources being thrown at the problem
b) the risk asymmetry involved
c) the privacy intrusion involved in the data collection (you knew I would bring it up, didn't you?) and not to mention
d) the inability of anyone to demand any kind of transparency from these AI pioneers
I can as well ask, are we as a society paying too high a cost for this progress? Could we really not do any better than this?
Certainly interesting ideas, but we only need an AI that is better than the current paradigm. If Teslas doesn't detect a white object, it is not as heavy as a human sleeping in the wheel. So if Tesla's kill 30k people a year, but humans die in the wheel at 120k people a year, an improvement would be nice, as it avoids 90k deaths a year. So if we are already using teleporting machines (and we accept the tradeoffs of using it, like cars) and 1 in a million fails, if AI makes a failure in 1 in 10 million, clearly we should use that technology right?
So as Elon musk said recently: "Whatever this thing is you are trying to create.. What would be the utility delta compared to the current state of the art times how many people it would affect?"
The wonderful thing about this AI algorithms is that we can rate them on their efficacy, they might be a black box, but the input and output are always known. If we see that google crashes 10x more cars, we wouldn't use their AI.
> only need an AI that is better than the current paradigm
This is fundamentally what is being debated here. While the current paradigm can seem fairly poor, let us consider a few things which are true for the human driver.
1. He/she puts himself ALSO at risk, as opposed to the self driving system (remember it is theoretically possible for the self-driving car to not have any occupants at all. It is potentially only a matter of time before it happily wades through stand-still traffic to go and buy grocery for you).
2. He/she is not, in the process of being/becoming a good driver, also taking away personal freedoms of other people - which is effectively what is happening when the megacorps collect any and every piece of data they encounter. In a recent article in the Economist, we hear about a system which augments the autonomous cars by mapping roads in extremely high resolution. [1] Remembering all the work Google does to occlude sensitive information from its maps, imagine how much more effort has to go into this system to have it occlude personal details completely. Now imagine this data (which is currently being collected by a third-party company) landing in the hands of Google/Tesla/Uber etc. who are going to combine it with other human oriented information (e.g. Bob always leaves his office at 5.00PM, and always swerves sharply to avoid the pothole at so and so corner street, let us add that info to our system and improve it).
3. If you think the above scenario is ridiculous, then the next thing you would probably ask for is accountability. In other words, at some point, you are going to ask these companies to open up their data collection processes and algorithms to the world. This is exactly what would happen if the entire thing were a completely OSS-based process. There isn't an equivalent problem for the human driver, because you have sufficient faith in a human's need for self-preservation that you will not demand a real-time thought reading machine which will warn oncoming traffic if the human driver is having an onset of road rage.
> So as Elon musk said recently: "Whatever this thing is you are trying to create.. What would be the utility delta compared to the current state of the art times how many people it would affect?"
This is also being debated. There are such things as side effects, and some of them are invisible. The current state of the art (i.e. the inefficiency, or rather the inadequacy, of humans to perform these tasks) does not, as a side effect, also rob society of its peace of mind. Imagine if, for every piece of information which is collected, you also had a tiny pebble placed somewhere in your neighborhood. Soon, by the time these systems have reached the utility delta that you are happy with, we might have a mountain the size of Everest. Will we? I don't really know. Because it is invisible. Some people would still be OK with it. But most people, hopefully, would want to see the size of the hill. Is it a molehill or is it really a mountain? The lack of accountability surrounding these questions is actually quite shocking to me. [2]
[2] Not to mention the other cascading side effects of the data collection process itself, such as your personal data, which you don't even know how it was collected, being collated to be made sense of and sold to the highest bidder
I am intrigued by the mention of "Monthly Q&A" at the end of the email. Is this Satya Nadella's version of Google's TGIF meetings? If so I heartily approve.
I left Microsoft for Google in 2010 and TGIF Q&A was one of the things I appreciated the most about Google culture (despite the occasional screwball live question). I think any company could benefit from a similar tradition.
mostly Seattle region. Since half of the division will be made up of Bing folks and there is a bunch of empty office space in Bellevue I would assume most would be clustered in the Bellevue office(as opposed to the regular Redmond Campus)
It's funny, the internet isn't old enough to find links to the CYC project in Austin that blew through hundreds of millions of DoD money in the 80's and early 90's.
In the past few years, both the hardware caught up to and new practical techniques have been developed for multilevel neural nets (deep learning). It turns out that these are very effective at many problems that were not as amenable to earlier techniques. The theory of neural nets has been around a long time, but there has been a recent increase in the practical applicability of them.
How do you really start a 5000 person division at once and expect to succeed?
I assume AI development is a niche field. And you would want smaller dedicated teams of brilliant researchers and practitioners focusing on a single problem.
I can't imagine the overhead in maintaining and operating such a large division. I hope they know what they are doing.
The bulk of 'practical AI' don't come from massive products, like 'Office' - but from very specific neural algorithms, applied to very specific things - done by a few people.
That said - this kind of research may benefit from a lot of concurrent research.
Also - there are a couple of strategic issues:
1) Prestige - it's important to be recognized as 'a leader' to maintain brand cachet among tech talent
2) Talent Hoarding - Google, FB and MS are each big enough to tilt the landscape in any specific field. It's actually economically viable for them to pay the best talent to sit in a lab and fiddle, even while accomplishing little, over letting the talent go to competitors.
5000 more people working hard to make themselves obsolete. I wonder how long it will take those working on AI to figure out that they're doing to themselves and the rest of the IT industry what the IT industry has already done to many others. If and when they succeed we'll finally know the true meaning of the term disruption.
So OpenAI is spending a billion dollars over the next several years.
Microsoft is spending a billion dollars per year.
Google, etc. do the math.
There's literally billions of dollars now being spent on moving deep learning forward. Pretty amazing when I think back to 2011 and there were machine learning conferences where literally no one I spoke to had heard of deep learning.
People worrying about a second AI winter are like the people that have been worrying about an "internet bubble" since 2004. It's fine to be worried, there will be bubbles, but this time it's different and there are many reasons for that. There is no "internet industry" anymore; it's become more segmented and just plain bigger. Similarly, there will be no "AI industry"; it will branch out, and there are more potential applications of "understanding data and automating decisions" than there were in the 80s.
> something that they are branding AI - I might wait and see how much of it is on real AI - what ever that is.
Yeah, I wouldn't get hung up on that, they're calling it AI because that's easier to explain to reporters than deep learning. The difference is deep learning techniques are already being used in released products and companies are looking to do more of that, so there is a very real definition and set of goals associated with what these groups are doing, it's not just, "hey everyone, let's make an ai!"