The key point I kept picking up was the extent to which a press willing to laud a discovery was reticent about owning the clinb-down.
Peer review in ML journals should be tighter maybe? If you solve a limited subset of the three body problem you can't claim to solve "the three body problem" and if you apply a well known Rubik's cube model solution you didn't learn it, you had it baked in.
Some of the Google searches I have done recently turn up mostly automatically written articles that plug in numbers and facts but make no sense at all, because there is no real understanding. (Yes, this is probably not "real" AI, but it's the sort of thing people are and will be trying to apply AI to)
It seems to me that there is no way to get to a serene, well-functioning society that utilizes intelligent machines, because in the process of developing software that can digest and write intelligently about topics humans are interested in, we will inevitably come up with less intelligent software first, that produces plausible misinformation far cheaper than humans. And economics will drive out, is already driving out, real research and journalism. Human society would be drowned by automatically generated misinformation by the time a machine is truly intelligent, whether that is in a couple years or a decade or longer.
I'm increasingly thinking that the people who are worried about artificial intelligence ending the world as we know it have slightly missed the mark - superhuman intelligence is a singularity that we are approaching, but we are going to be ripped apart by the tides beforehand and maybe that will prevent the endpoint from ever happening.
I think this is not only inevitable, but necessary! This time around it has been a lot more useful, due in no small part to the advance in hardware since 1970s.
Unfortunately this has caused many important people to believe far too much of the hype and not see it's current limitations. As a result they have started integrating it into important part of our societies - i find this alarming - not for the reason most people find it alarming i.e "because it's too smart", but because it's far far too dumb in combination with people assuming it's very smart. I think a lot of this problem stems from inappropriately anthropomorphising ML with terms like "AI" when we are no where near the stage that we need to have the philosophical debate about where something is sentient or "intelligent". The ML we are doing with NNs is still at the "tiny-chunk-of-very-specifically-engineered-piece-of-brain" stage. It's important people understand this before we start integrating what are essentially basic statistical mechanisms into our societies.
For those in pursuit of better ML and things like real AI aka AGI, I also think having the hype blow away will do more good in the form of clarity and lack of noise than it will harm in lack of funding.
The radiology thing, is it really the case that there are no startups in that area with useful AI software? Seems like he overstated that.
Part of this is a worldview difference. Many people truly believe that AGI is just around the corner, and that even before then, the narrow AI applications will significantly alter the world as they are deployed. And since no one can actually predict the future, it's an area where it's easy to have different worldviews.
Personally I think that it's true that there is a lot of poor reporting and companies that overhype results, but it also seems like people like Gary Marcus are really not keeping up to date with the true capabilities of DL systems. If he was up to date, why would he be so pessimistic about applications like radiology? There seem to already be a lot of strong results.
The conversation that I posted at:
https://medium.com/@scottlegrand/my-interview-with-the-world...
is performed in one shot. And I think it shows both the abilities and the limitations of GPT-2 and similar models. I am 100% role playing with the language model and prompting it to go in the directions it goes, but it surprised me several times, and eventually it all fell apart because I didn't perform any transfer learning on the model I just used the raw GPT-2 XL model to measure whether further work would be worth the effort and I would conclude yes it would be.
The first thing I need to do is dramatically increase the length of its input context. It's pretty good at running with an ensuing script because I suspect much of its training data was formatted that way. But since I ran out of context symbols, it eventually suffers from several incidents of amnesia and eventually effective multiple personality disorder. It also contradicts itself several times but then no more than the typical thought leader or politician does IMO.
What The Economist did was effectively erase the thing's memory between questions. So they were starting fresh with each question. And I think that's why they had to do the best of 5.
Hype lasts the next quarter but is replaced with distrust. Science is a long game of incremental discoveries. Breakthroughs usually are an understanding of some interesting outcome that needed more interpretation.
I don't know if GCE is better, but the temptation to overload the hardware is always there: hardware rental is fundamentally a business with low barriers to entry. Anyone can buy some machines, bring up a Kubernetes or OpenShift cluster and start renting it out. So the big 3 are always looking for proprietary advantage and dedicated AI chips are something other firms can't easily do at the moment, making it a good source of lockin.
Do many people need it though? Deep learning is pretty useless for most business apps, unless you happen to need an image classifier or something else pre-canned. Classical ML is often sufficient, or better, human written logic. The latter can be explained, debugged, rapidly improved and in the best case requires no training data at all!
GCE perf is significantly better (and more consistent) than Azure. Even with Windows instances. :/
But I agree that “cloud is slow” when compared to bare metal- it’s also most financially costly especially for the same performance due to it being slower. But the gains in flexibility are immeasurable.
No matter what DL or DRL does, or how little his own preferred paradigm does, Marcus will never ever admit anything. AlphaGo beats humans? Well, it just copied humans. AlphaZero learns from scratch? Well, he wrote a whole paper explaining how akshully it still copies humans because the tree search encodes the rules. MuZero throws out even the tree search's knowledge of rules? Crickets and essays about 'misinformation'.
Regarding the lack of transfer, yes, AlphaGo, AlphaZero and most of their variants have boards of fixed size and shape hard-coded in their architecture (as they have the types of piece moves-hard coded) and need architectural modifications and re-training before they can play on different boards or with different pieces (e.g. AlphaGo can't play Chess and Shoggi unmodified). The KataGo paper (the paper you linked) is one exception to this. Personally, I don't know others. Anyway general game-playing is a hard task and nobody claims it's solved by AlphaGo.
Regarding KataGo its main contribution is a significant reduction to the cost of training an AlpahGo variant while maintaining a competitive performance. This is very promising- after DeepBlue, creating a chess engine became cheaper and cheaper until they could run on a smartphone. We are far from that with Go computer players.
However, in the KataGo paper, major gains are claimed to come from a) game-playing specific or MCTS-specific improvements (playout cap randomisation, forced playouts and policy target pruning) or architecture-specific improvements (global pooling) or, b) domain-specific improvements (auxiliary ownership and score targets). Finally, KataGo has a few game-specific features (liberties, pass-alive regions and ladder features).
The KataGo paper itself says it very clearly. I quote, snipping for brevity:
Second, our work serves as a case study that there is still a significant efficiency gap between AlphaZero's methods and what is possible from self-play. We find nontrivial further gains from some domain-specific methods (...) We also find that a set of standard game-specific input features still significantly accelerates learning, showing that AlphaZero does not yet obsolete even simple additional tuning.
Finally, "it would obviously work so nobody tried" would make sense if it wasn't for the extremely competitive nature of machine learning research where every novel result is presented as a big breakthrough. Also, if something is obvious but never seems to make it to publication the chances are someone has tried and it didn't work as expected so they shelved the paper. We all know what happens to negative results in machine learning.
Specifically, MuZero uses MCTS and MCTS needs to have at the very least a move generator in order to produce actions that can then be evaluated for their results. The trained MuZero model learns the transition function and evaluation function but I don't see in the paper where it learns what actions are legal in the domain. And I don't understand how any architecture could model the possible moves in a game without observing examples of external play (i.e. not self-play).
MuZero reuses the AlphaZero architecture so most likely the moves of the pieces for Chess, Shoggi and Go are hard-coded in the architecture, as they are in AlphaZero. There's also probably some similar hard-coding of Atari actions, which I'm probably missing in the paper.
I wonder if it's because it's so vague and fuzzy, or because the techniques are so general. Like with self driving cars. The will someday probably be safer than people and that's huge. People get excited. We want that! Self driving cars save lives! But in those four sentences we went from "will someday probably" to let's do it now. The class of thing we're talking about now and the class of thing in the future are one and the same, so it's hard to talk about the future versions as distinctly separate from today's.
AI will probably be able to talk to people well. We have AI today that talks to people. They're not the same thing, but these two sentences don't make that clear, because it's all AI.
In principle, polynomials can learn any function! And we have polynomials today! We can learn anything! Rinse and repeat with fourier series or (as a totally random example) deep learning and it sounds like tomorrow's techniques are the same as today's, so we're done, right?
Or maybe it's on lay people's poor math and stats skills and lack of understanding of the simple stuff. If I tell lay people I do stats, they think I'm taking an average with a lot of bureaucracy they don't really get. They won't think I'm using simple logistic regression to do really cool stuff like classify documents. They didn't know "stats" could do that! So they might be even more misled about what I do if I call it "statistics" than if I call it "AI." If they're mislead whatever I call it, we're already screwed.
i think the trend started from the industry , but you re right warrantless self-promotion very pervasive in academia , and it's sad that it works!
Yes, in a way it must suck to always act like the grumpy party-crasher but Gary Marcus is absolutely spot on when it comes to the facts. That interview with the economist had me shaking my head, everyone had to know within five seconds that the chance that this is uncurated is zero.
He's not negative in this article.
He's right.
Shameless but relevant plug warning, I run this effort Skynet Today with the aim to let AI researchers/experts present various advancements/topics to a general audience without hype and with context. Everyone on the team is volunteering their time for free and we can always use more people to help us tackle various subjects, so if you care about this issue feel free to take a look here: https://www.skynettoday.com/contribute
Writing good articles- there's no other way to combat bad coverage. Keep it up please and all the best.
(To the extent that I have kept up with it) modern AI skips the 'knowledge base' part of ES, in favor of pattern-recognition based on 'training'.
Today's (Indeterministic, trained, n-net) AI has clearly saved a lot of time/effort in creating 'knowledge bases'. I suspect it appeals more to singular fantasies about 'more human than human' intelligence. (Sorry Ray)
Question is: Is today's AI even a magnitude-better than (deterministic) ES insofar as extensibility and verifiability? What if we had spent those decades refining the ES approach instead?
There's now discussion about how neural networks can succumb to data poisoning / adversarial attacks, because there are no immutable facts. Adding a mostly immutable fact table can help keep things grounded in reason. Most of these engines support complex inference abilities that can lead to unexpected connections.
ES is not really dead. It feels like many rules engines changed their names to "AI Intelligent Agents"-type wording to describe their product. Rete algorithm is similar rule based calculation, is still used to calculate FICO score, which you could say fits into the problems that may be better served by the latest Neural network models. Allegro graph lets you query using prolog and is often used for governance and compliance tools. RDFox is one of the latest inference engines that made major advancements in turning first order logic in datalog into parallel computation.
I'd imagine if you can build a neural network that can successfully interact with a ES knowledge base you could easily make a neural network as good as the one that won in jeopardy
We do still rely on expert systems for things that we want to be carefully parsed, verified and analyzed by people though. Such as rules that oversee most self-driving cars based on perception handled by neural networks. However, not all self-driving systems lean as heavily on rules at the top level as others.
Also, an hypothesis: if (big if) you somehow managed to have an expert system outperform deep learning for vision, I bet that it won't be any more verifiable than a deep neural net is today.
To me the complaint that modern deep learning is unverifiable is a bit dumb, in the sense that any perception algorithm working with low level signals (vision, sound) will not be transparent to a human. 15 years ago, an image classification pipeline looked like: bag of SIFT features + SVM classifier. Try explaining the decision made by that algorithm in an intuitive way!
The only risk is misappropriation of investment, but that has already been smoothed over quickly.
I’d get the complaint if AI was not doing anything useful as it wasn’t in ‘74, but it simply is.
For example, plate recognition, drones and facial recognition. These are brutally efficient technologies. To imagine any government under investing in these is a total misunderstanding of the value that AI brings to military, security, and economic dominance.
It’s a race between nations and the only danger is the computers taking over (not in a general AI way, but in a way that everything is run by algorithms no human can understand)
Without a doubt, there's an investment bubble for AI companies in industry: a huge amount of companies have added "AI" to their proposition, where the AI is only marginally useful if not useful at all, for the purpose of marketing and inflating valuation.
On research side, it's now clear the current AI paradigm will not produce the kind of massive, society-shifting promises that were made to investors, which is the focus of the article.
Yes deep learning is here to stay for narrow tasks, but the investment contraction for the rest could certainly feel like a Winter.
It's correct in that it won't produce the society-shifting promises. But there is a huge amount of money on non-society-shifting innovation.
Besides, I really doubt the media predictions are what actually was told to investors.
[1] https://www.wired.com/story/deepminds-losses-future-artifici...
I don't mean just AI, but transparency on everything: Transparency from researchers, news outlets, governments, retailers, and YouTubers. It's a great shame that a lot of IT has been used for bringing more confusion and obfuscation to the people, whereas it could be used to provide a more coherent/honest view of the world. We should demand more transparency on every matter.
>In 1966, the MIT AI lab famously assigned Gerald Sussman the problem of solving vision in a summer; as we all know, machine vision still hasn't been solved over five decades later.
It certainly took five extra decades, but it would be a massive shift of goalposts to say the problem of vision hasn't been sufficiently solved today.
>In November 2016, in the pages of Harvard Business Review, Andrew Ng, another well-known figure in deep learning, wrote that “If a typical person can do a mental task with less than one second of thought, we can probably automate it using AI either now or in the near future.” A more realistic appraisal is that whether or not something can be automated depends very much on the nature of the problem, and the data that can be gathered, and the relation between the two. For closed-end problems like board games, in which a massive amount of data can be gathered through simulation, Ng’s claim has proven prophetic; in open-ended problems, like conversational understanding, which cannot be fully simulated, Ng’s claim thus far has proven incorrect. Business leaders and policy-makers would be well-served to understand the difference between those problems that are amenable to current techniques and those that are not; Ng’s words obscured this. (Rebooting AI gives some discussion.)
It takes significantly longer than a second to actually understand spoken conversation (rather than provide a conditioned response or match against expected statements, both of which computers are fully capable of doing).
>I just wish that were the norm rather than the exception. When it’s not, policy-makers and the general public can easily find themselves confused; because the bias tends to be towards overreporting rather than underreporting results, the public starts fearing a kind of AI (replacing many jobs) that does not and will not exist in the foreseeable future.
Robotic manufacturing has already eliminated massive swaths of high paying jobs. Likewise, software has eliminated massive swaths of data entry and customer service jobs (with software being a particularly poor replacement for the latter, but still being put into widespread use to cut costs). And contrary to beliefs that new jobs will be created in IT, software is able to massively eliminate low skilled tech jobs as well, as e.g. automated testing did to India's IT industry.
As with existing jobs that have been automated away, companies won't need generalized AI to eliminate many more jobs. Many jobs don't rely on unconstrained complex deduction and thinking, and will be ripe for replacement with deep learning algorithms. And we can be reliably assured that corporations will engage in such replacements even when the outcomes are not up to par.
> The Economist [...] said that GPT-2’s answers were “unedited”, when in reality each answer that was published was selected from five options
> [Erik Bryjngjolffson] tweeted that the interview was “impressive” and that “the answers are more coherent than those of many humans.” In fact the apparent coherence of the interview stemmed from (a) the enormous corpus of human writing that the system drew from and (b) the filtering for coherence that was done by the human journalist.
If your success rate is ≥20%, the coherence is coming from the model, not the selection process. This is just basic statistics.
> OpenAI created a pair of neural networks that allowed a robot to learn to manipulate a custom-built Rubik's cube
Jeez, I've already corrected you here... well, why not have to do it again?
> publicized it with a somewhat misleading video and blog that led many to think that the system had learned the cognitive aspects of cube-solving
The side not stated: OpenAI said explicitly in the blog that they used an unlearned algorithm for this, and sent a correction to a publisher that got this wrong.
> the cube was instrumented with Bluetooth sensors
During training, but they ended up with a fully vision-based system.
> even in the best case only 20% of fully-scrambled cubes were solved
No, 60% of fully scrambled cubes were solved. 20% of maximally difficult scrambles were solved.
> one report claimed that “A neural net solves the three-body problem 100 million times faster” [...] but the network did no solving in the classical sense, it did approximation
All solvers for this problem are approximators, and vice-versa. The article you complain about states the accuracy (“error of just 10^(-5)”) in the body of text.
> and it approximated only a highly simplified two degree-of-freedom problem
As reported: “Breen and co first simplify the problem by limiting it to those involving three equal-mass particles in a plane, each with zero velocity to start with.”
> MIT AI lab famously assigned Gerald Sussman the problem of solving vision in a summer [https://dspace.mit.edu/handle/1721.1/6125]
I... sigh
“The original document outlined a plan to do some kind of basic foreground/background segmentation, followed by a subgoal of analysing scenes with simple non-overlapping objects, with distinct uniform colour and texture and homogeneous backgrounds. A further subgoal was to extend the system to more complex objects.
So it would seem that Computer Vision was never a summer project for a single student, nor did it aim to make a complete working vision system.”
http://www.lyndonhill.com/opinion-cvlegends.html
> Geoff Hinton [said] that the company (again The Guardian’s paraphrase), “is on the brink of developing algorithms with the capacity for logic, natural conversation and even flirtation.” Four years later, we are still a long way from machines that can hold natural conversations absent human intervention
‘Four years later’ to natural conversation is not a reasonable point of criticism when the only timeline given was ‘within a decade’ for a specified subset of the problem.
> [In 2016 Hinton said] “We should stop training radiologists now. It’s just completely obvious that within five years, deep learning is going to do better than radiologists.” [...] but thus far no actual radiologists have been replaced
So Hinton actually said “People should stop training radiologists now. It’s just completely obvious that within five years, deep learning is going to do better than radiologists, because it's going to be able to get a lot more experience. It might be 10 years, but we've got plenty of radiologists already.”
2019 is not 2026. “thus far no actual radiologists have been replaced” is thus not a counterargument.
> Andrew Ng, another well-known figure in deep learning, wrote that “If a typical person can do a mental task with less than one second of thought, we can probably automate it using AI either now or in the near future.” [...] Ng’s claim thus far has proven incorrect.
I agree. This quote captures the wrong nuance of the issue.
Well, finally finding one point by Gary Marcus that isn't misleading, I think I'm going to call this a day.
https://www.reddit.com/r/MachineLearning/comments/e453kl/d_a...