Embrace the random

One of the strongest opening lines is from this seminal book on pattern recognition (aka machine learning):

“Life is just a long random walk.”

It stuck into my mind. All I could add was “perhaps, maybe, with a drift” that you control, but that’s it. Our life is full of random events that we later try to fit into a well-constructed narrative, with more or less success.

There are several systems that proudly embrace the random: the US green card lottery, the Vietnam war draft lottery, or, more recently, the Paris medical student selection draw, not without controversy of course. There are others which cling to the deterministic narrative, even at the price of several person-years of work. One of these is our selection procedures of papers in high-profile scientific conferences.

I was an area chair of ICML this year. I had sixteen papers. Two of them were strong accepts with three positive reviews, some enthusiastic. Eight of them were clear rejects, half-written, technically flawed, irrelevant, or without novelty (yes, a surprisingly high fraction). The remaining six were what we call borderline. They were all technically correct, novel, relevant, and clearly written. Reviews varied because of “significance”, a highly subjective criteria based on 80% reviewer taste and 20% experience. The unofficial target acceptance rate (dictated by two factors, largely irrelevant to paper quality: the number of submitted papers and the capacity of the conference program) was 25%, which meant that roughly one third of the borderline papers could be accepted. How did we make the decision? With a lot of hard head-scratching, reading and re-reading the paper and the reviews, initiating reviewer discussions, and a lot of work of the program chairs (I don’t think Nina and Kilian slept a lot that week). What I found the most ludicrous is the realigning of reviewer score: to sweeten the bitter pill of rejection, we pressure the positive reviewers to lower their marks, to make it look like the decision was non-random.

I think we should make a random decision on borderline papers (possibly with a biased coin for authors with multiple papers or with other penalties to discourage flooding the system), or accept all of them but randomly select those to be presented. By borderline papers, again, I mean: technically correct, clearly written, relevant, and novel, but with questionable significance.

Here are my arguments.

  1. We have known it’s random since the now famous NIPS experiment. The numbers above (12% clear accept, 12% random out of 40% borderline) are actually fitting the data. My suggestion does not make it more or less random, it makes us to embrace its randomness.
  2. It is cost effective. We spend about 25 to 50 person-years to review the yearly batch of ICML and NIPS papers. That is a full research carrier. Any savings there can go into other production.
  3. Reviews can be more objective. We could concentrate on clarity, technical quality, relevance (which is not the same as significance), and novelty. Significance is hard to judge and highly subjective. As a reviewer, I’m there to keep up paper quality. But who am I to decide what people should work on?
  4. It’s fair and psychologically easier to accept.
  5. It randomizes search. We are arguably fixating too much and too fast on certain topics. This generates our famous waves of Hollywoodesque hypes, bubbles, and stars. Randomization alone will not solve this but it will alleviate it.

If you like what you read, follow me on Medium, LinkedIn, & Twitter.

--

--

--

Head of AI research, Huawei France, previously head of the Paris-Saclay Center for Data Science, co-creator of RAMP (http://www.ramp.studio).

Love podcasts or audiobooks? Learn on the go with our new app.

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Balázs Kégl

Balázs Kégl

Head of AI research, Huawei France, previously head of the Paris-Saclay Center for Data Science, co-creator of RAMP (http://www.ramp.studio).

More from Medium

Aim from Zero to Hero

Mathematicians’ relations and communities with each other

Quality Control of surface defects and inclusions using Image Segmentation and Deep Learning

Google’s LaMDA engineer got it wrong — and that is exactly what was meant to happen