Rene Bidart

NIPS 2017 Highlights

Like many of the attendees, this was my first NIPS conference. Overall a great experience, and I doubt there is any substitute for meeting interesting people and learning about the current state of machine learning. Here I’ll mention a few of the highlights of the conference for me, and some of the weirdness.

Money is flooding in from everywhere in AI, with desperate investors trying to get a piece of the pie. Free alcohol and parties every night was unexpected for an academic conference:

Flo Rida
If this doesn't signal a bubble I don't know what does

With this level of hype there can be some downsides:

Conference lineup
Over 8000 people attended, and they all needed to register

Rigor vs. Accuracy

In the last few years the field of machine learning has become obsessed with the performance provided by larger and less interpretable deep neural networks. Not everyone is in love with the hype. According to Ali Rahimi:

Machine learning isn’t the new electricity, it is alchemy. We should start from simple, easy to understand things and then move onto more complex things. Currently we apply brittle optimization techniques to loss surfaces we don’t understand

Yann LeCun’s response:

We often use methods before we fully understand them in engineering and it has made the world a better place. Don’t throw the baby out with the bathwater

There was also a debate about model interpretability, where many of the same arguments are thrown around:

Do you really want to be diagnosed by something you don’t understand? vs.

We will choose an accurate model over an interpretable one when it is our body on the line. I think when deploying a model for real use we are facing a trade off between a complex model, and an ensemble of a simple model and human judgement. We know the complex model performs better on the test data, but we assume in real use the data distribution will change. The human/simple model will be more robust to changes in data distribution, because of the human’s debugging ability.

We must decide if the increased accuracy of the complex model makes up for the danger of it generalizing less well than the human/simple model ensemble. Unfortunately I have no idea how to quantify how much better one model may generalize than the other, so we will continue arguing.


ML for Healthcare

Fei-Fei Li and the Stanford group noted that there are plenty of issues throughout the hospital — not just in pathology and radiology. They proposed some interesting problems and solutions:

Greg Corrado gave a interesting talk about how ML will change medicine:

  1. Diagnostics — We know there will be a massive expansion in the use of imaging and the information that we get out of it. The question now is how do doctors and machines collaborate? There is always the problem of doctors not using the technology if it disrupts their workflow. We have lots of advances in pathology and radiology so far, and this will start saving lives very soon.
  2. Care management and decision support — Electronic medical records are a mess with many coding systems and inconsistencies, and most data isn’t used by humans or Ml methods. Unstructured data is more difficult than something like radiology, but promising results have been shown. There is a lot of promise in making EHR smart, but this may take decades. Medical records ->fihr resources
  3. Personalized medicine — We are just getting started on this, but there are promising results in genetics and eventually this will be a crucial part of healthcare. Eventually we will use some kind of omics for all routine medical prescription, but this may be a long time away.

Medical Imaging

Practical Concerns


Fairness

Now that we have seen so many successes in machine learning there’s increased interest in promoting fairness in AI. Here are some take aways from the presentations on bias in AI:

These discussions were useful, but in the end we were left with more questions than answers:

Even for someone who isn’t particularly interested in fairness, it is important to pay attention to this work. Now that these discussions are happening, ignorance will not be an excuse for mistakes like the Google made, and there will be bigger repercussions for someone who harms marginalized groups through machine learning.


Meta Learning/Reinforcement Learning/Self-Play/Generalization

I knew nothing of this field before the conference, but this is the topic I found most interesting. I will be cracking open the deep RL book soon, but since that hasn’t happened yet don’t trust anything written here.

Josh Tenenbaum explained that human learning doesn’t happen at an single level. There are different processes in our brain, operating at different time scales:

  1. Perception (<1 second)
  2. Thinking (<1min)
  3. Learning (>1minute)
  4. Development(lifetime)
  5. Evolution(many lifetimes)

A lot of work seems to be trying take the idea that we learn in different ways in different time scales into account in order to make models generalize better or train faster:


Random Stuff: