Part of my series of notes from ICLR 2019 in New Orleans.
Meta-Learning Update Rules for Unsupervised Representation Learning
- want representations that capture high-level attributes of data, without using labels
- currrent approaches:
- hand-design loss functions (autoencoders)
- hand-design target statistics (sparsity)
- hand-design updates (GANS)
- mismatch between objective & desired task
- instead of designing by hand, metalearn what to learn with help from the supervised task directly
- want algorithm to be transferrable to different architectures
- cf. learning transferrable features
- metaloss to update learning rule which is used in inner loop
- inner loop is unsupervised, but outer meta-learning loop can use labeled data
- wow this is so obvious and wonderful
- don’t actually need inner loss at all, use a different learning rule
- learning rule parameterised by MLP, these parameters are what we learn to update
- generalises to new datasets, architectures, modalities
- e.g. metalearn on images and evaluate on text (!)
- main limitation is scale (of course)
- question asker: “just to be clear, there’s no way your method would find something like Deep InfoMax”
- (an impassioned defense of ML scientists in the face of meta-learning )
Temporal Difference Variational Auto-Encoder
- environment models for agents in RL
- how to model temporal data?
- want world state abstracted from observations
- make temporally extended predictions
- include uncertainty
- I… stopped paying attention, again, cut me some slack I’ve been doing pretty well
Transferring Knowledge across Learning Processes
- AI should leverage prior knowledge maximally
- one way is transfer learning – use source model
- can have information loss since model doesn’t know what the future task might be
- meta-learn instead!
- Leap: inductive bias over learning process
- takes into account (approximate) length of learning trajectory
- updates initialization to minimise trajectory length
- get meta-gradients (almost) for free
- more computationally efficient, scales beyond few-shot learning