NLP Papers at ICML2022
Summarising ideas (usually methods) from some of the NLP papers at ICML.
Canβt comment much about trends as they are mostly one to two papers in each sub NLP category. Although a common thread is building off your own work. Many times I wondered why a paper used some particular less well-known model or method and the answer almost always can be found in the list of authors. Apparently happens alot in ICML in general.
Relying on GPT3 prompt paradigm
- Co-training improves prompt-based learning for Large LMs
- Blackbox tuning for Language Models as a service
Fancy Inference
- Latent Diffusion Energy-based model for interpretable Text Modeling
- Controlling conditional language models without catastrophic forgetting
Probably Novel
Empirical ++
Architecture
- Towards Coherent and Consistent Use of Entities in Narrative Generation
- Improving Language Models by Retrieving from Trillions of Tokens
- Improving Transformers with Probabilistic Attention Keys
Intepretability
High Level non NLP Impressions
Neural Architecture papers appear to be at the level of controllable multi-task mixture of expert stuff. People are trying to combine meta-learning and multi-task and subspace learning.
Approximate Inference I feel like there were few variational inference advances, and also very few Bayesian NN at the conference. Variational Inference papers seemed to be mostly for specific architectures, and people seem to be working on sampling methods again.
NN Theory Empirical NN papers without theory were more popular at poster sessions than the two-layer NN theory proof papers. Simply more digestable/convincing/practical?!
Robustness, fairness, differential privacy Very popular. Big in the main conference and also in the workshops.
Acknowledgements
Conference attendance was generously supported by my advisor Kevin Duh. Also shout out to my ICML twin David Mueller who has his more ML focused coverage β> here!