In-Context Learning for Extreme Multi-Label Classification

Abstract

Multi-label classification problems with thousands of classes are hard to solve with in-context learning alone, as language models (LMs) might lack prior knowledge about the precise classes or how to assign them, and it is generally infeasible to demonstrate every class in a prompt. We propose a general program, π™Έπš—πšπšŽπš›--πšπšŽπšπš›πš’πšŽπšŸπšŽ--πšπšŠπš—πš”, that defines multi-step interactions between LMs and retrievers to efficiently tackle such problems. We implement this program using the π™³πš‚π™Ώπš’ programming model, which specifies in-context systems in a declarative manner, and use π™³πš‚π™Ώπš’ optimizers to tune it towards specific datasets by bootstrapping only tens of few-shot examples. Our primary extreme classification program, optimized separately for each task, attains state-of-the-art results across three benchmarks (HOUSE, TECH, TECHWOLF). We apply the same program to a benchmark with vastly different characteristics and attain competitive performance as well (BioDEX). Unlike prior work, our proposed solution requires no finetuning, is easily applicable to new tasks, alleviates prompt engineering, and requires only tens of labeled examples.


Research paper below links to GitHub repo
Link
We care about your privacy so we do not store nor use any cookie unless it is stricly necessary to make the website to work
Got it
Learn more