On the Relation between Sensitivity and Accuracy in In-context Learning
In-context learning (ICL) suffers from oversensitivity to the prompt, which makes it unreliable in real-world scenarios. We study the sensitivity of ICL with respect to multiple types of perturbations. First, we find that label bias obscures true ICL sensitivity, and hence prior work may have significantly underestimated the true ICL sensitivity. Second, we observe a strong negative correlation between ICL sensitivity and accuracy, with sensitive predictions less likely to be correct. Motivated by these observations, we propose SenSel, a few-shot selective prediction method based on ICL sensitivity. Experiments on ten classification benchmarks show that SenSel consistently outperforms a commonly used confidence-based selective prediction baseline.
READ FULL TEXT