DS1 spectrogram: Product-of-Experts Training Reduces Dataset Artifacts in Natural Language Inference

Product-of-Experts Training Reduces Dataset Artifacts in Natural Language Inference

April 21, 20262604.19069

Authors

Aby Mammen Mathew

Abstract

Neural NLI models overfit dataset artifacts instead of truly reasoning. A hypothesis-only model gets 57.7% in SNLI, showing strong spurious correlations, and 38.6% of the baseline errors are the result of these artifacts.

We propose Product-of-Experts (PoE) training, which downweights examples where biased models are overconfident. PoE nearly preserves accuracy (89.10% vs.

89.30%) while cutting bias reliance by 4.71% (bias agreement 49.85% to 45%). An ablation finds lambda = 1.5 that best balances debiasing and accuracy.

Behavioral tests still reveal issues with negation and numerical reasoning.

Resources

Stay in the loop

Every AI paper that matters, free in your inbox daily.

Details

  • © 2026 takara.ai Ltd
  • Content is sourced from third-party publications.