Slide 1
Slide 1 text
Background & Research Question:
The performance of models are degrading by the lapse of time. One of
the solutions is re-training, but it requires huge computational cost. Can
we estimate the performance before re-training?
Key idea:
We use an efficiently computable metric Semantic Shift Stability (SSS)
based on the methodology of semantic shift analysis.
Contributions:
- We created models (RoBERTa and word2vec) that vary by time-series
and revealed the performance degradation via experiments on Nikkei
(Japanese) and NOW (English).
- Our experiments reported that a large time-series performance
degradation occurs in the years when SSS is smaller.
Future work:
More diverse dataset and model, and discussion in a persuasive manner.
Resources:
GitHub: https://github.com/Nikkei/semantic-shift-stability
Semantic Shift Stability: Efficient Way to Detect
Performance Degradation of Word Embeddings
and Pre-trained Language Models
Shotaro Ishihara, Hiromu Takahashi, and Hono Shirai (AACL-IJCNLP 2022, Long)
Fig. 1: Procedure to calculate SSS.
Fig. 2: Word2vec performance improvement
vs SSS of Nikkei (upper) and NOW.
Fig. 3: Nikkei RoBERTa performance degradation (not improvement) vs SSS.
coef: -0.4855
coef: -0.8861
Finding: Performance of models gets
worse with the corpus 2016 and 2020.