ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.04624
81
0

Static Word Embeddings for Sentence Semantic Representation

5 June 2025
Takashi Wada
Yuki Hirakawa
Ryotaro Shimizu
Takahiro Kawashima
Yuki Saito
ArXiv (abs)PDFHTML
Abstract

We propose new static word embeddings optimised for sentence semantic representation. We first extract word embeddings from a pre-trained Sentence Transformer, and improve them with sentence-level principal component analysis, followed by either knowledge distillation or contrastive learning. During inference, we represent sentences by simply averaging word embeddings, which requires little computational cost. We evaluate models on both monolingual and cross-lingual tasks and show that our model substantially outperforms existing static models on sentence semantic tasks, and even rivals a basic Sentence Transformer model (SimCSE) on some data sets. Lastly, we perform a variety of analyses and show that our method successfully removes word embedding components that are irrelevant to sentence semantics, and adjusts the vector norms based on the influence of words on sentence semantics.

View on arXiv
@article{wada2025_2506.04624,
  title={ Static Word Embeddings for Sentence Semantic Representation },
  author={ Takashi Wada and Yuki Hirakawa and Ryotaro Shimizu and Takahiro Kawashima and Yuki Saito },
  journal={arXiv preprint arXiv:2506.04624},
  year={ 2025 }
}
Comments on this paper