ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2105.13003
  4. Cited By
Rethinking InfoNCE: How Many Negative Samples Do You Need?

Rethinking InfoNCE: How Many Negative Samples Do You Need?

27 May 2021
Chuhan Wu
Fangzhao Wu
Yongfeng Huang
ArXivPDFHTML

Papers citing "Rethinking InfoNCE: How Many Negative Samples Do You Need?"

5 / 5 papers shown
Title
Can LLM-Driven Hard Negative Sampling Empower Collaborative Filtering? Findings and Potentials
Can LLM-Driven Hard Negative Sampling Empower Collaborative Filtering? Findings and Potentials
Chu Zhao
Enneng Yang
Yuting Liu
Jianzhe Zhao
G. Guo
Xingwei Wang
28
0
0
07 Apr 2025
Learning Temporal Distances: Contrastive Successor Features Can Provide a Metric Structure for Decision-Making
Learning Temporal Distances: Contrastive Successor Features Can Provide a Metric Structure for Decision-Making
Vivek Myers
Chongyi Zheng
Anca Dragan
Sergey Levine
Benjamin Eysenbach
OffRL
45
7
0
24 Jun 2024
CLCE: An Approach to Refining Cross-Entropy and Contrastive Learning for
  Optimized Learning Fusion
CLCE: An Approach to Refining Cross-Entropy and Contrastive Learning for Optimized Learning Fusion
Zijun Long
George Killick
Lipeng Zhuang
Gerardo Aragon Camarasa
Zaiqiao Meng
R. McCreadie
VLM
50
2
0
22 Feb 2024
Progressive Domain Adaptation with Contrastive Learning for Object
  Detection in the Satellite Imagery
Progressive Domain Adaptation with Contrastive Learning for Object Detection in the Satellite Imagery
Debojyoti Biswas
Jelena Tevsić
ObjD
16
4
0
06 Sep 2022
A Mutual Information Maximization Perspective of Language Representation
  Learning
A Mutual Information Maximization Perspective of Language Representation Learning
Lingpeng Kong
Cyprien de Masson dÁutume
Wang Ling
Lei Yu
Zihang Dai
Dani Yogatama
SSL
217
165
0
18 Oct 2019
1