ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2501.00562
  4. Cited By
An Overview and Discussion on Using Large Language Models for Implementation Generation of Solutions to Open-Ended Problems
v1v2 (latest)

An Overview and Discussion on Using Large Language Models for Implementation Generation of Solutions to Open-Ended Problems

31 December 2024
Hashmath Shaik
Alex Doboli
    OffRLELM
ArXiv (abs)PDFHTML

Papers citing "An Overview and Discussion on Using Large Language Models for Implementation Generation of Solutions to Open-Ended Problems"

50 / 105 papers shown
Title
SimRAG: Self-Improving Retrieval-Augmented Generation for Adapting Large Language Models to Specialized Domains
SimRAG: Self-Improving Retrieval-Augmented Generation for Adapting Large Language Models to Specialized DomainsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2024
Ran Xu
Hui Liu
Jiapeng Liu
Zhenwei Dai
Yaochen Xie
...
Chen Luo
Yang Li
Joyce C. Ho
Carl Yang
Qi He
RALM
402
21
0
28 Jan 2025
RAG-Star: Enhancing Deliberative Reasoning with Retrieval Augmented
  Verification and Refinement
RAG-Star: Enhancing Deliberative Reasoning with Retrieval Augmented Verification and Refinement
Jinhao Jiang
Jiayi Chen
Junyi Li
Ruiyang Ren
Shijie Wang
Wayne Xin Zhao
Yang Song
Tao Zhang
LRM
186
28
0
17 Dec 2024
LLMs as Debate Partners: Utilizing Genetic Algorithms and Adversarial
  Search for Adaptive Arguments
LLMs as Debate Partners: Utilizing Genetic Algorithms and Adversarial Search for Adaptive Arguments
Prakash Aryan
231
5
0
09 Dec 2024
Controlling the Mutation in Large Language Models for the Efficient
  Evolution of Algorithms
Controlling the Mutation in Large Language Models for the Efficient Evolution of Algorithms
Haoran Yin
Anna V. Kononova
Thomas Bäck
Niki van Stein
155
7
0
04 Dec 2024
Skywork-Reward: Bag of Tricks for Reward Modeling in LLMs
Skywork-Reward: Bag of Tricks for Reward Modeling in LLMs
Chris Yuhao Liu
Liang Zeng
Qingbin Liu
Rui Yan
Jujie He
Chaojie Wang
Shuicheng Yan
Yang Liu
Yahui Zhou
AI4TS
228
205
0
24 Oct 2024
Optimal Design for Reward Modeling in RLHF
Optimal Design for Reward Modeling in RLHF
Antoine Scheid
Etienne Boursier
Alain Durmus
Michael I. Jordan
Pierre Ménard
Eric Moulines
Michal Valko
OffRL
317
16
0
22 Oct 2024
SBI-RAG: Enhancing Math Word Problem Solving for Students through
  Schema-Based Instruction and Retrieval-Augmented Generation
SBI-RAG: Enhancing Math Word Problem Solving for Students through Schema-Based Instruction and Retrieval-Augmented Generation
Prakhar Dixit
Tim Oates
145
4
0
17 Oct 2024
Preference Optimization with Multi-Sample Comparisons
Preference Optimization with Multi-Sample Comparisons
Chaoqi Wang
Zhuokai Zhao
Chen Zhu
Karthik Abinav Sankararaman
Michal Valko
...
Zhaorun Chen
Madian Khabsa
Yuxin Chen
Hao Ma
Sinong Wang
265
13
0
16 Oct 2024
Enhancing Multi-Step Reasoning Abilities of Language Models through Direct Q-Function Optimization
Enhancing Multi-Step Reasoning Abilities of Language Models through Direct Q-Function Optimization
Guanlin Liu
Kaixuan Ji
Ning Dai
Zheng Wu
Chen Dun
Q. Gu
Lin Yan
Quanquan Gu
Lin Yan
OffRLLRM
263
17
0
11 Oct 2024
TableRAG: Million-Token Table Understanding with Language Models
TableRAG: Million-Token Table Understanding with Language ModelsNeural Information Processing Systems (NeurIPS), 2024
Si-An Chen
Lesly Miculicich
Julian Martin Eisenschlos
Zifeng Wang
Zilong Wang
Yanfei Chen
Yasuhisa Fujii
Hsuan-Tien Lin
Zifeng Wang
Tomas Pfister
LMTDRALM
192
41
0
07 Oct 2024
Can We Further Elicit Reasoning in LLMs? Critic-Guided Planning with
  Retrieval-Augmentation for Solving Challenging Tasks
Can We Further Elicit Reasoning in LLMs? Critic-Guided Planning with Retrieval-Augmentation for Solving Challenging TasksAnnual Meeting of the Association for Computational Linguistics (ACL), 2024
Xingxuan Li
Weiwen Xu
Ruochen Zhao
Fangkai Jiao
Shafiq Joty
Lidong Bing
LRM
198
24
0
02 Oct 2024
HelpSteer2-Preference: Complementing Ratings with Preferences
HelpSteer2-Preference: Complementing Ratings with PreferencesInternational Conference on Learning Representations (ICLR), 2024
Zhilin Wang
Alexander Bukharin
Olivier Delalleau
Daniel Egert
Gerald Shen
Jiaqi Zeng
Oleksii Kuchaiev
Yi Dong
ALM
352
94
0
02 Oct 2024
EPO: Hierarchical LLM Agents with Environment Preference Optimization
EPO: Hierarchical LLM Agents with Environment Preference OptimizationConference on Empirical Methods in Natural Language Processing (EMNLP), 2024
Qi Zhao
Haotian Fu
Chen Sun
George Konidaris
216
21
0
28 Aug 2024
Speculative RAG: Enhancing Retrieval Augmented Generation through Drafting
Speculative RAG: Enhancing Retrieval Augmented Generation through Drafting
Zilong Wang
Zifeng Wang
Long Le
Huaixiu Steven Zheng
Swaroop Mishra
...
Anush Mattapalli
Ankur Taly
Jingbo Shang
Zifeng Wang
Tomas Pfister
RALM
254
68
0
11 Jul 2024
SeaKR: Self-aware Knowledge Retrieval for Adaptive Retrieval Augmented
  Generation
SeaKR: Self-aware Knowledge Retrieval for Adaptive Retrieval Augmented Generation
Zijun Yao
Weijian Qi
Liangming Pan
S. Cao
Linmei Hu
Weichuan Liu
Lei Hou
Juanzi Li
RALM
120
19
0
27 Jun 2024
Large Language Models as Surrogate Models in Evolutionary Algorithms: A
  Preliminary Study
Large Language Models as Surrogate Models in Evolutionary Algorithms: A Preliminary StudySwarm and Evolutionary Computation (Swarm Evol. Comput.), 2024
Hao Hao
Xiaoqun Zhang
Aimin Zhou
ELM
174
25
0
15 Jun 2024
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs
  with Nothing
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing
Zhangchen Xu
Fengqing Jiang
Luyao Niu
Yuntian Deng
Radha Poovendran
Yejin Choi
Bill Yuchen Lin
SyDa
263
237
0
12 Jun 2024
Enhancing Long-Term Memory using Hierarchical Aggregate Tree for
  Retrieval Augmented Generation
Enhancing Long-Term Memory using Hierarchical Aggregate Tree for Retrieval Augmented Generation
A. AadharshAadhithya
S. SachinKumar
Soman K. P
RALM
173
3
0
10 Jun 2024
GRAM: Generative Retrieval Augmented Matching of Data Schemas in the
  Context of Data Security
GRAM: Generative Retrieval Augmented Matching of Data Schemas in the Context of Data Security
Xuanqing Liu
Luyang Kong
Runhui Wang
Patrick Song
Austin Nevins
Henrik Johnson
Nimish Amlathe
Davor Golac
127
4
0
04 Jun 2024
Similarity is Not All You Need: Endowing Retrieval Augmented Generation
  with Multi Layered Thoughts
Similarity is Not All You Need: Endowing Retrieval Augmented Generation with Multi Layered Thoughts
Chunjing Gan
Dan Yang
Binbin Hu
Hanxiao Zhang
Siyuan Li
...
Lin Ju
Qing Cui
Jinjie Gu
Lei Liang
Jun Zhou
212
15
0
30 May 2024
Robust Preference Optimization through Reward Model Distillation
Robust Preference Optimization through Reward Model Distillation
Adam Fisch
Jacob Eisenstein
Vicky Zayats
Alekh Agarwal
Ahmad Beirami
Chirag Nagpal
Peter Shaw
Jonathan Berant
349
56
0
29 May 2024
HippoRAG: Neurobiologically Inspired Long-Term Memory for Large Language Models
HippoRAG: Neurobiologically Inspired Long-Term Memory for Large Language ModelsNeural Information Processing Systems (NeurIPS), 2024
Bernal Jiménez Gutiérrez
Yiheng Shu
Yu Gu
Michihiro Yasunaga
Yu-Chuan Su
RALMCLL
264
101
0
23 May 2024
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via
  Reinforcement Learning
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2024
Yuexiang Zhai
Hao Bai
Zipeng Lin
Jiayi Pan
Shengbang Tong
...
Alane Suhr
Saining Xie
Yann LeCun
Yi-An Ma
Sergey Levine
LLMAGLRM
283
128
0
16 May 2024
Self-Play Preference Optimization for Language Model Alignment
Self-Play Preference Optimization for Language Model Alignment
Yue Wu
Zhiqing Sun
Huizhuo Yuan
Kaixuan Ji
Yiming Yang
Quanquan Gu
466
193
0
01 May 2024
Reducing hallucination in structured outputs via Retrieval-Augmented
  Generation
Reducing hallucination in structured outputs via Retrieval-Augmented Generation
Patrice Béchard
Orlando Marquez Ayala
LLMAG
194
107
0
12 Apr 2024
WESE: Weak Exploration to Strong Exploitation for LLM Agents
WESE: Weak Exploration to Strong Exploitation for LLM Agents
Xu Huang
Weiwen Liu
Xiaolong Chen
Xingmei Wang
Defu Lian
Yasheng Wang
Ruiming Tang
Enhong Chen
LLMAG
141
9
0
11 Apr 2024
sDPO: Don't Use Your Data All at Once
sDPO: Don't Use Your Data All at Once
Dahyun Kim
Yungi Kim
Wonho Song
Hyeonwoo Kim
Yunsu Kim
Sanghoon Kim
Chanjun Park
134
40
0
28 Mar 2024
ExploRLLM: Guiding Exploration in Reinforcement Learning with Large Language Models
ExploRLLM: Guiding Exploration in Reinforcement Learning with Large Language ModelsIEEE International Conference on Robotics and Automation (ICRA), 2024
Runyu Ma
Jelle Luijkx
Zlatan Ajanović
Jens Kober
LM&RoLRM
321
18
0
14 Mar 2024
HRLAIF: Improvements in Helpfulness and Harmlessness in Open-domain
  Reinforcement Learning From AI Feedback
HRLAIF: Improvements in Helpfulness and Harmlessness in Open-domain Reinforcement Learning From AI Feedback
Ang Li
Qiugen Xiao
Peng Cao
Jian Tang
Yi Yuan
...
Weidong Guo
Yukang Gan
Jeffrey Xu Yu
D. Wang
Ying Shan
VLMALM
214
13
0
13 Mar 2024
Large Language Model-Based Evolutionary Optimizer: Reasoning with
  elitism
Large Language Model-Based Evolutionary Optimizer: Reasoning with elitism
Shuvayan Brahmachary
Subodh M. Joshi
Aniruddha Panda
K. Koneripalli
A. Sagotra
Harshil Patel
Ankush Sharma
Ameya Dilip Jagtap
Kaushic Kalyanaraman
LRM
266
41
0
04 Mar 2024
Balancing Exploration and Exploitation in LLM using Soft RLLF for
  Enhanced Negation Understanding
Balancing Exploration and Exploitation in LLM using Soft RLLF for Enhanced Negation Understanding
Ha-Thanh Nguyen
Ken Satoh
185
3
0
02 Mar 2024
Provably Robust DPO: Aligning Language Models with Noisy Feedback
Provably Robust DPO: Aligning Language Models with Noisy Feedback
Sayak Ray Chowdhury
Anush Kini
Nagarajan Natarajan
257
94
0
01 Mar 2024
Can GPT Improve the State of Prior Authorization via Guideline Based
  Automated Question Answering?
Can GPT Improve the State of Prior Authorization via Guideline Based Automated Question Answering?
Shubham Vatsal
Ayush Singh
Shabnam Tafreshi
LM&MAAI4MH
90
4
0
28 Feb 2024
Smaug: Fixing Failure Modes of Preference Optimisation with DPO-Positive
Smaug: Fixing Failure Modes of Preference Optimisation with DPO-Positive
Arka Pal
Deep Karkhanis
Samuel Dooley
Manley Roberts
Siddartha Naidu
Colin White
OSLM
296
202
0
20 Feb 2024
MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop
  Queries
MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries
Yixuan Tang
Yi Yang
RALM
214
172
0
27 Jan 2024
When Large Language Models Meet Evolutionary Algorithms: Potential Enhancements and Challenges
When Large Language Models Meet Evolutionary Algorithms: Potential Enhancements and Challenges
Wang Chao
Jiaxuan Zhao
Licheng Jiao
Lingling Li
Fang Liu
Shuyuan Yang
347
18
0
19 Jan 2024
Evolutionary Computation in the Era of Large Language Model: Survey and
  Roadmap
Evolutionary Computation in the Era of Large Language Model: Survey and Roadmap
Xingyu Wu
Sheng-hao Wu
Jibin Wu
Liang Feng
Kay Chen Tan
ELM
381
113
0
18 Jan 2024
Self-Rewarding Language Models
Self-Rewarding Language Models
Weizhe Yuan
Richard Yuanzhe Pang
Kyunghyun Cho
Xian Li
Sainbayar Sukhbaatar
Jing Xu
Jason Weston
ReLMSyDaALMLRM
718
436
0
18 Jan 2024
Secrets of RLHF in Large Language Models Part II: Reward Modeling
Secrets of RLHF in Large Language Models Part II: Reward Modeling
Bing Wang
Rui Zheng
Luyao Chen
Yan Liu
Jiajun Sun
...
Tao Gui
Xipeng Qiu
Xuanjing Huang
Zuxuan Wu
Yuanyuan Jiang
ALM
265
134
0
11 Jan 2024
Chain-of-Table: Evolving Tables in the Reasoning Chain for Table
  Understanding
Chain-of-Table: Evolving Tables in the Reasoning Chain for Table UnderstandingInternational Conference on Learning Representations (ICLR), 2024
Zilong Wang
Hao Zhang
Chun-Liang Li
Julian Martin Eisenschlos
Vincent Perot
...
Lesly Miculicich
Yasuhisa Fujii
Jingbo Shang
Chen-Yu Lee
Tomas Pfister
ReLMLMTDLRM
175
183
0
09 Jan 2024
A Survey of Reinforcement Learning from Human Feedback
A Survey of Reinforcement Learning from Human Feedback
Timo Kaufmann
Paul Weng
Viktor Bengs
Eyke Hüllermeier
OffRL
190
253
0
22 Dec 2023
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator
Chengshu Li
Jacky Liang
Andy Zeng
Xinyun Chen
Karol Hausman
Dorsa Sadigh
Sergey Levine
Fei-Fei Li
Fei Xia
Brian Ichter
LLMAGLRM
230
116
0
07 Dec 2023
Retrieval-augmented Multi-modal Chain-of-Thoughts Reasoning for Large
  Language Models
Retrieval-augmented Multi-modal Chain-of-Thoughts Reasoning for Large Language Models
Bingshuai Liu
Chenyang Lyu
Zijun Min
Zhanyu Wang
Jinsong Su
Longyue Wang
LRM
215
9
0
04 Dec 2023
Nash Learning from Human Feedback
Nash Learning from Human FeedbackInternational Conference on Machine Learning (ICML), 2023
Rémi Munos
Michal Valko
Daniele Calandriello
M. G. Azar
Mark Rowland
...
Nikola Momchev
Olivier Bachem
D. Mankowitz
Doina Precup
Bilal Piot
377
180
0
01 Dec 2023
System 2 Attention (is something you might need too)
System 2 Attention (is something you might need too)
Jason Weston
Sainbayar Sukhbaatar
RALMOffRLLRM
161
75
0
20 Nov 2023
Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2
Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2
Michal Guerquin
Yizhong Wang
Valentina Pyatkin
Nathan Lambert
Matthew E. Peters
...
Joel Jang
Aman Rangapur
Noah A. Smith
Iz Beltagy
Hanna Hajishirzi
ALMELM
266
230
0
17 Nov 2023
Contrastive Chain-of-Thought Prompting
Contrastive Chain-of-Thought Prompting
Yew Ken Chia
Guizhen Chen
Anh Tuan Luu
Soujanya Poria
Lidong Bing
LRMAI4CE
254
45
0
15 Nov 2023
Thread of Thought Unraveling Chaotic Contexts
Thread of Thought Unraveling Chaotic Contexts
Yucheng Zhou
Xiubo Geng
Tao Shen
Chongyang Tao
Guodong Long
Jian-Guang Lou
Jianbing Shen
LRM
174
69
0
15 Nov 2023
Eureka: Human-Level Reward Design via Coding Large Language Models
Eureka: Human-Level Reward Design via Coding Large Language Models
Yecheng Jason Ma
William Liang
Guanzhi Wang
De-An Huang
Osbert Bastani
Dinesh Jayaraman
Yuke Zhu
Linxi Fan
A. Anandkumar
192
447
0
19 Oct 2023
A General Theoretical Paradigm to Understand Learning from Human
  Preferences
A General Theoretical Paradigm to Understand Learning from Human PreferencesInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2023
M. G. Azar
Mark Rowland
Bilal Piot
Daniel Guo
Daniele Calandriello
Michal Valko
Rémi Munos
465
808
0
18 Oct 2023
123
Next