Delve into the most influential research papers on LLMs and uncover key insights into language learning models. Our handpicked selection provides a comprehensive overview, making it easy for researchers and enthusiasts to stay updated with the latest advancements in the field. Whether you're a beginner or an expert, these papers will provide valuable knowledge and inspire new ideas.
Looking for research-backed answers?Try AI Search
Xinyin Ma, Gongfan Fang, Xinchao Wang
ArXiv
This work explores LLM compression in a task-agnostic manner, which aims to preserve the multi-task solving and language generation ability of the original LLM, and adopts structural pruning that selectively removes non-critical coupled structures based on gradient information, maximally preserving the majority of the LLM's functionality.
Juan Manuel Zambrano Chaves, Eric Wang, Tao Tu + 7 more
ArXiv
This work introduces Tx-LLM, a generalist large language model (LLM) fine-tuned from PaLM-2 which encodes knowledge about diverse therapeutic modalities and believes it represents an important step towards LLMs encoding biochemical knowledge and could have a future role as an end-to-end tool across the drug discovery development pipeline.
Wenqi Fan, Zihuai Zhao, Jiatong Li + 5 more
IEEE Transactions on Knowledge and Data Engineering
This survey comprehensively review LLM-empowered recommender systems from various perspectives including pre-training, fine-tuning, and prompting paradigms, and comprehensively discusses the promising future directions in this emerging field.
Yuzhang Shang, Zhihang Yuan, Qiang Wu + 1 more
ArXiv
This paper explores network binarization, a radical form of quantization, compressing model weights to a single bit, specifically for Large Language Models (LLMs) compression and proposes a novel approach, Partially-Binarized LLM (PB-LLM), which can achieve extreme low-bit quantization while maintaining the linguistic reasoning capacity of quantized LLMs.
MD Vera Sorin, MD Danna Brin, MD Yiftach Barash + 4 more
journal unavailable
Purpose: Empathy, a cornerstone of human interaction, is a unique quality to humans that Large Language Models (LLMs) are believed to lack. Our study aims to review the literature on the capacity of LLMs in demonstrating empathy Methods: We conducted a literature search on MEDLINE up to July 2023. Seven publications ultimately met the inclusion criteria. Results: All studies included in this review were published in 2023. All studies but one focused on ChatGPT-3.5 by OpenAI. Only one study evaluated empathy based on objective metrics, and all others used subjective human assessment. The studie...
Duzhen Zhang, Yahan Yu, Chenxing Li + 4 more
journal unavailable
A taxonomy encompassing 126 MM-LLMs, each characterized by its specific formulations is introduced, and the performance of selected MM-LLMs on mainstream benchmarks and key training recipes to enhance the potency of MM-LLMs are reviewed.
S. Routray, A. Javali, K. Sharmila + 3 more
2023 International Conference on Computer Science and Emerging Technologies (CSET)
The basic principles and features of large language models, a type of AI model that is trained on vast amounts of text data to understand and generate human-like language outputs, are studied.
Zhikai Chen, Haitao Mao, Hang Li + 8 more
ACM SIGKDD Explorations Newsletter
This paper aims to explore the potential of LLMs in graph machine learning, especially the node classification task, and investigates two possible pipelines: LLMs-as-Enhancers and LLMs-as-Predictors.
Yining Hong, Haoyu Zhen, Peihao Chen + 4 more
ArXiv
This work proposes to inject the 3D world into large language models and introduce a whole new family of 3D-LLMs that can take 3D point clouds and their features as input and perform a diverse set of 3D-related tasks, including captioning, dense captioning, 3D question answering, task decomposition, 3D grounding, 3D-assisted dialog, navigation, and so on.
Daniel P. Jeong, Zachary Chase Lipton, Pradeep Ravikumar
ArXiv
It is found that the latest models, such as GPT-4, can consistently identify the most predictive features regardless of the query mechanism and across various prompting strategies, which suggests that LLM-based feature selection consistently achieves strong performance competitive with data-driven methods such as the LASSO.
Ruyang Liu, Chen Li, Haoran Tang + 3 more
ArXiv
This paper proposes ST-LLM, an effective video-LLM baseline with Spatial-Temporal sequence modeling inside LLM, and develops a dynamic masking strategy with tailor-made training objectives to address the overhead and stability issues introduced by uncompressed video tokens within LLMs.
Biwei Yan, Kun Li, Minghui Xu + 4 more
ArXiv
This paper conducts an assessment of the privacy protection mechanisms employed by LLMs at various stages, followed by a detailed examination of their efficacy and constraints, and delineates the spectrum of data privacy threats.
Yadong Zhang, Shaoguang Mao, Tao Ge + 7 more
ArXiv
This paper explores the scopes, applications, methodologies, and evaluation metrics related to strategic reasoning with LLMs, highlighting the burgeoning development in this area and the interdisciplinary approaches enhancing their decision-making performance.
Tosin P. Adewumi, Nudrat Habib, Lama Alkhaled + 1 more
ArXiv
This work empirically evaluate the power of 3 open SotA LLMs in zero-shot setting (LLaMA-2-13B, Mixtral 8x7B, and Gemma-7B), and introduces a new hallucination metric - Simple Hallucination Index (SHI).
Bernardo Magnini, Roberto Zanoli, Michele Resta + 4 more
ArXiv
Evalita-LLM, a new benchmark designed to evaluate Large Language Models on Italian tasks, is described, and an iterative methodology, where candidate tasks and candidate prompts are validated against a set of LLMs used for development is proposed.
Haiwei Dong, Shuang Xie
ArXiv
This paper explored the deployment strategies, economic considerations, and sustainability challenges associated with the state-of-the-art LLMs, and discussed the deployment debate between Retrieval-Augmented Generation and fine-tuning, highlighting their respective advantages and limitations.
Kaiqi Yang, Hang Li, Hongzhi Wen + 3 more
journal unavailable
It is revealed that LLMs cannot work as expected on social prediction when given general input features without shortcuts, and possible reasons for this phenomenon are investigated, which suggest potential ways to enhance LLMs for social prediction.
Rajesh Pasupuleti, Ravi Vadapalli, Christopher Mader + 1 more
2024 2nd International Conference on Foundation and Large Language Models (FLLM)
The paper aims to provide a comprehensive analysis of the transformative impact of LLMs across various enterprise sectors and provides a comprehensive overview of current popular LLMs in Enterprise applications, in various domains, and discusses the Ethical, Technical, and Regulatory challenges, future trends, and developments in this dynamic field.
Abiodun Finbarrs Oketunji, Muhammad Anas, Deepthi Saina
ArXiv
The research reveals LLMs, whilst demonstrating impressive capabilities in text generation, exhibit varying degrees of bias across different dimensions, offering a quantifiable measure to compare biases across models and over time, offering a vital tool for systems engineers, researchers and regulators in enhancing the fairness and reliability of LLMs.
Jiongnan Liu, Jiajie Jin, Zihan Wang + 3 more
ArXiv
RETA-LLM provides more plug-and-play modules to support better interaction between IR systems and LLMs, including {request rewriting, document retrieval, passage extraction, answer generation, and fact checking} modules.
G. P. Reddy, Y. V. Pavan Kumar, K. P. Prakash
2024 IEEE Open Conference of Electrical, Electronic and Information Sciences (eStream)
The causes of hallucinations in Large Language Models are understood, the implications are explored, and potential strategies for mitigation are discussed to enhance the reliability of AI-generated content.
Wenqi Fan, Shijie Wang, Jiani Huang + 8 more
ArXiv
This survey reviews the recent developments in Graph ML and explores how LLMs can be utilized to enhance the quality of graph features, alleviate the reliance on labeled data, and address challenges such as graph heterogeneity and out-of-distribution (OOD) generalization.
B. Liu, Yuqian Jiang, Xiaohan Zhang + 4 more
ArXiv
LLM+P is the first framework that incorporates the strengths of classical planners into large language models, and is able to provide optimal solutions for most problems, while LLMs fail to provide even feasible plans for most Problems.
Shih-Chieh Dai, Aiping Xiong, Lun-Wei Ku
journal unavailable
This work proposes a human-LLM collaboration framework (i.e., LLM-in-the-loop) to conduct TA with in-context learning (ICL), which yields similar coding quality to that of human coders but reduces TA's labor and time demands.
Tianyu Du, Ayush Kanodia, Herman Brunborg + 2 more
ArXiv
The value of fine-tuning is demonstrated and it is shown that by adding more career data from a different population, fine-tuning smaller LLMs surpasses the performance of fine-tuning larger models.
Mengting Wan, Tara Safavi, S. Jauhar + 11 more
journal unavailable
TnT-LLM, a two-phase framework that employs LLMs to automate the process of end-to-end label generation and assignment with minimal human effort, and generates more accurate and relevant label taxonomies when compared against state-of-the-art baselines, and achieves a favorable balance between accuracy and efficiency for classification at scale.
Yang Liu, Yuanshun Yao, Jean-François Ton + 6 more
ArXiv
A comprehensive survey of key dimensions that are crucial to consider when assessing LLM trustworthiness is presented, which indicates that, in general, more aligned models tend to perform better in terms of overall trustworthiness.
Ming Jin, Shiyu Wang, Lintao Ma + 8 more
ArXiv
Time-LLM is a reprogramming framework to repurpose LLMs for general time series forecasting with the backbone language models kept intact and is demonstrated to be a powerful time series learner that outperforms state-of-the-art, specialized forecasting models.
Avinash Maurya, Robert Underwood, M. Rafique + 2 more
journal unavailable
This paper introduces a lazy asynchronous multi-level approach that takes advantage of the fact that the tensors making up the model and optimizer state shards remain immutable for extended periods of time, which makes it possible to copy their content in the background with minimal interference during the training process.
Kai Sun, Y. Xu, Hanwen Zha + 2 more
ArXiv
This paper constructed Head-to-Tail, a benchmark that consists of 18K question-answer pairs regarding head, torso, and tail facts in terms of popularity and designed an automated evaluation method and a set of metrics that closely approximate the knowledge an LLM confidently internalizes.
Qinbin Li, Junyuan Hong, Chulin Xie + 10 more
Proc. VLDB Endow.
LLM-PBE is a toolkit crafted specifically for the systematic evaluation of data privacy risks in LLMs, designed to analyze privacy across the entire lifecycle of LLMs, incorporating diverse attack and defense strategies, and handling various data types and metrics.
Saurabh Pahune, Manoj Chandrasekharan
ArXiv
The purpose of this study is to provide readers, developers, academics, and users interested in LLM-based chatbots and virtual intelligent assistant technologies with use full information and future directions.
Shahriar Golchin, M. Surdeanu
ArXiv
The best method achieves an accuracy between 92% and 100% in detecting if an LLM is contaminated with seven datasets, containing train and test/validation partitions, when contrasted with manual evaluation by human experts.
M. Treder, Sojin Lee, K. Tsvetanov
Frontiers in Dementia
Overall, this review corroborates the promising utilization of LLMs to positively impact dementia care by boosting cognitive abilities, enriching social interaction, and supporting caregivers.
Rajvardhan Patil, V. Gudivada
Applied Sciences
This paper extensively discusses different pretraining objectives, benchmarks, and transfer learning methods used in LLMs, and explores how LLMs can perform well across many domains and datasets if sufficiently trained on a large and diverse dataset.
Callie Y. Kim, Christine P. Lee, Bilge Mutlu
2024 19th ACM/IEEE International Conference on Human-Robot Interaction (HRI)
The findings show that LLM-powered robots elevate expectations for sophisticated non-verbal cues and excel in connection-building and deliberation, but fall short in logical communication and may induce anxiety.
Apurv Verma, Satyapriya Krishna, Sebastian Gehrmann + 7 more
ArXiv
A detailed threat model and systematization of knowledge of red-teaming attacks on LLMs are presented and a taxonomy of attacks based on the stages of the LLM development and deployment process is developed to improve the security and robustness of LLM-based systems.
Shuming Ma, Hongyu Wang, Lingxiao Ma + 7 more
ArXiv
This work introduces a 1-bit LLM variant, namely BitNet b1.58, in which every single parameter of the LLM is ternary, which defines a new scaling law and recipe for training new generations of LLMs that are both high-performance and cost-effective.
Keivan Alizadeh-Vahid, Iman Mirzadeh, Dmitry Belenko + 5 more
ArXiv
The integration of sparsity awareness, context-adaptive loading, and a hardware-oriented design paves the way for effective inference of LLMs on devices with limited memory.
Mahi Kolla, Siddharth Salunkhe, Eshwar Chandrasekharan + 1 more
Extended Abstracts of the CHI Conference on Human Factors in Computing Systems
This work explores the feasibility of using large language understanding capabilities (LLMs) to identify rule violations on Reddit and examines how an LLM-based moderator (LLM-Mod) reasons about 744 posts across 9 subreddits that violate different types of rules.
Zhikai Chen, Haitao Mao, Hongzhi Wen + 5 more
ArXiv
This work introduces a label-free node classification on graphs with LLMs pipeline, LLM-GNN, which amalgamates the strengths of both GNNs and LLMs while mitigating their limitations while leveraging the confidence scores derived from LLMs to advanced node selection.
Jiaxin Zhang, Yiqi Wang, Xihong Yang + 6 more
ArXiv
This paper designs a novel Test-Time Training pipeline, LLMTTT, which conducts the test-time adaptation under the annotations by LLMs on a carefully-selected node set and introduces a hybrid active node selection strategy that considers not only node diversity and representativeness, but also prediction signals from the pre-trained model.
Pinaki Raj
international journal of advanced research in computer science
The need for emotional intelligence in transformer-based LLMs and the various existing studies that have evaluated this aspect will be discussed.
Jiahao Yu, Xingwei Lin, Zheng Yu + 1 more
journal unavailable
An automated solution for large-scale LLM jailbreak susceptibility assessment called LLM-F UZZER, inspired by fuzz testing, which generates additional jailbreak prompts tailored to specific LLMs and highlights that many open-source and commercial LLMs suffer from severe jailbreak issues, even after safety fine-tuning.
Eleftheria Papageorgiou, Christos Chronis, Iraklis Varlamis + 1 more
Future Internet
The article delves into the capabilities of LLMs in generating both fake news and fake profiles, highlighting their dual role as both a tool for disinformation and a powerful means of detection.
Geng Sun, Yixian Wang, D. Niyato + 4 more
ArXiv
A novel framework of LLM-enabled graphs for networking optimization is proposed, and a case study on UAV networking is presented, concentrating on optimizing UAV trajectory and communication resource allocation to validate the effectiveness of the proposed framework.
Ali Maatouk, Kenny Chirino Ampudia, Rex Ying + 1 more
ArXiv
Tele-LLMs, the first series of language models ranging from 1B to 8B parameters, specifically tailored for telecommunications are developed and open-source, demonstrating that these models outperform their general-purpose counterparts on Tele-Eval while retaining their previously acquired capabilities, thus avoiding the catastrophic forgetting phenomenon.
Tiankai Yang, Yi Nian, Shawn Li + 9 more
ArXiv
AD-LLM is introduced, the first benchmark that evaluates how LLMs can help with NLP anomaly detection and finds that LLMs can work well in zero-shot AD, that carefully designed augmentation methods are useful, and that explaining model selection for specific datasets remains challenging.
Dong Shu, Tianle Chen, Mingyu Jin + 4 more
ArXiv
The Knowledge Graph Large Language Model (KG-LLM) is introduced, a novel framework that leverages large language models (LLMs) for knowledge graph tasks and significantly improves the models' generalization capabilities, leading to more accurate predictions in unfamiliar scenarios.
Hanjia Lyu, Song Jiang, Hanqing Zeng + 2 more
journal unavailable
This study introduces a novel approach, coined LLM-Rec, which incorporates four distinct prompting strategies of text enrichment for improving personalized text-based recommendations, and empirical experiments reveal that using LLM-augmented text significantly enhances recommendation quality.