Delve into the most influential research papers on LLMs and uncover key insights into language learning models. Our handpicked selection provides a comprehensive overview, making it easy for researchers and enthusiasts to stay updated with the latest advancements in the field. Whether you're a beginner or an expert, these papers will provide valuable knowledge and inspire new ideas.
Looking for research-backed answers?Try AI Search
G. P. Reddy, Y. V. Pavan Kumar, K. P. Prakash
2024 IEEE Open Conference of Electrical, Electronic and Information Sciences (eStream)
The causes of hallucinations in Large Language Models are understood, the implications are explored, and potential strategies for mitigation are discussed to enhance the reliability of AI-generated content.
S. Routray, A. Javali, K. Sharmila + 3 more
2023 International Conference on Computer Science and Emerging Technologies (CSET)
The basic principles and features of large language models, a type of AI model that is trained on vast amounts of text data to understand and generate human-like language outputs, are studied.
Stanford CS224N CustomProject, Alex Wang, Calvin Laughlin + 1 more
journal unavailable
Our project introduces a multifaceted approach to generating novel LEGO instruction manuals in a text-based format. We leverage the vision capabilities of GPT-4o and fine-tune models such as GPT-3.5-turbo, Llama-2-7B-chat-hf, and Mistral-7B using a corpus of 90 existing text-based LEGO manuals. We detail our methodology, which includes fine-tuning these models on both existing and synthetically generated manuals from GPT-4o vision prompt engineering. Our contributions include a novel vision-to-text agent and the generation of new, small-scale LEGO instructions. Using our custom dataset compris...
Zhiping Zhang, Chenxinran Shen, Bingsheng Yao + 2 more
journal unavailable
It is found that secretive behavior is often triggered by certain tasks, transcending demographic and personality differences among users, and task types were found to affect users' intentions to use secretive behavior.
This study represents the first attempt to comprehend the novel metric brainscore within this interdisciplinary domain and reveals distinctive feature combinations conducive to interpreting existing brainscores across various brain regions of interest (ROIs) and hemispheres, thereby significantly contributing to advancing interpretable machine learning studies.
Juan Manuel Zambrano Chaves, Eric Wang, Tao Tu + 7 more
ArXiv
This work introduces Tx-LLM, a generalist large language model (LLM) fine-tuned from PaLM-2 which encodes knowledge about diverse therapeutic modalities and believes it represents an important step towards LLMs encoding biochemical knowledge and could have a future role as an end-to-end tool across the drug discovery development pipeline.
Cecilia N. Arighi, Steven Brenner, Zhiyong Lu
Pacific Symposium on Biocomputing. Pacific Symposium on Biocomputing
This workshop aims to introduce the attendees to an in-depth understanding of the rise of LLMs in biomedicine, and how they are being used to drive innovation and improve outcomes in the field, along with associated challenges and pitfalls.
Carlin Soos, Levon Haroutunian
NASKO
This paper examines the theoretical and practical issues introduced by LLMs and describes how their use erodes the supposedly firm boundaries separating specific works and creators, and encourages a reevaluation of reductive work/creator associations and advocate for the adoption of a more expansive approach to authorship.
Angelos Antikatzidis, M. Feidakis, Konstantina Marathaki + 3 more
2024 IEEE Global Engineering Education Conference (EDUCON)
A solution to enrich Softbank NAO6 with A.I. capacity, in order to act as an LLM vessel, and a NAO6 acts as the ancient Greek Philosopher Plato that “guides the one who seeks wisdom” based on his theory is presented.
Bo Liang
Proceedings of the 2024 International Symposium on Physical Design
Issues of large parameter size, trends and new usage scenarios will shape future computing architecture design, and especially their impacts on mobile processor design are discussed.
Xinyin Ma, Gongfan Fang, Xinchao Wang
ArXiv
This work explores LLM compression in a task-agnostic manner, which aims to preserve the multi-task solving and language generation ability of the original LLM, and adopts structural pruning that selectively removes non-critical coupled structures based on gradient information, maximally preserving the majority of the LLM's functionality.
Tosin P. Adewumi, Nudrat Habib, Lama Alkhaled + 1 more
ArXiv
This work empirically evaluate the power of 3 open SotA LLMs in zero-shot setting (LLaMA-2-13B, Mixtral 8x7B, and Gemma-7B), and introduces a new hallucination metric - Simple Hallucination Index (SHI).
Yuzhang Shang, Zhihang Yuan, Qiang Wu + 1 more
ArXiv
This paper explores network binarization, a radical form of quantization, compressing model weights to a single bit, specifically for Large Language Models (LLMs) compression and proposes a novel approach, Partially-Binarized LLM (PB-LLM), which can achieve extreme low-bit quantization while maintaining the linguistic reasoning capacity of quantized LLMs.
Narendra Nadh Vema
journal unavailable
Artificial intelligence has been revolutionized due to the integration the vector data with LLMs and enhancing the similarity search as well as boosting semantic analysis, which can provide the path for further advancements.
This paper systematically explores the multifaceted roles of LLMs in tourism, from generating dynamic travel itineraries and culturally rich site descriptions to providing real-time assistance and multilingual support for global travelers.
Wenqi Fan, Zihuai Zhao, Jiatong Li + 5 more
IEEE Transactions on Knowledge and Data Engineering
This survey comprehensively review LLM-empowered recommender systems from various perspectives including pre-training, fine-tuning, and prompting paradigms, and comprehensively discusses the promising future directions in this emerging field.
Abiodun Finbarrs Oketunji, Muhammad Anas, Deepthi Saina
ArXiv
The research reveals LLMs, whilst demonstrating impressive capabilities in text generation, exhibit varying degrees of bias across different dimensions, offering a quantifiable measure to compare biases across models and over time, offering a vital tool for systems engineers, researchers and regulators in enhancing the fairness and reliability of LLMs.
Kaiqi Yang, Hang Li, Hongzhi Wen + 3 more
journal unavailable
It is revealed that LLMs cannot work as expected on social prediction when given general input features without shortcuts, and possible reasons for this phenomenon are investigated, which suggest potential ways to enhance LLMs for social prediction.
Alison Fang, Jana Perkins
MIT Science Policy Review
A historical overview of LLMs is offered and some of the most pressing concerns involving LLMs today are highlighted, and legislative attempts to address these concerns are discussed and potential complicating factors are outlined.
By combining local and global dependencies over latent representations using causal convolutional filters and Transformer, this work achieves significant gains in performance and showcases a robust speech architecture that can be integrated and adapted in a causal setup beyond speech applications for large-scale language modeling.
MD Vera Sorin, MD Danna Brin, MD Yiftach Barash + 4 more
journal unavailable
Purpose: Empathy, a cornerstone of human interaction, is a unique quality to humans that Large Language Models (LLMs) are believed to lack. Our study aims to review the literature on the capacity of LLMs in demonstrating empathy Methods: We conducted a literature search on MEDLINE up to July 2023. Seven publications ultimately met the inclusion criteria. Results: All studies included in this review were published in 2023. All studies but one focused on ChatGPT-3.5 by OpenAI. Only one study evaluated empathy based on objective metrics, and all others used subjective human assessment. The studie...
Daniel P. Jeong, Zachary Chase Lipton, Pradeep Ravikumar
ArXiv
It is found that the latest models, such as GPT-4, can consistently identify the most predictive features regardless of the query mechanism and across various prompting strategies, which suggests that LLM-based feature selection consistently achieves strong performance competitive with data-driven methods such as the LASSO.
Haiwei Dong, Shuang Xie
ArXiv
This paper explored the deployment strategies, economic considerations, and sustainability challenges associated with the state-of-the-art LLMs, and discussed the deployment debate between Retrieval-Augmented Generation and fine-tuning, highlighting their respective advantages and limitations.
Samuel Cahyawijaya
ArXiv
This thesis proposes data-and-compute-efficient methods to mitigate the disparity in LLM ability in underrepresented languages, allowing better generalization on underrepresented languages without the loss of task generalization ability.
Yazi Gholami
World Journal of Advanced Engineering Technology and Sciences
A comprehensive review of the current research on the application of Large Language Models in cybersecurity, using a systematic literature review (SLR) to synthesize key findings on how LLMs have been employed in tasks such as vulnerability detection, malware analysis, and phishing detection.
Sifan Wu, A. Khasahmadi, Mor Katz + 4 more
journal unavailable
This work develops generative models for CAD by leveraging pre-trained language models and apply them to manipulate engineering sketches and demonstrate that models pre-trained on natural language can be fine- tuned on engineering sketches and achieve remarkable performance in various CAD generation scenarios.
Tianyu Du, Ayush Kanodia, Herman Brunborg + 2 more
ArXiv
This paper considers an alternative where the fine-tuning of the CAREER foundation model is replaced by fine-tuning LLMs, and demonstrates that models trained with this approach outperform several alternatives in terms of predictive performance on the survey data, including traditional econometric models, CAREER, and LLMs with in-context learning.
Apurv Verma, Satyapriya Krishna, Sebastian Gehrmann + 7 more
ArXiv
A detailed threat model and systematization of knowledge of red-teaming attacks on LLMs are presented and a taxonomy of attacks based on the stages of the LLM development and deployment process is developed to improve the security and robustness of LLM-based systems.
Jiongnan Liu, Jiajie Jin, Zihan Wang + 3 more
ArXiv
RETA-LLM provides more plug-and-play modules to support better interaction between IR systems and LLMs, including {request rewriting, document retrieval, passage extraction, answer generation, and fact checking} modules.
Yining Hong, Haoyu Zhen, Peihao Chen + 4 more
ArXiv
This work proposes to inject the 3D world into large language models and introduce a whole new family of 3D-LLMs that can take 3D point clouds and their features as input and perform a diverse set of 3D-related tasks, including captioning, dense captioning, 3D question answering, task decomposition, 3D grounding, 3D-assisted dialog, navigation, and so on.
Duzhen Zhang, Yahan Yu, Chenxing Li + 4 more
journal unavailable
A taxonomy encompassing 126 MM-LLMs, each characterized by its specific formulations is introduced, and the performance of selected MM-LLMs on mainstream benchmarks and key training recipes to enhance the potency of MM-LLMs are reviewed.
Saurabh Pahune, Manoj Chandrasekharan
ArXiv
The purpose of this study is to provide readers, developers, academics, and users interested in LLM-based chatbots and virtual intelligent assistant technologies with use full information and future directions.
Xiaoliang Chen, Liangbin Li, Le Chang + 4 more
ArXiv
It's suggested to diversify training data, fine-tune models, enhance transparency and interpretability, and incorporate ethics and fairness training to address issues of domain specificity and knowledge forgetting.
Shengwei Tian, Lifeng Han, Erick Mendez Guzman + 1 more
ArXiv
This work is to develop a pioneer system by using Named Entity Recognition and Relation Extraction methods that automatically extract key information about large language models from the papers, helping researchers to efficiently access information about LLMs.
Ali Maatouk, Kenny Chirino Ampudia, Rex Ying + 1 more
ArXiv
Tele-LLMs, the first series of language models ranging from 1B to 8B parameters, specifically tailored for telecommunications are developed and open-source, demonstrating that these models outperform their general-purpose counterparts on Tele-Eval while retaining their previously acquired capabilities, thus avoiding the catastrophic forgetting phenomenon.
Rafael Jace Sunico, Shubh Pachchigar, Vinay Kumar + 3 more
2023 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS)
A novel resume building application that employs the Large Language Model (LLM) to aid students in composing their first resumes, which demonstrates the effectiveness of these functionalities while emphasizing ease-of-use.
Moses Oluoke Omopekunola, E. Kardanova
REID (Research and Evaluation in Education)
This study fills a research gap by exploring how AI-generated items in secondary/high school physics aligned with educational taxonomy, and focuses on a preliminary assessment of LLMs ability to generate physics items that match the Bloom's taxonomy application level.
Guang Lin, Qibin Zhao
journal unavailable
A novel defense technique named Large LAnguage MOdel Sentinel (LLAMOS), which is designed to enhance the adversarial robustness of LLMs by purifying the adversarial textual examples before feeding them into the target LLM.
Vuk Jakovljevic, Barbara Gallina, A. Cicchetti + 2 more
journal unavailable
The primary objective is to reduce the manual labor of organizations by creating a framework that automises the processing of unstructured data to create useful data insights.
Jiahao Yu, Xingwei Lin, Zheng Yu + 1 more
journal unavailable
An automated solution for large-scale LLM jailbreak susceptibility assessment called LLM-F UZZER, inspired by fuzz testing, which generates additional jailbreak prompts tailored to specific LLMs and highlights that many open-source and commercial LLMs suffer from severe jailbreak issues, even after safety fine-tuning.
Ruyang Liu, Chen Li, Haoran Tang + 3 more
ArXiv
This paper proposes ST-LLM, an effective video-LLM baseline with Spatial-Temporal sequence modeling inside LLM, and develops a dynamic masking strategy with tailor-made training objectives to address the overhead and stability issues introduced by uncompressed video tokens within LLMs.
Biwei Yan, Kun Li, Minghui Xu + 4 more
ArXiv
This paper conducts an assessment of the privacy protection mechanisms employed by LLMs at various stages, followed by a detailed examination of their efficacy and constraints, and delineates the spectrum of data privacy threats.
Shih-Chieh Dai, Aiping Xiong, Lun-Wei Ku
journal unavailable
This work proposes a human-LLM collaboration framework (i.e., LLM-in-the-loop) to conduct TA with in-context learning (ICL), which yields similar coding quality to that of human coders but reduces TA's labor and time demands.
Yadong Zhang, Shaoguang Mao, Tao Ge + 7 more
ArXiv
This paper explores the scopes, applications, methodologies, and evaluation metrics related to strategic reasoning with LLMs, highlighting the burgeoning development in this area and the interdisciplinary approaches enhancing their decision-making performance.
Qinbin Li, Junyuan Hong, Chulin Xie + 10 more
Proc. VLDB Endow.
LLM-PBE is a toolkit crafted specifically for the systematic evaluation of data privacy risks in LLMs, designed to analyze privacy across the entire lifecycle of LLMs, incorporating diverse attack and defense strategies, and handling various data types and metrics.
M. Treder, Sojin Lee, K. Tsvetanov
Frontiers in Dementia
Overall, this review corroborates the promising utilization of LLMs to positively impact dementia care by boosting cognitive abilities, enriching social interaction, and supporting caregivers.
Eleftheria Papageorgiou, Christos Chronis, Iraklis Varlamis + 1 more
Future Internet
The article delves into the capabilities of LLMs in generating both fake news and fake profiles, highlighting their dual role as both a tool for disinformation and a powerful means of detection.
Jiaxin Zhang, Yiqi Wang, Xihong Yang + 6 more
ArXiv
This paper designs a novel Test-Time Training pipeline, LLMTTT, which conducts the test-time adaptation under the annotations by LLMs on a carefully-selected node set and introduces a hybrid active node selection strategy that considers not only node diversity and representativeness, but also prediction signals from the pre-trained model.
Pinaki Raj
international journal of advanced research in computer science
The need for emotional intelligence in transformer-based LLMs and the various existing studies that have evaluated this aspect will be discussed.
A. J. Titus
journal unavailable
ChatGPT can facilitate biostatistical analyses in healthcare research, making statistical methods more accessible and underscores the importance of independent verification mechanisms to ensure the accuracy of LLM-assisted analyses.