Language models (LMs) face significant challenges related to privacy and copyright concerns due to their training on vast amounts of text data. The inadvertent inclusion of private and copyrighted content…
As LLMs become increasingly integral to various AI tasks, their massive parameter sizes lead to high memory requirements and bandwidth consumption. While quantization-aware training (QAT) offers a potential solution by…
Evaluating large language models (LLMs) has become increasingly challenging due to their complexity and versatility. Ensuring the reliability and quality of these models’ outputs is crucial for advancing AI technologies…
High-dimensional clinical data (HDCD) refers to datasets in healthcare where the number of variables (or features) is significantly larger than the number of patients (or observations). As the number of…
Large language models (LLMs) have showcased remarkable capabilities in generating content and solving complex problems across various domains. However, a notable challenge persists in their ability to perform multi-step deductive…
Using offline web apps and AI apps often comes with challenges. Users typically need to navigate multiple steps to get an app running. These steps can be confusing and time-consuming,…
Evaluating the retrieval and reasoning capabilities of large language models (LLMs) in extremely long contexts, extending up to 1 million tokens, is a significant challenge. Efficiently processing long texts is…
Despite their expanding capabilities, large language models (LLMs) need help with processing extensive contexts. These limitations stem from Transformer-based architectures struggling to extrapolate beyond their training window size. Processing long…
Innovation and the artistic, musical, and literary expression of human experiences and emotions depend on creativity. However, the idea that material created by humans is inherently better is coming under…
LLMs excel in natural language processing tasks but face deployment challenges due to high computational and memory demands during inference. Recent research [MWM+24, WMD+23, SXZ+24, XGZC23, LKM23] aims to enhance…