• Mon. Nov 25th, 2024

NavGPT-2: Integrating LLMs and Navigation Policy Networks for Smarter Agents

LLMs excel in processing textual data, while VLN primarily involves visual information. Effectively combining these modalities requires sophisticated techniques to align and correlate visual and textual representations. Despite significant advancements…

Tencent AI Team Introduces Patch-Level Training for Large Language Models LLMs: Reducing the Sequence Length by Compressing Multiple Tokens into a Single Patch

The enormous increase in the training data needed by Large Language Models, along with their exceptional model capability, has allowed them to accomplish outstanding language understanding and generation advancements. The…

Arcee AI Introduces Arcee-Nova: A New Open-Sourced Language Model based on Qwen2-72B and Approaches GPT-4 Performance Level

Arcee AI introduced Arcee-Nova, a groundbreaking achievement in open-source artificial intelligence. Following their previous release, Arcee-Scribe, Arcee-Nova has quickly established itself as the highest-performing model within the open-source domain. Evaluated…

LOTUS: A Query Engine for Reasoning over Large Corpora of Unstructured and Structured Data with LLMs

The semantic capabilities of modern language models offer the potential for advanced analytics and reasoning over extensive knowledge corpora. However, current systems need more high-level abstractions for large-scale semantic queries.…

Monitoring AI-Modified Content at Scale: Impact of ChatGPT on Peer Reviews in AI Conferences

Large Language Models (LLMs) have been widely discussed in several domains, such as global media, science, and education. Even with this focus, measuring exactly how much LLM is used or…

Athene-Llama3-70B Released: An Open-Weight LLM Trained through RLHF based on Llama-3-70B-Instruct

Nexusflow has released Athene-Llama3-70B, an open-weight chat model fine-tuned from Meta AI’s Llama-3-70B. Athene-70B has achieved an Arena-Hard-Auto score of 77.8%, rivaling proprietary models like GPT-4o and Claude-3.5-Sonnet. This marks…

Apple AI Released a 7B Open-Source Language Model Trained on 2.5T Tokens on Open Datasets

Language models (LMs) have become fundamental in natural language processing (NLP), enabling text generation, translation, and sentiment analysis tasks. These models demand vast amounts of training data to function accurately…

Nephilim v3 8B Released: An Innovative AI Approach to Merging Models for Enhanced Roleplay and Creativity

Llama-3-Nephilim-v3-8B and llama-3-Nephilim-v3-8B-GGUF are two innovative models released on Hugging Face. Although these models were never explicitly trained for roleplay, they exhibit remarkable capability in this domain, highlighting the potential…

The Neo4j LLM Knowledge Graph Builder: An AI Tool that Creates Knowledge Graphs from Unstructured Data

In the rapidly developing field of Artificial Intelligence, it is more important than ever to convert unstructured data into organized, useful information efficiently. Recently, a team of researchers introduced the…

Reinforcing Robust Refusal Training in LLMs: A Past Tense Reformulation Attack and Potential Defenses

Large Language Models (LLMs) like GPT-3.5 and GPT-4 are advanced artificial intelligence systems capable of generating human-like text. These models are trained on vast amounts of data to perform various…