Search Results for "llm"
Enhancing LLM Workflows with NVIDIA NeMo-Skills
NVIDIA's NeMo-Skills library offers seamless integration for improving LLM workflows, addressing challenges in synthetic data generation, model training, and evaluation.
Exploring Open Source Reinforcement Learning Libraries for LLMs
An in-depth analysis of leading open-source reinforcement learning libraries for large language models, comparing frameworks like TRL, Verl, and RAGEN.
Exploring Context Engineering in AI Agent Development
Discover how context engineering is transforming AI agent development by optimizing information management through strategies like writing, selecting, compressing, and isolating context.
Optimizing LLM Inference with TensorRT: A Comprehensive Guide
Explore how TensorRT-LLM enhances large language model inference by optimizing performance through benchmarking and tuning, offering developers a robust toolset for efficient deployment.
Character.AI Unveils pipeling-sft: A New Framework for Fine-Tuning MoE LLMs
Character.AI introduces pipeling-sft, an open-source framework designed to enhance fine-tuning of Mixture-of-Experts large language models, facilitating scalability and efficiency in AI research.
Together AI Introduces Flexible Benchmarking for LLMs
Together AI unveils Together Evaluations, a framework for benchmarking large language models using open-source models as judges, offering customizable insights into model performance.
NVIDIA's NeMo Framework Enables Weekend Training of Reasoning-Capable LLMs
NVIDIA introduces an efficient method to train reasoning-capable language models over a weekend using the NeMo framework, leveraging the Llama Nemotron dataset and LoRA adapters.
NVIDIA's ProRL v2 Advances LLM Reinforcement Learning with Extended Training
NVIDIA unveils ProRL v2, a significant leap in reinforcement learning for large language models (LLMs), enhancing performance through extended training and innovative algorithms.
Open-Source LLMs Surpass Proprietary Models in Specialized Tasks
Parsed's fine-tuning of a 27B open-source LLM outperforms Claude Sonnet 4 by 60% in healthcare tasks, offering significant cost savings and performance gains.
Together AI Leverages AI Agents for Complex Engineering Automation
Together AI utilizes AI agents to automate intricate engineering tasks, optimizing LLM inference systems and reducing manual intervention, according to Together AI.
Solana (SOL) Bench: Evaluating LLMs' Competence in Crypto Transactions
Solana (SOL) introduces Solana Bench, a tool to assess the effectiveness of LLMs in executing complex crypto transactions on the Solana blockchain.
Enhancing LLM Inference with CPU-GPU Memory Sharing
NVIDIA introduces a unified memory architecture to optimize large language model inference, addressing memory constraints and improving performance.