LLM News - Blockchain.News

DEEPSEEK

AutoJudge Revolutionizes LLM Inference with Enhanced Token Processing
deepseek

AutoJudge Revolutionizes LLM Inference with Enhanced Token Processing

AutoJudge introduces a novel method to accelerate large language model inference by optimizing token processing, reducing human annotation needs, and improving processing speed with minimal accuracy loss.

Unsloth Simplifies LLM Training on NVIDIA Blackwell GPUs
deepseek

Unsloth Simplifies LLM Training on NVIDIA Blackwell GPUs

Unsloth's open-source framework enables efficient LLM training on NVIDIA Blackwell GPUs, democratizing AI development with faster throughput and reduced VRAM usage.

ATLAS: Revolutionizing LLM Inference with Adaptive Learning
deepseek

ATLAS: Revolutionizing LLM Inference with Adaptive Learning

Together.ai introduces ATLAS, a system enhancing LLM inference speed by adapting to workloads, achieving 500 TPS on DeepSeek-V3.1.

Enhancing LLM Inference with NVIDIA Run:ai and Dynamo Integration
deepseek

Enhancing LLM Inference with NVIDIA Run:ai and Dynamo Integration

NVIDIA's Run:ai v2.23 integrates with Dynamo to address large language model inference challenges, offering gang scheduling and topology-aware placement for efficient, scalable deployments.

NVIDIA's Run:ai Model Streamer Enhances LLM Inference Speed
deepseek

NVIDIA's Run:ai Model Streamer Enhances LLM Inference Speed

NVIDIA introduces the Run:ai Model Streamer, significantly reducing cold start latency for large language models in GPU environments, enhancing user experience and scalability.

Enhancing LLM Inference with CPU-GPU Memory Sharing
deepseek

Enhancing LLM Inference with CPU-GPU Memory Sharing

NVIDIA introduces a unified memory architecture to optimize large language model inference, addressing memory constraints and improving performance.

NVIDIA's ProRL v2 Advances LLM Reinforcement Learning with Extended Training
deepseek

NVIDIA's ProRL v2 Advances LLM Reinforcement Learning with Extended Training

NVIDIA unveils ProRL v2, a significant leap in reinforcement learning for large language models (LLMs), enhancing performance through extended training and innovative algorithms.

Together AI Introduces Flexible Benchmarking for LLMs
deepseek

Together AI Introduces Flexible Benchmarking for LLMs

Together AI unveils Together Evaluations, a framework for benchmarking large language models using open-source models as judges, offering customizable insights into model performance.

NVIDIA's NeMo Framework Enables Weekend Training of Reasoning-Capable LLMs
deepseek

NVIDIA's NeMo Framework Enables Weekend Training of Reasoning-Capable LLMs

NVIDIA introduces an efficient method to train reasoning-capable language models over a weekend using the NeMo framework, leveraging the Llama Nemotron dataset and LoRA adapters.

Optimizing LLM Inference with TensorRT: A Comprehensive Guide
deepseek

Optimizing LLM Inference with TensorRT: A Comprehensive Guide

Explore how TensorRT-LLM enhances large language model inference by optimizing performance through benchmarking and tuning, offering developers a robust toolset for efficient deployment.

Trending topics