DEEPSEEK
NVIDIA's ComputeEval 2025.2 Challenges LLMs with Advanced CUDA Tasks
NVIDIA expands ComputeEval with 232 new CUDA challenges, testing LLMs' capabilities in complex programming tasks. Discover the impact on AI-assisted coding.
Large Reasoning Models Struggle with Instruction Adherence, Study Reveals
A recent study by Together AI unveils that large reasoning models often fail to comply with instructions during reasoning, highlighting significant challenges in AI model adherence.
NVIDIA Blackwell Outshines in InferenceMAX™ v1 Benchmarks
NVIDIA's Blackwell architecture demonstrates significant performance and efficiency gains in SemiAnalysis's InferenceMAX™ v1 benchmarks, setting new standards for AI hardware.
NVIDIA Blackwell Dominates InferenceMAX Benchmarks with Unmatched AI Efficiency
NVIDIA's Blackwell platform excels in the latest InferenceMAX v1 benchmarks, showcasing superior AI performance and efficiency, promising significant return on investment for AI factories.
Together AI Introduces Flexible Benchmarking for LLMs
Together AI unveils Together Evaluations, a framework for benchmarking large language models using open-source models as judges, offering customizable insights into model performance.
Optimizing LLM Inference with TensorRT: A Comprehensive Guide
Explore how TensorRT-LLM enhances large language model inference by optimizing performance through benchmarking and tuning, offering developers a robust toolset for efficient deployment.
Optimizing LLM Inference Costs: A Comprehensive Guide
Explore strategies for benchmarking large language model (LLM) inference costs, enabling smarter scaling and deployment in the AI landscape, as detailed by NVIDIA's latest insights.
Evaluating Multi-Agent Architectures: A Performance Benchmark
LangChain's new study benchmarks various multi-agent architectures, focusing on their performance and scalability using the Tau-bench dataset, highlighting the advantages of modular systems.
NVIDIA MLPerf v5.0: Reproducing Training Scores for LLM Benchmarks
NVIDIA outlines the process to replicate MLPerf v5.0 training scores for LLM benchmarks, emphasizing hardware prerequisites and step-by-step execution.
NVIDIA Blackwell Achieves 2.6x Performance Boost in MLPerf Training v5.0
NVIDIA's Blackwell architecture showcases significant performance improvements in MLPerf Training v5.0, delivering up to 2.6x faster training times across various benchmarks.