✕
  • News ▸

    • ◂ Back
    • Crypto News
      • Bitcoin News
      • Ethereum News
      • Cardano News
      • Stablecoin News
      • CBDC News
      • DeFi News
      Regulatory News
      • Regulation
      • Legal
      • Cybercrime
      Industry News
      • Exchanges
      • Mining
      • Data Service
      Technology News
      • Enterprise
      • Blockchain Application
  • Analysis
  • Interview
  • Wiki
  • Press Release
  • Price
  • Events
  • Opinion
  • Join Us
  • Tools
  • About
  • Contact
  • Privacy
  • Terms & Conditions
  • Advertise
  • 中文
  •      

DEEPSEEK

 

DeepSeek is an AI company and a family of large language models based in Hangzhou, China. It was founded in 2023 and funded by High-Flyer, a well - known quantitative asset management giant. DeepSeek is dedicated to developing advanced large language models and related technologies. It has released several models, including DeepSeek LLM, DeepSeek Coder, DeepSeekMath, and DeepSeek - VL. The latest version, DeepSeek - V3, which was launched in December 2024, has 67.1 billion parameters and was trained on a dataset of 14.8 trillion tokens. It uses FP8 training and open - sources the native FP8 weights. Benchmark tests show that it outperforms Llama 3.1 and Qwen 2.5 while matching GPT - 4O and Claude 3.5 Sonnet. In addition, DeepSeek - R1, which was officially released on January 20, 2025, performs on a par with OpenAI O1 in terms of mathematics, code, and natural language reasoning tasks. DeepSeek's models have a wide range of applications, such as chat and coding scenarios, multilingual automatic translation, image generation, and AI painting. With their high performance and low cost, DeepSeek's models have quickly gained popularity. For example, on February 2, 2025, the DeepSeek app climbed to the top of the download charts in 140 countries on the Apple App Store and also topped the Android Play Store in the United States

NVIDIA Blackwell Leads MLPerf Training v5.1 with Record-Breaking Performance

Joerg Hiller     Nov 13, 2025

#NVIDIA #AI TRAINING #MLPERF #BLACKWELL ARCHITECTURE
NVIDIA Dominates MLPerf Training v5.1 with Blackwell Ultra GPUs

Rongchai Wang     Nov 13, 2025

#NVIDIA #MLPERF #AI TRAINING #BLACKWELL ULTRA
Boosting Model Training with CUDA-X: An In-Depth Look at GPU Acceleration

Joerg Hiller     Sep 26, 2025

#GPU ACCELERATION #CUDA-X #MODEL TRAINING
Enhancing AI Training: NVIDIA's NCCL Advances Cross-Data Center Communication

Luisa Crawford     Jul 15, 2025

#NVIDIA #NCCL #DATA CENTER #AI TRAINING
Effective FP8 Training: Exploring Per-Tensor and Per-Block Scaling Strategies

Alvin Lang     Jul 02, 2025

#FP8 TRAINING #NVIDIA #AI MODELS
NVIDIA and AWS Join Forces to Enhance AI Training Scalability

Iris Coleman     Jun 24, 2025

#NVIDIA #AWS #AI TRAINING #CLOUD COMPUTING
Floating-Point 8: Revolutionizing AI Training with Lower Precision

Felix Pinkston     Jun 05, 2025

#AI TRAINING #FLOATING-POINT 8 #NVIDIA
NVIDIA Expands AI Training with Multilingual Workshop at GTC Paris

Felix Pinkston     May 31, 2025

#AI TRAINING #NVIDIA #GTC PARIS #MULTILINGUAL AI
Open-Source AI: Mixture-of-Agents Alignment Revolutionizes Post-Training for LLMs

Felix Pinkston     May 29, 2025

#AI #LLMS #OPEN-SOURCE #POST-TRAINING
Anyscale Introduces Comprehensive Ray Training Programs

Peter Zhang     Mar 20, 2025

#ANYSCALE #RAY #AI TRAINING #MACHINE LEARNING



  • Write
  • About
  • Contact
  • Privacy
  • Terms & Condition
  • Advertise
Copyright © 2020 Blockchain News. All Rights Reserved.