Search Results for "llm"
NVIDIA Megatron-LM Powers 172 Billion Parameter LLM for Japanese Language Proficiency
NVIDIA's Megatron-LM aids in developing a 172 billion parameter large language model focusing on Japanese language capabilities, enhancing AI's multilingual proficiency.
NVIDIA's TensorRT-LLM Multiblock Attention Enhances AI Inference on HGX H200
NVIDIA's TensorRT-LLM introduces multiblock attention, significantly boosting AI inference throughput by up to 3.5x on the HGX H200, tackling challenges of long-sequence lengths.
NVIDIA NIM Revolutionizes AI Model Deployment with Optimized Microservices
NVIDIA NIM streamlines the deployment of fine-tuned AI models, offering performance-optimized microservices for seamless inference, enhancing enterprise AI applications.
Enhancing LLMs for Domain-Specific Multi-Turn Conversations
Explore the challenges and solutions in fine-tuning Large Language Models (LLMs) for effective domain-specific multi-turn conversations, as detailed by together.ai.
NVIDIA TensorRT-LLM Enhances Encoder-Decoder Models with In-Flight Batching
NVIDIA's TensorRT-LLM now supports encoder-decoder models with in-flight batching, offering optimized inference for AI applications. Discover the enhancements for generative AI on NVIDIA GPUs.
Enhancing AI Workflow Security with WebAssembly Sandboxing
Explore how WebAssembly provides a secure environment for executing AI-generated code, mitigating risks and enhancing application security.
Exploring the Impact of LLM Integration on Conversation Intelligence Platforms
Discover how integrating Large Language Models (LLMs) revolutionizes Conversation Intelligence platforms, enhancing user experience, customer understanding, and decision-making processes.
NVIDIA Enhances TensorRT-LLM with KV Cache Optimization Features
NVIDIA introduces new KV cache optimizations in TensorRT-LLM, enhancing performance and efficiency for large language models on GPUs by managing memory and computational resources.
NVIDIA Introduces Nemotron-CC: A Massive Dataset for LLM Pretraining
NVIDIA debuts Nemotron-CC, a 6.3-trillion-token English dataset, enhancing pretraining for large language models with innovative data curation methods.
LangSmith Enhances LLM Evaluations with Pytest and Vitest Integrations
LangSmith introduces Pytest and Vitest integrations to enhance LLM application evaluations, offering improved testing frameworks for developers.
Efficient Meeting Summaries with LLMs Using Python
Learn how to create detailed meeting summaries using AssemblyAI's LeMUR framework and large language models (LLMs) with just five lines of Python code.
Exploring LLM Red Teaming: A Crucial Aspect of AI Security
LLM red teaming involves testing AI models to identify vulnerabilities and ensure security. Learn about its practices, motivations, and significance in AI development.