Odaily Planet Daily reports that NVIDIA is about to launch a dedicated processing chip designed for artificial intelligence inference scenarios, aiming to improve AI query processing speed and enhance model responsiveness. The report states that OpenAI is expected to be one of the main customers for this new processor.
The chip focuses on optimizing AI model inference performance in practical applications and is expected to be officially released next month at the NVIDIA GTC developer conference in San Jose. (The Wall Street Journal)
View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
NVIDIA will launch AI inference-specific chips, with OpenAI potentially becoming one of its main customers
Odaily Planet Daily reports that NVIDIA is about to launch a dedicated processing chip designed for artificial intelligence inference scenarios, aiming to improve AI query processing speed and enhance model responsiveness. The report states that OpenAI is expected to be one of the main customers for this new processor.
The chip focuses on optimizing AI model inference performance in practical applications and is expected to be officially released next month at the NVIDIA GTC developer conference in San Jose. (The Wall Street Journal)