Advanced Computing in the Age of AI | Tuesday, March 19, 2024

inference

Nvidia’s Speedy New Inference Engine Keeps BERT Latency Within a Millisecond

Disappointment abounds when your data scientists dial in the accuracy on deep learning models to a high degree but are then eventually forced to gut the model for inference ...Full Article

IBM’s Latest Prototype Low-Power AI Chip Offers ‘Precision Scaling’

IBM has released details of a prototype AI chip geared toward low-precision training and inference across different AI model types while retaining model quality within AI applications. In a ...Full Article

Nvidia Probes Accelerators, Photons, GPU Scaling

Nvidia spotlighted an AI inference accelerator, emerging optical interconnects and a new programming framework designed to scale GPU performance during this week’s GTC China virtual event. In a keynote, ...Full Article

Xilinx Keeps Pace in AI Accelerator Race

FPGAs are increasingly used to accelerate AI workloads in datacenters for tasks like machine learning inference. A growing list of FPGA accelerators are challenging datacenter GPU deployments, promising to ...Full Article

NeoML Released as TensorFlow Alternative

A new open source library for training machine learning models is billed as rivaling the performance of AI models trained with established libraries like TensorFlow, especially models running on ...Full Article

SiFive Adds Tools for Cloud-Based Chip Design

Chip designers are drawing on new cloud resources along with conventional electronic design automation (EDA) tools to accelerate IC templates from tape-out to custom silicon. Among the challengers to ...Full Article

AI Inference Benchmark Bake-off Puts Nvidia on Top

MLPerf.org, the young AI-benchmarking consortium, has issued the first round of results for its inference test suite. Among organizations with submissions were Nvidia, Intel, Alibaba, Supermicro, Google, Huawei, Dell ...Full Article

AWS Upgrades Nvidia GPU Cloud Instances for Inferencing, Graphics

Graphics processor acceleration in the form of G4 cloud instances have been unleashed by Amazon Web Services for machine learning applications. AWS (NASDAQ: AMZN) on Friday (Sept. 20) announced ...Full Article

AI Used to Convert Brain Signals to Speech

A deep learning framework developed by university researchers aims to convert brain signals recorded by an implant into synthesized speech, aiding those who have lost the ability to speak ...Full Article

Google Cloud Goes Global with Nvidia T4 GPUs

Nvidia’s T4 GPUs unveiled earlier this year for accelerating workloads such as AI inference and training are making their “global” debut as cloud instances on Google Cloud. Google (NASDAQ: ...Full Article
Page 1 of 212
EnterpriseAI