GPU

Recent Articles

Sort Options:

How Nvidia Built the Ultimate AI Engine: A Look at GPU Core Architecture

How Nvidia Built the Ultimate AI Engine: A Look at GPU Core Architecture

The article delves into the evolution of GPUs from graphics accelerators to AI powerhouses, highlighting advancements in architecture and semiconductor technology. It emphasizes Nvidia's innovations, including CUDA, which transformed GPUs into essential tools for deep learning and high-performance computing.


What is CUDA and how did it transform Nvidia GPUs for AI applications?
CUDA is a parallel computing platform and programming model developed by Nvidia that allows developers to use GPUs for general purpose processing beyond graphics. It transformed Nvidia GPUs into essential tools for deep learning and high-performance computing by enabling efficient parallel processing of complex AI algorithms.
What architectural features make Nvidia GPUs particularly suited for AI and deep learning?
Nvidia GPUs feature thousands of efficient cores designed for parallel processing, high-bandwidth memory to minimize data transfer bottlenecks, and specialized tensor cores optimized for deep learning operations like matrix multiplication. These features enable faster computations, reduced training times, and efficient handling of large datasets typical in AI workloads.
Sources: [1]

06 August, 2025
Embedded

Lisuan G100 GPU shows promise, at least in OpenCL — homegrown Chinese chip outguns Arc A770 and RTX 4060 in new benchmark, 10% slower than RTX 5060

Lisuan G100 GPU shows promise, at least in OpenCL — homegrown Chinese chip outguns Arc A770 and RTX 4060 in new benchmark, 10% slower than RTX 5060

A new variant of Lisuan Technology's G100 graphics card has been spotted in the Geekbench database, sparking excitement among tech enthusiasts. This development hints at potential advancements in performance and capabilities for the popular graphics card line.


What are the current performance levels of the Lisuan G100 GPU compared to other GPUs?
The Lisuan G100 has shown significant improvement in recent benchmarks, scoring 111,290 points in OpenCL, which places it above the NVIDIA RTX 4060 and close to the RTX 5060. However, earlier tests indicated performance similar to the 13-year-old GTX 660 Ti, suggesting ongoing development and optimization.
Sources: [1], [2]
What are the specifications and potential issues with the Lisuan G100 GPU?
Initial specifications listed for the Lisuan G100 included 32 Compute Units, 256 MB of VRAM, and a 300 MHz clock speed, which are likely placeholder values due to unoptimized firmware and drivers. The GPU is built on a 6nm process, and while it shows promise, it faces challenges in achieving stable performance and game optimizations.
Sources: [1], [2]

24 July, 2025
Tom's Hardware

Exla FLOPs

Exla FLOPs

A recent article highlights the availability of on-demand GPU clusters featuring the most affordable H100 models. This development promises to enhance accessibility for users seeking powerful computing solutions, making advanced technology more attainable for various applications.


What does FLOPS mean and why is it important in computing?
FLOPS stands for Floating Point Operations Per Second, a unit measuring a computer's ability to perform floating-point calculations, which involve decimal numbers. It is important because it quantifies computational performance, especially for tasks requiring complex mathematical calculations such as scientific simulations, graphics rendering, and machine learning.
Sources: [1], [2]
What is an exaflop and how does it relate to FLOPS?
An exaflop is a measure of computing performance equal to one quintillion (10^18) floating-point operations per second. It represents an extremely high level of computational power, used to describe the capabilities of the most advanced supercomputers, enabling them to solve complex problems in science and technology.
Sources: [1], [2]

05 July, 2025
Product Hunt

An unhandled error has occurred. Reload 🗙