Skip to main content
  1. Data Science Blog/

Demystifying NVIDIA GPUs

·1690 words·8 mins· loading · ·
AI Hardware & Infrastructure IT Infrastructure AI Hardware Deep Learning (DL)

On This Page

Table of Contents
Share with :

Demystifying NVIDIA GPUs

Demystifying NVIDIA GPUs
#

Introduction
#

NVIDIA has been in the GPU manufacturing business since 1993. They offer hundreds of different types of GPUs for various segments and purposes. For those not in the GPU infrastructure business, it can be confusing to understand even their naming conventions. In this article, I will do my best to help you understand the different types of NVIDIA GPUs and their naming conventions.

Any GPU can be identified by asking following questions.

  • What architecture that GPU is using?
  • For what purpose that GPU is used?
  • What type of VRAM it has?
  • Whether it has ECC (Error Correction Code) memory?
  • What is the brand classification of the GPU?
  • Whether it has Ray Tracing support?
  • What generation of Tensor Cores it has?
  • How many CUDA Cores it has?
  • How much FP64 support it has?
  • Does that GPU has Multi-GPU support?
  • How much power that GPU consumes?

What are these terms?
#

First let’s understand these terms.

  1. Architecture: The fundamental design and technology behind a GPU, determining its performance and capabilities. Example: NVIDIA’s Ampere architecture powers the RTX 3000 series GPUs.

  2. Purpose: The primary application or use case for the GPU. Example: NVIDIA GeForce GPUs are designed for gaming, while Tesla GPUs are optimized for AI and data centers.

  3. VRAM Type: The type of memory used in the GPU to store graphical data, affecting speed and bandwidth. Example: GDDR6 is commonly used in gaming GPUs like the NVIDIA RTX 3080.

  4. ECC Memory: Error Correction Code memory, which detects and corrects data errors, crucial for reliability in professional workstations. Example: NVIDIA Quadro GPUs often include ECC memory.

  5. Brand Classification: The category or series of a GPU within its brand, indicating its target market. Example: NVIDIA GeForce (gaming), AMD Radeon (gaming), or NVIDIA Quadro (professional).

  6. Ray Tracing Support: A feature that enables realistic lighting, shadows, and reflections in real-time rendering. Example: NVIDIA RTX 2060 supports ray tracing.

  7. Tensor Core Generation: Specialized cores in a GPU designed for AI and machine learning tasks, with newer generations offering better performance. Example: NVIDIA Ampere GPUs feature 3rd-gen Tensor Cores.

  8. CUDA Cores: Parallel processing units in NVIDIA GPUs that handle multiple tasks simultaneously. Example: The NVIDIA RTX 3090 has 10,496 CUDA cores.

  9. FP64 Support: The GPU’s ability to perform double-precision floating-point calculations, important for scientific computing. Example: NVIDIA A100 has strong FP64 performance.

  10. Multi-GPU Support: The ability to combine multiple GPUs for increased performance, often used in high-end setups. Example: NVIDIA SLI or AMD CrossFire technologies enable multi-GPU configurations.

  11. Power Consumption: The amount of electrical power a GPU requires, measured in watts (W). Example: The NVIDIA RTX 3080 has a power consumption of around 320W.

NVIDIA GPU Classification – Variables & Values
#

VariablePossible ValuesDescription
ArchitectureKepler, Maxwell, Pascal, Volta, Turing, Ampere, Hopper, Ada Lovelace, BlackwellCore design used in a GPU generation
GPU UsageConsumer Gaming, AI Training, AI Inference, Data Center, Workstation, HPC (High-Performance Computing), Embedded SystemsMain purpose of the GPU
Product LineGeForce GTX, GeForce RTX, Tesla, Quadro, TITAN, RTX Workstation, Jetson, A-Series, H-Series, B-SeriesBrand classification based on usage
VRAM TypeGDDR5, GDDR6, GDDR6X, HBM2, HBM3, HBM3eType of memory used
ECC MemoryYes, NoError Correction Code (important for AI research & HPC)
Tensor CoresNone, 1st Gen, 2nd Gen, 3rd Gen, 4th Gen, 5th GenAI acceleration cores
Ray Tracing (RTX)Yes, NoHardware-accelerated ray tracing for graphics
CUDA CoresFew (GTX Series), Many (RTX Series), Extreme (Data Center GPUs)Parallel processing units for compute tasks
FP64 (Double Precision)Low, Medium, HighImportant for scientific calculations
Multi-GPU SupportNo, NVLink, PCIe ScalingAbility to connect multiple GPUs
Power ConsumptionLow (<100W), Medium (150-300W), High (350-450W), Extreme (600W+)GPU power requirements

Example Entries for Different GPUs
#

1. RTX 4090 (High-End Gaming + AI)
#

VariableValue
ArchitectureAda Lovelace
GPU UsageConsumer Gaming, AI Training, AI Inference
Product LineGeForce RTX
VRAM TypeGDDR6X
ECC MemoryNo
Tensor Cores4th Gen
Ray Tracing (RTX)Yes
CUDA CoresMany
FP64 (Double Precision)Low
Multi-GPU SupportNo
Power ConsumptionHigh (450W)

2. Tesla K80 (Old AI Compute GPU)
#

VariableValue
ArchitectureKepler
GPU UsageAI Training, Data Center, HPC
Product LineTesla
VRAM TypeGDDR5
ECC MemoryYes
Tensor CoresNone
Ray Tracing (RTX)No
CUDA CoresMedium
FP64 (Double Precision)High
Multi-GPU SupportPCIe Scaling
Power ConsumptionMedium (150-300W)

3. H100 (Latest AI GPU for Data Centers)
#

VariableValue
ArchitectureHopper
GPU UsageAI Training, AI Inference, Data Center, HPC
Product LineH-Series
VRAM TypeHBM3
ECC MemoryYes
Tensor Cores4th Gen
Ray Tracing (RTX)No
CUDA CoresExtreme
FP64 (Double Precision)High
Multi-GPU SupportNVLink
Power ConsumptionExtreme (600W+)

We have heard words like Tesla, Kepler and GTX. What are they and how they are related?#

What are these terms?
#

  • Kepler is an architecture (2012–2014) used in multiple GPU series.
  • GTX is a consumer gaming GPU lineup that included Kepler-based GPUs.
  • Tesla is a professional & AI/HPC (High-Performance Computing) GPU lineup, which also used the Kepler architecture.

What are Kepler based GPUS?
#

Kepler-based GPUs in Different NVIDIA Product Lines

GPU SeriesKepler-Based ModelsPurpose
GTX (Gaming GPUs)GTX 780 Ti, GTX 780, GTX 770, GTX 760, GTX 750 TiHigh-performance gaming & general computing
Tesla (Data Center GPUs)Tesla K40, Tesla K80, Tesla K20AI, scientific computing, HPC workloads
Quadro (Workstation GPUs)Quadro K6000, Quadro K5000Professional 3D rendering, CAD, content creation

Key Differences and Relationship Between Tesla, Kepler, and GTX
#

FeatureGTX (Gaming GPUs)Tesla (AI/HPC GPUs)
Target AudienceGamers, general usersAI researchers, scientific computing
CUDA CoresOptimized for gamingOptimized for AI & compute
Double Precision (FP64)LimitedHigh precision for AI/HPC
ECC Memory (Error Correction)NoYes (important for research)
Driver OptimizationGaming, DirectX, OpenGLAI, deep learning, CUDA, HPC
VRAM TypeGDDR5GDDR5, ECC support

How Tesla & GTX Differ in AI Workloads
#

  • GTX GPUs (e.g., GTX 780 Ti) can run AI models but are not optimized for deep learning or scientific accuracy.
  • Tesla GPUs (e.g., Tesla K80) have higher FP64 performance, ECC memory, and better CUDA support for AI research.
  • Today, Tesla has been replaced by modern AI-focused GPUs like A100, H100, and Blackwell (B100, B200).

Why Does This Matter Today?
#

  • If you’re doing AI freelance work, modern RTX GPUs (4090, 4070, etc.) have replaced old Tesla/Kepler GPUs for deep learning.
  • If you’re working with data center-level AI, then you’d look at modern A100, H100, or Blackwell GPUs instead of old Tesla cards.

What are different GPU architecture from nvidia?
#

NVIDIA has released multiple GPU architectures over the years, each bringing improvements in performance, efficiency, and AI capabilities. Here’s a breakdown of the major architectures:

Latest & Current Architectures
#

ArchitectureYearNotable GPUsKey Features
Blackwell2024B100, B200, RTX 50 Series (upcoming)AI-focused, HBM3e memory, improved efficiency, NVLink 5.0
Ada Lovelace2022RTX 4090, 4080, 4070, 4060, 40504th-gen Tensor Cores, DLSS 3, better ray tracing
Hopper2022H100, H200Data center AI GPUs, FP8 support, Transformer Engine
Ampere2020RTX 3090, 3080, A100, GA1023rd-gen Tensor Cores, PCIe 4.0, AI optimizations
Turing2018RTX 2080 Ti, 2080, 2060, GTX 16 SeriesFirst with RTX (Ray Tracing), DLSS 1.0

Older Architectures (Legacy & Gaming Focused)
#

ArchitectureYearNotable GPUsKey Features
Volta2017Tesla V100, Titan VFirst with Tensor Cores, HBM2 memory
Pascal2016GTX 1080 Ti, 1080, 1070, 1060GDDR5X, power-efficient, great gaming performance
Maxwell2014GTX 980, 970, 960Big power efficiency improvements
Kepler2012GTX 780 Ti, 780, 770First with GPU Boost, CUDA advancements
Fermi2010GTX 580, 480, Tesla GPUsEarly CUDA support, power-hungry

Specialized Architectures for AI & Data Centers
#

  • Tesla (Fermi, Kepler, Maxwell) – Early AI & HPC focus
  • Pascal & Volta – Introduced deep learning accelerations
  • Ampere, Hopper, Blackwell – Modern AI training GPUs

Which Architecture Should You Consider?
#

  • If you want gaming + AI workloads, go for Ada Lovelace (RTX 40) or wait for Blackwell (RTX 50).
  • If you need AI research & LLM training, Hopper (H100) or Blackwell (B100, B200) are the best.
  • If you’re using an RTX 4070, you already have Ada Lovelace, which is quite powerful for AI freelancing.

How much power GeForce RTX 4070 consumes?
#

The power consumption of an RTX 4070 in a day depends on usage, power draw, and duration. Let’s break it down:

RTX 4070 Power Draw (Approximate)
#

Usage TypePower Consumption (Wattage, W)
Idle / Low Usage~15-20W
Web Browsing / Light Work~40-50W
Gaming / AI Inference~150-180W
AI Training / Full Load~200-220W

Power Consumption Formula
#

Energy Consumption (kWh) = (GPU Wattage × Hours Used) ÷ 1000

Example Scenarios (for 24 hours)

Usage ScenarioHours Per DayWattage (W)Energy Used (kWh)Cost (at $0.15/kWh)
Idle / Light Work24 hrs20W0.48 kWh~$0.07
Gaming (3 hrs) + Light Work (5 hrs)3 hrs (180W) + 5 hrs (50W)~~0.69 kWh~$0.10
AI Inference (8 hrs) + Light Work (8 hrs)8 hrs (180W) + 8 hrs (50W)~~1.84 kWh~$0.28
Full AI Training (12 hrs) + Idle (12 hrs)12 hrs (220W) + 12 hrs (20W)~~3.0 kWh~$0.45

Note: Cost of electricity is different in the differnt countries. It depends upon country, state of the country, provider, consumer type and how much they consume. In India, electricity is provided by various companies like Reliance, TATA, BPL, etc. and the price is different according to the company and the state. The rate of electricity in India varies between 12 - 13 cents per kilowatt-hour for industrial users, and roughly 8 - 9 cents per kilowatt-hour for residential users.

The average electric rate in the United States is 16.54 cents per kWh. Louisiana has the cheapest electric rate (on average) with customers paying 11.23 cents per kWh for electricity. Electricity rates in Hawaii are the highest with rates 42.10 cents per kWh for electricity.

Monthly & Yearly Estimates (AI Workload)
#

If you run AI workloads for 8 hours/day at ~180W:

  • Daily: ~1.44 kWh → $0.22/day
  • Monthly: ~43.2 kWh → $6.48/month
  • Yearly: ~525.6 kWh → $78.84/year
Dr. Hari Thapliyaal's avatar

Dr. Hari Thapliyaal

Dr. Hari Thapliyal is a seasoned professional and prolific blogger with a multifaceted background that spans the realms of Data Science, Project Management, and Advait-Vedanta Philosophy. Holding a Doctorate in AI/NLP from SSBM (Geneva, Switzerland), Hari has earned Master's degrees in Computers, Business Management, Data Science, and Economics, reflecting his dedication to continuous learning and a diverse skill set. With over three decades of experience in management and leadership, Hari has proven expertise in training, consulting, and coaching within the technology sector. His extensive 16+ years in all phases of software product development are complemented by a decade-long focus on course design, training, coaching, and consulting in Project Management. In the dynamic field of Data Science, Hari stands out with more than three years of hands-on experience in software development, training course development, training, and mentoring professionals. His areas of specialization include Data Science, AI, Computer Vision, NLP, complex machine learning algorithms, statistical modeling, pattern identification, and extraction of valuable insights. Hari's professional journey showcases his diverse experience in planning and executing multiple types of projects. He excels in driving stakeholders to identify and resolve business problems, consistently delivering excellent results. Beyond the professional sphere, Hari finds solace in long meditation, often seeking secluded places or immersing himself in the embrace of nature.

Comments:

Share with :

Related

What is a Digital Twin?
·805 words·4 mins· loading
Industry Applications Technology Trends & Future Computer Vision (CV) Digital Twin Internet of Things (IoT) Manufacturing Technology Artificial Intelligence (AI) Graphics
What is a digital twin? # A digital twin is a virtual representation of a real-world entity or …
Frequencies in Time and Space: Understanding Nyquist Theorem & its Applications
·4103 words·20 mins· loading
Data Analysis & Visualization Computer Vision (CV) Mathematics Signal Processing Space Exploration Statistics
Applications of Nyquists theorem # Can the Nyquist-Shannon sampling theorem applies to light …
The Real Story of Nyquist, Shannon, and the Science of Sampling
·1146 words·6 mins· loading
Technology Trends & Future Interdisciplinary Topics Signal Processing Remove Statistics Technology Concepts
The Story of Nyquist, Shannon, and the Science of Sampling # In the early days of the 20th century, …
BitNet b1.58-2B4T: Revolutionary Binary Neural Network for Efficient AI
·2637 words·13 mins· loading
AI/ML Models Artificial Intelligence (AI) AI Hardware & Infrastructure Neural Network Architectures AI Model Optimization Language Models (LLMs) Business Concepts Data Privacy Remove
Archive Paper Link BitNet b1.58-2B4T: The Future of Efficient AI Processing # A History of 1 bit …
Ollama Setup and Running Models
·1753 words·9 mins· loading
AI and NLP Ollama Models Ollama Large Language Models Local Models Cost Effective AI Models
Ollama: Running Large Language Models Locally # The landscape of Artificial Intelligence (AI) and …