In Python, How Can I Verify If PyTorch Is Utilizing The GPU?

To check if PyTorch is using the GPU, you can use the torch.cuda.is_available() function to verify if a GPU is available, and then utilize the torch.cuda.current_device() function to display the current device being used by PyTorch. Here is an example:

import torch

# Check if GPU is available
if torch.cuda.is_available():
    print("GPU is available")
    # Get the current device being used
    print("Current device:", torch.cuda.current_device())
    print("GPU is not available")

If a GPU is available, it will display "GPU is available" along with the current device. Otherwise, it will display "GPU is not available".

About the Author Rex

I'm a passionate tech blogger with an insatiable love for programming! From my early days tinkering with code, I've delved into web dev, mobile apps, and AI. Sharing insights and tutorials with the world is my joy, connecting me to a global community of like-minded tech enthusiasts. Python holds a special place in my heart, but I embrace all challenges. Constantly learning, I attend tech conferences, contribute to open-source projects, and engage in code review sessions. My ultimate goal is to inspire the next generation of developers and contribute positively to the ever-evolving tech landscape. Let's code together!