How do a selection of GPUs from NVIDIA’s professional lineup compare to each other in the llama.cpp benchmark?
LLM Inference – Consumer GPU performance
How do a selection of GPUs from NVIDIA’s GeForce series compare to each other in the llama.cpp benchmark?
Tech Primer: What hardware do you need to run a local LLM?
What considerations need to be made when starting off running LLMs locally?
Effects of CPU speed on GPU inference in llama.cpp
What effect, if any, does a system’s CPU speed have on GPU inference with CUDA in llama.cpp?
Stable Diffusion Linux vs. Windows
How does the choice of Operating System affect image generation performance in Stable Diffusion?
Puget Systems Hardware Trends of 2023
With 2023 at a close, we wanted to look back at the sales trends we saw for CPU, GPU, storage, and OS.
Stable Diffusion LoRA Training – Consumer GPU Analysis
How does performance compare across a variety of consumer-grade GPUs in regard to SDXL LoRA training?
Stable Diffusion LoRA Training – Professional GPU Analysis
How does performance compare across a variety of professional-grade GPUs in regard to SDXL LoRA training?
AMD Microsoft Olive Optimizations for Stable Diffusion Performance Analysis
AMD has published a guide outlining how to use Microsoft Olive for Stable Diffusion to get up to a 9.9x improvement in performance. But is that enough to catch up to NVIDIA?
NVIDIA TensorRT Extension for Stable Diffusion Performance Analysis
NVIDIA has released a TensorRT extension for Stable Diffusion using Automatic 1111, promising significant performance gains. But does it work as advertised?