I am keeping this post very short, with mostly photos. I tested the cooling performance with different games. The GPU’s max power is 72W, though during my tests, it exceeded 75W. It’s also possible to limit it to 30W. I tested the GPU by running games like Black Myth: Wukong, Cyberpunk 2077, Uncharted 4: A…
Category: AI/ML & vGPU
Nvidia L4: Powerful Low-Power GPU for Nvidia AI Enterprise and Virtual GPU
I’ve been searching the internet for a long time to find a versatile GPU for AI and video graphics workloads that also supports vGPU and Nvidia AI Enterprise. Some of the GPUs I considered were the RTX 6000 Ada, A2, A10, L4, T4, A40, and A16. I was most drawn to the RTX 6000 Ada…
Deploying and Configuring Nvidia DLS for AI Enterprise and vGPU: Step-by-Step Guide
NB! At the end of the blog post, there is a YouTube video and an eBook – a photo-based step-by-step guide Download Nvidia vGPU Drivers for ESXi Download Nvidia vGPU License Server: Installing Nvidia vGPU Drivers on ESXi Deploying the NvidiaDLS OVA to vSphere Configuring Nvidia DLS (License) Server Installing Nvidia Drivers on a Windows…
Overcoming PCIe Slot Compatibility Challenges for Nvidia Tesla P4 GPU Installation
I bought an Nvidia Tesla P4. It was an unused GPU and came with a 3D-printed cooler and fan. I played around with this GPU on my AI/ML server, and it worked fine. Then I decided to move it to my other server, which runs 24/7. The reason is simple: I have jump hosts and…
Quick and Easy Guide to Installing Meta Llama 3.1 405B, 70B, 8B Language Models with Ollama, Docker, and OpenWebUI
I will show how easy and quick it is to install Llama 3.1 405B, 70B, 8B, or another language model on your computer or VM using Ollama, Docker, and OpenWebUI. It is so simple to install that even a grandmother or grandfather could do it. This is private AI, not cloud-based. All data is on…
Meta Llama 3.1 405B: GPU vs. CPU Performance Evaluation and RAM Considerations
It’s time to start testing various Private AI models, and fortunately, the timing is just right. Meta has just released six new AI language models. These models run on-premises and do not interact with the cloud or OpenAI’s ChatGPT. Llama 3.1 405B competes with leading models like GPT-4, GPT-4o, and Claude 3.5 Sonnet, while smaller…
Solving Real Business Problems with Private AI: Unlocking Efficiency and Productivity
I’m on a mission to find a company facing a real problem that can be solved using Private AI—AI that operates entirely within your own data center, without relying on cloud services. I need your help to identify the most challenging, time-consuming, or tedious issues within your company that could greatly benefit from an AI…
Struggling with Crucial T705 NVMe and VMware ESXi 8. U3 Compatibility
I was so confident that the Crucial T705 4TB PCIe Gen5 NVMe M.2 SSD would work with VMware ESXi without any issues that I didn’t even bother to Google or check. I wanted to get and test what it feels like and how much of a performance boost it gives when using the fastest (R:…
AI, ML, and vGPU Server Build – Part 3: The Server Case as a Double-Decker Bus
Those two mining frames in the photo illustrate my idea well, but I plan to use a PC case and build a frame on top of it where I can easily attach different GPUs, essentially connecting 2 PC cases together. The server motherboard is secured inside the PC case, and the GPUs are openly mounted…
AI, ML, and vGPU Server Build – Part 2: Server Noise, Temperature, and Power
Noise: Server not operating: I used the Tadeto SL720 Sound Level Meter for measurement. When only the ESX-Core Server, router, and switch are turned on in my room, I measured the sound and obtained the following results. My computer desk, where I sit: 34.7 dBCabinet on the other side of the wall: 34.7 dB Right…