PCI-E slot utilisation - graphics

What is the PCI-E slot utilisation? During gaming, desktop work, 4k video watching, video converting or AI training? E.g. on my Dell Precision 7520 laptop with PCI-E gen3 x16 GPU is max. 10%. That means on PCI-E gen4 it would be 5%?
Thinking to use e.g. RTX 3060 or Arc770 for AI computing, connected to laptop thru miniPcie->pciex16 riser.
As far as I know from benchmarks, the gaming would be up to 20% fps slower, but AI IMHO doesn't load slot so much.

Related

Computer Problems (CPU maybe?)

first some specs about my machine.
CPU: Ryzen 5 1600
RAM: Corsair Vengeance LPX 16 GB (2x8GB) DDR4 3200MHz C16 (and 8gb corsair with 3000mhz I think)
SSD: Force MP500 (Windows on it), SanDisk SSD PLUS 480gb (only games etc.)
HDD: Western Digital 1 GB
Graphics Card: Nvidia GeForce GTX 1070 Ti
Motherboard: GA-AB350-Gaming 3
Power supply: Thermaltake Hamburg 530W
I sometimes play Garry's Mod where my game crashes at a specific add-on, this was never a problem before. If I code a WPF Program in Visual Studio 2022, the Player Movement lags hard, and it's not processing it right. If I've opened a stream (netflix or amazon, something like that) in fullscreen on my second screen AND I'm tabbing out of my Fullscreen Game, the whole PC will crash and always ends with a weird screen at the end > https://imgur.com/a/9WU8251 (those are 2 crashes, same screen).
I think the problem is produced from the same Component and I think, it's the CPU. What are your thoughts?

CPU bottleneck Xeon 3104 with RTX A4000 when using PyTorch?

I am looking to upgrade an older machine we have at our lab to use for deep learning (PyTorch) in addition to my personal work station. Its an older Dell work station but the relevant specs are as follows:
PSU: 950W
RAM: 64 GB DDR4 ECC
CPU: Xeon Bronze 3104 #1.7 GHz
It even has an older NVIDIA GPU I can use for display output when the A4000 is fully loaded like I currently do on my personal setup.
Through the university we can acquire a RTX A4000 (I know not best price to performance), which is basically a 3070ti with more VRAM. I am concerned that the low clock speeds may cause a bottle neck. Does anyone have experience with a similar configuration?
Thank you for the help!

Why is slither.io CPU intensive?

From a programmer's point of view, how can a snakes and dots game consume much more resources than many other more advanced games?
Lazy programming.
Considering you could run the original Quake, a fully 3D game with texture mapping on a Pentium 66 Mhz with 16 Mb RAM without a 3D accelerator card, there is no reason for a silly 2D browser snake game to stall on a modern multi-Ghz CPU other than programmer incompetence.
All the calculations are done by the CPU, whereas other computation-intensive games tend to make use of the GPU. You can observe this by looking at your CPU monitor/task manager, and your graphics driver configuration interface (if you have a fancy card).
When I load the slither game my GPU temp doesn't increase at all, whereas the CPU temp climbs steadily, and the game immediately swallows half of my cores. CPU use further increases in "busy" periods when there are many other players interacting in a small area.
I suspect Slither.io is calculating block chain while we play the game. in such cases the game would run smoother on an old graphics card where their secondary intention is not possible. The greater the GPU the more time of the game is split for their nefarious purposes.

Confusion over compute units and expected cores on nvidia GPU

I have an nvidia GTX 750 Ti card, which is advertised as having 640 CUDA cores. Indeed, the nvidia settings application also reports this.
I'm trying to use this card to do OpenCL development on Linux. Now, I have reported from the OpenCL environment (through PyOpenCL if it makes a difference) that the number of compute units is 5. My understanding is that one compute unit on an nvidia device maps to one multiprocessor, which I understand to be 32 SIMD units (which I assume is the a CUDA core).
Clearly, 5 * 32 is not 640 (rather a quarter of what is expected).
Am I missing something as regards the meaning of a a work unit on nvidia? The card is also driving the graphics output which will be using some of the computational capability - is a proportion of the processing capability reserved for graphics use? (if so, can I change this?).
NVIDIA have a whitepaper for the NVIDIA GeForce GTX 750 Ti, which is worth a read.
An OpenCL compute unit translates to a streaming multiprocessor in NVIDIA GPU terms. Each Maxwell SMM in your GPU contains 128 processing elements ("CUDA cores") - and 128*5 = 640. The SIMD width of the device is still 32, but each compute unit (SMM) can issue instructions to four different warps at once.

Something Similar to RAPL for non Sandy Bridge/xeon processors

First post ever here.
I wanted to know if there was something similar to the Running Average Power Limit for other processors(Intel i7) that aren't Sandy Bridge or Xeon Processors as the machine im working on in the lab.
For those who do not know. I pulled this description to bring you up to speed.
"RAPL(Running Average Power Limit) interface provides platform software
with the ability to monitor, control, and get notifications on SOC
power consumptions."
What I am looking for in particular is to acquire energy consumption measurements on a processor's individual cores after running some code like Matrix Multiplication or Vector Addition. Temperature would be excellent too but that's another question for another day(lm-sensors is a bit puzzling to me)
Thanks and Take Care.
Late answer on this: There's PowerTOP on Linux, but that works for Laptops only as it needs the battery discharge rate for that. It can display Watts per process, but don't ask me how accurate that is (personally I think there might be some problems with that). IIRC it counts the number of CPU wakeups from a CPU sleep state to calculate the energy consumption per process. Also, for AMD processors there's the fam15h_power driver in the lm-sensors software package. For rather new (2011 and newer) Bulldozer AMD CPUs you can get the energy consumption that way.
Note that RAPL does not provide energy consumption per core on a multicore CPU, but only for the whole CPU. You can get the energy consumption of core and non-core (like integrated graphics) separately, but per-core is not possible.

Resources