New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MEM1
and MEM2
are both zero on AMD 9654
#613
Comments
I assume |
The above situation also occurs on AMD 9554. (note: I made a sum statistics data output, so the runtime is 384) Runtime (RDTSC) [s]: 384.015717
Runtime unhalted [s]: 0.058734
Clock [MHz]: 199629.250000
CPI: nan
Memory bandwidth (channels 0-3) [MBytes/s]: 0.000000
Memory data volume (channels 0-3) [GBytes]: 0.000000
----------------------------
Runtime (RDTSC) [s]: 384.044250
Runtime unhalted [s]: 0.041370
Clock [MHz]: 186808.812500
CPI: nan
Memory bandwidth (channels 0-3) [MBytes/s]: 0.000000
Memory data volume (channels 0-3) [GBytes]: 0.000000
----------------------------
Runtime (RDTSC) [s]: 384.012970
Runtime unhalted [s]: 0.113669
Clock [MHz]: 187998.656250
CPI: nan
Memory bandwidth (channels 0-3) [MBytes/s]: 0.000000
Memory data volume (channels 0-3) [GBytes]: 0.000000
----------------------------
Runtime (RDTSC) [s]: 384.045624
Runtime unhalted [s]: 0.561691
Clock [MHz]: 191052.828125
CPI: nan
Memory bandwidth (channels 0-3) [MBytes/s]: 0.000000
Memory data volume (channels 0-3) [GBytes]: 0.000000 |
I attempted this, but it seems to have been ineffective. |
What has been ineffective? Setting the value to zero or getting messages? LIKWID with |
I encountered the same problem, [root@localhost bin]# grep -i perf_event /boot/config-4.18.0-372.9.1.el8.x86_64 [root@localhost bin]# likwid-perfctr -f -V 1 -g MEM2 /home/pcadmin/stream CPU name: AMD EPYC 9554 64-Core Processor DEBUG - [access_client_startDaemon:157] Starting daemon /usr/local/sbin/likwid-accessD |
zen4 cpu has 12 memory channels(https://www.amd.com/en/products/cpu/amd-epyc-9554),but why likwid library only support 8 memory channels for profmon datas? |
@marquis-wang Yes, you found it. I fixed it yesterday night. Please test it: 7027aa6 I will add additional memory channels to the branch. Yes it should be 12. |
@TomTheBear Great ! I test branch amd_zen4 :44cf4ca it works well. |
It works but it is not done. I did some major updates yesterday to the branch but the branch cannot be merged, so I create a new one only with the fixes. The events currently configured in MEM1 and MEM2 do no exist for Zen4 anymore, so unclear whether they actually count memory traffic. The updated version will not have MEM1 and MEM2 anymore but MEMREAD and MEMWRITE and use the officially documented metrics for memory traffic.. |
I want to using likwid library to develop collect tools for our's Cluster(Zen4), the memory bandwidth data of 7027aa6 missing 4 memory channls。 I look at the newest commit (44cf4ca) had add full channls ,so I test it ,I compare the likwid-perfctr‘s output(MEMREAD and MEMWRITE) and stream’s output,the results is no big difference。In he officially documented (AMD PPR Family 19h),i found a new event (DATA_BW)maybe helperful moniter the memory bandwidth, I will test the event . |
I'm glad that it works for you now. Please be careful with the PPRs, you have to use the one for the family & model: AMD Family 19h Model 11h should be the right one. In the third document, it documents a |
The UMC performance counters would be of interest to count at the memory controller instead of the DataFabric but they seem quite complicated to add. There is already infrastructure for MMIO based counters but some effort would be required. Unfortunately, they are never exposed by perf_event, so they can be added for accessdaemon/direct only. |
I am trying to measure memory bandwidth for a stencil application that runs on both sockets of a two socket AMD 9654 system.
I am getting zero as the memory bandwidth as seen below. Is there an issue with
DFC
counters on zen4 architecture? Is it fully supported? I tried with and without-f
.The text was updated successfully, but these errors were encountered: