mirror of
https://github.com/torvalds/linux.git
synced 2024-11-21 19:41:42 +00:00
f849ce6baf
Alibaba's T-Head ali_drw PMU supports DDR bandwidth metrics. Update its usage in the documentation. Reviewed-by: John Garry <john.g.garry@oracle.com> Signed-off-by: Jing Zhang <renyu.zj@linux.alibaba.com> Acked-by: Ian Rogers <irogers@google.com> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Shuai Xue <xueshuai@linux.alibaba.com> Cc: Will Deacon <will@kernel.org> Cc: Zhuo Song <zhuo.song@linux.alibaba.com> Cc: linux-arm-kernel@lists.infradead.org Cc: linux-doc@vger.kernel.org Link: https://lore.kernel.org/r/1690528175-2499-4-git-send-email-renyu.zj@linux.alibaba.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
106 lines
4.2 KiB
ReStructuredText
106 lines
4.2 KiB
ReStructuredText
=============================================================
|
|
Alibaba's T-Head SoC Uncore Performance Monitoring Unit (PMU)
|
|
=============================================================
|
|
|
|
The Yitian 710, custom-built by Alibaba Group's chip development business,
|
|
T-Head, implements uncore PMU for performance and functional debugging to
|
|
facilitate system maintenance.
|
|
|
|
DDR Sub-System Driveway (DRW) PMU Driver
|
|
=========================================
|
|
|
|
Yitian 710 employs eight DDR5/4 channels, four on each die. Each DDR5 channel
|
|
is independent of others to service system memory requests. And one DDR5
|
|
channel is split into two independent sub-channels. The DDR Sub-System Driveway
|
|
implements separate PMUs for each sub-channel to monitor various performance
|
|
metrics.
|
|
|
|
The Driveway PMU devices are named as ali_drw_<sys_base_addr> with perf.
|
|
For example, ali_drw_21000 and ali_drw_21080 are two PMU devices for two
|
|
sub-channels of the same channel in die 0. And the PMU device of die 1 is
|
|
prefixed with ali_drw_400XXXXX, e.g. ali_drw_40021000.
|
|
|
|
Each sub-channel has 36 PMU counters in total, which is classified into
|
|
four groups:
|
|
|
|
- Group 0: PMU Cycle Counter. This group has one pair of counters
|
|
pmu_cycle_cnt_low and pmu_cycle_cnt_high, that is used as the cycle count
|
|
based on DDRC core clock.
|
|
|
|
- Group 1: PMU Bandwidth Counters. This group has 8 counters that are used
|
|
to count the total access number of either the eight bank groups in a
|
|
selected rank, or four ranks separately in the first 4 counters. The base
|
|
transfer unit is 64B.
|
|
|
|
- Group 2: PMU Retry Counters. This group has 10 counters, that intend to
|
|
count the total retry number of each type of uncorrectable error.
|
|
|
|
- Group 3: PMU Common Counters. This group has 16 counters, that are used
|
|
to count the common events.
|
|
|
|
For now, the Driveway PMU driver only uses counters in group 0 and group 3.
|
|
|
|
The DDR Controller (DDRCTL) and DDR PHY combine to create a complete solution
|
|
for connecting an SoC application bus to DDR memory devices. The DDRCTL
|
|
receives transactions Host Interface (HIF) which is custom-defined by Synopsys.
|
|
These transactions are queued internally and scheduled for access while
|
|
satisfying the SDRAM protocol timing requirements, transaction priorities, and
|
|
dependencies between the transactions. The DDRCTL in turn issues commands on
|
|
the DDR PHY Interface (DFI) to the PHY module, which launches and captures data
|
|
to and from the SDRAM. The driveway PMUs have hardware logic to gather
|
|
statistics and performance logging signals on HIF, DFI, etc.
|
|
|
|
By counting the READ, WRITE and RMW commands sent to the DDRC through the HIF
|
|
interface, we could calculate the bandwidth. Example usage of counting memory
|
|
data bandwidth::
|
|
|
|
perf stat \
|
|
-e ali_drw_21000/hif_wr/ \
|
|
-e ali_drw_21000/hif_rd/ \
|
|
-e ali_drw_21000/hif_rmw/ \
|
|
-e ali_drw_21000/cycle/ \
|
|
-e ali_drw_21080/hif_wr/ \
|
|
-e ali_drw_21080/hif_rd/ \
|
|
-e ali_drw_21080/hif_rmw/ \
|
|
-e ali_drw_21080/cycle/ \
|
|
-e ali_drw_23000/hif_wr/ \
|
|
-e ali_drw_23000/hif_rd/ \
|
|
-e ali_drw_23000/hif_rmw/ \
|
|
-e ali_drw_23000/cycle/ \
|
|
-e ali_drw_23080/hif_wr/ \
|
|
-e ali_drw_23080/hif_rd/ \
|
|
-e ali_drw_23080/hif_rmw/ \
|
|
-e ali_drw_23080/cycle/ \
|
|
-e ali_drw_25000/hif_wr/ \
|
|
-e ali_drw_25000/hif_rd/ \
|
|
-e ali_drw_25000/hif_rmw/ \
|
|
-e ali_drw_25000/cycle/ \
|
|
-e ali_drw_25080/hif_wr/ \
|
|
-e ali_drw_25080/hif_rd/ \
|
|
-e ali_drw_25080/hif_rmw/ \
|
|
-e ali_drw_25080/cycle/ \
|
|
-e ali_drw_27000/hif_wr/ \
|
|
-e ali_drw_27000/hif_rd/ \
|
|
-e ali_drw_27000/hif_rmw/ \
|
|
-e ali_drw_27000/cycle/ \
|
|
-e ali_drw_27080/hif_wr/ \
|
|
-e ali_drw_27080/hif_rd/ \
|
|
-e ali_drw_27080/hif_rmw/ \
|
|
-e ali_drw_27080/cycle/ -- sleep 10
|
|
|
|
Example usage of counting all memory read/write bandwidth by metric::
|
|
|
|
perf stat -M ddr_read_bandwidth.all -- sleep 10
|
|
perf stat -M ddr_write_bandwidth.all -- sleep 10
|
|
|
|
The average DRAM bandwidth can be calculated as follows:
|
|
|
|
- Read Bandwidth = perf_hif_rd * DDRC_WIDTH * DDRC_Freq / DDRC_Cycle
|
|
- Write Bandwidth = (perf_hif_wr + perf_hif_rmw) * DDRC_WIDTH * DDRC_Freq / DDRC_Cycle
|
|
|
|
Here, DDRC_WIDTH = 64 bytes.
|
|
|
|
The current driver does not support sampling. So "perf record" is
|
|
unsupported. Also attach to a task is unsupported as the events are all
|
|
uncore.
|