File: | dev/pci/drm/amd/amdgpu/df_v3_6.c |
Warning: | line 470, column 21 The left operand of '==' is a garbage value |
Press '?' to see keyboard shortcuts
Keyboard shortcuts:
1 | /* | |||
2 | * Copyright 2018 Advanced Micro Devices, Inc. | |||
3 | * | |||
4 | * Permission is hereby granted, free of charge, to any person obtaining a | |||
5 | * copy of this software and associated documentation files (the "Software"), | |||
6 | * to deal in the Software without restriction, including without limitation | |||
7 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | |||
8 | * and/or sell copies of the Software, and to permit persons to whom the | |||
9 | * Software is furnished to do so, subject to the following conditions: | |||
10 | * | |||
11 | * The above copyright notice and this permission notice shall be included in | |||
12 | * all copies or substantial portions of the Software. | |||
13 | * | |||
14 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |||
15 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |||
16 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |||
17 | * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR | |||
18 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, | |||
19 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | |||
20 | * OTHER DEALINGS IN THE SOFTWARE. | |||
21 | * | |||
22 | */ | |||
23 | #include "amdgpu.h" | |||
24 | #include "df_v3_6.h" | |||
25 | ||||
26 | #include "df/df_3_6_default.h" | |||
27 | #include "df/df_3_6_offset.h" | |||
28 | #include "df/df_3_6_sh_mask.h" | |||
29 | ||||
30 | #define DF_3_6_SMN_REG_INST_DIST0x8 0x8 | |||
31 | #define DF_3_6_INST_CNT8 8 | |||
32 | ||||
33 | static u32 df_v3_6_channel_number[] = {1, 2, 0, 4, 0, 8, 0, | |||
34 | 16, 32, 0, 0, 0, 2, 4, 8}; | |||
35 | ||||
36 | #ifdef __linux__ | |||
37 | /* init df format attrs */ | |||
38 | AMDGPU_PMU_ATTR(event, "config:0-7")static ssize_t event_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[(!(sizeof("config:0-7" ) >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__ )); return sprintf(page, "config:0-7" "\n"); } static struct device_attribute pmu_attr_event = __ATTR_RO(event); | |||
39 | AMDGPU_PMU_ATTR(instance, "config:8-15")static ssize_t instance_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[(!(sizeof("config:8-15" ) >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__ )); return sprintf(page, "config:8-15" "\n"); } static struct device_attribute pmu_attr_instance = __ATTR_RO(instance); | |||
40 | AMDGPU_PMU_ATTR(umask, "config:16-23")static ssize_t umask_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[(!(sizeof("config:16-23" ) >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__ )); return sprintf(page, "config:16-23" "\n"); } static struct device_attribute pmu_attr_umask = __ATTR_RO(umask); | |||
41 | ||||
42 | /* df format attributes */ | |||
43 | static struct attribute *df_v3_6_format_attrs[] = { | |||
44 | &pmu_attr_event.attr, | |||
45 | &pmu_attr_instance.attr, | |||
46 | &pmu_attr_umask.attr, | |||
47 | NULL((void *)0) | |||
48 | }; | |||
49 | ||||
50 | /* df format attribute group */ | |||
51 | static struct attribute_group df_v3_6_format_attr_group = { | |||
52 | .name = "format", | |||
53 | .attrs = df_v3_6_format_attrs, | |||
54 | }; | |||
55 | ||||
56 | /* df event attrs */ | |||
57 | AMDGPU_PMU_ATTR(cake0_pcsout_txdata,static ssize_t cake0_pcsout_txdata_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x46,umask=0x2") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x46,umask=0x2" "\n"); } static struct device_attribute pmu_attr_cake0_pcsout_txdata = __ATTR_RO(cake0_pcsout_txdata ) | |||
58 | "event=0x7,instance=0x46,umask=0x2")static ssize_t cake0_pcsout_txdata_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x46,umask=0x2") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x46,umask=0x2" "\n"); } static struct device_attribute pmu_attr_cake0_pcsout_txdata = __ATTR_RO(cake0_pcsout_txdata ); | |||
59 | AMDGPU_PMU_ATTR(cake1_pcsout_txdata,static ssize_t cake1_pcsout_txdata_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x47,umask=0x2") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x47,umask=0x2" "\n"); } static struct device_attribute pmu_attr_cake1_pcsout_txdata = __ATTR_RO(cake1_pcsout_txdata ) | |||
60 | "event=0x7,instance=0x47,umask=0x2")static ssize_t cake1_pcsout_txdata_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x47,umask=0x2") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x47,umask=0x2" "\n"); } static struct device_attribute pmu_attr_cake1_pcsout_txdata = __ATTR_RO(cake1_pcsout_txdata ); | |||
61 | AMDGPU_PMU_ATTR(cake0_pcsout_txmeta,static ssize_t cake0_pcsout_txmeta_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x46,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x46,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake0_pcsout_txmeta = __ATTR_RO(cake0_pcsout_txmeta ) | |||
62 | "event=0x7,instance=0x46,umask=0x4")static ssize_t cake0_pcsout_txmeta_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x46,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x46,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake0_pcsout_txmeta = __ATTR_RO(cake0_pcsout_txmeta ); | |||
63 | AMDGPU_PMU_ATTR(cake1_pcsout_txmeta,static ssize_t cake1_pcsout_txmeta_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x47,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x47,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake1_pcsout_txmeta = __ATTR_RO(cake1_pcsout_txmeta ) | |||
64 | "event=0x7,instance=0x47,umask=0x4")static ssize_t cake1_pcsout_txmeta_show(struct device *dev, struct device_attribute *attr, char *page) { extern char _ctassert[ (!(sizeof("event=0x7,instance=0x47,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0x7,instance=0x47,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake1_pcsout_txmeta = __ATTR_RO(cake1_pcsout_txmeta ); | |||
65 | AMDGPU_PMU_ATTR(cake0_ftiinstat_reqalloc,static ssize_t cake0_ftiinstat_reqalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x46,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x46,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake0_ftiinstat_reqalloc = __ATTR_RO (cake0_ftiinstat_reqalloc) | |||
66 | "event=0xb,instance=0x46,umask=0x4")static ssize_t cake0_ftiinstat_reqalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x46,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x46,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake0_ftiinstat_reqalloc = __ATTR_RO (cake0_ftiinstat_reqalloc); | |||
67 | AMDGPU_PMU_ATTR(cake1_ftiinstat_reqalloc,static ssize_t cake1_ftiinstat_reqalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x47,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x47,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake1_ftiinstat_reqalloc = __ATTR_RO (cake1_ftiinstat_reqalloc) | |||
68 | "event=0xb,instance=0x47,umask=0x4")static ssize_t cake1_ftiinstat_reqalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x47,umask=0x4") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x47,umask=0x4" "\n"); } static struct device_attribute pmu_attr_cake1_ftiinstat_reqalloc = __ATTR_RO (cake1_ftiinstat_reqalloc); | |||
69 | AMDGPU_PMU_ATTR(cake0_ftiinstat_rspalloc,static ssize_t cake0_ftiinstat_rspalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x46,umask=0x8") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x46,umask=0x8" "\n"); } static struct device_attribute pmu_attr_cake0_ftiinstat_rspalloc = __ATTR_RO (cake0_ftiinstat_rspalloc) | |||
70 | "event=0xb,instance=0x46,umask=0x8")static ssize_t cake0_ftiinstat_rspalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x46,umask=0x8") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x46,umask=0x8" "\n"); } static struct device_attribute pmu_attr_cake0_ftiinstat_rspalloc = __ATTR_RO (cake0_ftiinstat_rspalloc); | |||
71 | AMDGPU_PMU_ATTR(cake1_ftiinstat_rspalloc,static ssize_t cake1_ftiinstat_rspalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x47,umask=0x8") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x47,umask=0x8" "\n"); } static struct device_attribute pmu_attr_cake1_ftiinstat_rspalloc = __ATTR_RO (cake1_ftiinstat_rspalloc) | |||
72 | "event=0xb,instance=0x47,umask=0x8")static ssize_t cake1_ftiinstat_rspalloc_show(struct device *dev , struct device_attribute *attr, char *page) { extern char _ctassert [(!(sizeof("event=0xb,instance=0x47,umask=0x8") >= (1 << 12) - 1)) ? 1 : -1 ] __attribute__((__unused__)); return sprintf (page, "event=0xb,instance=0x47,umask=0x8" "\n"); } static struct device_attribute pmu_attr_cake1_ftiinstat_rspalloc = __ATTR_RO (cake1_ftiinstat_rspalloc); | |||
73 | ||||
74 | /* df event attributes */ | |||
75 | static struct attribute *df_v3_6_event_attrs[] = { | |||
76 | &pmu_attr_cake0_pcsout_txdata.attr, | |||
77 | &pmu_attr_cake1_pcsout_txdata.attr, | |||
78 | &pmu_attr_cake0_pcsout_txmeta.attr, | |||
79 | &pmu_attr_cake1_pcsout_txmeta.attr, | |||
80 | &pmu_attr_cake0_ftiinstat_reqalloc.attr, | |||
81 | &pmu_attr_cake1_ftiinstat_reqalloc.attr, | |||
82 | &pmu_attr_cake0_ftiinstat_rspalloc.attr, | |||
83 | &pmu_attr_cake1_ftiinstat_rspalloc.attr, | |||
84 | NULL((void *)0) | |||
85 | }; | |||
86 | ||||
87 | /* df event attribute group */ | |||
88 | static struct attribute_group df_v3_6_event_attr_group = { | |||
89 | .name = "events", | |||
90 | .attrs = df_v3_6_event_attrs | |||
91 | }; | |||
92 | ||||
93 | /* df event attr groups */ | |||
94 | const struct attribute_group *df_v3_6_attr_groups[] = { | |||
95 | &df_v3_6_format_attr_group, | |||
96 | &df_v3_6_event_attr_group, | |||
97 | NULL((void *)0) | |||
98 | }; | |||
99 | ||||
100 | #endif /* __linux__ */ | |||
101 | ||||
102 | static uint64_t df_v3_6_get_fica(struct amdgpu_device *adev, | |||
103 | uint32_t ficaa_val) | |||
104 | { | |||
105 | unsigned long flags, address, data; | |||
106 | uint32_t ficadl_val, ficadh_val; | |||
107 | ||||
108 | address = adev->nbio.funcs->get_pcie_index_offset(adev); | |||
109 | data = adev->nbio.funcs->get_pcie_data_offset(adev); | |||
110 | ||||
111 | spin_lock_irqsave(&adev->pcie_idx_lock, flags)do { flags = 0; mtx_enter(&adev->pcie_idx_lock); } while (0); | |||
112 | WREG32(address, smnDF_PIE_AON_FabricIndirectConfigAccessAddress3)amdgpu_device_wreg(adev, (address), (0x1d05cUL), 0); | |||
113 | WREG32(data, ficaa_val)amdgpu_device_wreg(adev, (data), (ficaa_val), 0); | |||
114 | ||||
115 | WREG32(address, smnDF_PIE_AON_FabricIndirectConfigAccessDataLo3)amdgpu_device_wreg(adev, (address), (0x1d098UL), 0); | |||
116 | ficadl_val = RREG32(data)amdgpu_device_rreg(adev, (data), 0); | |||
117 | ||||
118 | WREG32(address, smnDF_PIE_AON_FabricIndirectConfigAccessDataHi3)amdgpu_device_wreg(adev, (address), (0x1d09cUL), 0); | |||
119 | ficadh_val = RREG32(data)amdgpu_device_rreg(adev, (data), 0); | |||
120 | ||||
121 | spin_unlock_irqrestore(&adev->pcie_idx_lock, flags)do { (void)(flags); mtx_leave(&adev->pcie_idx_lock); } while (0); | |||
122 | ||||
123 | return (((ficadh_val & 0xFFFFFFFFFFFFFFFF) << 32) | ficadl_val); | |||
124 | } | |||
125 | ||||
126 | static void df_v3_6_set_fica(struct amdgpu_device *adev, uint32_t ficaa_val, | |||
127 | uint32_t ficadl_val, uint32_t ficadh_val) | |||
128 | { | |||
129 | unsigned long flags, address, data; | |||
130 | ||||
131 | address = adev->nbio.funcs->get_pcie_index_offset(adev); | |||
132 | data = adev->nbio.funcs->get_pcie_data_offset(adev); | |||
133 | ||||
134 | spin_lock_irqsave(&adev->pcie_idx_lock, flags)do { flags = 0; mtx_enter(&adev->pcie_idx_lock); } while (0); | |||
135 | WREG32(address, smnDF_PIE_AON_FabricIndirectConfigAccessAddress3)amdgpu_device_wreg(adev, (address), (0x1d05cUL), 0); | |||
136 | WREG32(data, ficaa_val)amdgpu_device_wreg(adev, (data), (ficaa_val), 0); | |||
137 | ||||
138 | WREG32(address, smnDF_PIE_AON_FabricIndirectConfigAccessDataLo3)amdgpu_device_wreg(adev, (address), (0x1d098UL), 0); | |||
139 | WREG32(data, ficadl_val)amdgpu_device_wreg(adev, (data), (ficadl_val), 0); | |||
140 | ||||
141 | WREG32(address, smnDF_PIE_AON_FabricIndirectConfigAccessDataHi3)amdgpu_device_wreg(adev, (address), (0x1d09cUL), 0); | |||
142 | WREG32(data, ficadh_val)amdgpu_device_wreg(adev, (data), (ficadh_val), 0); | |||
143 | ||||
144 | spin_unlock_irqrestore(&adev->pcie_idx_lock, flags)do { (void)(flags); mtx_leave(&adev->pcie_idx_lock); } while (0); | |||
145 | } | |||
146 | ||||
147 | /* | |||
148 | * df_v3_6_perfmon_rreg - read perfmon lo and hi | |||
149 | * | |||
150 | * required to be atomic. no mmio method provided so subsequent reads for lo | |||
151 | * and hi require to preserve df finite state machine | |||
152 | */ | |||
153 | static void df_v3_6_perfmon_rreg(struct amdgpu_device *adev, | |||
154 | uint32_t lo_addr, uint32_t *lo_val, | |||
155 | uint32_t hi_addr, uint32_t *hi_val) | |||
156 | { | |||
157 | unsigned long flags, address, data; | |||
158 | ||||
159 | address = adev->nbio.funcs->get_pcie_index_offset(adev); | |||
160 | data = adev->nbio.funcs->get_pcie_data_offset(adev); | |||
161 | ||||
162 | spin_lock_irqsave(&adev->pcie_idx_lock, flags)do { flags = 0; mtx_enter(&adev->pcie_idx_lock); } while (0); | |||
163 | WREG32(address, lo_addr)amdgpu_device_wreg(adev, (address), (lo_addr), 0); | |||
164 | *lo_val = RREG32(data)amdgpu_device_rreg(adev, (data), 0); | |||
165 | WREG32(address, hi_addr)amdgpu_device_wreg(adev, (address), (hi_addr), 0); | |||
166 | *hi_val = RREG32(data)amdgpu_device_rreg(adev, (data), 0); | |||
167 | spin_unlock_irqrestore(&adev->pcie_idx_lock, flags)do { (void)(flags); mtx_leave(&adev->pcie_idx_lock); } while (0); | |||
168 | } | |||
169 | ||||
170 | /* | |||
171 | * df_v3_6_perfmon_wreg - write to perfmon lo and hi | |||
172 | * | |||
173 | * required to be atomic. no mmio method provided so subsequent reads after | |||
174 | * data writes cannot occur to preserve data fabrics finite state machine. | |||
175 | */ | |||
176 | static void df_v3_6_perfmon_wreg(struct amdgpu_device *adev, uint32_t lo_addr, | |||
177 | uint32_t lo_val, uint32_t hi_addr, uint32_t hi_val) | |||
178 | { | |||
179 | unsigned long flags, address, data; | |||
180 | ||||
181 | address = adev->nbio.funcs->get_pcie_index_offset(adev); | |||
182 | data = adev->nbio.funcs->get_pcie_data_offset(adev); | |||
183 | ||||
184 | spin_lock_irqsave(&adev->pcie_idx_lock, flags)do { flags = 0; mtx_enter(&adev->pcie_idx_lock); } while (0); | |||
185 | WREG32(address, lo_addr)amdgpu_device_wreg(adev, (address), (lo_addr), 0); | |||
186 | WREG32(data, lo_val)amdgpu_device_wreg(adev, (data), (lo_val), 0); | |||
187 | WREG32(address, hi_addr)amdgpu_device_wreg(adev, (address), (hi_addr), 0); | |||
188 | WREG32(data, hi_val)amdgpu_device_wreg(adev, (data), (hi_val), 0); | |||
189 | spin_unlock_irqrestore(&adev->pcie_idx_lock, flags)do { (void)(flags); mtx_leave(&adev->pcie_idx_lock); } while (0); | |||
190 | } | |||
191 | ||||
192 | /* same as perfmon_wreg but return status on write value check */ | |||
193 | static int df_v3_6_perfmon_arm_with_status(struct amdgpu_device *adev, | |||
194 | uint32_t lo_addr, uint32_t lo_val, | |||
195 | uint32_t hi_addr, uint32_t hi_val) | |||
196 | { | |||
197 | unsigned long flags, address, data; | |||
198 | uint32_t lo_val_rb, hi_val_rb; | |||
199 | ||||
200 | address = adev->nbio.funcs->get_pcie_index_offset(adev); | |||
201 | data = adev->nbio.funcs->get_pcie_data_offset(adev); | |||
202 | ||||
203 | spin_lock_irqsave(&adev->pcie_idx_lock, flags)do { flags = 0; mtx_enter(&adev->pcie_idx_lock); } while (0); | |||
204 | WREG32(address, lo_addr)amdgpu_device_wreg(adev, (address), (lo_addr), 0); | |||
205 | WREG32(data, lo_val)amdgpu_device_wreg(adev, (data), (lo_val), 0); | |||
206 | WREG32(address, hi_addr)amdgpu_device_wreg(adev, (address), (hi_addr), 0); | |||
207 | WREG32(data, hi_val)amdgpu_device_wreg(adev, (data), (hi_val), 0); | |||
208 | ||||
209 | WREG32(address, lo_addr)amdgpu_device_wreg(adev, (address), (lo_addr), 0); | |||
210 | lo_val_rb = RREG32(data)amdgpu_device_rreg(adev, (data), 0); | |||
211 | WREG32(address, hi_addr)amdgpu_device_wreg(adev, (address), (hi_addr), 0); | |||
212 | hi_val_rb = RREG32(data)amdgpu_device_rreg(adev, (data), 0); | |||
213 | spin_unlock_irqrestore(&adev->pcie_idx_lock, flags)do { (void)(flags); mtx_leave(&adev->pcie_idx_lock); } while (0); | |||
214 | ||||
215 | if (!(lo_val == lo_val_rb && hi_val == hi_val_rb)) | |||
216 | return -EBUSY16; | |||
217 | ||||
218 | return 0; | |||
219 | } | |||
220 | ||||
221 | ||||
222 | /* | |||
223 | * retry arming counters every 100 usecs within 1 millisecond interval. | |||
224 | * if retry fails after time out, return error. | |||
225 | */ | |||
226 | #define ARM_RETRY_USEC_TIMEOUT1000 1000 | |||
227 | #define ARM_RETRY_USEC_INTERVAL100 100 | |||
228 | static int df_v3_6_perfmon_arm_with_retry(struct amdgpu_device *adev, | |||
229 | uint32_t lo_addr, uint32_t lo_val, | |||
230 | uint32_t hi_addr, uint32_t hi_val) | |||
231 | { | |||
232 | int countdown = ARM_RETRY_USEC_TIMEOUT1000; | |||
233 | ||||
234 | while (countdown) { | |||
235 | ||||
236 | if (!df_v3_6_perfmon_arm_with_status(adev, lo_addr, lo_val, | |||
237 | hi_addr, hi_val)) | |||
238 | break; | |||
239 | ||||
240 | countdown -= ARM_RETRY_USEC_INTERVAL100; | |||
241 | udelay(ARM_RETRY_USEC_INTERVAL100); | |||
242 | } | |||
243 | ||||
244 | return countdown > 0 ? 0 : -ETIME60; | |||
245 | } | |||
246 | ||||
247 | /* get the number of df counters available */ | |||
248 | static ssize_t df_v3_6_get_df_cntr_avail(struct device *dev, | |||
249 | struct device_attribute *attr, | |||
250 | char *buf) | |||
251 | { | |||
252 | struct amdgpu_device *adev; | |||
253 | struct drm_device *ddev; | |||
254 | int i, count; | |||
255 | ||||
256 | ddev = dev_get_drvdata(dev)((void *)0); | |||
257 | adev = drm_to_adev(ddev); | |||
258 | count = 0; | |||
259 | ||||
260 | for (i = 0; i < DF_V3_6_MAX_COUNTERS4; i++) { | |||
261 | if (adev->df_perfmon_config_assign_mask[i] == 0) | |||
262 | count++; | |||
263 | } | |||
264 | ||||
265 | return snprintf(buf, PAGE_SIZE(1 << 12), "%i\n", count); | |||
266 | } | |||
267 | ||||
268 | /* device attr for available perfmon counters */ | |||
269 | static DEVICE_ATTR(df_cntr_avail, S_IRUGO, df_v3_6_get_df_cntr_avail, NULL)struct device_attribute dev_attr_df_cntr_avail; | |||
270 | ||||
271 | static void df_v3_6_query_hashes(struct amdgpu_device *adev) | |||
272 | { | |||
273 | u32 tmp; | |||
274 | ||||
275 | adev->df.hash_status.hash_64k = false0; | |||
276 | adev->df.hash_status.hash_2m = false0; | |||
277 | adev->df.hash_status.hash_1g = false0; | |||
278 | ||||
279 | if (adev->asic_type != CHIP_ARCTURUS) | |||
280 | return; | |||
281 | ||||
282 | /* encoding for hash-enabled on Arcturus */ | |||
283 | if (adev->df.funcs->get_fb_channel_number(adev) == 0xe) { | |||
284 | tmp = RREG32_SOC15(DF, 0, mmDF_CS_UMC_AON0_DfGlobalCtrl)amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x00fe), 0); | |||
285 | adev->df.hash_status.hash_64k = REG_GET_FIELD(tmp,(((tmp) & 0x00100000L) >> 0x14) | |||
286 | DF_CS_UMC_AON0_DfGlobalCtrl,(((tmp) & 0x00100000L) >> 0x14) | |||
287 | GlbHashIntlvCtl64K)(((tmp) & 0x00100000L) >> 0x14); | |||
288 | adev->df.hash_status.hash_2m = REG_GET_FIELD(tmp,(((tmp) & 0x00200000L) >> 0x15) | |||
289 | DF_CS_UMC_AON0_DfGlobalCtrl,(((tmp) & 0x00200000L) >> 0x15) | |||
290 | GlbHashIntlvCtl2M)(((tmp) & 0x00200000L) >> 0x15); | |||
291 | adev->df.hash_status.hash_1g = REG_GET_FIELD(tmp,(((tmp) & 0x00400000L) >> 0x16) | |||
292 | DF_CS_UMC_AON0_DfGlobalCtrl,(((tmp) & 0x00400000L) >> 0x16) | |||
293 | GlbHashIntlvCtl1G)(((tmp) & 0x00400000L) >> 0x16); | |||
294 | } | |||
295 | } | |||
296 | ||||
297 | /* init perfmons */ | |||
298 | static void df_v3_6_sw_init(struct amdgpu_device *adev) | |||
299 | { | |||
300 | int i, ret; | |||
301 | ||||
302 | ret = device_create_file(adev->dev, &dev_attr_df_cntr_avail)0; | |||
303 | if (ret) | |||
304 | DRM_ERROR("failed to create file for available df counters\n")__drm_err("failed to create file for available df counters\n" ); | |||
305 | ||||
306 | for (i = 0; i < AMDGPU_MAX_DF_PERFMONS4; i++) | |||
307 | adev->df_perfmon_config_assign_mask[i] = 0; | |||
308 | ||||
309 | df_v3_6_query_hashes(adev); | |||
310 | } | |||
311 | ||||
312 | static void df_v3_6_sw_fini(struct amdgpu_device *adev) | |||
313 | { | |||
314 | ||||
315 | device_remove_file(adev->dev, &dev_attr_df_cntr_avail); | |||
316 | ||||
317 | } | |||
318 | ||||
319 | static void df_v3_6_enable_broadcast_mode(struct amdgpu_device *adev, | |||
320 | bool_Bool enable) | |||
321 | { | |||
322 | u32 tmp; | |||
323 | ||||
324 | if (enable) { | |||
325 | tmp = RREG32_SOC15(DF, 0, mmFabricConfigAccessControl)amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x0410), 0); | |||
326 | tmp &= ~FabricConfigAccessControl__CfgRegInstAccEn_MASK0x00000001L; | |||
327 | WREG32_SOC15(DF, 0, mmFabricConfigAccessControl, tmp)amdgpu_device_wreg(adev, ((adev->reg_offset[DF_HWIP][0][0] + 0x0410)), (tmp), 0); | |||
328 | } else | |||
329 | WREG32_SOC15(DF, 0, mmFabricConfigAccessControl,amdgpu_device_wreg(adev, ((adev->reg_offset[DF_HWIP][0][0] + 0x0410)), (0x00000000), 0) | |||
330 | mmFabricConfigAccessControl_DEFAULT)amdgpu_device_wreg(adev, ((adev->reg_offset[DF_HWIP][0][0] + 0x0410)), (0x00000000), 0); | |||
331 | } | |||
332 | ||||
333 | static u32 df_v3_6_get_fb_channel_number(struct amdgpu_device *adev) | |||
334 | { | |||
335 | u32 tmp; | |||
336 | ||||
337 | tmp = RREG32_SOC15(DF, 0, mmDF_CS_UMC_AON0_DramBaseAddress0)amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x0044), 0); | |||
338 | tmp &= DF_CS_UMC_AON0_DramBaseAddress0__IntLvNumChan_MASK0x0000003CL; | |||
339 | tmp >>= DF_CS_UMC_AON0_DramBaseAddress0__IntLvNumChan__SHIFT0x2; | |||
340 | ||||
341 | return tmp; | |||
342 | } | |||
343 | ||||
344 | static u32 df_v3_6_get_hbm_channel_number(struct amdgpu_device *adev) | |||
345 | { | |||
346 | int fb_channel_number; | |||
347 | ||||
348 | fb_channel_number = adev->df.funcs->get_fb_channel_number(adev); | |||
349 | if (fb_channel_number >= ARRAY_SIZE(df_v3_6_channel_number)(sizeof((df_v3_6_channel_number)) / sizeof((df_v3_6_channel_number )[0]))) | |||
350 | fb_channel_number = 0; | |||
351 | ||||
352 | return df_v3_6_channel_number[fb_channel_number]; | |||
353 | } | |||
354 | ||||
355 | static void df_v3_6_update_medium_grain_clock_gating(struct amdgpu_device *adev, | |||
356 | bool_Bool enable) | |||
357 | { | |||
358 | u32 tmp; | |||
359 | ||||
360 | if (adev->cg_flags & AMD_CG_SUPPORT_DF_MGCG(1 << 23)) { | |||
361 | /* Put DF on broadcast mode */ | |||
362 | adev->df.funcs->enable_broadcast_mode(adev, true1); | |||
363 | ||||
364 | if (enable) { | |||
365 | tmp = RREG32_SOC15(DF, 0,amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x00fc), 0) | |||
366 | mmDF_PIE_AON0_DfGlobalClkGater)amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x00fc), 0); | |||
367 | tmp &= ~DF_PIE_AON0_DfGlobalClkGater__MGCGMode_MASK0x0000000FL; | |||
368 | tmp |= DF_V3_6_MGCG_ENABLE_15_CYCLE_DELAY; | |||
369 | WREG32_SOC15(DF, 0,amdgpu_device_wreg(adev, ((adev->reg_offset[DF_HWIP][0][0] + 0x00fc)), (tmp), 0) | |||
370 | mmDF_PIE_AON0_DfGlobalClkGater, tmp)amdgpu_device_wreg(adev, ((adev->reg_offset[DF_HWIP][0][0] + 0x00fc)), (tmp), 0); | |||
371 | } else { | |||
372 | tmp = RREG32_SOC15(DF, 0,amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x00fc), 0) | |||
373 | mmDF_PIE_AON0_DfGlobalClkGater)amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x00fc), 0); | |||
374 | tmp &= ~DF_PIE_AON0_DfGlobalClkGater__MGCGMode_MASK0x0000000FL; | |||
375 | tmp |= DF_V3_6_MGCG_DISABLE; | |||
376 | WREG32_SOC15(DF, 0,amdgpu_device_wreg(adev, ((adev->reg_offset[DF_HWIP][0][0] + 0x00fc)), (tmp), 0) | |||
377 | mmDF_PIE_AON0_DfGlobalClkGater, tmp)amdgpu_device_wreg(adev, ((adev->reg_offset[DF_HWIP][0][0] + 0x00fc)), (tmp), 0); | |||
378 | } | |||
379 | ||||
380 | /* Exit broadcast mode */ | |||
381 | adev->df.funcs->enable_broadcast_mode(adev, false0); | |||
382 | } | |||
383 | } | |||
384 | ||||
385 | static void df_v3_6_get_clockgating_state(struct amdgpu_device *adev, | |||
386 | u32 *flags) | |||
387 | { | |||
388 | u32 tmp; | |||
389 | ||||
390 | /* AMD_CG_SUPPORT_DF_MGCG */ | |||
391 | tmp = RREG32_SOC15(DF, 0, mmDF_PIE_AON0_DfGlobalClkGater)amdgpu_device_rreg(adev, (adev->reg_offset[DF_HWIP][0][0] + 0x00fc), 0); | |||
392 | if (tmp & DF_V3_6_MGCG_ENABLE_15_CYCLE_DELAY) | |||
393 | *flags |= AMD_CG_SUPPORT_DF_MGCG(1 << 23); | |||
394 | } | |||
395 | ||||
396 | /* get assigned df perfmon ctr as int */ | |||
397 | static int df_v3_6_pmc_config_2_cntr(struct amdgpu_device *adev, | |||
398 | uint64_t config) | |||
399 | { | |||
400 | int i; | |||
401 | ||||
402 | for (i = 0; i < DF_V3_6_MAX_COUNTERS4; i++) { | |||
403 | if ((config & 0x0FFFFFFUL) == | |||
404 | adev->df_perfmon_config_assign_mask[i]) | |||
405 | return i; | |||
406 | } | |||
407 | ||||
408 | return -EINVAL22; | |||
409 | } | |||
410 | ||||
411 | /* get address based on counter assignment */ | |||
412 | static void df_v3_6_pmc_get_addr(struct amdgpu_device *adev, | |||
413 | uint64_t config, | |||
414 | int is_ctrl, | |||
415 | uint32_t *lo_base_addr, | |||
416 | uint32_t *hi_base_addr) | |||
417 | { | |||
418 | int target_cntr = df_v3_6_pmc_config_2_cntr(adev, config); | |||
419 | ||||
420 | if (target_cntr < 0) | |||
421 | return; | |||
422 | ||||
423 | switch (target_cntr) { | |||
424 | ||||
425 | case 0: | |||
426 | *lo_base_addr = is_ctrl ? smnPerfMonCtlLo40x01d880UL : smnPerfMonCtrLo40x01d790UL; | |||
427 | *hi_base_addr = is_ctrl ? smnPerfMonCtlHi40x01d884UL : smnPerfMonCtrHi40x01d794UL; | |||
428 | break; | |||
429 | case 1: | |||
430 | *lo_base_addr = is_ctrl ? smnPerfMonCtlLo50x01d888UL : smnPerfMonCtrLo50x01d798UL; | |||
431 | *hi_base_addr = is_ctrl ? smnPerfMonCtlHi50x01d88cUL : smnPerfMonCtrHi50x01d79cUL; | |||
432 | break; | |||
433 | case 2: | |||
434 | *lo_base_addr = is_ctrl ? smnPerfMonCtlLo60x01d890UL : smnPerfMonCtrLo60x01d7a0UL; | |||
435 | *hi_base_addr = is_ctrl ? smnPerfMonCtlHi60x01d894UL : smnPerfMonCtrHi60x01d7a4UL; | |||
436 | break; | |||
437 | case 3: | |||
438 | *lo_base_addr = is_ctrl ? smnPerfMonCtlLo70x01d898UL : smnPerfMonCtrLo70x01d7a8UL; | |||
439 | *hi_base_addr = is_ctrl ? smnPerfMonCtlHi70x01d89cUL : smnPerfMonCtrHi70x01d7acUL; | |||
440 | break; | |||
441 | ||||
442 | } | |||
443 | ||||
444 | } | |||
445 | ||||
446 | /* get read counter address */ | |||
447 | static void df_v3_6_pmc_get_read_settings(struct amdgpu_device *adev, | |||
448 | uint64_t config, | |||
449 | uint32_t *lo_base_addr, | |||
450 | uint32_t *hi_base_addr) | |||
451 | { | |||
452 | df_v3_6_pmc_get_addr(adev, config, 0, lo_base_addr, hi_base_addr); | |||
453 | } | |||
454 | ||||
455 | /* get control counter settings i.e. address and values to set */ | |||
456 | static int df_v3_6_pmc_get_ctrl_settings(struct amdgpu_device *adev, | |||
457 | uint64_t config, | |||
458 | uint32_t *lo_base_addr, | |||
459 | uint32_t *hi_base_addr, | |||
460 | uint32_t *lo_val, | |||
461 | uint32_t *hi_val, | |||
462 | bool_Bool is_enable) | |||
463 | { | |||
464 | ||||
465 | uint32_t eventsel, instance, unitmask; | |||
466 | uint32_t instance_10, instance_5432, instance_76; | |||
467 | ||||
468 | df_v3_6_pmc_get_addr(adev, config, 1, lo_base_addr, hi_base_addr); | |||
469 | ||||
470 | if ((*lo_base_addr == 0) || (*hi_base_addr == 0)) { | |||
| ||||
471 | DRM_ERROR("[DF PMC] addressing not retrieved! Lo: %x, Hi: %x",__drm_err("[DF PMC] addressing not retrieved! Lo: %x, Hi: %x" , *lo_base_addr, *hi_base_addr) | |||
472 | *lo_base_addr, *hi_base_addr)__drm_err("[DF PMC] addressing not retrieved! Lo: %x, Hi: %x" , *lo_base_addr, *hi_base_addr); | |||
473 | return -ENXIO6; | |||
474 | } | |||
475 | ||||
476 | eventsel = DF_V3_6_GET_EVENT(config)(config & 0xFFUL) & 0x3f; | |||
477 | unitmask = DF_V3_6_GET_UNITMASK(config)((config >> 16) & 0xFFUL) & 0xf; | |||
478 | instance = DF_V3_6_GET_INSTANCE(config)((config >> 8) & 0xFFUL); | |||
479 | ||||
480 | instance_10 = instance & 0x3; | |||
481 | instance_5432 = (instance >> 2) & 0xf; | |||
482 | instance_76 = (instance >> 6) & 0x3; | |||
483 | ||||
484 | *lo_val = (unitmask << 8) | (instance_10 << 6) | eventsel; | |||
485 | *lo_val = is_enable ? *lo_val | (1 << 22) : *lo_val & ~(1 << 22); | |||
486 | *hi_val = (instance_76 << 29) | instance_5432; | |||
487 | ||||
488 | DRM_DEBUG_DRIVER("config=%llx addr=%08x:%08x val=%08x:%08x",__drm_dbg(DRM_UT_DRIVER, "config=%llx addr=%08x:%08x val=%08x:%08x" , config, *lo_base_addr, *hi_base_addr, *lo_val, *hi_val) | |||
489 | config, *lo_base_addr, *hi_base_addr, *lo_val, *hi_val)__drm_dbg(DRM_UT_DRIVER, "config=%llx addr=%08x:%08x val=%08x:%08x" , config, *lo_base_addr, *hi_base_addr, *lo_val, *hi_val); | |||
490 | ||||
491 | return 0; | |||
492 | } | |||
493 | ||||
494 | /* add df performance counters for read */ | |||
495 | static int df_v3_6_pmc_add_cntr(struct amdgpu_device *adev, | |||
496 | uint64_t config) | |||
497 | { | |||
498 | int i, target_cntr; | |||
499 | ||||
500 | target_cntr = df_v3_6_pmc_config_2_cntr(adev, config); | |||
501 | ||||
502 | if (target_cntr >= 0) | |||
503 | return 0; | |||
504 | ||||
505 | for (i = 0; i < DF_V3_6_MAX_COUNTERS4; i++) { | |||
506 | if (adev->df_perfmon_config_assign_mask[i] == 0U) { | |||
507 | adev->df_perfmon_config_assign_mask[i] = | |||
508 | config & 0x0FFFFFFUL; | |||
509 | return 0; | |||
510 | } | |||
511 | } | |||
512 | ||||
513 | return -ENOSPC28; | |||
514 | } | |||
515 | ||||
516 | #define DEFERRED_ARM_MASK(1 << 31) (1 << 31) | |||
517 | static int df_v3_6_pmc_set_deferred(struct amdgpu_device *adev, | |||
518 | uint64_t config, bool_Bool is_deferred) | |||
519 | { | |||
520 | int target_cntr; | |||
521 | ||||
522 | target_cntr = df_v3_6_pmc_config_2_cntr(adev, config); | |||
523 | ||||
524 | if (target_cntr < 0) | |||
525 | return -EINVAL22; | |||
526 | ||||
527 | if (is_deferred) | |||
528 | adev->df_perfmon_config_assign_mask[target_cntr] |= | |||
529 | DEFERRED_ARM_MASK(1 << 31); | |||
530 | else | |||
531 | adev->df_perfmon_config_assign_mask[target_cntr] &= | |||
532 | ~DEFERRED_ARM_MASK(1 << 31); | |||
533 | ||||
534 | return 0; | |||
535 | } | |||
536 | ||||
537 | static bool_Bool df_v3_6_pmc_is_deferred(struct amdgpu_device *adev, | |||
538 | uint64_t config) | |||
539 | { | |||
540 | int target_cntr; | |||
541 | ||||
542 | target_cntr = df_v3_6_pmc_config_2_cntr(adev, config); | |||
543 | ||||
544 | /* | |||
545 | * we never get target_cntr < 0 since this funciton is only called in | |||
546 | * pmc_count for now but we should check anyways. | |||
547 | */ | |||
548 | return (target_cntr >= 0 && | |||
549 | (adev->df_perfmon_config_assign_mask[target_cntr] | |||
550 | & DEFERRED_ARM_MASK(1 << 31))); | |||
551 | ||||
552 | } | |||
553 | ||||
554 | /* release performance counter */ | |||
555 | static void df_v3_6_pmc_release_cntr(struct amdgpu_device *adev, | |||
556 | uint64_t config) | |||
557 | { | |||
558 | int target_cntr = df_v3_6_pmc_config_2_cntr(adev, config); | |||
559 | ||||
560 | if (target_cntr >= 0) | |||
561 | adev->df_perfmon_config_assign_mask[target_cntr] = 0ULL; | |||
562 | } | |||
563 | ||||
564 | ||||
565 | static void df_v3_6_reset_perfmon_cntr(struct amdgpu_device *adev, | |||
566 | uint64_t config) | |||
567 | { | |||
568 | uint32_t lo_base_addr = 0, hi_base_addr = 0; | |||
569 | ||||
570 | df_v3_6_pmc_get_read_settings(adev, config, &lo_base_addr, | |||
571 | &hi_base_addr); | |||
572 | ||||
573 | if ((lo_base_addr == 0) || (hi_base_addr == 0)) | |||
574 | return; | |||
575 | ||||
576 | df_v3_6_perfmon_wreg(adev, lo_base_addr, 0, hi_base_addr, 0); | |||
577 | } | |||
578 | ||||
579 | static int df_v3_6_pmc_start(struct amdgpu_device *adev, uint64_t config, | |||
580 | int is_add) | |||
581 | { | |||
582 | uint32_t lo_base_addr, hi_base_addr, lo_val, hi_val; | |||
583 | int err = 0, ret = 0; | |||
584 | ||||
585 | switch (adev->asic_type) { | |||
586 | case CHIP_VEGA20: | |||
587 | if (is_add) | |||
588 | return df_v3_6_pmc_add_cntr(adev, config); | |||
589 | ||||
590 | df_v3_6_reset_perfmon_cntr(adev, config); | |||
591 | ||||
592 | ret = df_v3_6_pmc_get_ctrl_settings(adev, | |||
593 | config, | |||
594 | &lo_base_addr, | |||
595 | &hi_base_addr, | |||
596 | &lo_val, | |||
597 | &hi_val, | |||
598 | true1); | |||
599 | ||||
600 | if (ret) | |||
601 | return ret; | |||
602 | ||||
603 | err = df_v3_6_perfmon_arm_with_retry(adev, | |||
604 | lo_base_addr, | |||
605 | lo_val, | |||
606 | hi_base_addr, | |||
607 | hi_val); | |||
608 | ||||
609 | if (err) | |||
610 | ret = df_v3_6_pmc_set_deferred(adev, config, true1); | |||
611 | ||||
612 | break; | |||
613 | default: | |||
614 | break; | |||
615 | } | |||
616 | ||||
617 | return ret; | |||
618 | } | |||
619 | ||||
620 | static int df_v3_6_pmc_stop(struct amdgpu_device *adev, uint64_t config, | |||
621 | int is_remove) | |||
622 | { | |||
623 | uint32_t lo_base_addr, hi_base_addr, lo_val, hi_val; | |||
| ||||
624 | int ret = 0; | |||
625 | ||||
626 | switch (adev->asic_type) { | |||
627 | case CHIP_VEGA20: | |||
628 | ret = df_v3_6_pmc_get_ctrl_settings(adev, | |||
629 | config, | |||
630 | &lo_base_addr, | |||
631 | &hi_base_addr, | |||
632 | &lo_val, | |||
633 | &hi_val, | |||
634 | false0); | |||
635 | ||||
636 | if (ret) | |||
637 | return ret; | |||
638 | ||||
639 | ||||
640 | if (is_remove) { | |||
641 | df_v3_6_reset_perfmon_cntr(adev, config); | |||
642 | df_v3_6_pmc_release_cntr(adev, config); | |||
643 | } | |||
644 | ||||
645 | break; | |||
646 | default: | |||
647 | break; | |||
648 | } | |||
649 | ||||
650 | return ret; | |||
651 | } | |||
652 | ||||
653 | static void df_v3_6_pmc_get_count(struct amdgpu_device *adev, | |||
654 | uint64_t config, | |||
655 | uint64_t *count) | |||
656 | { | |||
657 | uint32_t lo_base_addr = 0, hi_base_addr = 0, lo_val = 0, hi_val = 0; | |||
658 | *count = 0; | |||
659 | ||||
660 | switch (adev->asic_type) { | |||
661 | case CHIP_VEGA20: | |||
662 | df_v3_6_pmc_get_read_settings(adev, config, &lo_base_addr, | |||
663 | &hi_base_addr); | |||
664 | ||||
665 | if ((lo_base_addr == 0) || (hi_base_addr == 0)) | |||
666 | return; | |||
667 | ||||
668 | /* rearm the counter or throw away count value on failure */ | |||
669 | if (df_v3_6_pmc_is_deferred(adev, config)) { | |||
670 | int rearm_err = df_v3_6_perfmon_arm_with_status(adev, | |||
671 | lo_base_addr, lo_val, | |||
672 | hi_base_addr, hi_val); | |||
673 | ||||
674 | if (rearm_err) | |||
675 | return; | |||
676 | ||||
677 | df_v3_6_pmc_set_deferred(adev, config, false0); | |||
678 | } | |||
679 | ||||
680 | df_v3_6_perfmon_rreg(adev, lo_base_addr, &lo_val, | |||
681 | hi_base_addr, &hi_val); | |||
682 | ||||
683 | *count = ((hi_val | 0ULL) << 32) | (lo_val | 0ULL); | |||
684 | ||||
685 | if (*count >= DF_V3_6_PERFMON_OVERFLOW0xFFFFFFFFFFFFULL) | |||
686 | *count = 0; | |||
687 | ||||
688 | DRM_DEBUG_DRIVER("config=%llx addr=%08x:%08x val=%08x:%08x",__drm_dbg(DRM_UT_DRIVER, "config=%llx addr=%08x:%08x val=%08x:%08x" , config, lo_base_addr, hi_base_addr, lo_val, hi_val) | |||
689 | config, lo_base_addr, hi_base_addr, lo_val, hi_val)__drm_dbg(DRM_UT_DRIVER, "config=%llx addr=%08x:%08x val=%08x:%08x" , config, lo_base_addr, hi_base_addr, lo_val, hi_val); | |||
690 | ||||
691 | break; | |||
692 | default: | |||
693 | break; | |||
694 | } | |||
695 | } | |||
696 | ||||
697 | const struct amdgpu_df_funcs df_v3_6_funcs = { | |||
698 | .sw_init = df_v3_6_sw_init, | |||
699 | .sw_fini = df_v3_6_sw_fini, | |||
700 | .enable_broadcast_mode = df_v3_6_enable_broadcast_mode, | |||
701 | .get_fb_channel_number = df_v3_6_get_fb_channel_number, | |||
702 | .get_hbm_channel_number = df_v3_6_get_hbm_channel_number, | |||
703 | .update_medium_grain_clock_gating = | |||
704 | df_v3_6_update_medium_grain_clock_gating, | |||
705 | .get_clockgating_state = df_v3_6_get_clockgating_state, | |||
706 | .pmc_start = df_v3_6_pmc_start, | |||
707 | .pmc_stop = df_v3_6_pmc_stop, | |||
708 | .pmc_get_count = df_v3_6_pmc_get_count, | |||
709 | .get_fica = df_v3_6_get_fica, | |||
710 | .set_fica = df_v3_6_set_fica, | |||
711 | }; |