| File: | dev/pci/drm/amd/pm/swsmu/smu_cmn.c |
| Warning: | line 121, column 2 Undefined or garbage value returned to caller |
Press '?' to see keyboard shortcuts
Keyboard shortcuts:
| 1 | /* | |||
| 2 | * Copyright 2020 Advanced Micro Devices, Inc. | |||
| 3 | * | |||
| 4 | * Permission is hereby granted, free of charge, to any person obtaining a | |||
| 5 | * copy of this software and associated documentation files (the "Software"), | |||
| 6 | * to deal in the Software without restriction, including without limitation | |||
| 7 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | |||
| 8 | * and/or sell copies of the Software, and to permit persons to whom the | |||
| 9 | * Software is furnished to do so, subject to the following conditions: | |||
| 10 | * | |||
| 11 | * The above copyright notice and this permission notice shall be included in | |||
| 12 | * all copies or substantial portions of the Software. | |||
| 13 | * | |||
| 14 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |||
| 15 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |||
| 16 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |||
| 17 | * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR | |||
| 18 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, | |||
| 19 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | |||
| 20 | * OTHER DEALINGS IN THE SOFTWARE. | |||
| 21 | */ | |||
| 22 | ||||
| 23 | #define SWSMU_CODE_LAYER_L4 | |||
| 24 | ||||
| 25 | #include "amdgpu.h" | |||
| 26 | #include "amdgpu_smu.h" | |||
| 27 | #include "smu_cmn.h" | |||
| 28 | #include "soc15_common.h" | |||
| 29 | ||||
| 30 | /* | |||
| 31 | * DO NOT use these for err/warn/info/debug messages. | |||
| 32 | * Use dev_err, dev_warn, dev_info and dev_dbg instead. | |||
| 33 | * They are more MGPU friendly. | |||
| 34 | */ | |||
| 35 | #undef pr_err | |||
| 36 | #undef pr_warn | |||
| 37 | #undef pr_info | |||
| 38 | #undef pr_debug | |||
| 39 | ||||
| 40 | #define MP1_C2PMSG_90__CONTENT_MASK0xFFFFFFFFL 0xFFFFFFFFL | |||
| 41 | ||||
| 42 | #undef __SMU_DUMMY_MAP | |||
| 43 | #define __SMU_DUMMY_MAP(type)"type" #type | |||
| 44 | static const char * const __smu_message_names[] = { | |||
| 45 | SMU_MESSAGE_TYPES"TestMessage", "GetSmuVersion", "GetDriverIfVersion", "SetAllowedFeaturesMaskLow" , "SetAllowedFeaturesMaskHigh", "EnableAllSmuFeatures", "DisableAllSmuFeatures" , "EnableSmuFeaturesLow", "EnableSmuFeaturesHigh", "DisableSmuFeaturesLow" , "DisableSmuFeaturesHigh", "GetEnabledSmuFeatures", "GetEnabledSmuFeaturesLow" , "GetEnabledSmuFeaturesHigh", "SetWorkloadMask", "SetPptLimit" , "SetDriverDramAddrHigh", "SetDriverDramAddrLow", "SetToolsDramAddrHigh" , "SetToolsDramAddrLow", "TransferTableSmu2Dram", "TransferTableDram2Smu" , "UseDefaultPPTable", "UseBackupPPTable", "RunBtc", "RequestI2CBus" , "ReleaseI2CBus", "SetFloorSocVoltage", "SoftReset", "StartBacoMonitor" , "CancelBacoMonitor", "EnterBaco", "SetSoftMinByFreq", "SetSoftMaxByFreq" , "SetHardMinByFreq", "SetHardMaxByFreq", "GetMinDpmFreq", "GetMaxDpmFreq" , "GetDpmFreqByIndex", "GetDpmClockFreq", "GetSsVoltageByDpm" , "SetMemoryChannelConfig", "SetGeminiMode", "SetGeminiApertureHigh" , "SetGeminiApertureLow", "SetMinLinkDpmByIndex", "OverridePcieParameters" , "OverDriveSetPercentage", "SetMinDeepSleepDcefclk", "ReenableAcDcInterrupt" , "AllowIHHostInterrupt", "NotifyPowerSource", "SetUclkFastSwitch" , "SetUclkDownHyst", "GfxDeviceDriverReset", "GetCurrentRpm", "SetVideoFps", "SetTjMax", "SetFanTemperatureTarget", "PrepareMp1ForUnload" , "DramLogSetDramAddrHigh", "DramLogSetDramAddrLow", "DramLogSetDramSize" , "SetFanMaxRpm", "SetFanMinPwm", "ConfigureGfxDidt", "NumOfDisplays" , "RemoveMargins", "ReadSerialNumTop32", "ReadSerialNumBottom32" , "SetSystemVirtualDramAddrHigh", "SetSystemVirtualDramAddrLow" , "WaflTest", "SetFclkGfxClkRatio", "AllowGfxOff", "DisallowGfxOff" , "GetPptLimit", "GetDcModeMaxDpmFreq", "GetDebugData", "SetXgmiMode" , "RunAfllBtc", "ExitBaco", "PrepareMp1ForReset", "PrepareMp1ForShutdown" , "SetMGpuFanBoostLimitRpm", "GetAVFSVoltageByDpm", "PowerUpVcn" , "PowerDownVcn", "PowerUpJpeg", "PowerDownJpeg", "BacoAudioD3PME" , "ArmD3", "RunDcBtc", "RunGfxDcBtc", "RunSocDcBtc", "SetMemoryChannelEnable" , "SetDfSwitchType", "GetVoltageByDpm", "GetVoltageByDpmOverdrive" , "PowerUpVcn0", "PowerDownVcn0", "PowerUpVcn1", "PowerDownVcn1" , "PowerUpGfx", "PowerDownIspByTile", "PowerUpIspByTile", "PowerDownSdma" , "PowerUpSdma", "SetHardMinIspclkByFreq", "SetHardMinVcn", "SetAllowFclkSwitch" , "SetMinVideoGfxclkFreq", "ActiveProcessNotify", "SetCustomPolicy" , "QueryPowerLimit", "SetGfxclkOverdriveByFreqVid", "SetHardMinDcfclkByFreq" , "SetHardMinSocclkByFreq", "ControlIgpuATS", "SetMinVideoFclkFreq" , "SetMinDeepSleepDcfclk", "ForcePowerDownGfx", "SetPhyclkVoltageByFreq" , "SetDppclkVoltageByFreq", "SetSoftMinVcn", "EnablePostCode" , "GetGfxclkFrequency", "GetFclkFrequency", "GetMinGfxclkFrequency" , "GetMaxGfxclkFrequency", "SetGfxCGPG", "SetSoftMaxGfxClk", "SetHardMinGfxClk" , "SetSoftMaxSocclkByFreq", "SetSoftMaxFclkByFreq", "SetSoftMaxVcn" , "PowerGateMmHub", "UpdatePmeRestore", "GpuChangeState", "SetPowerLimitPercentage" , "ForceGfxContentSave", "EnableTmdp48MHzRefclkPwrDown", "PowerGateAtHub" , "SetSoftMinJpeg", "SetHardMinFclkByFreq", "DFCstateControl" , "GmiPwrDnControl", "spare", "SetNumBadHbmPagesRetired", "GetGmiPwrDnHyst" , "SetGmiPwrDnHyst", "EnterGfxoff", "ExitGfxoff", "SetExecuteDMATest" , "DAL_DISABLE_DUMMY_PSTATE_CHANGE", "DAL_ENABLE_DUMMY_PSTATE_CHANGE" , "SET_DRIVER_DUMMY_TABLE_DRAM_ADDR_HIGH", "SET_DRIVER_DUMMY_TABLE_DRAM_ADDR_LOW" , "GET_UMC_FW_WA", "Mode1Reset", "RlcPowerNotify", "SetHardMinIspiclkByFreq" , "SetHardMinIspxclkByFreq", "SetSoftMinSocclkByFreq", "PowerUpCvip" , "PowerDownCvip", "EnableGfxOff", "DisableGfxOff", "SetSoftMinGfxclk" , "SetSoftMinFclk", "GetThermalLimit", "GetCurrentTemperature" , "GetCurrentPower", "GetCurrentVoltage", "GetCurrentCurrent" , "GetAverageCpuActivity", "GetAverageGfxActivity", "GetAveragePower" , "GetAverageTemperature", "SetAveragePowerTimeConstant", "SetAverageActivityTimeConstant" , "SetAverageTemperatureTimeConstant", "SetMitigationEndHysteresis" , "GetCurrentFreq", "SetReducedPptLimit", "SetReducedThermalLimit" , "DramLogSetDramAddr", "StartDramLogging", "StopDramLogging" , "SetSoftMinCclk", "SetSoftMaxCclk", "SetGpoFeaturePMask", "DisallowGpo" , "Enable2ndUSB20Port", "RequestActiveWgp", "SetFastPPTLimit" , "SetSlowPPTLimit", "GetFastPPTLimit", "GetSlowPPTLimit", "EnableDeterminism" , "DisableDeterminism", "SetUclkDpmMode", "LightSBR", "GfxDriverResetRecovery" , "BoardPowerCalibration", "RequestGfxclk", "ForceGfxVid", "Spare0" , "UnforceGfxVid", "HeavySBR", "SetBadHBMPagesRetiredFlagsPerChannel" , "EnableGfxImu", "DriverMode2Reset", "GetGfxOffStatus", "GetGfxOffEntryCount" , "LogGfxOffResidency", "SetNumBadMemoryPagesRetired", "SetBadMemoryPagesRetiredFlagsPerChannel" , "AllowGpo", | |||
| 46 | }; | |||
| 47 | ||||
| 48 | #define smu_cmn_call_asic_func(intf, smu, args...)((smu)->ppt_funcs ? ((smu)->ppt_funcs->intf ? (smu)-> ppt_funcs->intf(smu, args...) : -91) : -22) \ | |||
| 49 | ((smu)->ppt_funcs ? ((smu)->ppt_funcs->intf ? \ | |||
| 50 | (smu)->ppt_funcs->intf(smu, ##args) : \ | |||
| 51 | -ENOTSUPP91) : \ | |||
| 52 | -EINVAL22) | |||
| 53 | ||||
| 54 | static const char *smu_get_message_name(struct smu_context *smu, | |||
| 55 | enum smu_message_type type) | |||
| 56 | { | |||
| 57 | if (type < 0 || type >= SMU_MSG_MAX_COUNT) | |||
| 58 | return "unknown smu message"; | |||
| 59 | ||||
| 60 | return __smu_message_names[type]; | |||
| 61 | } | |||
| 62 | ||||
| 63 | static void smu_cmn_read_arg(struct smu_context *smu, | |||
| 64 | uint32_t *arg) | |||
| 65 | { | |||
| 66 | struct amdgpu_device *adev = smu->adev; | |||
| 67 | ||||
| 68 | *arg = RREG32(smu->param_reg)amdgpu_device_rreg(adev, (smu->param_reg), 0); | |||
| 69 | } | |||
| 70 | ||||
| 71 | /* Redefine the SMU error codes here. | |||
| 72 | * | |||
| 73 | * Note that these definitions are redundant and should be removed | |||
| 74 | * when the SMU has exported a unified header file containing these | |||
| 75 | * macros, which header file we can just include and use the SMU's | |||
| 76 | * macros. At the moment, these error codes are defined by the SMU | |||
| 77 | * per-ASIC unfortunately, yet we're a one driver for all ASICs. | |||
| 78 | */ | |||
| 79 | #define SMU_RESP_NONE0 0 | |||
| 80 | #define SMU_RESP_OK1 1 | |||
| 81 | #define SMU_RESP_CMD_FAIL0xFF 0xFF | |||
| 82 | #define SMU_RESP_CMD_UNKNOWN0xFE 0xFE | |||
| 83 | #define SMU_RESP_CMD_BAD_PREREQ0xFD 0xFD | |||
| 84 | #define SMU_RESP_BUSY_OTHER0xFC 0xFC | |||
| 85 | #define SMU_RESP_DEBUG_END0xFB 0xFB | |||
| 86 | ||||
| 87 | /** | |||
| 88 | * __smu_cmn_poll_stat -- poll for a status from the SMU | |||
| 89 | * @smu: a pointer to SMU context | |||
| 90 | * | |||
| 91 | * Returns the status of the SMU, which could be, | |||
| 92 | * 0, the SMU is busy with your command; | |||
| 93 | * 1, execution status: success, execution result: success; | |||
| 94 | * 0xFF, execution status: success, execution result: failure; | |||
| 95 | * 0xFE, unknown command; | |||
| 96 | * 0xFD, valid command, but bad (command) prerequisites; | |||
| 97 | * 0xFC, the command was rejected as the SMU is busy; | |||
| 98 | * 0xFB, "SMC_Result_DebugDataDumpEnd". | |||
| 99 | * | |||
| 100 | * The values here are not defined by macros, because I'd rather we | |||
| 101 | * include a single header file which defines them, which is | |||
| 102 | * maintained by the SMU FW team, so that we're impervious to firmware | |||
| 103 | * changes. At the moment those values are defined in various header | |||
| 104 | * files, one for each ASIC, yet here we're a single ASIC-agnostic | |||
| 105 | * interface. Such a change can be followed-up by a subsequent patch. | |||
| 106 | */ | |||
| 107 | static u32 __smu_cmn_poll_stat(struct smu_context *smu) | |||
| 108 | { | |||
| 109 | struct amdgpu_device *adev = smu->adev; | |||
| 110 | int timeout = adev->usec_timeout * 20; | |||
| 111 | u32 reg; | |||
| 112 | ||||
| 113 | for ( ; timeout > 0; timeout--) { | |||
| 114 | reg = RREG32(smu->resp_reg)amdgpu_device_rreg(adev, (smu->resp_reg), 0); | |||
| 115 | if ((reg & MP1_C2PMSG_90__CONTENT_MASK0xFFFFFFFFL) != 0) | |||
| 116 | break; | |||
| 117 | ||||
| 118 | udelay(1); | |||
| 119 | } | |||
| 120 | ||||
| 121 | return reg; | |||
| ||||
| 122 | } | |||
| 123 | ||||
| 124 | static void __smu_cmn_reg_print_error(struct smu_context *smu, | |||
| 125 | u32 reg_c2pmsg_90, | |||
| 126 | int msg_index, | |||
| 127 | u32 param, | |||
| 128 | enum smu_message_type msg) | |||
| 129 | { | |||
| 130 | struct amdgpu_device *adev = smu->adev; | |||
| 131 | const char *message = smu_get_message_name(smu, msg); | |||
| 132 | u32 msg_idx, prm; | |||
| 133 | ||||
| 134 | switch (reg_c2pmsg_90) { | |||
| 135 | case SMU_RESP_NONE0: { | |||
| 136 | msg_idx = RREG32(smu->msg_reg)amdgpu_device_rreg(adev, (smu->msg_reg), 0); | |||
| 137 | prm = RREG32(smu->param_reg)amdgpu_device_rreg(adev, (smu->param_reg), 0); | |||
| 138 | dev_err_ratelimited(adev->dev,printf("drm:pid%d:%s *ERROR* " "SMU: I'm not done with your previous command: SMN_C2PMSG_66:0x%08X SMN_C2PMSG_82:0x%08X" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_idx , prm) | |||
| 139 | "SMU: I'm not done with your previous command: SMN_C2PMSG_66:0x%08X SMN_C2PMSG_82:0x%08X",printf("drm:pid%d:%s *ERROR* " "SMU: I'm not done with your previous command: SMN_C2PMSG_66:0x%08X SMN_C2PMSG_82:0x%08X" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_idx , prm) | |||
| 140 | msg_idx, prm)printf("drm:pid%d:%s *ERROR* " "SMU: I'm not done with your previous command: SMN_C2PMSG_66:0x%08X SMN_C2PMSG_82:0x%08X" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_idx , prm); | |||
| 141 | } | |||
| 142 | break; | |||
| 143 | case SMU_RESP_OK1: | |||
| 144 | /* The SMU executed the command. It completed with a | |||
| 145 | * successful result. | |||
| 146 | */ | |||
| 147 | break; | |||
| 148 | case SMU_RESP_CMD_FAIL0xFF: | |||
| 149 | /* The SMU executed the command. It completed with an | |||
| 150 | * unsuccessful result. | |||
| 151 | */ | |||
| 152 | break; | |||
| 153 | case SMU_RESP_CMD_UNKNOWN0xFE: | |||
| 154 | dev_err_ratelimited(adev->dev,printf("drm:pid%d:%s *ERROR* " "SMU: unknown command: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message) | |||
| 155 | "SMU: unknown command: index:%d param:0x%08X message:%s",printf("drm:pid%d:%s *ERROR* " "SMU: unknown command: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message) | |||
| 156 | msg_index, param, message)printf("drm:pid%d:%s *ERROR* " "SMU: unknown command: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message); | |||
| 157 | break; | |||
| 158 | case SMU_RESP_CMD_BAD_PREREQ0xFD: | |||
| 159 | dev_err_ratelimited(adev->dev,printf("drm:pid%d:%s *ERROR* " "SMU: valid command, bad prerequisites: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message) | |||
| 160 | "SMU: valid command, bad prerequisites: index:%d param:0x%08X message:%s",printf("drm:pid%d:%s *ERROR* " "SMU: valid command, bad prerequisites: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message) | |||
| 161 | msg_index, param, message)printf("drm:pid%d:%s *ERROR* " "SMU: valid command, bad prerequisites: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message); | |||
| 162 | break; | |||
| 163 | case SMU_RESP_BUSY_OTHER0xFC: | |||
| 164 | dev_err_ratelimited(adev->dev,printf("drm:pid%d:%s *ERROR* " "SMU: I'm very busy for your command: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message) | |||
| 165 | "SMU: I'm very busy for your command: index:%d param:0x%08X message:%s",printf("drm:pid%d:%s *ERROR* " "SMU: I'm very busy for your command: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message) | |||
| 166 | msg_index, param, message)printf("drm:pid%d:%s *ERROR* " "SMU: I'm very busy for your command: index:%d param:0x%08X message:%s" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , msg_index , param, message); | |||
| 167 | break; | |||
| 168 | case SMU_RESP_DEBUG_END0xFB: | |||
| 169 | dev_err_ratelimited(adev->dev,printf("drm:pid%d:%s *ERROR* " "SMU: I'm debugging!", ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci ) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci; })->ci_curproc->p_p->ps_pid, __func__) | |||
| 170 | "SMU: I'm debugging!")printf("drm:pid%d:%s *ERROR* " "SMU: I'm debugging!", ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci ) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci; })->ci_curproc->p_p->ps_pid, __func__); | |||
| 171 | break; | |||
| 172 | default: | |||
| 173 | dev_err_ratelimited(adev->dev,printf("drm:pid%d:%s *ERROR* " "SMU: response:0x%08X for index:%d param:0x%08X message:%s?" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , reg_c2pmsg_90 , msg_index, param, message) | |||
| 174 | "SMU: response:0x%08X for index:%d param:0x%08X message:%s?",printf("drm:pid%d:%s *ERROR* " "SMU: response:0x%08X for index:%d param:0x%08X message:%s?" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , reg_c2pmsg_90 , msg_index, param, message) | |||
| 175 | reg_c2pmsg_90, msg_index, param, message)printf("drm:pid%d:%s *ERROR* " "SMU: response:0x%08X for index:%d param:0x%08X message:%s?" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__ , reg_c2pmsg_90 , msg_index, param, message); | |||
| 176 | break; | |||
| 177 | } | |||
| 178 | } | |||
| 179 | ||||
| 180 | static int __smu_cmn_reg2errno(struct smu_context *smu, u32 reg_c2pmsg_90) | |||
| 181 | { | |||
| 182 | int res; | |||
| 183 | ||||
| 184 | switch (reg_c2pmsg_90) { | |||
| 185 | case SMU_RESP_NONE0: | |||
| 186 | /* The SMU is busy--still executing your command. | |||
| 187 | */ | |||
| 188 | res = -ETIME60; | |||
| 189 | break; | |||
| 190 | case SMU_RESP_OK1: | |||
| 191 | res = 0; | |||
| 192 | break; | |||
| 193 | case SMU_RESP_CMD_FAIL0xFF: | |||
| 194 | /* Command completed successfully, but the command | |||
| 195 | * status was failure. | |||
| 196 | */ | |||
| 197 | res = -EIO5; | |||
| 198 | break; | |||
| 199 | case SMU_RESP_CMD_UNKNOWN0xFE: | |||
| 200 | /* Unknown command--ignored by the SMU. | |||
| 201 | */ | |||
| 202 | res = -EOPNOTSUPP45; | |||
| 203 | break; | |||
| 204 | case SMU_RESP_CMD_BAD_PREREQ0xFD: | |||
| 205 | /* Valid command--bad prerequisites. | |||
| 206 | */ | |||
| 207 | res = -EINVAL22; | |||
| 208 | break; | |||
| 209 | case SMU_RESP_BUSY_OTHER0xFC: | |||
| 210 | /* The SMU is busy with other commands. The client | |||
| 211 | * should retry in 10 us. | |||
| 212 | */ | |||
| 213 | res = -EBUSY16; | |||
| 214 | break; | |||
| 215 | default: | |||
| 216 | /* Unknown or debug response from the SMU. | |||
| 217 | */ | |||
| 218 | res = -EREMOTEIO5; | |||
| 219 | break; | |||
| 220 | } | |||
| 221 | ||||
| 222 | return res; | |||
| 223 | } | |||
| 224 | ||||
| 225 | static void __smu_cmn_send_msg(struct smu_context *smu, | |||
| 226 | u16 msg, | |||
| 227 | u32 param) | |||
| 228 | { | |||
| 229 | struct amdgpu_device *adev = smu->adev; | |||
| 230 | ||||
| 231 | WREG32(smu->resp_reg, 0)amdgpu_device_wreg(adev, (smu->resp_reg), (0), 0); | |||
| 232 | WREG32(smu->param_reg, param)amdgpu_device_wreg(adev, (smu->param_reg), (param), 0); | |||
| 233 | WREG32(smu->msg_reg, msg)amdgpu_device_wreg(adev, (smu->msg_reg), (msg), 0); | |||
| 234 | } | |||
| 235 | ||||
| 236 | static int __smu_cmn_send_debug_msg(struct smu_context *smu, | |||
| 237 | u32 msg, | |||
| 238 | u32 param) | |||
| 239 | { | |||
| 240 | struct amdgpu_device *adev = smu->adev; | |||
| 241 | ||||
| 242 | WREG32(smu->debug_param_reg, param)amdgpu_device_wreg(adev, (smu->debug_param_reg), (param), 0 ); | |||
| 243 | WREG32(smu->debug_msg_reg, msg)amdgpu_device_wreg(adev, (smu->debug_msg_reg), (msg), 0); | |||
| 244 | WREG32(smu->debug_resp_reg, 0)amdgpu_device_wreg(adev, (smu->debug_resp_reg), (0), 0); | |||
| 245 | ||||
| 246 | return 0; | |||
| 247 | } | |||
| 248 | /** | |||
| 249 | * smu_cmn_send_msg_without_waiting -- send the message; don't wait for status | |||
| 250 | * @smu: pointer to an SMU context | |||
| 251 | * @msg_index: message index | |||
| 252 | * @param: message parameter to send to the SMU | |||
| 253 | * | |||
| 254 | * Send a message to the SMU with the parameter passed. Do not wait | |||
| 255 | * for status/result of the message, thus the "without_waiting". | |||
| 256 | * | |||
| 257 | * Return 0 on success, -errno on error if we weren't able to _send_ | |||
| 258 | * the message for some reason. See __smu_cmn_reg2errno() for details | |||
| 259 | * of the -errno. | |||
| 260 | */ | |||
| 261 | int smu_cmn_send_msg_without_waiting(struct smu_context *smu, | |||
| 262 | uint16_t msg_index, | |||
| 263 | uint32_t param) | |||
| 264 | { | |||
| 265 | struct amdgpu_device *adev = smu->adev; | |||
| 266 | u32 reg; | |||
| 267 | int res; | |||
| 268 | ||||
| 269 | if (adev->no_hw_access) | |||
| 270 | return 0; | |||
| 271 | ||||
| 272 | reg = __smu_cmn_poll_stat(smu); | |||
| 273 | res = __smu_cmn_reg2errno(smu, reg); | |||
| 274 | if (reg == SMU_RESP_NONE0 || | |||
| 275 | res == -EREMOTEIO5) | |||
| 276 | goto Out; | |||
| 277 | __smu_cmn_send_msg(smu, msg_index, param); | |||
| 278 | res = 0; | |||
| 279 | Out: | |||
| 280 | if (unlikely(adev->pm.smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR)__builtin_expect(!!(adev->pm.smu_debug_mask & 0x1), 0) && | |||
| 281 | res && (res != -ETIME60)) { | |||
| 282 | amdgpu_device_halt(adev); | |||
| 283 | WARN_ON(1)({ int __ret = !!(1); if (__ret) printf("WARNING %s failed at %s:%d\n" , "1", "/usr/src/sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c", 283 ); __builtin_expect(!!(__ret), 0); }); | |||
| 284 | } | |||
| 285 | ||||
| 286 | return res; | |||
| 287 | } | |||
| 288 | ||||
| 289 | /** | |||
| 290 | * smu_cmn_wait_for_response -- wait for response from the SMU | |||
| 291 | * @smu: pointer to an SMU context | |||
| 292 | * | |||
| 293 | * Wait for status from the SMU. | |||
| 294 | * | |||
| 295 | * Return 0 on success, -errno on error, indicating the execution | |||
| 296 | * status and result of the message being waited for. See | |||
| 297 | * __smu_cmn_reg2errno() for details of the -errno. | |||
| 298 | */ | |||
| 299 | int smu_cmn_wait_for_response(struct smu_context *smu) | |||
| 300 | { | |||
| 301 | u32 reg; | |||
| 302 | int res; | |||
| 303 | ||||
| 304 | reg = __smu_cmn_poll_stat(smu); | |||
| 305 | res = __smu_cmn_reg2errno(smu, reg); | |||
| 306 | ||||
| 307 | if (unlikely(smu->adev->pm.smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR)__builtin_expect(!!(smu->adev->pm.smu_debug_mask & 0x1 ), 0) && | |||
| 308 | res && (res != -ETIME60)) { | |||
| 309 | amdgpu_device_halt(smu->adev); | |||
| 310 | WARN_ON(1)({ int __ret = !!(1); if (__ret) printf("WARNING %s failed at %s:%d\n" , "1", "/usr/src/sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c", 310 ); __builtin_expect(!!(__ret), 0); }); | |||
| 311 | } | |||
| 312 | ||||
| 313 | return res; | |||
| 314 | } | |||
| 315 | ||||
| 316 | /** | |||
| 317 | * smu_cmn_send_smc_msg_with_param -- send a message with parameter | |||
| 318 | * @smu: pointer to an SMU context | |||
| 319 | * @msg: message to send | |||
| 320 | * @param: parameter to send to the SMU | |||
| 321 | * @read_arg: pointer to u32 to return a value from the SMU back | |||
| 322 | * to the caller | |||
| 323 | * | |||
| 324 | * Send the message @msg with parameter @param to the SMU, wait for | |||
| 325 | * completion of the command, and return back a value from the SMU in | |||
| 326 | * @read_arg pointer. | |||
| 327 | * | |||
| 328 | * Return 0 on success, -errno when a problem is encountered sending | |||
| 329 | * message or receiving reply. If there is a PCI bus recovery or | |||
| 330 | * the destination is a virtual GPU which does not allow this message | |||
| 331 | * type, the message is simply dropped and success is also returned. | |||
| 332 | * See __smu_cmn_reg2errno() for details of the -errno. | |||
| 333 | * | |||
| 334 | * If we weren't able to send the message to the SMU, we also print | |||
| 335 | * the error to the standard log. | |||
| 336 | * | |||
| 337 | * Command completion status is printed only if the -errno is | |||
| 338 | * -EREMOTEIO, indicating that the SMU returned back an | |||
| 339 | * undefined/unknown/unspecified result. All other cases are | |||
| 340 | * well-defined, not printed, but instead given back to the client to | |||
| 341 | * decide what further to do. | |||
| 342 | * | |||
| 343 | * The return value, @read_arg is read back regardless, to give back | |||
| 344 | * more information to the client, which on error would most likely be | |||
| 345 | * @param, but we can't assume that. This also eliminates more | |||
| 346 | * conditionals. | |||
| 347 | */ | |||
| 348 | int smu_cmn_send_smc_msg_with_param(struct smu_context *smu, | |||
| 349 | enum smu_message_type msg, | |||
| 350 | uint32_t param, | |||
| 351 | uint32_t *read_arg) | |||
| 352 | { | |||
| 353 | struct amdgpu_device *adev = smu->adev; | |||
| 354 | int res, index; | |||
| 355 | u32 reg; | |||
| 356 | ||||
| 357 | if (adev->no_hw_access) | |||
| 358 | return 0; | |||
| 359 | ||||
| 360 | index = smu_cmn_to_asic_specific_index(smu, | |||
| 361 | CMN2ASIC_MAPPING_MSG, | |||
| 362 | msg); | |||
| 363 | if (index < 0) | |||
| 364 | return index == -EACCES13 ? 0 : index; | |||
| 365 | ||||
| 366 | mutex_lock(&smu->message_lock)rw_enter_write(&smu->message_lock); | |||
| 367 | reg = __smu_cmn_poll_stat(smu); | |||
| 368 | res = __smu_cmn_reg2errno(smu, reg); | |||
| 369 | if (reg == SMU_RESP_NONE0 || | |||
| 370 | res == -EREMOTEIO5) { | |||
| 371 | __smu_cmn_reg_print_error(smu, reg, index, param, msg); | |||
| 372 | goto Out; | |||
| 373 | } | |||
| 374 | __smu_cmn_send_msg(smu, (uint16_t) index, param); | |||
| 375 | reg = __smu_cmn_poll_stat(smu); | |||
| 376 | res = __smu_cmn_reg2errno(smu, reg); | |||
| 377 | if (res != 0) | |||
| 378 | __smu_cmn_reg_print_error(smu, reg, index, param, msg); | |||
| 379 | if (read_arg) | |||
| 380 | smu_cmn_read_arg(smu, read_arg); | |||
| 381 | Out: | |||
| 382 | if (unlikely(adev->pm.smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR)__builtin_expect(!!(adev->pm.smu_debug_mask & 0x1), 0) && res) { | |||
| 383 | amdgpu_device_halt(adev); | |||
| 384 | WARN_ON(1)({ int __ret = !!(1); if (__ret) printf("WARNING %s failed at %s:%d\n" , "1", "/usr/src/sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c", 384 ); __builtin_expect(!!(__ret), 0); }); | |||
| 385 | } | |||
| 386 | ||||
| 387 | mutex_unlock(&smu->message_lock)rw_exit_write(&smu->message_lock); | |||
| 388 | return res; | |||
| 389 | } | |||
| 390 | ||||
| 391 | int smu_cmn_send_smc_msg(struct smu_context *smu, | |||
| 392 | enum smu_message_type msg, | |||
| 393 | uint32_t *read_arg) | |||
| 394 | { | |||
| 395 | return smu_cmn_send_smc_msg_with_param(smu, | |||
| 396 | msg, | |||
| 397 | 0, | |||
| 398 | read_arg); | |||
| 399 | } | |||
| 400 | ||||
| 401 | int smu_cmn_send_debug_smc_msg(struct smu_context *smu, | |||
| 402 | uint32_t msg) | |||
| 403 | { | |||
| 404 | return __smu_cmn_send_debug_msg(smu, msg, 0); | |||
| 405 | } | |||
| 406 | ||||
| 407 | int smu_cmn_to_asic_specific_index(struct smu_context *smu, | |||
| 408 | enum smu_cmn2asic_mapping_type type, | |||
| 409 | uint32_t index) | |||
| 410 | { | |||
| 411 | struct cmn2asic_msg_mapping msg_mapping; | |||
| 412 | struct cmn2asic_mapping mapping; | |||
| 413 | ||||
| 414 | switch (type) { | |||
| 415 | case CMN2ASIC_MAPPING_MSG: | |||
| 416 | if (index >= SMU_MSG_MAX_COUNT || | |||
| 417 | !smu->message_map) | |||
| 418 | return -EINVAL22; | |||
| 419 | ||||
| 420 | msg_mapping = smu->message_map[index]; | |||
| 421 | if (!msg_mapping.valid_mapping) | |||
| 422 | return -EINVAL22; | |||
| 423 | ||||
| 424 | if (amdgpu_sriov_vf(smu->adev)((smu->adev)->virt.caps & (1 << 2)) && | |||
| 425 | !msg_mapping.valid_in_vf) | |||
| 426 | return -EACCES13; | |||
| 427 | ||||
| 428 | return msg_mapping.map_to; | |||
| 429 | ||||
| 430 | case CMN2ASIC_MAPPING_CLK: | |||
| 431 | if (index >= SMU_CLK_COUNT || | |||
| 432 | !smu->clock_map) | |||
| 433 | return -EINVAL22; | |||
| 434 | ||||
| 435 | mapping = smu->clock_map[index]; | |||
| 436 | if (!mapping.valid_mapping) | |||
| 437 | return -EINVAL22; | |||
| 438 | ||||
| 439 | return mapping.map_to; | |||
| 440 | ||||
| 441 | case CMN2ASIC_MAPPING_FEATURE: | |||
| 442 | if (index >= SMU_FEATURE_COUNT || | |||
| 443 | !smu->feature_map) | |||
| 444 | return -EINVAL22; | |||
| 445 | ||||
| 446 | mapping = smu->feature_map[index]; | |||
| 447 | if (!mapping.valid_mapping) | |||
| 448 | return -EINVAL22; | |||
| 449 | ||||
| 450 | return mapping.map_to; | |||
| 451 | ||||
| 452 | case CMN2ASIC_MAPPING_TABLE: | |||
| 453 | if (index >= SMU_TABLE_COUNT || | |||
| 454 | !smu->table_map) | |||
| 455 | return -EINVAL22; | |||
| 456 | ||||
| 457 | mapping = smu->table_map[index]; | |||
| 458 | if (!mapping.valid_mapping) | |||
| 459 | return -EINVAL22; | |||
| 460 | ||||
| 461 | return mapping.map_to; | |||
| 462 | ||||
| 463 | case CMN2ASIC_MAPPING_PWR: | |||
| 464 | if (index >= SMU_POWER_SOURCE_COUNT || | |||
| 465 | !smu->pwr_src_map) | |||
| 466 | return -EINVAL22; | |||
| 467 | ||||
| 468 | mapping = smu->pwr_src_map[index]; | |||
| 469 | if (!mapping.valid_mapping) | |||
| 470 | return -EINVAL22; | |||
| 471 | ||||
| 472 | return mapping.map_to; | |||
| 473 | ||||
| 474 | case CMN2ASIC_MAPPING_WORKLOAD: | |||
| 475 | if (index > PP_SMC_POWER_PROFILE_WINDOW3D || | |||
| 476 | !smu->workload_map) | |||
| 477 | return -EINVAL22; | |||
| 478 | ||||
| 479 | mapping = smu->workload_map[index]; | |||
| 480 | if (!mapping.valid_mapping) | |||
| 481 | return -EINVAL22; | |||
| 482 | ||||
| 483 | return mapping.map_to; | |||
| 484 | ||||
| 485 | default: | |||
| 486 | return -EINVAL22; | |||
| 487 | } | |||
| 488 | } | |||
| 489 | ||||
| 490 | int smu_cmn_feature_is_supported(struct smu_context *smu, | |||
| 491 | enum smu_feature_mask mask) | |||
| 492 | { | |||
| 493 | struct smu_feature *feature = &smu->smu_feature; | |||
| 494 | int feature_id; | |||
| 495 | ||||
| 496 | feature_id = smu_cmn_to_asic_specific_index(smu, | |||
| 497 | CMN2ASIC_MAPPING_FEATURE, | |||
| 498 | mask); | |||
| 499 | if (feature_id < 0) | |||
| 500 | return 0; | |||
| 501 | ||||
| 502 | WARN_ON(feature_id > feature->feature_num)({ int __ret = !!(feature_id > feature->feature_num); if (__ret) printf("WARNING %s failed at %s:%d\n", "feature_id > feature->feature_num" , "/usr/src/sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c", 502); __builtin_expect (!!(__ret), 0); }); | |||
| 503 | ||||
| 504 | return test_bit(feature_id, feature->supported); | |||
| 505 | } | |||
| 506 | ||||
| 507 | static int __smu_get_enabled_features(struct smu_context *smu, | |||
| 508 | uint64_t *enabled_features) | |||
| 509 | { | |||
| 510 | return smu_cmn_call_asic_func(get_enabled_mask, smu, enabled_features)((smu)->ppt_funcs ? ((smu)->ppt_funcs->get_enabled_mask ? (smu)->ppt_funcs->get_enabled_mask(smu, enabled_features ) : -91) : -22); | |||
| 511 | } | |||
| 512 | ||||
| 513 | int smu_cmn_feature_is_enabled(struct smu_context *smu, | |||
| 514 | enum smu_feature_mask mask) | |||
| 515 | { | |||
| 516 | struct amdgpu_device *adev = smu->adev; | |||
| 517 | uint64_t enabled_features; | |||
| 518 | int feature_id; | |||
| 519 | ||||
| 520 | if (__smu_get_enabled_features(smu, &enabled_features)) { | |||
| 521 | dev_err(adev->dev, "Failed to retrieve enabled ppfeatures!\n")printf("drm:pid%d:%s *ERROR* " "Failed to retrieve enabled ppfeatures!\n" , ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_p->ps_pid, __func__); | |||
| 522 | return 0; | |||
| 523 | } | |||
| 524 | ||||
| 525 | /* | |||
| 526 | * For Renoir and Cyan Skillfish, they are assumed to have all features | |||
| 527 | * enabled. Also considering they have no feature_map available, the | |||
| 528 | * check here can avoid unwanted feature_map check below. | |||
| 529 | */ | |||
| 530 | if (enabled_features == ULLONG_MAX0xffffffffffffffffULL) | |||
| 531 | return 1; | |||
| 532 | ||||
| 533 | feature_id = smu_cmn_to_asic_specific_index(smu, | |||
| 534 | CMN2ASIC_MAPPING_FEATURE, | |||
| 535 | mask); | |||
| 536 | if (feature_id < 0) | |||
| 537 | return 0; | |||
| 538 | ||||
| 539 | return test_bit(feature_id, (unsigned long *)&enabled_features); | |||
| 540 | } | |||
| 541 | ||||
| 542 | bool_Bool smu_cmn_clk_dpm_is_enabled(struct smu_context *smu, | |||
| 543 | enum smu_clk_type clk_type) | |||
| 544 | { | |||
| 545 | enum smu_feature_mask feature_id = 0; | |||
| 546 | ||||
| 547 | switch (clk_type) { | |||
| 548 | case SMU_MCLK: | |||
| 549 | case SMU_UCLK: | |||
| 550 | feature_id = SMU_FEATURE_DPM_UCLK_BIT; | |||
| 551 | break; | |||
| 552 | case SMU_GFXCLK: | |||
| 553 | case SMU_SCLK: | |||
| 554 | feature_id = SMU_FEATURE_DPM_GFXCLK_BIT; | |||
| 555 | break; | |||
| 556 | case SMU_SOCCLK: | |||
| 557 | feature_id = SMU_FEATURE_DPM_SOCCLK_BIT; | |||
| 558 | break; | |||
| 559 | case SMU_VCLK: | |||
| 560 | case SMU_VCLK1: | |||
| 561 | feature_id = SMU_FEATURE_DPM_VCLK_BIT; | |||
| 562 | break; | |||
| 563 | case SMU_DCLK: | |||
| 564 | case SMU_DCLK1: | |||
| 565 | feature_id = SMU_FEATURE_DPM_DCLK_BIT; | |||
| 566 | break; | |||
| 567 | case SMU_FCLK: | |||
| 568 | feature_id = SMU_FEATURE_DPM_FCLK_BIT; | |||
| 569 | break; | |||
| 570 | default: | |||
| 571 | return true1; | |||
| 572 | } | |||
| 573 | ||||
| 574 | if (!smu_cmn_feature_is_enabled(smu, feature_id)) | |||
| 575 | return false0; | |||
| 576 | ||||
| 577 | return true1; | |||
| 578 | } | |||
| 579 | ||||
| 580 | int smu_cmn_get_enabled_mask(struct smu_context *smu, | |||
| 581 | uint64_t *feature_mask) | |||
| 582 | { | |||
| 583 | uint32_t *feature_mask_high; | |||
| 584 | uint32_t *feature_mask_low; | |||
| 585 | int ret = 0, index = 0; | |||
| 586 | ||||
| 587 | if (!feature_mask) | |||
| 588 | return -EINVAL22; | |||
| 589 | ||||
| 590 | feature_mask_low = &((uint32_t *)feature_mask)[0]; | |||
| 591 | feature_mask_high = &((uint32_t *)feature_mask)[1]; | |||
| 592 | ||||
| 593 | index = smu_cmn_to_asic_specific_index(smu, | |||
| 594 | CMN2ASIC_MAPPING_MSG, | |||
| 595 | SMU_MSG_GetEnabledSmuFeatures); | |||
| 596 | if (index > 0) { | |||
| 597 | ret = smu_cmn_send_smc_msg_with_param(smu, | |||
| 598 | SMU_MSG_GetEnabledSmuFeatures, | |||
| 599 | 0, | |||
| 600 | feature_mask_low); | |||
| 601 | if (ret) | |||
| 602 | return ret; | |||
| 603 | ||||
| 604 | ret = smu_cmn_send_smc_msg_with_param(smu, | |||
| 605 | SMU_MSG_GetEnabledSmuFeatures, | |||
| 606 | 1, | |||
| 607 | feature_mask_high); | |||
| 608 | } else { | |||
| 609 | ret = smu_cmn_send_smc_msg(smu, | |||
| 610 | SMU_MSG_GetEnabledSmuFeaturesHigh, | |||
| 611 | feature_mask_high); | |||
| 612 | if (ret) | |||
| 613 | return ret; | |||
| 614 | ||||
| 615 | ret = smu_cmn_send_smc_msg(smu, | |||
| 616 | SMU_MSG_GetEnabledSmuFeaturesLow, | |||
| 617 | feature_mask_low); | |||
| 618 | } | |||
| 619 | ||||
| 620 | return ret; | |||
| 621 | } | |||
| 622 | ||||
| 623 | uint64_t smu_cmn_get_indep_throttler_status( | |||
| 624 | const unsigned long dep_status, | |||
| 625 | const uint8_t *throttler_map) | |||
| 626 | { | |||
| 627 | uint64_t indep_status = 0; | |||
| 628 | uint8_t dep_bit = 0; | |||
| 629 | ||||
| 630 | for_each_set_bit(dep_bit, (unsigned long *)&dep_status, 32)for ((dep_bit) = find_first_bit(((unsigned long *)&dep_status ), (32)); (dep_bit) < (32); (dep_bit) = find_next_bit(((unsigned long *)&dep_status), (32), (dep_bit) + 1)) | |||
| 631 | indep_status |= 1ULL << throttler_map[dep_bit]; | |||
| 632 | ||||
| 633 | return indep_status; | |||
| 634 | } | |||
| 635 | ||||
| 636 | int smu_cmn_feature_update_enable_state(struct smu_context *smu, | |||
| 637 | uint64_t feature_mask, | |||
| 638 | bool_Bool enabled) | |||
| 639 | { | |||
| 640 | int ret = 0; | |||
| 641 | ||||
| 642 | if (enabled) { | |||
| 643 | ret = smu_cmn_send_smc_msg_with_param(smu, | |||
| 644 | SMU_MSG_EnableSmuFeaturesLow, | |||
| 645 | lower_32_bits(feature_mask)((u32)(feature_mask)), | |||
| 646 | NULL((void *)0)); | |||
| 647 | if (ret) | |||
| 648 | return ret; | |||
| 649 | ret = smu_cmn_send_smc_msg_with_param(smu, | |||
| 650 | SMU_MSG_EnableSmuFeaturesHigh, | |||
| 651 | upper_32_bits(feature_mask)((u32)(((feature_mask) >> 16) >> 16)), | |||
| 652 | NULL((void *)0)); | |||
| 653 | } else { | |||
| 654 | ret = smu_cmn_send_smc_msg_with_param(smu, | |||
| 655 | SMU_MSG_DisableSmuFeaturesLow, | |||
| 656 | lower_32_bits(feature_mask)((u32)(feature_mask)), | |||
| 657 | NULL((void *)0)); | |||
| 658 | if (ret) | |||
| 659 | return ret; | |||
| 660 | ret = smu_cmn_send_smc_msg_with_param(smu, | |||
| 661 | SMU_MSG_DisableSmuFeaturesHigh, | |||
| 662 | upper_32_bits(feature_mask)((u32)(((feature_mask) >> 16) >> 16)), | |||
| 663 | NULL((void *)0)); | |||
| 664 | } | |||
| 665 | ||||
| 666 | return ret; | |||
| 667 | } | |||
| 668 | ||||
| 669 | int smu_cmn_feature_set_enabled(struct smu_context *smu, | |||
| 670 | enum smu_feature_mask mask, | |||
| 671 | bool_Bool enable) | |||
| 672 | { | |||
| 673 | int feature_id; | |||
| 674 | ||||
| 675 | feature_id = smu_cmn_to_asic_specific_index(smu, | |||
| 676 | CMN2ASIC_MAPPING_FEATURE, | |||
| 677 | mask); | |||
| 678 | if (feature_id < 0) | |||
| 679 | return -EINVAL22; | |||
| 680 | ||||
| 681 | return smu_cmn_feature_update_enable_state(smu, | |||
| 682 | 1ULL << feature_id, | |||
| 683 | enable); | |||
| 684 | } | |||
| 685 | ||||
| 686 | #undef __SMU_DUMMY_MAP | |||
| 687 | #define __SMU_DUMMY_MAP(fea)"fea" #fea | |||
| 688 | static const char* __smu_feature_names[] = { | |||
| 689 | SMU_FEATURE_MASKS"DPM_PREFETCHER", "DPM_GFXCLK", "DPM_UCLK", "DPM_SOCCLK", "DPM_UVD" , "DPM_VCE", "DPM_LCLK", "ULV", "DPM_MP0CLK", "DPM_LINK", "DPM_DCEFCLK" , "DPM_XGMI", "DS_GFXCLK", "DS_SOCCLK", "DS_LCLK", "PPT", "TDC" , "THERMAL", "GFX_PER_CU_CG", "DATA_CALCULATIONS", "RM", "DS_DCEFCLK" , "ACDC", "VR0HOT", "VR1HOT", "FW_CTF", "LED_DISPLAY", "FAN_CONTROL" , "GFX_EDC", "GFXOFF", "CG", "DPM_FCLK", "DS_FCLK", "DS_MP1CLK" , "DS_MP0CLK", "XGMI_PER_LINK_PWR_DWN", "DPM_GFX_PACE", "MEM_VDDCI_SCALING" , "MEM_MVDD_SCALING", "DS_UCLK", "GFX_ULV", "FW_DSTATE", "BACO" , "VCN_PG", "MM_DPM_PG", "JPEG_PG", "USB_PG", "RSMU_SMN_CG", "APCC_PLUS" , "GTHR", "GFX_DCS", "GFX_SS", "OUT_OF_BAND_MONITOR", "TEMP_DEPENDENT_VMIN" , "MMHUB_PG", "ATHUB_PG", "APCC_DFLL", "DF_CSTATE", "DPM_GFX_GPO" , "WAFL_CG", "CCLK_DPM", "FAN_CONTROLLER", "VCN_DPM", "LCLK_DPM" , "SHUBCLK_DPM", "DCFCLK_DPM", "DS_DCFCLK", "S0I2", "SMU_LOW_POWER" , "GFX_DEM", "PSI", "PROCHOT", "CPUOFF", "STAPM", "S0I3", "DF_CSTATES" , "PERF_LIMIT", "CORE_DLDO", "RSMU_LOW_POWER", "SMN_LOW_POWER" , "THM_LOW_POWER", "SMUIO_LOW_POWER", "MP1_LOW_POWER", "DS_VCN" , "CPPC", "OS_CSTATES", "ISP_DPM", "A55_DPM", "CVIP_DSP_DPM", "MSMU_LOW_POWER", "FUSE_CG", "MP1_CG", "SMUIO_CG", "THM_CG", "CLK_CG", "DATA_CALCULATION", "DPM_VCLK", "DPM_DCLK", "FW_DATA_READ" , "DPM_GFX_POWER_OPTIMIZER", "DPM_DCN", "VMEMP_SCALING", "VDDIO_MEM_SCALING" , "MM_DPM", "SOC_MPCLK_DS", "BACO_MPCLK_DS", "THROTTLERS", "SMARTSHIFT" , "GFX_READ_MARGIN", "GFX_IMU", "GFX_PCC_DFLL", "BOOT_TIME_CAL" , "BOOT_POWER_OPT", "GFXCLK_SPREAD_SPECTRUM", "SOC_PCC", "OPTIMIZED_VMIN" , "CLOCK_POWER_DOWN_BYPASS", "MEM_TEMP_READ", "ATHUB_MMHUB_PG" , "BACO_CG", "SOC_CG", | |||
| 690 | }; | |||
| 691 | ||||
| 692 | static const char *smu_get_feature_name(struct smu_context *smu, | |||
| 693 | enum smu_feature_mask feature) | |||
| 694 | { | |||
| 695 | if (feature < 0 || feature >= SMU_FEATURE_COUNT) | |||
| 696 | return "unknown smu feature"; | |||
| 697 | return __smu_feature_names[feature]; | |||
| 698 | } | |||
| 699 | ||||
| 700 | size_t smu_cmn_get_pp_feature_mask(struct smu_context *smu, | |||
| 701 | char *buf) | |||
| 702 | { | |||
| 703 | int8_t sort_feature[max(SMU_FEATURE_COUNT, SMU_FEATURE_MAX)(((SMU_FEATURE_COUNT)>((64)))?(SMU_FEATURE_COUNT):((64)))]; | |||
| 704 | uint64_t feature_mask; | |||
| 705 | int i, feature_index; | |||
| 706 | uint32_t count = 0; | |||
| 707 | size_t size = 0; | |||
| 708 | ||||
| 709 | if (__smu_get_enabled_features(smu, &feature_mask)) | |||
| 710 | return 0; | |||
| 711 | ||||
| 712 | size = sysfs_emit_at(buf, size, "features high: 0x%08x low: 0x%08x\n", | |||
| 713 | upper_32_bits(feature_mask)((u32)(((feature_mask) >> 16) >> 16)), lower_32_bits(feature_mask)((u32)(feature_mask))); | |||
| 714 | ||||
| 715 | memset(sort_feature, -1, sizeof(sort_feature))__builtin_memset((sort_feature), (-1), (sizeof(sort_feature)) ); | |||
| 716 | ||||
| 717 | for (i = 0; i < SMU_FEATURE_COUNT; i++) { | |||
| 718 | feature_index = smu_cmn_to_asic_specific_index(smu, | |||
| 719 | CMN2ASIC_MAPPING_FEATURE, | |||
| 720 | i); | |||
| 721 | if (feature_index < 0) | |||
| 722 | continue; | |||
| 723 | ||||
| 724 | sort_feature[feature_index] = i; | |||
| 725 | } | |||
| 726 | ||||
| 727 | size += sysfs_emit_at(buf, size, "%-2s. %-20s %-3s : %-s\n", | |||
| 728 | "No", "Feature", "Bit", "State"); | |||
| 729 | ||||
| 730 | for (feature_index = 0; feature_index < SMU_FEATURE_MAX(64); feature_index++) { | |||
| 731 | if (sort_feature[feature_index] < 0) | |||
| 732 | continue; | |||
| 733 | ||||
| 734 | size += sysfs_emit_at(buf, size, "%02d. %-20s (%2d) : %s\n", | |||
| 735 | count++, | |||
| 736 | smu_get_feature_name(smu, sort_feature[feature_index]), | |||
| 737 | feature_index, | |||
| 738 | !!test_bit(feature_index, (unsigned long *)&feature_mask) ? | |||
| 739 | "enabled" : "disabled"); | |||
| 740 | } | |||
| 741 | ||||
| 742 | return size; | |||
| 743 | } | |||
| 744 | ||||
| 745 | int smu_cmn_set_pp_feature_mask(struct smu_context *smu, | |||
| 746 | uint64_t new_mask) | |||
| 747 | { | |||
| 748 | int ret = 0; | |||
| 749 | uint64_t feature_mask; | |||
| 750 | uint64_t feature_2_enabled = 0; | |||
| 751 | uint64_t feature_2_disabled = 0; | |||
| 752 | ||||
| 753 | ret = __smu_get_enabled_features(smu, &feature_mask); | |||
| 754 | if (ret) | |||
| 755 | return ret; | |||
| 756 | ||||
| 757 | feature_2_enabled = ~feature_mask & new_mask; | |||
| 758 | feature_2_disabled = feature_mask & ~new_mask; | |||
| 759 | ||||
| 760 | if (feature_2_enabled) { | |||
| 761 | ret = smu_cmn_feature_update_enable_state(smu, | |||
| 762 | feature_2_enabled, | |||
| 763 | true1); | |||
| 764 | if (ret) | |||
| 765 | return ret; | |||
| 766 | } | |||
| 767 | if (feature_2_disabled) { | |||
| 768 | ret = smu_cmn_feature_update_enable_state(smu, | |||
| 769 | feature_2_disabled, | |||
| 770 | false0); | |||
| 771 | if (ret) | |||
| 772 | return ret; | |||
| 773 | } | |||
| 774 | ||||
| 775 | return ret; | |||
| 776 | } | |||
| 777 | ||||
| 778 | /** | |||
| 779 | * smu_cmn_disable_all_features_with_exception - disable all dpm features | |||
| 780 | * except this specified by | |||
| 781 | * @mask | |||
| 782 | * | |||
| 783 | * @smu: smu_context pointer | |||
| 784 | * @mask: the dpm feature which should not be disabled | |||
| 785 | * SMU_FEATURE_COUNT: no exception, all dpm features | |||
| 786 | * to disable | |||
| 787 | * | |||
| 788 | * Returns: | |||
| 789 | * 0 on success or a negative error code on failure. | |||
| 790 | */ | |||
| 791 | int smu_cmn_disable_all_features_with_exception(struct smu_context *smu, | |||
| 792 | enum smu_feature_mask mask) | |||
| 793 | { | |||
| 794 | uint64_t features_to_disable = U64_MAX0xffffffffffffffffULL; | |||
| 795 | int skipped_feature_id; | |||
| 796 | ||||
| 797 | if (mask != SMU_FEATURE_COUNT) { | |||
| 798 | skipped_feature_id = smu_cmn_to_asic_specific_index(smu, | |||
| 799 | CMN2ASIC_MAPPING_FEATURE, | |||
| 800 | mask); | |||
| 801 | if (skipped_feature_id < 0) | |||
| 802 | return -EINVAL22; | |||
| 803 | ||||
| 804 | features_to_disable &= ~(1ULL << skipped_feature_id); | |||
| 805 | } | |||
| 806 | ||||
| 807 | return smu_cmn_feature_update_enable_state(smu, | |||
| 808 | features_to_disable, | |||
| 809 | 0); | |||
| 810 | } | |||
| 811 | ||||
| 812 | int smu_cmn_get_smc_version(struct smu_context *smu, | |||
| 813 | uint32_t *if_version, | |||
| 814 | uint32_t *smu_version) | |||
| 815 | { | |||
| 816 | int ret = 0; | |||
| 817 | ||||
| 818 | if (!if_version && !smu_version) | |||
| 819 | return -EINVAL22; | |||
| 820 | ||||
| 821 | if (smu->smc_fw_if_version && smu->smc_fw_version) | |||
| 822 | { | |||
| 823 | if (if_version) | |||
| 824 | *if_version = smu->smc_fw_if_version; | |||
| 825 | ||||
| 826 | if (smu_version) | |||
| 827 | *smu_version = smu->smc_fw_version; | |||
| 828 | ||||
| 829 | return 0; | |||
| 830 | } | |||
| 831 | ||||
| 832 | if (if_version) { | |||
| 833 | ret = smu_cmn_send_smc_msg(smu, SMU_MSG_GetDriverIfVersion, if_version); | |||
| 834 | if (ret) | |||
| 835 | return ret; | |||
| 836 | ||||
| 837 | smu->smc_fw_if_version = *if_version; | |||
| 838 | } | |||
| 839 | ||||
| 840 | if (smu_version) { | |||
| 841 | ret = smu_cmn_send_smc_msg(smu, SMU_MSG_GetSmuVersion, smu_version); | |||
| 842 | if (ret) | |||
| 843 | return ret; | |||
| 844 | ||||
| 845 | smu->smc_fw_version = *smu_version; | |||
| 846 | } | |||
| 847 | ||||
| 848 | return ret; | |||
| 849 | } | |||
| 850 | ||||
| 851 | int smu_cmn_update_table(struct smu_context *smu, | |||
| 852 | enum smu_table_id table_index, | |||
| 853 | int argument, | |||
| 854 | void *table_data, | |||
| 855 | bool_Bool drv2smu) | |||
| 856 | { | |||
| 857 | struct smu_table_context *smu_table = &smu->smu_table; | |||
| 858 | struct amdgpu_device *adev = smu->adev; | |||
| 859 | struct smu_table *table = &smu_table->driver_table; | |||
| 860 | int table_id = smu_cmn_to_asic_specific_index(smu, | |||
| 861 | CMN2ASIC_MAPPING_TABLE, | |||
| 862 | table_index); | |||
| 863 | uint32_t table_size; | |||
| 864 | int ret = 0; | |||
| 865 | if (!table_data || table_id >= SMU_TABLE_COUNT || table_id < 0) | |||
| 866 | return -EINVAL22; | |||
| 867 | ||||
| 868 | table_size = smu_table->tables[table_index].size; | |||
| 869 | ||||
| 870 | if (drv2smu) { | |||
| 871 | memcpy(table->cpu_addr, table_data, table_size)__builtin_memcpy((table->cpu_addr), (table_data), (table_size )); | |||
| 872 | /* | |||
| 873 | * Flush hdp cache: to guard the content seen by | |||
| 874 | * GPU is consitent with CPU. | |||
| 875 | */ | |||
| 876 | amdgpu_asic_flush_hdp(adev, NULL)((adev)->asic_funcs->flush_hdp ? (adev)->asic_funcs-> flush_hdp((adev), (((void *)0))) : (adev)->hdp.funcs->flush_hdp ((adev), (((void *)0)))); | |||
| 877 | } | |||
| 878 | ||||
| 879 | ret = smu_cmn_send_smc_msg_with_param(smu, drv2smu ? | |||
| 880 | SMU_MSG_TransferTableDram2Smu : | |||
| 881 | SMU_MSG_TransferTableSmu2Dram, | |||
| 882 | table_id | ((argument & 0xFFFF) << 16), | |||
| 883 | NULL((void *)0)); | |||
| 884 | if (ret) | |||
| 885 | return ret; | |||
| 886 | ||||
| 887 | if (!drv2smu) { | |||
| 888 | amdgpu_asic_invalidate_hdp(adev, NULL)((adev)->asic_funcs->invalidate_hdp ? (adev)->asic_funcs ->invalidate_hdp((adev), (((void *)0))) : ((adev)->hdp. funcs->invalidate_hdp ? (adev)->hdp.funcs->invalidate_hdp ((adev), (((void *)0))) : 0)); | |||
| 889 | memcpy(table_data, table->cpu_addr, table_size)__builtin_memcpy((table_data), (table->cpu_addr), (table_size )); | |||
| 890 | } | |||
| 891 | ||||
| 892 | return 0; | |||
| 893 | } | |||
| 894 | ||||
| 895 | int smu_cmn_write_watermarks_table(struct smu_context *smu) | |||
| 896 | { | |||
| 897 | void *watermarks_table = smu->smu_table.watermarks_table; | |||
| 898 | ||||
| 899 | if (!watermarks_table) | |||
| 900 | return -EINVAL22; | |||
| 901 | ||||
| 902 | return smu_cmn_update_table(smu, | |||
| 903 | SMU_TABLE_WATERMARKS, | |||
| 904 | 0, | |||
| 905 | watermarks_table, | |||
| 906 | true1); | |||
| 907 | } | |||
| 908 | ||||
| 909 | int smu_cmn_write_pptable(struct smu_context *smu) | |||
| 910 | { | |||
| 911 | void *pptable = smu->smu_table.driver_pptable; | |||
| 912 | ||||
| 913 | return smu_cmn_update_table(smu, | |||
| 914 | SMU_TABLE_PPTABLE, | |||
| 915 | 0, | |||
| 916 | pptable, | |||
| 917 | true1); | |||
| 918 | } | |||
| 919 | ||||
| 920 | int smu_cmn_get_metrics_table(struct smu_context *smu, | |||
| 921 | void *metrics_table, | |||
| 922 | bool_Bool bypass_cache) | |||
| 923 | { | |||
| 924 | struct smu_table_context *smu_table= &smu->smu_table; | |||
| 925 | uint32_t table_size = | |||
| 926 | smu_table->tables[SMU_TABLE_SMU_METRICS].size; | |||
| 927 | int ret = 0; | |||
| 928 | ||||
| 929 | if (bypass_cache || | |||
| 930 | !smu_table->metrics_time || | |||
| 931 | time_after(jiffies, smu_table->metrics_time + msecs_to_jiffies(1)(((uint64_t)(1)) * hz / 1000))) { | |||
| 932 | ret = smu_cmn_update_table(smu, | |||
| 933 | SMU_TABLE_SMU_METRICS, | |||
| 934 | 0, | |||
| 935 | smu_table->metrics_table, | |||
| 936 | false0); | |||
| 937 | if (ret) { | |||
| 938 | dev_info(smu->adev->dev, "Failed to export SMU metrics table!\n")do { } while(0); | |||
| 939 | return ret; | |||
| 940 | } | |||
| 941 | smu_table->metrics_time = jiffies; | |||
| 942 | } | |||
| 943 | ||||
| 944 | if (metrics_table) | |||
| 945 | memcpy(metrics_table, smu_table->metrics_table, table_size)__builtin_memcpy((metrics_table), (smu_table->metrics_table ), (table_size)); | |||
| 946 | ||||
| 947 | return 0; | |||
| 948 | } | |||
| 949 | ||||
| 950 | int smu_cmn_get_combo_pptable(struct smu_context *smu) | |||
| 951 | { | |||
| 952 | void *pptable = smu->smu_table.combo_pptable; | |||
| 953 | ||||
| 954 | return smu_cmn_update_table(smu, | |||
| 955 | SMU_TABLE_COMBO_PPTABLE, | |||
| 956 | 0, | |||
| 957 | pptable, | |||
| 958 | false0); | |||
| 959 | } | |||
| 960 | ||||
| 961 | void smu_cmn_init_soft_gpu_metrics(void *table, uint8_t frev, uint8_t crev) | |||
| 962 | { | |||
| 963 | struct metrics_table_header *header = (struct metrics_table_header *)table; | |||
| 964 | uint16_t structure_size; | |||
| 965 | ||||
| 966 | #define METRICS_VERSION(a, b) ((a << 16) | b ) | |||
| 967 | ||||
| 968 | switch (METRICS_VERSION(frev, crev)) { | |||
| 969 | case METRICS_VERSION(1, 0): | |||
| 970 | structure_size = sizeof(struct gpu_metrics_v1_0); | |||
| 971 | break; | |||
| 972 | case METRICS_VERSION(1, 1): | |||
| 973 | structure_size = sizeof(struct gpu_metrics_v1_1); | |||
| 974 | break; | |||
| 975 | case METRICS_VERSION(1, 2): | |||
| 976 | structure_size = sizeof(struct gpu_metrics_v1_2); | |||
| 977 | break; | |||
| 978 | case METRICS_VERSION(1, 3): | |||
| 979 | structure_size = sizeof(struct gpu_metrics_v1_3); | |||
| 980 | break; | |||
| 981 | case METRICS_VERSION(2, 0): | |||
| 982 | structure_size = sizeof(struct gpu_metrics_v2_0); | |||
| 983 | break; | |||
| 984 | case METRICS_VERSION(2, 1): | |||
| 985 | structure_size = sizeof(struct gpu_metrics_v2_1); | |||
| 986 | break; | |||
| 987 | case METRICS_VERSION(2, 2): | |||
| 988 | structure_size = sizeof(struct gpu_metrics_v2_2); | |||
| 989 | break; | |||
| 990 | case METRICS_VERSION(2, 3): | |||
| 991 | structure_size = sizeof(struct gpu_metrics_v2_3); | |||
| 992 | break; | |||
| 993 | default: | |||
| 994 | return; | |||
| 995 | } | |||
| 996 | ||||
| 997 | #undef METRICS_VERSION | |||
| 998 | ||||
| 999 | memset(header, 0xFF, structure_size)__builtin_memset((header), (0xFF), (structure_size)); | |||
| 1000 | ||||
| 1001 | header->format_revision = frev; | |||
| 1002 | header->content_revision = crev; | |||
| 1003 | header->structure_size = structure_size; | |||
| 1004 | ||||
| 1005 | } | |||
| 1006 | ||||
| 1007 | int smu_cmn_set_mp1_state(struct smu_context *smu, | |||
| 1008 | enum pp_mp1_state mp1_state) | |||
| 1009 | { | |||
| 1010 | enum smu_message_type msg; | |||
| 1011 | int ret; | |||
| 1012 | ||||
| 1013 | switch (mp1_state) { | |||
| ||||
| 1014 | case PP_MP1_STATE_SHUTDOWN: | |||
| 1015 | msg = SMU_MSG_PrepareMp1ForShutdown; | |||
| 1016 | break; | |||
| 1017 | case PP_MP1_STATE_UNLOAD: | |||
| 1018 | msg = SMU_MSG_PrepareMp1ForUnload; | |||
| 1019 | break; | |||
| 1020 | case PP_MP1_STATE_RESET: | |||
| 1021 | msg = SMU_MSG_PrepareMp1ForReset; | |||
| 1022 | break; | |||
| 1023 | case PP_MP1_STATE_NONE: | |||
| 1024 | default: | |||
| 1025 | return 0; | |||
| 1026 | } | |||
| 1027 | ||||
| 1028 | ret = smu_cmn_send_smc_msg(smu, msg, NULL((void *)0)); | |||
| 1029 | if (ret) | |||
| 1030 | dev_err(smu->adev->dev, "[PrepareMp1] Failed!\n")printf("drm:pid%d:%s *ERROR* " "[PrepareMp1] Failed!\n", ({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci ) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci; })->ci_curproc->p_p->ps_pid, __func__); | |||
| 1031 | ||||
| 1032 | return ret; | |||
| 1033 | } | |||
| 1034 | ||||
| 1035 | bool_Bool smu_cmn_is_audio_func_enabled(struct amdgpu_device *adev) | |||
| 1036 | { | |||
| 1037 | STUB()do { printf("%s: stub\n", __func__); } while(0); | |||
| 1038 | return false0; | |||
| 1039 | #ifdef notyet | |||
| 1040 | struct pci_dev *p = NULL((void *)0); | |||
| 1041 | bool_Bool snd_driver_loaded; | |||
| 1042 | ||||
| 1043 | /* | |||
| 1044 | * If the ASIC comes with no audio function, we always assume | |||
| 1045 | * it is "enabled". | |||
| 1046 | */ | |||
| 1047 | p = pci_get_domain_bus_and_slot(pci_domain_nr(adev->pdev->bus), | |||
| 1048 | adev->pdev->bus->number, 1); | |||
| 1049 | if (!p) | |||
| 1050 | return true1; | |||
| 1051 | ||||
| 1052 | snd_driver_loaded = pci_is_enabled(p) ? true1 : false0; | |||
| 1053 | ||||
| 1054 | pci_dev_put(p); | |||
| 1055 | ||||
| 1056 | return snd_driver_loaded; | |||
| 1057 | #endif | |||
| 1058 | } |