return 0;
 
        debugfs_create_x32("amdgpu_smu_debug", 0600, root,
-                          &adev->smu.smu_debug_mask);
+                          &adev->pm.smu_debug_mask);
 
        ent = debugfs_create_file("amdgpu_preempt_ib", 0600, root, adev,
                                  &fops_ib_preempt);
 
        POWER_STATE_OFF,
 };
 
+/* Used to mask smu debug modes */
+#define SMU_DEBUG_HALT_ON_ERROR                0x1
+
 struct amdgpu_pm {
        struct mutex            mutex;
        u32                     current_sclk;
        struct list_head        pm_attr_list;
 
        atomic_t                pwr_state[AMD_IP_BLOCK_TYPE_NUM];
+
+       /*
+        * 0 = disabled (default), otherwise enable corresponding debug mode
+        */
+       uint32_t                smu_debug_mask;
 };
 
 #define R600_SSTU_DFLT                               0
 
 
 #define WORKLOAD_POLICY_MAX 7
 
-/* Used to mask smu debug modes */
-#define SMU_DEBUG_HALT_ON_ERROR                0x1
-
 struct smu_context
 {
        struct amdgpu_device            *adev;
        struct smu_user_dpm_profile user_dpm_profile;
 
        struct stb_context stb_context;
-
-       /*
-        * 0 = disabled (default), otherwise enable corresponding debug mode
-        */
-       uint32_t smu_debug_mask;
 };
 
 struct i2c_adapter;
 
                                     uint16_t msg_index,
                                     uint32_t param)
 {
+       struct amdgpu_device *adev = smu->adev;
        u32 reg;
        int res;
 
-       if (smu->adev->no_hw_access)
+       if (adev->no_hw_access)
                return 0;
 
        reg = __smu_cmn_poll_stat(smu);
        __smu_cmn_send_msg(smu, msg_index, param);
        res = 0;
 Out:
-       if (unlikely(smu->smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR) &&
+       if (unlikely(adev->pm.smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR) &&
            res && (res != -ETIME)) {
-               amdgpu_device_halt(smu->adev);
+               amdgpu_device_halt(adev);
                WARN_ON(1);
        }
 
        reg = __smu_cmn_poll_stat(smu);
        res = __smu_cmn_reg2errno(smu, reg);
 
-       if (unlikely(smu->smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR) &&
+       if (unlikely(smu->adev->pm.smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR) &&
            res && (res != -ETIME)) {
                amdgpu_device_halt(smu->adev);
                WARN_ON(1);
                                    uint32_t param,
                                    uint32_t *read_arg)
 {
+       struct amdgpu_device *adev = smu->adev;
        int res, index;
        u32 reg;
 
-       if (smu->adev->no_hw_access)
+       if (adev->no_hw_access)
                return 0;
 
        index = smu_cmn_to_asic_specific_index(smu,
        if (read_arg)
                smu_cmn_read_arg(smu, read_arg);
 Out:
-       if (unlikely(smu->smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR) && res) {
-               amdgpu_device_halt(smu->adev);
+       if (unlikely(adev->pm.smu_debug_mask & SMU_DEBUG_HALT_ON_ERROR) && res) {
+               amdgpu_device_halt(adev);
                WARN_ON(1);
        }