}
 
        for (i = 0; i < adev->vcn.num_vcn_inst; i++) {
+               volatile struct amdgpu_fw_shared *fw_shared;
                if (adev->vcn.harvest_config & (1 << i))
                        continue;
 
                        if (r)
                                return r;
                }
+
+               fw_shared = adev->vcn.inst[i].fw_shared_cpu_addr;
+               fw_shared->present_flag_0 |= cpu_to_le32(AMDGPU_VCN_SW_RING_FLAG);
+               fw_shared->sw_ring.is_enabled = cpu_to_le32(DEC_SW_RING_ENABLED);
        }
 
        if (amdgpu_sriov_vf(adev)) {
 static int vcn_v3_0_sw_fini(void *handle)
 {
        struct amdgpu_device *adev = (struct amdgpu_device *)handle;
-       int r;
+       int i, r;
+
+       for (i = 0; i < adev->vcn.num_vcn_inst; i++) {
+               volatile struct amdgpu_fw_shared *fw_shared;
+
+               if (adev->vcn.harvest_config & (1 << i))
+                       continue;
+               fw_shared = adev->vcn.inst[i].fw_shared_cpu_addr;
+               fw_shared->present_flag_0 = 0;
+               fw_shared->sw_ring.is_enabled = false;
+       }
 
        if (amdgpu_sriov_vf(adev))
                amdgpu_virt_free_mm_table(adev);
                upper_32_bits(adev->vcn.inst[inst].gpu_addr + offset + AMDGPU_VCN_STACK_SIZE));
        WREG32_SOC15(VCN, inst, mmUVD_VCPU_CACHE_OFFSET2, 0);
        WREG32_SOC15(VCN, inst, mmUVD_VCPU_CACHE_SIZE2, AMDGPU_VCN_CONTEXT_SIZE);
+
+       /* non-cache window */
+       WREG32_SOC15(VCN, inst, mmUVD_LMI_VCPU_NC0_64BIT_BAR_LOW,
+               lower_32_bits(adev->vcn.inst[inst].fw_shared_gpu_addr));
+       WREG32_SOC15(VCN, inst, mmUVD_LMI_VCPU_NC0_64BIT_BAR_HIGH,
+               upper_32_bits(adev->vcn.inst[inst].fw_shared_gpu_addr));
+       WREG32_SOC15(VCN, inst, mmUVD_VCPU_NONCACHE_OFFSET0, 0);
+       WREG32_SOC15(VCN, inst, mmUVD_VCPU_NONCACHE_SIZE0,
+               AMDGPU_GPU_PAGE_ALIGN(sizeof(struct amdgpu_fw_shared)));
 }
 
 static void vcn_v3_0_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect)
 
        /* non-cache window */
        WREG32_SOC15_DPG_MODE(inst_idx, SOC15_DPG_MODE_OFFSET(
-                       VCN, inst_idx, mmUVD_LMI_VCPU_NC0_64BIT_BAR_LOW), 0, 0, indirect);
+                       VCN, inst_idx, mmUVD_LMI_VCPU_NC0_64BIT_BAR_LOW),
+                       lower_32_bits(adev->vcn.inst[inst_idx].fw_shared_gpu_addr), 0, indirect);
        WREG32_SOC15_DPG_MODE(inst_idx, SOC15_DPG_MODE_OFFSET(
-                       VCN, inst_idx, mmUVD_LMI_VCPU_NC0_64BIT_BAR_HIGH), 0, 0, indirect);
+                       VCN, inst_idx, mmUVD_LMI_VCPU_NC0_64BIT_BAR_HIGH),
+                       upper_32_bits(adev->vcn.inst[inst_idx].fw_shared_gpu_addr), 0, indirect);
        WREG32_SOC15_DPG_MODE(inst_idx, SOC15_DPG_MODE_OFFSET(
                        VCN, inst_idx, mmUVD_VCPU_NONCACHE_OFFSET0), 0, 0, indirect);
        WREG32_SOC15_DPG_MODE(inst_idx, SOC15_DPG_MODE_OFFSET(
-                       VCN, inst_idx, mmUVD_VCPU_NONCACHE_SIZE0), 0, 0, indirect);
+                       VCN, inst_idx, mmUVD_VCPU_NONCACHE_SIZE0),
+                       AMDGPU_GPU_PAGE_ALIGN(sizeof(struct amdgpu_fw_shared)), 0, indirect);
 }
 
 static void vcn_v3_0_disable_static_power_gating(struct amdgpu_device *adev, int inst)