static void jpeg_v5_0_1_set_irq_funcs(struct amdgpu_device *adev);
 static int jpeg_v5_0_1_set_powergating_state(struct amdgpu_ip_block *ip_block,
                                             enum amd_powergating_state state);
+static void jpeg_v5_0_1_set_ras_funcs(struct amdgpu_device *adev);
 static void jpeg_v5_0_1_dec_ring_set_wptr(struct amdgpu_ring *ring);
 
 static int amdgpu_ih_srcid_jpeg[] = {
        adev->jpeg.num_jpeg_rings = AMDGPU_MAX_JPEG_RINGS;
        jpeg_v5_0_1_set_dec_ring_funcs(adev);
        jpeg_v5_0_1_set_irq_funcs(adev);
+       jpeg_v5_0_1_set_ras_funcs(adev);
 
        return 0;
 }
                if (r)
                        return r;
        }
+       /* JPEG DJPEG POISON EVENT */
+       r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_VCN,
+                       VCN_5_0__SRCID_DJPEG0_POISON, &adev->jpeg.inst->ras_poison_irq);
+       if (r)
+               return r;
+
+       /* JPEG EJPEG POISON EVENT */
+       r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_VCN,
+                       VCN_5_0__SRCID_EJPEG0_POISON, &adev->jpeg.inst->ras_poison_irq);
+       if (r)
+               return r;
 
        r = amdgpu_jpeg_sw_init(adev);
        if (r)
                        ret = jpeg_v5_0_1_set_powergating_state(ip_block, AMD_PG_STATE_GATE);
        }
 
+       if (amdgpu_ras_is_supported(adev, AMDGPU_RAS_BLOCK__JPEG))
+               amdgpu_irq_put(adev, &adev->jpeg.inst->ras_poison_irq, 0);
+
        return ret;
 }
 
        return 0;
 }
 
+static int jpeg_v5_0_1_set_ras_interrupt_state(struct amdgpu_device *adev,
+                                       struct amdgpu_irq_src *source,
+                                       unsigned int type,
+                                       enum amdgpu_interrupt_state state)
+{
+       return 0;
+}
+
+
+
 static int jpeg_v5_0_1_process_interrupt(struct amdgpu_device *adev,
                                         struct amdgpu_irq_src *source,
                                         struct amdgpu_iv_entry *entry)
        .process = jpeg_v5_0_1_process_interrupt,
 };
 
+static const struct amdgpu_irq_src_funcs jpeg_v5_0_1_ras_irq_funcs = {
+       .set = jpeg_v5_0_1_set_ras_interrupt_state,
+       .process = amdgpu_jpeg_process_poison_irq,
+};
+
 static void jpeg_v5_0_1_set_irq_funcs(struct amdgpu_device *adev)
 {
        int i;
                adev->jpeg.inst->irq.num_types += adev->jpeg.num_jpeg_rings;
 
        adev->jpeg.inst->irq.funcs = &jpeg_v5_0_1_irq_funcs;
+
+       adev->jpeg.inst->ras_poison_irq.num_types = 1;
+       adev->jpeg.inst->ras_poison_irq.funcs = &jpeg_v5_0_1_ras_irq_funcs;
+
 }
 
 const struct amdgpu_ip_block_version jpeg_v5_0_1_ip_block = {
        .rev = 1,
        .funcs = &jpeg_v5_0_1_ip_funcs,
 };
+
+static uint32_t jpeg_v5_0_1_query_poison_by_instance(struct amdgpu_device *adev,
+               uint32_t instance, uint32_t sub_block)
+{
+       uint32_t poison_stat = 0, reg_value = 0;
+
+       switch (sub_block) {
+       case AMDGPU_JPEG_V5_0_1_JPEG0:
+               reg_value = RREG32_SOC15(JPEG, instance, regUVD_RAS_JPEG0_STATUS);
+               poison_stat = REG_GET_FIELD(reg_value, UVD_RAS_JPEG0_STATUS, POISONED_PF);
+               break;
+       case AMDGPU_JPEG_V5_0_1_JPEG1:
+               reg_value = RREG32_SOC15(JPEG, instance, regUVD_RAS_JPEG1_STATUS);
+               poison_stat = REG_GET_FIELD(reg_value, UVD_RAS_JPEG1_STATUS, POISONED_PF);
+               break;
+       default:
+               break;
+       }
+
+       if (poison_stat)
+               dev_info(adev->dev, "Poison detected in JPEG%d sub_block%d\n",
+                       instance, sub_block);
+
+       return poison_stat;
+}
+
+static bool jpeg_v5_0_1_query_ras_poison_status(struct amdgpu_device *adev)
+{
+       uint32_t inst = 0, sub = 0, poison_stat = 0;
+
+       for (inst = 0; inst < adev->jpeg.num_jpeg_inst; inst++)
+               for (sub = 0; sub < AMDGPU_JPEG_V5_0_1_MAX_SUB_BLOCK; sub++)
+                       poison_stat +=
+                       jpeg_v5_0_1_query_poison_by_instance(adev, inst, sub);
+
+       return !!poison_stat;
+}
+
+static const struct amdgpu_ras_block_hw_ops jpeg_v5_0_1_ras_hw_ops = {
+       .query_poison_status = jpeg_v5_0_1_query_ras_poison_status,
+};
+
+static int jpeg_v5_0_1_aca_bank_parser(struct aca_handle *handle, struct aca_bank *bank,
+                                     enum aca_smu_type type, void *data)
+{
+       struct aca_bank_info info;
+       u64 misc0;
+       int ret;
+
+       ret = aca_bank_info_decode(bank, &info);
+       if (ret)
+               return ret;
+
+       misc0 = bank->regs[ACA_REG_IDX_MISC0];
+       switch (type) {
+       case ACA_SMU_TYPE_UE:
+               bank->aca_err_type = ACA_ERROR_TYPE_UE;
+               ret = aca_error_cache_log_bank_error(handle, &info, ACA_ERROR_TYPE_UE,
+                                                    1ULL);
+               break;
+       case ACA_SMU_TYPE_CE:
+               bank->aca_err_type = ACA_ERROR_TYPE_CE;
+               ret = aca_error_cache_log_bank_error(handle, &info, bank->aca_err_type,
+                                                    ACA_REG__MISC0__ERRCNT(misc0));
+               break;
+       default:
+               return -EINVAL;
+       }
+
+       return ret;
+}
+
+/* reference to smu driver if header file */
+static int jpeg_v5_0_1_err_codes[] = {
+       16, 17, 18, 19, 20, 21, 22, 23, /* JPEG[0-7][S|D] */
+       24, 25, 26, 27, 28, 29, 30, 31
+};
+
+static bool jpeg_v5_0_1_aca_bank_is_valid(struct aca_handle *handle, struct aca_bank *bank,
+                                        enum aca_smu_type type, void *data)
+{
+       u32 instlo;
+
+       instlo = ACA_REG__IPID__INSTANCEIDLO(bank->regs[ACA_REG_IDX_IPID]);
+       instlo &= GENMASK(31, 1);
+
+       if (instlo != mmSMNAID_AID0_MCA_SMU)
+               return false;
+
+       if (aca_bank_check_error_codes(handle->adev, bank,
+                                      jpeg_v5_0_1_err_codes,
+                                      ARRAY_SIZE(jpeg_v5_0_1_err_codes)))
+               return false;
+
+       return true;
+}
+
+static const struct aca_bank_ops jpeg_v5_0_1_aca_bank_ops = {
+       .aca_bank_parser = jpeg_v5_0_1_aca_bank_parser,
+       .aca_bank_is_valid = jpeg_v5_0_1_aca_bank_is_valid,
+};
+
+static const struct aca_info jpeg_v5_0_1_aca_info = {
+       .hwip = ACA_HWIP_TYPE_SMU,
+       .mask = ACA_ERROR_UE_MASK,
+       .bank_ops = &jpeg_v5_0_1_aca_bank_ops,
+};
+
+static int jpeg_v5_0_1_ras_late_init(struct amdgpu_device *adev, struct ras_common_if *ras_block)
+{
+       int r;
+
+       r = amdgpu_ras_block_late_init(adev, ras_block);
+       if (r)
+               return r;
+
+       if (amdgpu_ras_is_supported(adev, ras_block->block) &&
+               adev->jpeg.inst->ras_poison_irq.funcs) {
+               r = amdgpu_irq_get(adev, &adev->jpeg.inst->ras_poison_irq, 0);
+               if (r)
+                       goto late_fini;
+       }
+
+       r = amdgpu_ras_bind_aca(adev, AMDGPU_RAS_BLOCK__JPEG,
+                               &jpeg_v5_0_1_aca_info, NULL);
+       if (r)
+               goto late_fini;
+
+       return 0;
+
+late_fini:
+       amdgpu_ras_block_late_fini(adev, ras_block);
+
+       return r;
+}
+
+static struct amdgpu_jpeg_ras jpeg_v5_0_1_ras = {
+       .ras_block = {
+               .hw_ops = &jpeg_v5_0_1_ras_hw_ops,
+               .ras_late_init = jpeg_v5_0_1_ras_late_init,
+       },
+};
+
+static void jpeg_v5_0_1_set_ras_funcs(struct amdgpu_device *adev)
+{
+       adev->jpeg.ras = &jpeg_v5_0_1_ras;
+}