*/
 void handle_untracked_irq(struct irq_desc *desc)
 {
-       unsigned int flags = 0;
-
        raw_spin_lock(&desc->lock);
 
        if (!irq_may_run(desc))
        irqd_set(&desc->irq_data, IRQD_IRQ_INPROGRESS);
        raw_spin_unlock(&desc->lock);
 
-       __handle_irq_event_percpu(desc, &flags);
+       __handle_irq_event_percpu(desc);
 
        raw_spin_lock(&desc->lock);
        irqd_clear(&desc->irq_data, IRQD_IRQ_INPROGRESS);
 
        wake_up_process(action->thread);
 }
 
-irqreturn_t __handle_irq_event_percpu(struct irq_desc *desc, unsigned int *flags)
+irqreturn_t __handle_irq_event_percpu(struct irq_desc *desc)
 {
        irqreturn_t retval = IRQ_NONE;
        unsigned int irq = desc->irq_data.irq;
                        }
 
                        __irq_wake_thread(desc, action);
-
-                       fallthrough;    /* to add to randomness */
-               case IRQ_HANDLED:
-                       *flags |= action->flags;
                        break;
 
                default:
 irqreturn_t handle_irq_event_percpu(struct irq_desc *desc)
 {
        irqreturn_t retval;
-       unsigned int flags = 0;
 
-       retval = __handle_irq_event_percpu(desc, &flags);
+       retval = __handle_irq_event_percpu(desc);
 
        add_interrupt_randomness(desc->irq_data.irq);
 
 
 
 extern void init_kstat_irqs(struct irq_desc *desc, int node, int nr);
 
-irqreturn_t __handle_irq_event_percpu(struct irq_desc *desc, unsigned int *flags);
+irqreturn_t __handle_irq_event_percpu(struct irq_desc *desc);
 irqreturn_t handle_irq_event_percpu(struct irq_desc *desc);
 irqreturn_t handle_irq_event(struct irq_desc *desc);