Message ID | 169657717635.1491153.10040969889268437662.stgit@dwillia2-xfh.jf.intel.com |
---|---|
State | Accepted |
Commit | 08b8a8c05423174e3ef4fb0bd514de20088cf5ac |
Headers | show |
Series | cxl/mem: Fix shutdown order | expand |
On Fri, 06 Oct 2023, Dan Williams wrote: >Now that all callers of cxl_request_irq() are using threaded irqs, drop >the hardirq handler option. > >Signed-off-by: Dan Williams <dan.j.williams@intel.com> Reviewed-by: Davidlohr Bueso <dave@stgolabs.net>
Dan Williams wrote: > Now that all callers of cxl_request_irq() are using threaded irqs, drop > the hardirq handler option. > > Signed-off-by: Dan Williams <dan.j.williams@intel.com> Reviewed-by: Ira Weiny <ira.weiny@intel.com>
On Fri, 06 Oct 2023 00:26:16 -0700 Dan Williams <dan.j.williams@intel.com> wrote: > Now that all callers of cxl_request_irq() are using threaded irqs, drop > the hardirq handler option. > > Signed-off-by: Dan Williams <dan.j.williams@intel.com> Makes sense. Reviewed-by: Jonathan Cameron <Jonathan.Cameron@huawei.com> > --- > drivers/cxl/pci.c | 12 ++++++------ > 1 file changed, 6 insertions(+), 6 deletions(-) > > diff --git a/drivers/cxl/pci.c b/drivers/cxl/pci.c > index 49d9b2ef5c5c..dc665b12be8f 100644 > --- a/drivers/cxl/pci.c > +++ b/drivers/cxl/pci.c > @@ -90,7 +90,7 @@ struct cxl_dev_id { > }; > > static int cxl_request_irq(struct cxl_dev_state *cxlds, int irq, > - irq_handler_t handler, irq_handler_t thread_fn) > + irq_handler_t thread_fn) > { > struct device *dev = cxlds->dev; > struct cxl_dev_id *dev_id; > @@ -101,9 +101,9 @@ static int cxl_request_irq(struct cxl_dev_state *cxlds, int irq, > return -ENOMEM; > dev_id->cxlds = cxlds; > > - return devm_request_threaded_irq(dev, irq, handler, thread_fn, > - IRQF_SHARED | IRQF_ONESHOT, > - NULL, dev_id); > + return devm_request_threaded_irq(dev, irq, NULL, thread_fn, > + IRQF_SHARED | IRQF_ONESHOT, NULL, > + dev_id); > } > > static bool cxl_mbox_background_complete(struct cxl_dev_state *cxlds) > @@ -440,7 +440,7 @@ static int cxl_pci_setup_mailbox(struct cxl_memdev_state *mds) > if (irq < 0) > return 0; > > - if (cxl_request_irq(cxlds, irq, NULL, cxl_pci_mbox_irq)) > + if (cxl_request_irq(cxlds, irq, cxl_pci_mbox_irq)) > return 0; > > dev_dbg(cxlds->dev, "Mailbox interrupts enabled\n"); > @@ -638,7 +638,7 @@ static int cxl_event_req_irq(struct cxl_dev_state *cxlds, u8 setting) > if (irq < 0) > return irq; > > - return cxl_request_irq(cxlds, irq, NULL, cxl_event_thread); > + return cxl_request_irq(cxlds, irq, cxl_event_thread); > } > > static int cxl_event_get_int_policy(struct cxl_memdev_state *mds, > >
On 10/6/23 00:26, Dan Williams wrote: > Now that all callers of cxl_request_irq() are using threaded irqs, drop > the hardirq handler option. > > Signed-off-by: Dan Williams <dan.j.williams@intel.com> Reviewed-by: Dave Jiang <dave.jiang@intel.com> > --- > drivers/cxl/pci.c | 12 ++++++------ > 1 file changed, 6 insertions(+), 6 deletions(-) > > diff --git a/drivers/cxl/pci.c b/drivers/cxl/pci.c > index 49d9b2ef5c5c..dc665b12be8f 100644 > --- a/drivers/cxl/pci.c > +++ b/drivers/cxl/pci.c > @@ -90,7 +90,7 @@ struct cxl_dev_id { > }; > > static int cxl_request_irq(struct cxl_dev_state *cxlds, int irq, > - irq_handler_t handler, irq_handler_t thread_fn) > + irq_handler_t thread_fn) > { > struct device *dev = cxlds->dev; > struct cxl_dev_id *dev_id; > @@ -101,9 +101,9 @@ static int cxl_request_irq(struct cxl_dev_state *cxlds, int irq, > return -ENOMEM; > dev_id->cxlds = cxlds; > > - return devm_request_threaded_irq(dev, irq, handler, thread_fn, > - IRQF_SHARED | IRQF_ONESHOT, > - NULL, dev_id); > + return devm_request_threaded_irq(dev, irq, NULL, thread_fn, > + IRQF_SHARED | IRQF_ONESHOT, NULL, > + dev_id); > } > > static bool cxl_mbox_background_complete(struct cxl_dev_state *cxlds) > @@ -440,7 +440,7 @@ static int cxl_pci_setup_mailbox(struct cxl_memdev_state *mds) > if (irq < 0) > return 0; > > - if (cxl_request_irq(cxlds, irq, NULL, cxl_pci_mbox_irq)) > + if (cxl_request_irq(cxlds, irq, cxl_pci_mbox_irq)) > return 0; > > dev_dbg(cxlds->dev, "Mailbox interrupts enabled\n"); > @@ -638,7 +638,7 @@ static int cxl_event_req_irq(struct cxl_dev_state *cxlds, u8 setting) > if (irq < 0) > return irq; > > - return cxl_request_irq(cxlds, irq, NULL, cxl_event_thread); > + return cxl_request_irq(cxlds, irq, cxl_event_thread); > } > > static int cxl_event_get_int_policy(struct cxl_memdev_state *mds, > >
diff --git a/drivers/cxl/pci.c b/drivers/cxl/pci.c index 49d9b2ef5c5c..dc665b12be8f 100644 --- a/drivers/cxl/pci.c +++ b/drivers/cxl/pci.c @@ -90,7 +90,7 @@ struct cxl_dev_id { }; static int cxl_request_irq(struct cxl_dev_state *cxlds, int irq, - irq_handler_t handler, irq_handler_t thread_fn) + irq_handler_t thread_fn) { struct device *dev = cxlds->dev; struct cxl_dev_id *dev_id; @@ -101,9 +101,9 @@ static int cxl_request_irq(struct cxl_dev_state *cxlds, int irq, return -ENOMEM; dev_id->cxlds = cxlds; - return devm_request_threaded_irq(dev, irq, handler, thread_fn, - IRQF_SHARED | IRQF_ONESHOT, - NULL, dev_id); + return devm_request_threaded_irq(dev, irq, NULL, thread_fn, + IRQF_SHARED | IRQF_ONESHOT, NULL, + dev_id); } static bool cxl_mbox_background_complete(struct cxl_dev_state *cxlds) @@ -440,7 +440,7 @@ static int cxl_pci_setup_mailbox(struct cxl_memdev_state *mds) if (irq < 0) return 0; - if (cxl_request_irq(cxlds, irq, NULL, cxl_pci_mbox_irq)) + if (cxl_request_irq(cxlds, irq, cxl_pci_mbox_irq)) return 0; dev_dbg(cxlds->dev, "Mailbox interrupts enabled\n"); @@ -638,7 +638,7 @@ static int cxl_event_req_irq(struct cxl_dev_state *cxlds, u8 setting) if (irq < 0) return irq; - return cxl_request_irq(cxlds, irq, NULL, cxl_event_thread); + return cxl_request_irq(cxlds, irq, cxl_event_thread); } static int cxl_event_get_int_policy(struct cxl_memdev_state *mds,
Now that all callers of cxl_request_irq() are using threaded irqs, drop the hardirq handler option. Signed-off-by: Dan Williams <dan.j.williams@intel.com> --- drivers/cxl/pci.c | 12 ++++++------ 1 file changed, 6 insertions(+), 6 deletions(-)