} linear;
struct radix_tree_root tree;
} revmap_data;
- struct irq_domain_ops *ops;
+ const struct irq_domain_ops *ops;
void *host_data;
irq_hw_number_t inval_irq;
unsigned int size,
unsigned int first_irq,
irq_hw_number_t first_hwirq,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data);
struct irq_domain *irq_domain_add_linear(struct device_node *of_node,
unsigned int size,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data);
struct irq_domain *irq_domain_add_nomap(struct device_node *of_node,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data);
struct irq_domain *irq_domain_add_tree(struct device_node *of_node,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data);
extern struct irq_domain *irq_find_host(struct device_node *node);
static inline struct irq_domain *irq_domain_add_legacy_isa(
struct device_node *of_node,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data)
{
return irq_domain_add_legacy(of_node, NUM_ISA_INTERRUPTS, 0, 0, ops,
extern unsigned int irq_linear_revmap(struct irq_domain *host,
irq_hw_number_t hwirq);
-extern struct irq_domain_ops irq_domain_simple_ops;
+extern const struct irq_domain_ops irq_domain_simple_ops;
/* stock xlate functions */
int irq_domain_xlate_onecell(struct irq_domain *d, struct device_node *ctrlr,
static unsigned int irq_virq_count = NR_IRQS;
static struct irq_domain *irq_default_domain;
-static int default_irq_domain_match(struct irq_domain *d, struct device_node *np)
-{
- return d->of_node != NULL && d->of_node == np;
-}
-
/**
* irq_domain_alloc() - Allocate a new irq_domain data structure
* @of_node: optional device-tree node of the interrupt controller
*/
static struct irq_domain *irq_domain_alloc(struct device_node *of_node,
unsigned int revmap_type,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data)
{
struct irq_domain *domain;
domain->host_data = host_data;
domain->of_node = of_node_get(of_node);
- if (domain->ops->match == NULL)
- domain->ops->match = default_irq_domain_match;
-
return domain;
}
unsigned int size,
unsigned int first_irq,
irq_hw_number_t first_hwirq,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data)
{
struct irq_domain *domain;
*/
struct irq_domain *irq_domain_add_linear(struct device_node *of_node,
unsigned int size,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data)
{
struct irq_domain *domain;
}
struct irq_domain *irq_domain_add_nomap(struct device_node *of_node,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data)
{
struct irq_domain *domain = irq_domain_alloc(of_node,
* (the reverse mapping will use the slow path until that happens).
*/
struct irq_domain *irq_domain_add_tree(struct device_node *of_node,
- struct irq_domain_ops *ops,
+ const struct irq_domain_ops *ops,
void *host_data)
{
struct irq_domain *domain = irq_domain_alloc(of_node,
struct irq_domain *irq_find_host(struct device_node *node)
{
struct irq_domain *h, *found = NULL;
+ int rc;
/* We might want to match the legacy controller last since
* it might potentially be set to match all interrupts in
* yet though...
*/
mutex_lock(&irq_domain_mutex);
- list_for_each_entry(h, &irq_domain_list, link)
- if (h->ops->match(h, node)) {
+ list_for_each_entry(h, &irq_domain_list, link) {
+ if (h->ops->match)
+ rc = h->ops->match(h, node);
+ else
+ rc = (h->of_node != NULL) && (h->of_node == node);
+
+ if (rc) {
found = h;
break;
}
+ }
mutex_unlock(&irq_domain_mutex);
return found;
}
}
EXPORT_SYMBOL_GPL(irq_domain_xlate_onetwocell);
-struct irq_domain_ops irq_domain_simple_ops = {
+const struct irq_domain_ops irq_domain_simple_ops = {
.map = irq_domain_simple_map,
.xlate = irq_domain_xlate_onetwocell,
};