SUNRPC: Remove the bh-safe lock requirement on the rpc_wait_queue->lock
authorTrond Myklebust <trond.myklebust@hammerspace.com>
Thu, 2 May 2019 15:23:12 +0000 (11:23 -0400)
committerTrond Myklebust <trond.myklebust@hammerspace.com>
Sat, 6 Jul 2019 18:54:49 +0000 (14:54 -0400)
Signed-off-by: Trond Myklebust <trond.myklebust@hammerspace.com>
net/sunrpc/sched.c

index e0a0cf381ebac0aa0a161b33c469c1d6e5779487..191916168e671df089207b3eef9db7f7fc2517d4 100644 (file)
@@ -432,9 +432,9 @@ void rpc_sleep_on_timeout(struct rpc_wait_queue *q, struct rpc_task *task,
        /*
         * Protect the queue operations.
         */
-       spin_lock_bh(&q->lock);
+       spin_lock(&q->lock);
        __rpc_sleep_on_priority_timeout(q, task, timeout, task->tk_priority);
-       spin_unlock_bh(&q->lock);
+       spin_unlock(&q->lock);
 }
 EXPORT_SYMBOL_GPL(rpc_sleep_on_timeout);
 
@@ -450,9 +450,9 @@ void rpc_sleep_on(struct rpc_wait_queue *q, struct rpc_task *task,
        /*
         * Protect the queue operations.
         */
-       spin_lock_bh(&q->lock);
+       spin_lock(&q->lock);
        __rpc_sleep_on_priority(q, task, task->tk_priority);
-       spin_unlock_bh(&q->lock);
+       spin_unlock(&q->lock);
 }
 EXPORT_SYMBOL_GPL(rpc_sleep_on);
 
@@ -466,9 +466,9 @@ void rpc_sleep_on_priority_timeout(struct rpc_wait_queue *q,
        /*
         * Protect the queue operations.
         */
-       spin_lock_bh(&q->lock);
+       spin_lock(&q->lock);
        __rpc_sleep_on_priority_timeout(q, task, timeout, priority);
-       spin_unlock_bh(&q->lock);
+       spin_unlock(&q->lock);
 }
 EXPORT_SYMBOL_GPL(rpc_sleep_on_priority_timeout);
 
@@ -483,9 +483,9 @@ void rpc_sleep_on_priority(struct rpc_wait_queue *q, struct rpc_task *task,
        /*
         * Protect the queue operations.
         */
-       spin_lock_bh(&q->lock);
+       spin_lock(&q->lock);
        __rpc_sleep_on_priority(q, task, priority);
-       spin_unlock_bh(&q->lock);
+       spin_unlock(&q->lock);
 }
 EXPORT_SYMBOL_GPL(rpc_sleep_on_priority);
 
@@ -563,9 +563,9 @@ void rpc_wake_up_queued_task_on_wq(struct workqueue_struct *wq,
 {
        if (!RPC_IS_QUEUED(task))
                return;
-       spin_lock_bh(&queue->lock);
+       spin_lock(&queue->lock);
        rpc_wake_up_task_on_wq_queue_locked(wq, queue, task);
-       spin_unlock_bh(&queue->lock);
+       spin_unlock(&queue->lock);
 }
 
 /*
@@ -575,9 +575,9 @@ void rpc_wake_up_queued_task(struct rpc_wait_queue *queue, struct rpc_task *task
 {
        if (!RPC_IS_QUEUED(task))
                return;
-       spin_lock_bh(&queue->lock);
+       spin_lock(&queue->lock);
        rpc_wake_up_task_queue_locked(queue, task);
-       spin_unlock_bh(&queue->lock);
+       spin_unlock(&queue->lock);
 }
 EXPORT_SYMBOL_GPL(rpc_wake_up_queued_task);
 
@@ -610,9 +610,9 @@ rpc_wake_up_queued_task_set_status(struct rpc_wait_queue *queue,
 {
        if (!RPC_IS_QUEUED(task))
                return;
-       spin_lock_bh(&queue->lock);
+       spin_lock(&queue->lock);
        rpc_wake_up_task_queue_set_status_locked(queue, task, status);
-       spin_unlock_bh(&queue->lock);
+       spin_unlock(&queue->lock);
 }
 
 /*
@@ -675,12 +675,12 @@ struct rpc_task *rpc_wake_up_first_on_wq(struct workqueue_struct *wq,
 
        dprintk("RPC:       wake_up_first(%p \"%s\")\n",
                        queue, rpc_qname(queue));
-       spin_lock_bh(&queue->lock);
+       spin_lock(&queue->lock);
        task = __rpc_find_next_queued(queue);
        if (task != NULL)
                task = rpc_wake_up_task_on_wq_queue_action_locked(wq, queue,
                                task, func, data);
-       spin_unlock_bh(&queue->lock);
+       spin_unlock(&queue->lock);
 
        return task;
 }
@@ -719,7 +719,7 @@ void rpc_wake_up(struct rpc_wait_queue *queue)
 {
        struct list_head *head;
 
-       spin_lock_bh(&queue->lock);
+       spin_lock(&queue->lock);
        head = &queue->tasks[queue->maxpriority];
        for (;;) {
                while (!list_empty(head)) {
@@ -733,7 +733,7 @@ void rpc_wake_up(struct rpc_wait_queue *queue)
                        break;
                head--;
        }
-       spin_unlock_bh(&queue->lock);
+       spin_unlock(&queue->lock);
 }
 EXPORT_SYMBOL_GPL(rpc_wake_up);
 
@@ -748,7 +748,7 @@ void rpc_wake_up_status(struct rpc_wait_queue *queue, int status)
 {
        struct list_head *head;
 
-       spin_lock_bh(&queue->lock);
+       spin_lock(&queue->lock);
        head = &queue->tasks[queue->maxpriority];
        for (;;) {
                while (!list_empty(head)) {
@@ -763,7 +763,7 @@ void rpc_wake_up_status(struct rpc_wait_queue *queue, int status)
                        break;
                head--;
        }
-       spin_unlock_bh(&queue->lock);
+       spin_unlock(&queue->lock);
 }
 EXPORT_SYMBOL_GPL(rpc_wake_up_status);
 
@@ -775,7 +775,7 @@ static void __rpc_queue_timer_fn(struct work_struct *work)
        struct rpc_task *task, *n;
        unsigned long expires, now, timeo;
 
-       spin_lock_bh(&queue->lock);
+       spin_lock(&queue->lock);
        expires = now = jiffies;
        list_for_each_entry_safe(task, n, &queue->timer_list.list, u.tk_wait.timer_list) {
                timeo = task->tk_timeout;
@@ -790,7 +790,7 @@ static void __rpc_queue_timer_fn(struct work_struct *work)
        }
        if (!list_empty(&queue->timer_list.list))
                rpc_set_queue_timer(queue, expires);
-       spin_unlock_bh(&queue->lock);
+       spin_unlock(&queue->lock);
 }
 
 static void __rpc_atrun(struct rpc_task *task)
@@ -937,13 +937,13 @@ static void __rpc_execute(struct rpc_task *task)
                 * rpc_task pointer may still be dereferenced.
                 */
                queue = task->tk_waitqueue;
-               spin_lock_bh(&queue->lock);
+               spin_lock(&queue->lock);
                if (!RPC_IS_QUEUED(task)) {
-                       spin_unlock_bh(&queue->lock);
+                       spin_unlock(&queue->lock);
                        continue;
                }
                rpc_clear_running(task);
-               spin_unlock_bh(&queue->lock);
+               spin_unlock(&queue->lock);
                if (task_is_async)
                        return;