mirror of
https://github.com/tbsdtv/linux_media.git
synced 2025-07-23 12:43:29 +02:00
Please do not apply this to mainline directly, instead please re-run the coccinelle script shown below and apply its output. For several reasons, it is desirable to use {READ,WRITE}_ONCE() in preference to ACCESS_ONCE(), and new code is expected to use one of the former. So far, there's been no reason to change most existing uses of ACCESS_ONCE(), as these aren't harmful, and changing them results in churn. However, for some features, the read/write distinction is critical to correct operation. To distinguish these cases, separate read/write accessors must be used. This patch migrates (most) remaining ACCESS_ONCE() instances to {READ,WRITE}_ONCE(), using the following coccinelle script: ---- // Convert trivial ACCESS_ONCE() uses to equivalent READ_ONCE() and // WRITE_ONCE() // $ make coccicheck COCCI=/home/mark/once.cocci SPFLAGS="--include-headers" MODE=patch virtual patch @ depends on patch @ expression E1, E2; @@ - ACCESS_ONCE(E1) = E2 + WRITE_ONCE(E1, E2) @ depends on patch @ expression E; @@ - ACCESS_ONCE(E) + READ_ONCE(E) ---- Signed-off-by: Mark Rutland <mark.rutland@arm.com> Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: davem@davemloft.net Cc: linux-arch@vger.kernel.org Cc: mpe@ellerman.id.au Cc: shuah@kernel.org Cc: snitzer@redhat.com Cc: thor.thayer@linux.intel.com Cc: tj@kernel.org Cc: viro@zeniv.linux.org.uk Cc: will.deacon@arm.com Link: http://lkml.kernel.org/r/1508792849-3115-19-git-send-email-paulmck@linux.vnet.ibm.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
103 lines
2.0 KiB
C
103 lines
2.0 KiB
C
#include <linux/fs.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/slab.h>
|
|
#include "internal.h"
|
|
#include "mount.h"
|
|
|
|
static DEFINE_SPINLOCK(pin_lock);
|
|
|
|
void pin_remove(struct fs_pin *pin)
|
|
{
|
|
spin_lock(&pin_lock);
|
|
hlist_del_init(&pin->m_list);
|
|
hlist_del_init(&pin->s_list);
|
|
spin_unlock(&pin_lock);
|
|
spin_lock_irq(&pin->wait.lock);
|
|
pin->done = 1;
|
|
wake_up_locked(&pin->wait);
|
|
spin_unlock_irq(&pin->wait.lock);
|
|
}
|
|
|
|
void pin_insert_group(struct fs_pin *pin, struct vfsmount *m, struct hlist_head *p)
|
|
{
|
|
spin_lock(&pin_lock);
|
|
if (p)
|
|
hlist_add_head(&pin->s_list, p);
|
|
hlist_add_head(&pin->m_list, &real_mount(m)->mnt_pins);
|
|
spin_unlock(&pin_lock);
|
|
}
|
|
|
|
void pin_insert(struct fs_pin *pin, struct vfsmount *m)
|
|
{
|
|
pin_insert_group(pin, m, &m->mnt_sb->s_pins);
|
|
}
|
|
|
|
void pin_kill(struct fs_pin *p)
|
|
{
|
|
wait_queue_entry_t wait;
|
|
|
|
if (!p) {
|
|
rcu_read_unlock();
|
|
return;
|
|
}
|
|
init_wait(&wait);
|
|
spin_lock_irq(&p->wait.lock);
|
|
if (likely(!p->done)) {
|
|
p->done = -1;
|
|
spin_unlock_irq(&p->wait.lock);
|
|
rcu_read_unlock();
|
|
p->kill(p);
|
|
return;
|
|
}
|
|
if (p->done > 0) {
|
|
spin_unlock_irq(&p->wait.lock);
|
|
rcu_read_unlock();
|
|
return;
|
|
}
|
|
__add_wait_queue(&p->wait, &wait);
|
|
while (1) {
|
|
set_current_state(TASK_UNINTERRUPTIBLE);
|
|
spin_unlock_irq(&p->wait.lock);
|
|
rcu_read_unlock();
|
|
schedule();
|
|
rcu_read_lock();
|
|
if (likely(list_empty(&wait.entry)))
|
|
break;
|
|
/* OK, we know p couldn't have been freed yet */
|
|
spin_lock_irq(&p->wait.lock);
|
|
if (p->done > 0) {
|
|
spin_unlock_irq(&p->wait.lock);
|
|
break;
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
void mnt_pin_kill(struct mount *m)
|
|
{
|
|
while (1) {
|
|
struct hlist_node *p;
|
|
rcu_read_lock();
|
|
p = READ_ONCE(m->mnt_pins.first);
|
|
if (!p) {
|
|
rcu_read_unlock();
|
|
break;
|
|
}
|
|
pin_kill(hlist_entry(p, struct fs_pin, m_list));
|
|
}
|
|
}
|
|
|
|
void group_pin_kill(struct hlist_head *p)
|
|
{
|
|
while (1) {
|
|
struct hlist_node *q;
|
|
rcu_read_lock();
|
|
q = READ_ONCE(p->first);
|
|
if (!q) {
|
|
rcu_read_unlock();
|
|
break;
|
|
}
|
|
pin_kill(hlist_entry(q, struct fs_pin, s_list));
|
|
}
|
|
}
|