| // SPDX-License-Identifier: GPL-2.0-or-later | 
 | /* | 
 |  *  Copyright (C) 2008 Red Hat, Inc., Eric Paris <eparis@redhat.com> | 
 |  */ | 
 |  | 
 | #include <linux/list.h> | 
 | #include <linux/mutex.h> | 
 | #include <linux/slab.h> | 
 | #include <linux/srcu.h> | 
 | #include <linux/rculist.h> | 
 | #include <linux/wait.h> | 
 | #include <linux/memcontrol.h> | 
 |  | 
 | #include <linux/fsnotify_backend.h> | 
 | #include "fsnotify.h" | 
 |  | 
 | #include <linux/atomic.h> | 
 |  | 
 | /* | 
 |  * Final freeing of a group | 
 |  */ | 
 | static void fsnotify_final_destroy_group(struct fsnotify_group *group) | 
 | { | 
 | 	if (group->ops->free_group_priv) | 
 | 		group->ops->free_group_priv(group); | 
 |  | 
 | 	mem_cgroup_put(group->memcg); | 
 | 	mutex_destroy(&group->mark_mutex); | 
 |  | 
 | 	kfree(group); | 
 | } | 
 |  | 
 | /* | 
 |  * Stop queueing new events for this group. Once this function returns | 
 |  * fsnotify_add_event() will not add any new events to the group's queue. | 
 |  */ | 
 | void fsnotify_group_stop_queueing(struct fsnotify_group *group) | 
 | { | 
 | 	spin_lock(&group->notification_lock); | 
 | 	group->shutdown = true; | 
 | 	spin_unlock(&group->notification_lock); | 
 | } | 
 |  | 
 | /* | 
 |  * Trying to get rid of a group. Remove all marks, flush all events and release | 
 |  * the group reference. | 
 |  * Note that another thread calling fsnotify_clear_marks_by_group() may still | 
 |  * hold a ref to the group. | 
 |  */ | 
 | void fsnotify_destroy_group(struct fsnotify_group *group) | 
 | { | 
 | 	/* | 
 | 	 * Stop queueing new events. The code below is careful enough to not | 
 | 	 * require this but fanotify needs to stop queuing events even before | 
 | 	 * fsnotify_destroy_group() is called and this makes the other callers | 
 | 	 * of fsnotify_destroy_group() to see the same behavior. | 
 | 	 */ | 
 | 	fsnotify_group_stop_queueing(group); | 
 |  | 
 | 	/* Clear all marks for this group and queue them for destruction */ | 
 | 	fsnotify_clear_marks_by_group(group, FSNOTIFY_OBJ_TYPE_ANY); | 
 |  | 
 | 	/* | 
 | 	 * Some marks can still be pinned when waiting for response from | 
 | 	 * userspace. Wait for those now. fsnotify_prepare_user_wait() will | 
 | 	 * not succeed now so this wait is race-free. | 
 | 	 */ | 
 | 	wait_event(group->notification_waitq, !atomic_read(&group->user_waits)); | 
 |  | 
 | 	/* | 
 | 	 * Wait until all marks get really destroyed. We could actually destroy | 
 | 	 * them ourselves instead of waiting for worker to do it, however that | 
 | 	 * would be racy as worker can already be processing some marks before | 
 | 	 * we even entered fsnotify_destroy_group(). | 
 | 	 */ | 
 | 	fsnotify_wait_marks_destroyed(); | 
 |  | 
 | 	/* | 
 | 	 * Since we have waited for fsnotify_mark_srcu in | 
 | 	 * fsnotify_mark_destroy_list() there can be no outstanding event | 
 | 	 * notification against this group. So clearing the notification queue | 
 | 	 * of all events is reliable now. | 
 | 	 */ | 
 | 	fsnotify_flush_notify(group); | 
 |  | 
 | 	/* | 
 | 	 * Destroy overflow event (we cannot use fsnotify_destroy_event() as | 
 | 	 * that deliberately ignores overflow events. | 
 | 	 */ | 
 | 	if (group->overflow_event) | 
 | 		group->ops->free_event(group, group->overflow_event); | 
 |  | 
 | 	fsnotify_put_group(group); | 
 | } | 
 |  | 
 | /* | 
 |  * Get reference to a group. | 
 |  */ | 
 | void fsnotify_get_group(struct fsnotify_group *group) | 
 | { | 
 | 	refcount_inc(&group->refcnt); | 
 | } | 
 |  | 
 | /* | 
 |  * Drop a reference to a group.  Free it if it's through. | 
 |  */ | 
 | void fsnotify_put_group(struct fsnotify_group *group) | 
 | { | 
 | 	if (refcount_dec_and_test(&group->refcnt)) | 
 | 		fsnotify_final_destroy_group(group); | 
 | } | 
 | EXPORT_SYMBOL_GPL(fsnotify_put_group); | 
 |  | 
 | static struct fsnotify_group *__fsnotify_alloc_group( | 
 | 				const struct fsnotify_ops *ops, | 
 | 				int flags, gfp_t gfp) | 
 | { | 
 | 	struct fsnotify_group *group; | 
 |  | 
 | 	group = kzalloc(sizeof(struct fsnotify_group), gfp); | 
 | 	if (!group) | 
 | 		return ERR_PTR(-ENOMEM); | 
 |  | 
 | 	/* set to 0 when there a no external references to this group */ | 
 | 	refcount_set(&group->refcnt, 1); | 
 | 	atomic_set(&group->user_waits, 0); | 
 |  | 
 | 	spin_lock_init(&group->notification_lock); | 
 | 	INIT_LIST_HEAD(&group->notification_list); | 
 | 	init_waitqueue_head(&group->notification_waitq); | 
 | 	group->max_events = UINT_MAX; | 
 |  | 
 | 	mutex_init(&group->mark_mutex); | 
 | 	INIT_LIST_HEAD(&group->marks_list); | 
 |  | 
 | 	group->ops = ops; | 
 | 	group->flags = flags; | 
 |  | 
 | 	return group; | 
 | } | 
 |  | 
 | /* | 
 |  * Create a new fsnotify_group and hold a reference for the group returned. | 
 |  */ | 
 | struct fsnotify_group *fsnotify_alloc_group(const struct fsnotify_ops *ops, | 
 | 					    int flags) | 
 | { | 
 | 	gfp_t gfp = (flags & FSNOTIFY_GROUP_USER) ? GFP_KERNEL_ACCOUNT : | 
 | 						    GFP_KERNEL; | 
 |  | 
 | 	return __fsnotify_alloc_group(ops, flags, gfp); | 
 | } | 
 | EXPORT_SYMBOL_GPL(fsnotify_alloc_group); | 
 |  | 
 | int fsnotify_fasync(int fd, struct file *file, int on) | 
 | { | 
 | 	struct fsnotify_group *group = file->private_data; | 
 |  | 
 | 	return fasync_helper(fd, file, on, &group->fsn_fa) >= 0 ? 0 : -EIO; | 
 | } |