[Devel] [PATCH RH7 v4] cgroup: add export_operations to cgroup super block
Andrey Zhadchenko
andrey.zhadchenko at virtuozzo.com
Mon Oct 26 08:43:53 MSK 2020
This patch is not needed for vz8, because cgroup will use kernfs, which
has export_operations implemented.
Best regards,
Andrey Zhadchenko
On Thu, 24 Sep 2020 10:41:31 +0300
Konstantin Khorenko <khorenko at virtuozzo.com> wrote:
> Please, port this patch to vz8 as well, branch
> branch-rh8-4.18.0-193.6.3.vz8.4.x-ovz
>
> --
> Best regards,
>
> Konstantin Khorenko,
> Virtuozzo Linux Kernel Team
>
> On 07/30/2020 04:01 PM, Andrey Zhadchenko wrote:
> > criu uses fhandle from fdinfo to dump inotify objects. cgroup super
> > block has no export operations, but .encode_fh and .fh_to_dentry
> > are needed for inotify_fdinfo function and open_by_handle_at
> > syscall in order to correctly open files located on cgroupfs by
> > fhandle. Add hash table as a storage for inodes with exported
> > fhandle.
> >
> > v3: use inode->i_gen to protect from i_ino reusage. increase
> > fhandle size to 2 * u32.
> > Add an option to take reference of inode in cgroup_find_inode, so
> > no one can delete recently found inode.
> > v4: introduced hashtable helper functions to avoid races.
> > changed i_gen generation from get_seconds to prandom_u32.
> >
> > https://jira.sw.ru/browse/PSBM-105889
> > Signed-off-by: Andrey Zhadchenko <andrey.zhadchenko at virtuozzo.com>
> > ---
> > kernel/cgroup.c | 168
> > +++++++++++++++++++++++++++++++++++++++++++++++++++++++- 1 file
> > changed, 167 insertions(+), 1 deletion(-)
> >
> > diff --git a/kernel/cgroup.c b/kernel/cgroup.c
> > index 9fdba79..956a9ac 100644
> > --- a/kernel/cgroup.c
> > +++ b/kernel/cgroup.c
> > @@ -62,6 +62,8 @@
> > #include <linux/kthread.h>
> > #include <linux/ve.h>
> > #include <linux/stacktrace.h>
> > +#include <linux/exportfs.h>
> > +#include <linux/random.h>
> >
> > #include <linux/atomic.h>
> >
> > @@ -765,6 +767,7 @@ static struct inode *cgroup_new_inode(umode_t
> > mode, struct super_block *sb)
> >
> > if (inode) {
> > inode->i_ino = get_next_ino();
> > + inode->i_generation = prandom_u32();
> > inode->i_mode = mode;
> > inode->i_uid = current_fsuid();
> > inode->i_gid = current_fsgid();
> > @@ -1390,9 +1393,171 @@ out:
> > }
> > #endif
> >
> > +/*
> > + * hashtable for inodes that have exported fhandles.
> > + * When we export fhandle, we add it's inode into
> > + * hashtable so we can find it fast
> > + */
> > +
> > +#define CGROUP_INODE_HASH_BITS 10
> > +static DEFINE_HASHTABLE(cgroup_inode_table,
> > CGROUP_INODE_HASH_BITS); +static
> > DEFINE_SPINLOCK(cgroup_inode_table_lock); +
> > +struct cg_inode_hitem {
> > + struct inode *inode;
> > + struct hlist_node hlist;
> > +};
> > +
> > +static inline unsigned long cgroup_inode_get_hash(unsigned int
> > i_ino) +{
> > + return hash_32(i_ino, CGROUP_INODE_HASH_BITS);
> > +}
> > +
> > +static struct cg_inode_hitem *cgroup_find_item_no_lock(unsigned
> > long fh[2]) +{
> > + struct cg_inode_hitem *i;
> > + struct hlist_head *head = cgroup_inode_table
> > + + cgroup_inode_get_hash(fh[1]);
> > + struct cg_inode_hitem *found = NULL;
> > +
> > + hlist_for_each_entry(i, head, hlist) {
> > + if (i->inode->i_generation == fh[0] &&
> > + i->inode->i_ino == fh[1]) {
> > + found = i;
> > + break;
> > + }
> > + }
> > +
> > + return found;
> > +}
> > +
> > +static struct inode *cgroup_find_inode(unsigned long fh[2], char
> > take_ref) +{
> > + struct cg_inode_hitem *item;
> > + struct inode *ret = NULL;
> > +
> > + spin_lock(&cgroup_inode_table_lock);
> > + item = cgroup_find_item_no_lock(fh);
> > +
> > + /*
> > + * If we need to increase refcount, we should be aware of
> > possible
> > + * deadlock. Another thread may have started deleting this
> > inode:
> > + * iput->iput_final->cgroup_delete_inode->cgroup_hash_del
> > + * If we just call igrab, it will try to take i_lock and
> > this will
> > + * result in deadlock, because deleting thread has already
> > taken it
> > + * and waits on cgroup_inode_table_lock to find inode in
> > hashtable.
> > + *
> > + * If i_count is zero, someone is deleting it -> skip.
> > + */
> > + if (take_ref && item)
> > + if (!atomic_inc_not_zero(&item->inode->i_count))
> > + item = NULL;
> > +
> > + spin_unlock(&cgroup_inode_table_lock);
> > +
> > + if (item)
> > + ret = item->inode;
> > +
> > + return ret;
> > +}
> > +
> > +static int cgroup_hash_add(struct inode *inode)
> > +{
> > + unsigned long fh[2] = {inode->i_generation, inode->i_ino};
> > +
> > + if (!cgroup_find_inode(fh, 0)) {
> > + struct cg_inode_hitem *item;
> > + struct cg_inode_hitem *existing_item = 0;
> > + struct hlist_head *head = cgroup_inode_table
> > + + cgroup_inode_get_hash(inode->i_ino);
> > +
> > + item = kmalloc(sizeof(struct cg_inode_hitem),
> > GFP_KERNEL);
> > + if (!item)
> > + return -ENOMEM;
> > + item->inode = inode;
> > +
> > + spin_lock(&cgroup_inode_table_lock);
> > + existing_item = cgroup_find_item_no_lock(fh);
> > + if (!existing_item)
> > + hlist_add_head(&item->hlist, head);
> > + spin_unlock(&cgroup_inode_table_lock);
> > +
> > + if (existing_item)
> > + kfree(item);
> > + }
> > +
> > + return 0;
> > +}
> > +
> > +static void cgroup_hash_del(struct inode *inode)
> > +{
> > + struct cg_inode_hitem *item;
> > + unsigned long fh[2] = {inode->i_generation, inode->i_ino};
> > +
> > + spin_lock(&cgroup_inode_table_lock);
> > + item = cgroup_find_item_no_lock(fh);
> > + if (item)
> > + hlist_del(&item->hlist);
> > + spin_unlock(&cgroup_inode_table_lock);
> > +
> > + kfree(item);
> > + return;
> > +}
> > +
> > +static struct dentry *cgroup_fh_to_dentry(struct super_block *sb,
> > + struct fid *fid, int fh_len, int fh_type)
> > +{
> > + struct inode *inode;
> > + struct dentry *dentry = ERR_PTR(-ENOENT);
> > + unsigned long fhandle[2] = {fid->raw[0], fid->raw[1]};
> > +
> > + if (fh_len < 2)
> > + return NULL;
> > +
> > + inode = cgroup_find_inode(fhandle, 1);
> > + if (inode) {
> > + dentry = d_find_alias(inode);
> > + iput(inode);
> > + }
> > + return dentry;
> > +}
> > +
> > +static int cgroup_encode_fh(struct inode *inode, __u32 *fh, int
> > *len,
> > + struct inode *parent)
> > +{
> > + if (*len < 2) {
> > + *len = 2;
> > + return FILEID_INVALID;
> > + }
> > +
> > + /*
> > + * encode_fh is expected to return 255 (FILEID_INVALID)
> > + * in case of failure. We can't return ENOMEM, so
> > + * return FILEID_INVALID at least.
> > + */
> > + if (cgroup_hash_add(inode))
> > + return FILEID_INVALID;
> > +
> > + fh[0] = inode->i_generation;
> > + fh[1] = inode->i_ino;
> > + *len = 2;
> > + return 1;
> > +}
> > +
> > +static const struct export_operations cgroup_export_ops = {
> > + .encode_fh = cgroup_encode_fh,
> > + .fh_to_dentry = cgroup_fh_to_dentry,
> > +};
> > +
> > +static int cgroup_delete_inode(struct inode *inode)
> > +{
> > + cgroup_hash_del(inode);
> > + return generic_delete_inode(inode);
> > +}
> > +
> > static const struct super_operations cgroup_ops = {
> > .statfs = simple_statfs,
> > - .drop_inode = generic_delete_inode,
> > + .drop_inode = cgroup_delete_inode,
> > .show_options = cgroup_show_options,
> > #ifdef CONFIG_VE
> > .show_path = cgroup_show_path,
> > @@ -1539,6 +1704,7 @@ static int cgroup_set_super(struct
> > super_block *sb, void *data) sb->s_blocksize_bits =
> > PAGE_CACHE_SHIFT; sb->s_magic = CGROUP_SUPER_MAGIC;
> > sb->s_op = &cgroup_ops;
> > + sb->s_export_op = &cgroup_export_ops;
> >
> > return 0;
> > }
> >
More information about the Devel
mailing list