On Wed, Nov 04, 2020 at 05:44:45PM +0100, KP Singh wrote:
[ ... ]

> +static void *bpf_pid_task_storage_lookup_elem(struct bpf_map *map, void *key)
> +{
> +     struct bpf_local_storage_data *sdata;
> +     struct task_struct *task;
> +     unsigned int f_flags;
> +     struct pid *pid;
> +     int fd, err;
> +
> +     fd = *(int *)key;
> +     pid = pidfd_get_pid(fd, &f_flags);
> +     if (IS_ERR(pid))
> +             return ERR_CAST(pid);
> +
> +     /* We should be in an RCU read side critical section, it should be safe
> +      * to call pid_task.
> +      */
> +     WARN_ON_ONCE(!rcu_read_lock_held());
> +     task = pid_task(pid, PIDTYPE_PID);
> +     if (!task) {
> +             err = -ENOENT;
> +             goto out;
> +     }
> +
> +     sdata = task_storage_lookup(task, map, true);
> +     put_pid(pid);
> +     return sdata ? sdata->data : NULL;
> +out:
> +     put_pid(pid);
> +     return ERR_PTR(err);
> +}
> +
> +static int bpf_pid_task_storage_update_elem(struct bpf_map *map, void *key,
> +                                         void *value, u64 map_flags)
> +{
> +     struct bpf_local_storage_data *sdata;
> +     struct task_struct *task;
> +     unsigned int f_flags;
> +     struct pid *pid;
> +     int fd, err;
> +
> +     fd = *(int *)key;
> +     pid = pidfd_get_pid(fd, &f_flags);
> +     if (IS_ERR(pid))
> +             return PTR_ERR(pid);
> +
> +     /* We should be in an RCU read side critical section, it should be safe
> +      * to call pid_task.
> +      */
> +     WARN_ON_ONCE(!rcu_read_lock_held());
> +     task = pid_task(pid, PIDTYPE_PID);
> +     if (!task) {
> +             err = -ENOENT;
> +             goto out;
> +     }
> +
> +     sdata = bpf_local_storage_update(
> +             task, (struct bpf_local_storage_map *)map, value, map_flags);
It seems the task is protected by rcu here and the task may be going away.
Is it ok?

or the following comment in the later "BPF_CALL_4(bpf_task_storage_get, ...)"
is no longer valid?
        /* This helper must only called from where the task is guaranteed
         * to have a refcount and cannot be freed.
         */
         
> +
> +     err = PTR_ERR_OR_ZERO(sdata);
> +out:
> +     put_pid(pid);
> +     return err;
> +}
> +

[ ... ]

> +BPF_CALL_4(bpf_task_storage_get, struct bpf_map *, map, struct task_struct *,
> +        task, void *, value, u64, flags)
> +{
> +     struct bpf_local_storage_data *sdata;
> +
> +     if (flags & ~(BPF_LOCAL_STORAGE_GET_F_CREATE))
> +             return (unsigned long)NULL;
> +
> +     /* explicitly check that the task_storage_ptr is not
> +      * NULL as task_storage_lookup returns NULL in this case and
> +      * bpf_local_storage_update expects the owner to have a
> +      * valid storage pointer.
> +      */
> +     if (!task_storage_ptr(task))
> +             return (unsigned long)NULL;
> +
> +     sdata = task_storage_lookup(task, map, true);
> +     if (sdata)
> +             return (unsigned long)sdata->data;
> +
> +     /* This helper must only called from where the task is guaranteed
> +      * to have a refcount and cannot be freed.
> +      */
> +     if (flags & BPF_LOCAL_STORAGE_GET_F_CREATE) {
> +             sdata = bpf_local_storage_update(
> +                     task, (struct bpf_local_storage_map *)map, value,
> +                     BPF_NOEXIST);
> +             return IS_ERR(sdata) ? (unsigned long)NULL :
> +                                          (unsigned long)sdata->data;
> +     }
> +
> +     return (unsigned long)NULL;
> +}

[ ... ]

> diff --git a/kernel/bpf/syscall.c b/kernel/bpf/syscall.c
> index 8f50c9c19f1b..f3fe9f53f93c 100644
> --- a/kernel/bpf/syscall.c
> +++ b/kernel/bpf/syscall.c
> @@ -773,7 +773,8 @@ static int map_check_btf(struct bpf_map *map, const 
> struct btf *btf,
>                   map->map_type != BPF_MAP_TYPE_ARRAY &&
>                   map->map_type != BPF_MAP_TYPE_CGROUP_STORAGE &&
>                   map->map_type != BPF_MAP_TYPE_SK_STORAGE &&
> -                 map->map_type != BPF_MAP_TYPE_INODE_STORAGE)
> +                 map->map_type != BPF_MAP_TYPE_INODE_STORAGE &&
> +                 map->map_type != BPF_MAP_TYPE_TASK_STORAGE)
This is to enable spin lock support in a map's value.  Without peeking
patch 5, I was confused a bit here.  It seems patch 5 was missed when
inode storage was added.

>                       return -ENOTSUPP;
>               if (map->spin_lock_off + sizeof(struct bpf_spin_lock) >
>                   map->value_size) {

Reply via email to