[Devel] Re: [PATCH] c/r: Add UTS support (v4)

Oren Laadan orenl at cs.columbia.edu
Fri Mar 20 12:38:17 PDT 2009



Serge E. Hallyn wrote:
> Quoting Oren Laadan (orenl at cs.columbia.edu):
>>
>> Dan Smith wrote:
>>> OL> So what happens in the following scenario:
>>>
>>> OL> * task A is the container init(1)
>>> OL> * A calls fork() to create task B
>>> OL> * B calls unshare(CLONE_NEWUTS)
>>> OL> * B calls clone(CLONE_PARENT) to create task C
>>>
>>> In the previous version of the patch, I failed the checkpoint if this
>>> was the case by making sure that all tasks in the set had the same
>>> nsproxy.  You said in IRC that this was already done elsewhere in the
>>> infrastructure, but now that I look I don't see that anywhere.
>>>
>> in cr_may_checkpoint_task():
>>
>>  285         /* FIXME: change this for nested containers */
>>  286         if (task_nsproxy(t) != ctx->root_nsproxy)
>>  287                 return -EPERM;
>>
>>> The check I had was in response to Daniel's comments about avoiding
>>> the situation for the time being by making sure that all the tasks had
>>> the same set of namespaces (i.e. the same nsproxy at the time of
>>> checkpoint).
>>>
>>> OL> Two approaches to solve this are:
>>>
>>> OL> a) Identify, in mktree, that this was the case, and impose an
>>> OL> order on the forks/clones to recreate the same dependency (an
>>> OL> algorithm for this is described in [1])
>>>
>>> OL> b) Do it in the kernel: for each nsproxy (identified by an objref)
>>> OL> the first task that has it will create it during restart, in or
>>> OL> out of the kernel, and the next task will simply attach to the
>>> OL> existing one that will be deposited in the objhash.
>>>
>>> I think that prior discussion led to the conclusion that simplicity
>>> wins for the moment, but if you want to solve it now I can cook up
>>> some changes.
>>>
>> If we keep the assumption, for simplicity, that all tasks share the
>> same namespace, then the checkpoint code should check, once, how that
>> nsproxy differs from the container's parent (except for the obvious
>> pidns).
> 
> I disagree.  Whether the container had its own utsns doesn't
> affect whether it should have a private utsns on restart.

Right, I missed that...

>> If it does differ, e.g. in uts, then the checkpoint should save the
>> uts state _once_ - as in global data. Restart will restore the state
>> also _once_, for the init of the container (the first task restored),
>> _before_ it forks the rest of the tree.
>>
>> Otherwise, we don't get the same outcome.
> 
> Again I disagree.  If we were planning on never supporting nested
> uts namespaces it woudl be fine, but what you are talking about
> is making sure we have to break the checkpoint format later to support
> nested namespaces.

We don't know how we are to support nested namespaces. So either we solve
it now, or we do something that is bound to break later. The image format
is going to change anyways as we move along.

> 
> Rather, we should do:
> 
> 1. record the hostname for the container in global data.
> 2. The restart program can decide whether to honor the global
>    checkpoint image hostname or not.  It can either use a
>    command line option, or check whether the recorded hostname
>    is different from the restart host.  I prefer the former.

Sounds good.

> 3. for each task, leave an optional spot for hostname.  If
>    there is a hostname, then it will unshare(CLONE_NEWUTS)
>    and set its hostname before calling sys_restart() or
>    cloning any child tasks.

Doesn't this imply a a specific format that is bound to break later ?

Oren.
_______________________________________________
Containers mailing list
Containers at lists.linux-foundation.org
https://lists.linux-foundation.org/mailman/listinfo/containers




More information about the Devel mailing list