[CRIU] Dumping Process - lxc-ls -f Problem

Thouraya TH thouraya87 at gmail.com
Mon Jun 15 11:29:36 PDT 2015


Hello;

I have done new tests on my old container and the same problem:

restore.log
Warn  (cr-restore.c:1029): Set CLONE_PARENT | CLONE_NEWPID but it might
cause restore problem,because not all kernels support such clone flags
combinations!
RTNETLINK answers: File exists
RTNETLINK answers: File exists
RTNETLINK answers: File exists
   920: Error (sk-inet.c:610): Can't bind inet socket: Invalid argument


root at g-23:/tmp# lxc-ls -f
^CTraceback (most recent call last):
  File "/usr/bin/lxc-ls", line 432, in <module>
    containers = get_containers(root=True)
  File "/usr/bin/lxc-ls", line 261, in get_containers
    if container.controllable:
KeyboardInterrupt







*Do you have a way to reproduce this? I'm not sure what it means exactly,
perhaps Pavel can elaborate. However, it would be nice to have a small
testcase so I can try and fix it.*

In this old container,i have (lxc bridge network, jdk, some other
libraries, i have modified /etc/sudoers, etc... )
i can't understand the source of the problem !

Best Regards.



2015-06-15 14:45 GMT+01:00 Tycho Andersen <tycho.andersen at canonical.com>:

> On Sun, Jun 14, 2015 at 02:02:13PM +0100, Thouraya TH wrote:
> > Hello all;
> >
> > I have done two tests:  (criu 1.6)
> >
> > 1- *Test 1:*
> >    I have created a new container using:
> >                      lxc-create -t ubuntu -n worker2 http_proxy=True
> >
> >  *I have not installed any tool in this container.*
> >
> >   lxc-checkpoint -s -D /tmp/dire -n worker2
> >   lxc-checkpoint -r -D /tmp/dire -n worker2
> >
> > root at localhost:~# lxc-ls -f
> > NAME     STATE    IPV4       IPV6  GROUPS  AUTOSTART
> > ----------------------------------------------------
> > worker   STOPPED  -          -     -       NO
> > *worker2*  RUNNING  10.0.3.48  -     -       NO
> >
> > 2- *Test 2: *
> >
> > lxc-start -n worker  (it is an old worker: i have installed many tools in
> > this container, jdk, etc ......)
> > root at localhost:/tmp# lxc-ls -f
> > NAME     STATE    IPV4        IPV6  GROUPS  AUTOSTART
> > -----------------------------------------------------
> > worker   RUNNING  10.0.3.109  -     -       NO
> > worker2  RUNNING  10.0.3.48   -     -       NO
> >
> > lxc-checkpoint -s -D /tmp/direworker -n worker
> > lxc-checkpoint -r -D /tmp/direworker -n worker
> >
> > lxc-ls -f
> > ^CTraceback (most recent call last):
> >   File "/usr/bin/lxc-ls", line 432, in <module>
> >     containers = get_containers(root=True)
> >   File "/usr/bin/lxc-ls", line 261, in get_containers
> >     if container.controllable:
> > KeyboardInterrupt
>
> This means that the restore failed and lxc-checkpoint didn't
> understand that it failed, and is still waiting. I think we recently
> fixed a bug (there is a patch about SIGCHLD) that will cause this not
> to hang here.
>
> > dump.log:
> > Warn  (fsnotify.c:188): fsnotify:       Handle 800003:3c605 cannot be
> opened
> > Warn  (fsnotify.c:188): fsnotify:       Handle 800003:4a4ba cannot be
> opened
> > Warn  (arch/x86/crtools.c:132): Will restore 4271 with interrupted system
> > call
> > Warn  (arch/x86/crtools.c:132): Will restore 4454 with interrupted system
> > call
> > Warn  (arch/x86/crtools.c:132): Will restore 4455 with interrupted system
> > call
> > Warn  (arch/x86/crtools.c:132): Will restore 4460 with interrupted system
> > call
> > Warn  (arch/x86/crtools.c:132): Will restore 4461 with interrupted system
> > call
> > Warn  (arch/x86/crtools.c:132): Will restore 4463 with interrupted system
> > call
> >
> > restore.log:
> > Warn  (cr-restore.c:1029): Set CLONE_PARENT | CLONE_NEWPID but it might
> > cause restore problem,because not all kernels support such clone flags
> > combinations!
> > RTNETLINK answers: File exists
> > RTNETLINK answers: File exists
> > RTNETLINK answers: File exists
> >
> >
> > Can you please explain to me what is the problem?  is it because of some
> > tools installed in the container?
>
> Do you have a way to reproduce this? I'm not sure what it means
> exactly, perhaps Pavel can elaborate. However, it would be nice to
> have a small testcase so I can try and fix it.
>
> Thanks,
>
> Tycho
>
> >
> > Thanks a lot for help.
> > Best regards.
>
> > _______________________________________________
> > CRIU mailing list
> > CRIU at openvz.org
> > https://lists.openvz.org/mailman/listinfo/criu
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openvz.org/pipermail/criu/attachments/20150615/bed5c9c3/attachment.html>


More information about the CRIU mailing list