Hi Pradeep, Sorry I cannot get too much info from this log message, I see that you are using balloon_framework, can you try mesos-execute?
Can you please add the option of GLOG_v=1 when start master and append the whole log since the master start? Thanks, Guangya On Wed, Oct 7, 2015 at 6:17 PM, Pradeep Kiruvale <pradeepkiruv...@gmail.com> wrote: > Below are the logs from Master. > > -Pradeep > > 1007 12:16:28.257853 8005 leveldb.cpp:343] Persisting action (20 bytes) > to leveldb took 119428ns > I1007 12:16:28.257884 8005 leveldb.cpp:401] Deleting ~2 keys from leveldb > took 18847ns > I1007 12:16:28.257891 8005 replica.cpp:679] Persisted action at 1440 > I1007 12:16:28.257912 8005 replica.cpp:664] Replica learned TRUNCATE > action at position 1440 > I1007 12:16:36.666616 8002 http.cpp:336] HTTP GET for /master/state.json > from 192.168.0.102:40721 with User-Agent='Mozilla/5.0 (X11; Linux x86_64) > AppleWebKit/537.36 (KHTML, like Gecko) Chrome/44.0.2403.52 Safari/537.36' > I1007 12:16:39.126030 8001 master.cpp:2179] Received SUBSCRIBE call for > framework 'Balloon Framework (C++)' at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:39.126428 8001 master.cpp:2250] Subscribing framework Balloon > Framework (C++) with checkpointing disabled and capabilities [ ] > E1007 12:16:39.127459 8007 process.cpp:1912] Failed to shutdown socket > with fd 13: Transport endpoint is not connected > I1007 12:16:39.127535 8000 hierarchical.hpp:515] Added framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 > I1007 12:16:39.127734 8001 master.cpp:1119] Framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > disconnected > I1007 12:16:39.127765 8001 master.cpp:2475] Disconnecting framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > E1007 12:16:39.127768 8007 process.cpp:1912] Failed to shutdown socket > with fd 14: Transport endpoint is not connected > I1007 12:16:39.127789 8001 master.cpp:2499] Deactivating framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:39.127879 8006 hierarchical.hpp:599] Deactivated framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 > I1007 12:16:39.127913 8001 master.cpp:1143] Giving framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 0ns to > failover > I1007 12:16:39.129273 8005 master.cpp:4815] Framework failover timeout, > removing framework 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 (Balloon > Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:39.129312 8005 master.cpp:5571] Removing framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:39.129858 8003 hierarchical.hpp:552] Removed framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0000 > I1007 12:16:40.676519 8000 master.cpp:2179] Received SUBSCRIBE call for > framework 'Balloon Framework (C++)' at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:40.676678 8000 master.cpp:2250] Subscribing framework Balloon > Framework (C++) with checkpointing disabled and capabilities [ ] > I1007 12:16:40.677178 8006 hierarchical.hpp:515] Added framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 > E1007 12:16:40.677217 8007 process.cpp:1912] Failed to shutdown socket > with fd 13: Transport endpoint is not connected > I1007 12:16:40.677409 8000 master.cpp:1119] Framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > disconnected > I1007 12:16:40.677441 8000 master.cpp:2475] Disconnecting framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:40.677453 8000 master.cpp:2499] Deactivating framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > E1007 12:16:40.677459 8007 process.cpp:1912] Failed to shutdown socket > with fd 13: Transport endpoint is not connected > I1007 12:16:40.677501 8000 master.cpp:1143] Giving framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 0ns to > failover > I1007 12:16:40.677520 8005 hierarchical.hpp:599] Deactivated framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 > I1007 12:16:40.678864 8004 master.cpp:4815] Framework failover timeout, > removing framework 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 (Balloon > Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:40.678906 8004 master.cpp:5571] Removing framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:40.679147 8001 hierarchical.hpp:552] Removed framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0001 > I1007 12:16:41.853121 8002 master.cpp:2179] Received SUBSCRIBE call for > framework 'Balloon Framework (C++)' at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:41.853281 8002 master.cpp:2250] Subscribing framework Balloon > Framework (C++) with checkpointing disabled and capabilities [ ] > E1007 12:16:41.853806 8007 process.cpp:1912] Failed to shutdown socket > with fd 13: Transport endpoint is not connected > I1007 12:16:41.853833 8004 hierarchical.hpp:515] Added framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 > I1007 12:16:41.854032 8002 master.cpp:1119] Framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > disconnected > I1007 12:16:41.854063 8002 master.cpp:2475] Disconnecting framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:41.854076 8002 master.cpp:2499] Deactivating framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > E1007 12:16:41.854080 8007 process.cpp:1912] Failed to shutdown socket > with fd 13: Transport endpoint is not connected > I1007 12:16:41.854126 8005 hierarchical.hpp:599] Deactivated framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 > I1007 12:16:41.854121 8002 master.cpp:1143] Giving framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 0ns to > failover > I1007 12:16:41.855482 8006 master.cpp:4815] Framework failover timeout, > removing framework 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 (Balloon > Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:41.855515 8006 master.cpp:5571] Removing framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:41.855692 8001 hierarchical.hpp:552] Removed framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0002 > I1007 12:16:42.772830 8000 master.cpp:2179] Received SUBSCRIBE call for > framework 'Balloon Framework (C++)' at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:42.772974 8000 master.cpp:2250] Subscribing framework Balloon > Framework (C++) with checkpointing disabled and capabilities [ ] > I1007 12:16:42.773470 8004 hierarchical.hpp:515] Added framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 > E1007 12:16:42.773495 8007 process.cpp:1912] Failed to shutdown socket > with fd 13: Transport endpoint is not connected > I1007 12:16:42.773679 8000 master.cpp:1119] Framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > disconnected > I1007 12:16:42.773697 8000 master.cpp:2475] Disconnecting framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:42.773708 8000 master.cpp:2499] Deactivating framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > E1007 12:16:42.773710 8007 process.cpp:1912] Failed to shutdown socket > with fd 14: Transport endpoint is not connected > I1007 12:16:42.773761 8000 master.cpp:1143] Giving framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 0ns to > failover > I1007 12:16:42.773779 8001 hierarchical.hpp:599] Deactivated framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 > I1007 12:16:42.775089 8005 master.cpp:4815] Framework failover timeout, > removing framework 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 (Balloon > Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:42.775126 8005 master.cpp:5571] Removing framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 (Balloon Framework (C++)) at > scheduler-a8deafaa-cf10-401c-a61c-515340560c49@127.0.1.1:58843 > I1007 12:16:42.775324 8005 hierarchical.hpp:552] Removed framework > 0ccab17d-20e8-4ab8-9de4-ae60691f8c8e-0003 > I1007 12:16:47.665941 8001 http.cpp:336] HTTP GET for /master/state.json > from 192.168.0.102:40722 with User-Agent='Mozilla/5.0 (X11; Linux x86_64) > AppleWebKit/537.36 (KHTML, like Gecko) Chrome/44.0.2403.52 Safari/537.36' > > > On 7 October 2015 at 12:12, Guangya Liu <gyliu...@gmail.com> wrote: > >> Hi Pradeep, >> >> Can you please append more log for your master node? Just want to see >> what is wrong with your master, why the framework start to failover? >> >> Thanks, >> >> Guangya >> >> On Wed, Oct 7, 2015 at 5:27 PM, Pradeep Kiruvale < >> pradeepkiruv...@gmail.com> wrote: >> >>> Hi Guangya, >>> >>> I am running a frame work from some other physical node, which is part >>> of the same network. Still I am getting below messages and the framework >>> not getting registered. >>> >>> Any idea what is the reason? >>> >>> I1007 11:24:58.781914 32392 master.cpp:4815] Framework failover timeout, >>> removing framework 89b179d8-9fb7-4a61-ad03-a9a5525482ff-0019 (Balloon >>> Framework (C++)) at >>> scheduler-3848d80c-8d27-48e0-a6b7-7e1678d5401d@127.0.1.1:54203 >>> I1007 11:24:58.781968 32392 master.cpp:5571] Removing framework >>> 89b179d8-9fb7-4a61-ad03-a9a5525482ff-0019 (Balloon Framework (C++)) at >>> scheduler-3848d80c-8d27-48e0-a6b7-7e1678d5401d@127.0.1.1:54203 >>> I1007 11:24:58.782352 32392 hierarchical.hpp:552] Removed framework >>> 89b179d8-9fb7-4a61-ad03-a9a5525482ff-0019 >>> E1007 11:24:58.782577 32399 process.cpp:1912] Failed to shutdown socket >>> with fd 13: Transport endpoint is not connected >>> I1007 11:24:59.699587 32396 master.cpp:2179] Received SUBSCRIBE call for >>> framework 'Balloon Framework (C++)' at >>> scheduler-3848d80c-8d27-48e0-a6b7-7e1678d5401d@127.0.1.1:54203 >>> I1007 11:24:59.699717 32396 master.cpp:2250] Subscribing framework >>> Balloon Framework (C++) with checkpointing disabled and capabilities [ ] >>> I1007 11:24:59.700251 32393 hierarchical.hpp:515] Added framework >>> 89b179d8-9fb7-4a61-ad03-a9a5525482ff-0020 >>> E1007 11:24:59.700253 32399 process.cpp:1912] Failed to shutdown socket >>> with fd 13: Transport endpoint is not connected >>> >>> >>> Regards, >>> Pradeep >>> >>> >>> On 5 October 2015 at 13:51, Guangya Liu <gyliu...@gmail.com> wrote: >>> >>>> Hi Pradeep, >>>> >>>> I think that the problem might be caused by that you are running the >>>> lxc container on master node and not sure if there are any port conflict or >>>> what else wrong. >>>> >>>> For my case, I was running the client in a new node but not on master >>>> node, perhaps you can have a try to put your client on a new node but not >>>> on master node. >>>> >>>> Thanks, >>>> >>>> Guangya >>>> >>>> >>>> On Mon, Oct 5, 2015 at 7:30 PM, Pradeep Kiruvale < >>>> pradeepkiruv...@gmail.com> wrote: >>>> >>>>> Hi Guangya, >>>>> >>>>> Hmm!...That is strange in my case! >>>>> >>>>> If I run from the mesos-execute on one of the slave/master node then >>>>> the tasks get their resources and they get scheduled well. >>>>> But if I start the mesos-execute on another node which is neither >>>>> slave/master then I have this issue. >>>>> >>>>> I am using an lxc container on master as a client to launch the tasks. >>>>> This is also in the same network as master/slaves. >>>>> And I just launch the task as you did. But the tasks are not getting >>>>> scheduled. >>>>> >>>>> >>>>> On master the logs are same as I sent you before >>>>> >>>>> Deactivating framework 77539063-89ce-4efa-a20b-ca788abbd912-0066 >>>>> >>>>> On both of the slaves I can see the below logs >>>>> >>>>> I1005 13:23:32.547987 4831 slave.cpp:1980] Asked to shut down >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0060 by >>>>> master@192.168.0.102:5050 >>>>> W1005 13:23:32.548135 4831 slave.cpp:1995] Cannot shut down unknown >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0060 >>>>> I1005 13:23:33.697707 4833 slave.cpp:3926] Current disk usage 3.60%. >>>>> Max allowed age: 6.047984349521910days >>>>> I1005 13:23:34.098599 4829 slave.cpp:1980] Asked to shut down >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0061 by >>>>> master@192.168.0.102:5050 >>>>> W1005 13:23:34.098740 4829 slave.cpp:1995] Cannot shut down unknown >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0061 >>>>> I1005 13:23:35.274569 4831 slave.cpp:1980] Asked to shut down >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0062 by >>>>> master@192.168.0.102:5050 >>>>> W1005 13:23:35.274683 4831 slave.cpp:1995] Cannot shut down unknown >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0062 >>>>> I1005 13:23:36.193964 4829 slave.cpp:1980] Asked to shut down >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0063 by >>>>> master@192.168.0.102:5050 >>>>> W1005 13:23:36.194090 4829 slave.cpp:1995] Cannot shut down unknown >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0063 >>>>> I1005 13:24:01.914788 4827 slave.cpp:1980] Asked to shut down >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0064 by >>>>> master@192.168.0.102:5050 >>>>> W1005 13:24:01.914937 4827 slave.cpp:1995] Cannot shut down unknown >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0064 >>>>> I1005 13:24:03.469974 4833 slave.cpp:1980] Asked to shut down >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0065 by >>>>> master@192.168.0.102:5050 >>>>> W1005 13:24:03.470118 4833 slave.cpp:1995] Cannot shut down unknown >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0065 >>>>> I1005 13:24:04.642654 4826 slave.cpp:1980] Asked to shut down >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0066 by >>>>> master@192.168.0.102:5050 >>>>> W1005 13:24:04.642812 4826 slave.cpp:1995] Cannot shut down unknown >>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0066 >>>>> >>>>> >>>>> >>>>> On 5 October 2015 at 13:09, Guangya Liu <gyliu...@gmail.com> wrote: >>>>> >>>>>> Hi Pradeep, >>>>>> >>>>>> From your log, seems that the master process is exiting and this >>>>>> caused the framework fail over to another mesos master. Can you please >>>>>> show >>>>>> more detail for your issue reproduced steps? >>>>>> >>>>>> I did some test by running mesos-execute on a client host which does >>>>>> not have any mesos service and the task can schedule well. >>>>>> >>>>>> root@mesos008:~/src/mesos/m1/mesos/build# ./src/mesos-execute >>>>>> --master=192.168.0.107:5050 --name="cluster-test" >>>>>> --command="/bin/sleep 10" --resources="cpus(*):1;mem(*):256" >>>>>> I1005 18:59:47.974123 1233 sched.cpp:164] Version: 0.26.0 >>>>>> I1005 18:59:47.990890 1248 sched.cpp:262] New master detected at >>>>>> master@192.168.0.107:5050 >>>>>> I1005 18:59:47.993074 1248 sched.cpp:272] No credentials provided. >>>>>> Attempting to register without authentication >>>>>> I1005 18:59:48.001194 1249 sched.cpp:641] Framework registered with >>>>>> 04b9af5e-e9b6-4c59-8734-eba407163922-0002 >>>>>> Framework registered with 04b9af5e-e9b6-4c59-8734-eba407163922-0002 >>>>>> task cluster-test submitted to slave >>>>>> c0e5fdde-595e-4768-9d04-25901d4523b6-S0 >>>>>> Received status update TASK_RUNNING for task cluster-test >>>>>> Received status update TASK_FINISHED for task cluster-test >>>>>> I1005 18:59:58.431144 1249 sched.cpp:1771] Asked to stop the driver >>>>>> I1005 18:59:58.431591 1249 sched.cpp:1040] Stopping framework >>>>>> '04b9af5e-e9b6-4c59-8734-eba407163922-0002' >>>>>> root@mesos008:~/src/mesos/m1/mesos/build# ps -ef | grep mesos >>>>>> root 1259 1159 0 19:06 pts/0 00:00:00 grep --color=auto >>>>>> mesos >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Guangya >>>>>> >>>>>> >>>>>> On Mon, Oct 5, 2015 at 6:50 PM, Pradeep Kiruvale < >>>>>> pradeepkiruv...@gmail.com> wrote: >>>>>> >>>>>>> Hi Guangya, >>>>>>> >>>>>>> I am facing one more issue. If I try to schedule the tasks from some >>>>>>> external client system running the same cli mesos-execute. >>>>>>> The tasks are not getting launched. The tasks reach the Master and >>>>>>> it just drops the requests, below are the logs related to that >>>>>>> >>>>>>> I1005 11:33:35.025594 21369 master.cpp:2250] Subscribing framework >>>>>>> with checkpointing disabled and capabilities [ ] >>>>>>> E1005 11:33:35.026100 21373 process.cpp:1912] Failed to shutdown >>>>>>> socket with fd 14: Transport endpoint is not connected >>>>>>> I1005 11:33:35.026129 21372 hierarchical.hpp:515] Added framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 >>>>>>> I1005 11:33:35.026298 21369 master.cpp:1119] Framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 () at >>>>>>> scheduler-b1bc0243-b5be-44ae-894c-ca318c24ce6d@127.0.1.1:47259 >>>>>>> disconnected >>>>>>> I1005 11:33:35.026329 21369 master.cpp:2475] Disconnecting framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 () at >>>>>>> scheduler-b1bc0243-b5be-44ae-894c-ca318c24ce6d@127.0.1.1:47259 >>>>>>> I1005 11:33:35.026340 21369 master.cpp:2499] Deactivating framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 () at >>>>>>> scheduler-b1bc0243-b5be-44ae-894c-ca318c24ce6d@127.0.1.1:47259 >>>>>>> E1005 11:33:35.026345 21373 process.cpp:1912] Failed to shutdown >>>>>>> socket with fd 14: Transport endpoint is not connected >>>>>>> I1005 11:33:35.026376 21369 master.cpp:1143] Giving framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 () at >>>>>>> scheduler-b1bc0243-b5be-44ae-894c-ca318c24ce6d@127.0.1.1:47259 0ns >>>>>>> to failover >>>>>>> I1005 11:33:35.026743 21372 hierarchical.hpp:599] Deactivated >>>>>>> framework 77539063-89ce-4efa-a20b-ca788abbd912-0055 >>>>>>> W1005 11:33:35.026757 21368 master.cpp:4828] Master returning >>>>>>> resources offered to framework 77539063-89ce-4efa-a20b-ca788abbd912-0055 >>>>>>> because the framework has terminated or is inactive >>>>>>> I1005 11:33:35.027014 21371 hierarchical.hpp:1103] Recovered >>>>>>> cpus(*):8; mem(*):14868; disk(*):218835; ports(*):[31000-32000] (total: >>>>>>> cpus(*):8; mem(*):14868; disk(*):218835; ports(*):[31000-32000], >>>>>>> allocated: >>>>>>> ) on slave 77539063-89ce-4efa-a20b-ca788abbd912-S2 from framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 >>>>>>> I1005 11:33:35.027159 21371 hierarchical.hpp:1103] Recovered >>>>>>> cpus(*):8; mem(*):14930; disk(*):218578; ports(*):[31000-32000] (total: >>>>>>> cpus(*):8; mem(*):14930; disk(*):218578; ports(*):[31000-32000], >>>>>>> allocated: >>>>>>> ) on slave 77539063-89ce-4efa-a20b-ca788abbd912-S1 from framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 >>>>>>> I1005 11:33:35.027668 21366 master.cpp:4815] Framework failover >>>>>>> timeout, removing framework 77539063-89ce-4efa-a20b-ca788abbd912-0055 >>>>>>> () at >>>>>>> scheduler-b1bc0243-b5be-44ae-894c-ca318c24ce6d@127.0.1.1:47259 >>>>>>> I1005 11:33:35.027715 21366 master.cpp:5571] Removing framework >>>>>>> 77539063-89ce-4efa-a20b-ca788abbd912-0055 () at >>>>>>> scheduler-b1bc0243-b5be-44ae-894c-ca318c24ce6d@127.0.1.1:47259 >>>>>>> >>>>>>> >>>>>>> Can you please tell me what is the reason? The client is in the same >>>>>>> network as well. But it does not run any master or slave processes. >>>>>>> >>>>>>> Thanks & Regards, >>>>>>> Pradeeep >>>>>>> >>>>>>> On 5 October 2015 at 12:13, Guangya Liu <gyliu...@gmail.com> wrote: >>>>>>> >>>>>>>> Hi Pradeep, >>>>>>>> >>>>>>>> Glad it finally works! Not sure if you are using systemd.slice or >>>>>>>> not, are you running to this issue: >>>>>>>> https://issues.apache.org/jira/browse/MESOS-1195 >>>>>>>> >>>>>>>> Hope Jie Yu can give you some help on this ;-) >>>>>>>> >>>>>>>> Thanks, >>>>>>>> >>>>>>>> Guangya >>>>>>>> >>>>>>>> On Mon, Oct 5, 2015 at 5:25 PM, Pradeep Kiruvale < >>>>>>>> pradeepkiruv...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Hi Guangya, >>>>>>>>> >>>>>>>>> >>>>>>>>> Thanks for sharing the information. >>>>>>>>> >>>>>>>>> Now I could launch the tasks. The problem was with the permission. >>>>>>>>> If I start all the slaves and Master as root it works fine. >>>>>>>>> Else I have problem with launching the tasks. >>>>>>>>> >>>>>>>>> But on one of the slave I could not launch the slave as root, I am >>>>>>>>> facing the following issue. >>>>>>>>> >>>>>>>>> Failed to create a containerizer: Could not create >>>>>>>>> MesosContainerizer: Failed to create launcher: Failed to create Linux >>>>>>>>> launcher: Failed to mount cgroups hierarchy at >>>>>>>>> '/sys/fs/cgroup/freezer': >>>>>>>>> 'freezer' is already attached to another hierarchy >>>>>>>>> >>>>>>>>> I took that out from the cluster for now. The tasks are getting >>>>>>>>> scheduled on the other two slave nodes. >>>>>>>>> >>>>>>>>> Thanks for your timely help >>>>>>>>> >>>>>>>>> -Pradeep >>>>>>>>> >>>>>>>>> On 5 October 2015 at 10:54, Guangya Liu <gyliu...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi Pradeep, >>>>>>>>>> >>>>>>>>>> My steps was pretty simple just as >>>>>>>>>> https://github.com/apache/mesos/blob/master/docs/getting-started.md#examples >>>>>>>>>> >>>>>>>>>> On Master node: root@mesos1:~/src/mesos/m1/mesos/build# GLOG_v=1 >>>>>>>>>> ./bin/mesos-master.sh --ip=192.168.0.107 --work_dir=/var/lib/mesos >>>>>>>>>> On 3 Slave node: root@mesos007:~/src/mesos/m1/mesos/build# >>>>>>>>>> GLOG_v=1 ./bin/mesos-slave.sh --master=192.168.0.107:5050 >>>>>>>>>> >>>>>>>>>> Then schedule a task on any of the node, here I was using slave >>>>>>>>>> node mesos007, you can see that the two tasks was launched on >>>>>>>>>> different >>>>>>>>>> host. >>>>>>>>>> >>>>>>>>>> root@mesos007:~/src/mesos/m1/mesos/build# ./src/mesos-execute >>>>>>>>>> --master=192.168.0.107:5050 --name="cluster-test" >>>>>>>>>> --command="/bin/sleep 100" --resources="cpus(*):1;mem(*):256" >>>>>>>>>> I1005 16:49:11.013432 2971 sched.cpp:164] Version: 0.26.0 >>>>>>>>>> I1005 16:49:11.027802 2992 sched.cpp:262] New master detected at >>>>>>>>>> master@192.168.0.107:5050 >>>>>>>>>> I1005 16:49:11.029579 2992 sched.cpp:272] No credentials >>>>>>>>>> provided. Attempting to register without authentication >>>>>>>>>> I1005 16:49:11.038182 2985 sched.cpp:641] Framework registered >>>>>>>>>> with c0e5fdde-595e-4768-9d04-25901d4523b6-0002 >>>>>>>>>> Framework registered with >>>>>>>>>> c0e5fdde-595e-4768-9d04-25901d4523b6-0002 >>>>>>>>>> task cluster-test submitted to slave >>>>>>>>>> c0e5fdde-595e-4768-9d04-25901d4523b6-S0 <<<<<<<<<<<<<<<<<< >>>>>>>>>> Received status update TASK_RUNNING for task cluster-test >>>>>>>>>> ^C >>>>>>>>>> root@mesos007:~/src/mesos/m1/mesos/build# ./src/mesos-execute >>>>>>>>>> --master=192.168.0.107:5050 --name="cluster-test" >>>>>>>>>> --command="/bin/sleep 100" --resources="cpus(*):1;mem(*):256" >>>>>>>>>> I1005 16:50:18.346984 3036 sched.cpp:164] Version: 0.26.0 >>>>>>>>>> I1005 16:50:18.366114 3055 sched.cpp:262] New master detected at >>>>>>>>>> master@192.168.0.107:5050 >>>>>>>>>> I1005 16:50:18.368010 3055 sched.cpp:272] No credentials >>>>>>>>>> provided. Attempting to register without authentication >>>>>>>>>> I1005 16:50:18.376338 3056 sched.cpp:641] Framework registered >>>>>>>>>> with c0e5fdde-595e-4768-9d04-25901d4523b6-0003 >>>>>>>>>> Framework registered with >>>>>>>>>> c0e5fdde-595e-4768-9d04-25901d4523b6-0003 >>>>>>>>>> task cluster-test submitted to slave >>>>>>>>>> c0e5fdde-595e-4768-9d04-25901d4523b6-S1 <<<<<<<<<<<<<<<<<<<< >>>>>>>>>> Received status update TASK_RUNNING for task cluster-test >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> >>>>>>>>>> Guangya >>>>>>>>>> >>>>>>>>>> On Mon, Oct 5, 2015 at 4:21 PM, Pradeep Kiruvale < >>>>>>>>>> pradeepkiruv...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Guangya, >>>>>>>>>>> >>>>>>>>>>> Thanks for your reply. >>>>>>>>>>> >>>>>>>>>>> I just want to know how did you launch the tasks. >>>>>>>>>>> >>>>>>>>>>> 1. What processes you have started on Master? >>>>>>>>>>> 2. What are the processes you have started on Slaves? >>>>>>>>>>> >>>>>>>>>>> I am missing something here, otherwise all my slave have enough >>>>>>>>>>> memory and cpus to launch the tasks I mentioned. >>>>>>>>>>> What I am missing is some configuration steps. >>>>>>>>>>> >>>>>>>>>>> Thanks & Regards, >>>>>>>>>>> Pradeep >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On 3 October 2015 at 13:14, Guangya Liu <gyliu...@gmail.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi Pradeep, >>>>>>>>>>>> >>>>>>>>>>>> I did some test with your case and found that the task can run >>>>>>>>>>>> randomly on the three slave hosts, every time may have different >>>>>>>>>>>> result. >>>>>>>>>>>> The logic is here: >>>>>>>>>>>> https://github.com/apache/mesos/blob/master/src/master/allocator/mesos/hierarchical.hpp#L1263-#L1266 >>>>>>>>>>>> The allocator will help random shuffle the slaves every time >>>>>>>>>>>> when allocate resources for offers. >>>>>>>>>>>> >>>>>>>>>>>> I see that every of your task need the minimum resources as " >>>>>>>>>>>> resources="cpus(*):3;mem(*):2560", can you help check if all >>>>>>>>>>>> of your slaves have enough resources? If you want your task run on >>>>>>>>>>>> other >>>>>>>>>>>> slaves, then those slaves need to have at least 3 cpus and 2550M >>>>>>>>>>>> memory. >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> >>>>>>>>>>>> On Fri, Oct 2, 2015 at 9:26 PM, Pradeep Kiruvale < >>>>>>>>>>>> pradeepkiruv...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi Ondrej, >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks for your reply >>>>>>>>>>>>> >>>>>>>>>>>>> I did solve that issue, yes you are right there was an issue >>>>>>>>>>>>> with slave IP address setting. >>>>>>>>>>>>> >>>>>>>>>>>>> Now I am facing issue with the scheduling the tasks. When I >>>>>>>>>>>>> try to schedule a task using >>>>>>>>>>>>> >>>>>>>>>>>>> /src/mesos-execute --master=192.168.0.102:5050 >>>>>>>>>>>>> --name="cluster-test" --command="/usr/bin/hackbench -s 4096 -l >>>>>>>>>>>>> 10845760 -g >>>>>>>>>>>>> 2 -f 2 -P" --resources="cpus(*):3;mem(*):2560" >>>>>>>>>>>>> >>>>>>>>>>>>> The tasks always get scheduled on the same node. The resources >>>>>>>>>>>>> from the other nodes are not getting used to schedule the tasks. >>>>>>>>>>>>> >>>>>>>>>>>>> I just start the mesos slaves like below >>>>>>>>>>>>> >>>>>>>>>>>>> ./bin/mesos-slave.sh --master=192.168.0.102:5050/mesos >>>>>>>>>>>>> --hostname=slave1 >>>>>>>>>>>>> >>>>>>>>>>>>> If I submit the task using the above (mesos-execute) command >>>>>>>>>>>>> from same as one of the slave it runs on that system. >>>>>>>>>>>>> >>>>>>>>>>>>> But when I submit the task from some different system. It uses >>>>>>>>>>>>> just that system and queues the tasks not runs on the other >>>>>>>>>>>>> slaves. >>>>>>>>>>>>> Some times I see the message "Failed to getgid: unknown user" >>>>>>>>>>>>> >>>>>>>>>>>>> Do I need to start some process to push the task on all the >>>>>>>>>>>>> slaves equally? Am I missing something here? >>>>>>>>>>>>> >>>>>>>>>>>>> Regards, >>>>>>>>>>>>> Pradeep >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On 2 October 2015 at 15:07, Ondrej Smola < >>>>>>>>>>>>> ondrej.sm...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Pradeep, >>>>>>>>>>>>>> >>>>>>>>>>>>>> the problem is with IP your slave advertise - mesos by >>>>>>>>>>>>>> default resolves your hostname - there are several solutions >>>>>>>>>>>>>> (let say your >>>>>>>>>>>>>> node ip is 192.168.56.128) >>>>>>>>>>>>>> >>>>>>>>>>>>>> 1) export LIBPROCESS_IP=192.168.56.128 >>>>>>>>>>>>>> 2) set mesos options - ip, hostname >>>>>>>>>>>>>> >>>>>>>>>>>>>> one way to do this is to create files >>>>>>>>>>>>>> >>>>>>>>>>>>>> echo "192.168.56.128" > /etc/mesos-slave/ip >>>>>>>>>>>>>> echo "abc.mesos.com" > /etc/mesos-slave/hostname >>>>>>>>>>>>>> >>>>>>>>>>>>>> for more configuration options see >>>>>>>>>>>>>> http://mesos.apache.org/documentation/latest/configuration >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2015-10-02 10:06 GMT+02:00 Pradeep Kiruvale < >>>>>>>>>>>>>> pradeepkiruv...@gmail.com>: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hi Guangya, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks for reply. I found one interesting log message. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 7410 master.cpp:5977] Removed slave >>>>>>>>>>>>>>> 6a11063e-b8ff-43bd-86cf-e6eef0de06fd-S52 (192.168.0.178): a new >>>>>>>>>>>>>>> slave >>>>>>>>>>>>>>> registered at the same address >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Mostly because of this issue, the systems/slave nodes are >>>>>>>>>>>>>>> getting registered and de-registered to make a room for the >>>>>>>>>>>>>>> next node. I >>>>>>>>>>>>>>> can even see this on >>>>>>>>>>>>>>> the UI interface, for some time one node got added and after >>>>>>>>>>>>>>> some time that will be replaced with the new slave node. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> The above log is followed by the below log messages. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I1002 10:01:12.753865 7416 leveldb.cpp:343] Persisting >>>>>>>>>>>>>>> action (18 bytes) to leveldb took 104089ns >>>>>>>>>>>>>>> I1002 10:01:12.753885 7416 replica.cpp:679] Persisted >>>>>>>>>>>>>>> action at 384 >>>>>>>>>>>>>>> E1002 10:01:12.753891 7417 process.cpp:1912] Failed to >>>>>>>>>>>>>>> shutdown socket with fd 15: Transport endpoint is not connected >>>>>>>>>>>>>>> I1002 10:01:12.753988 7413 master.cpp:3930] Registered >>>>>>>>>>>>>>> slave 6a11063e-b8ff-43bd-86cf-e6eef0de06fd-S62 at slave(1)@ >>>>>>>>>>>>>>> 127.0.1.1:5051 (192.168.0.116) with cpus(*):8; >>>>>>>>>>>>>>> mem(*):14930; disk(*):218578; ports(*):[31000-32000] >>>>>>>>>>>>>>> I1002 10:01:12.754065 7413 master.cpp:1080] Slave >>>>>>>>>>>>>>> 6a11063e-b8ff-43bd-86cf-e6eef0de06fd-S62 at slave(1)@ >>>>>>>>>>>>>>> 127.0.1.1:5051 (192.168.0.116) disconnected >>>>>>>>>>>>>>> I1002 10:01:12.754072 7416 hierarchical.hpp:675] Added >>>>>>>>>>>>>>> slave 6a11063e-b8ff-43bd-86cf-e6eef0de06fd-S62 (192.168.0.116) >>>>>>>>>>>>>>> with >>>>>>>>>>>>>>> cpus(*):8; mem(*):14930; disk(*):218578; ports(*):[31000-32000] >>>>>>>>>>>>>>> (allocated: >>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>> I1002 10:01:12.754084 7413 master.cpp:2534] Disconnecting >>>>>>>>>>>>>>> slave 6a11063e-b8ff-43bd-86cf-e6eef0de06fd-S62 at slave(1)@ >>>>>>>>>>>>>>> 127.0.1.1:5051 (192.168.0.116) >>>>>>>>>>>>>>> E1002 10:01:12.754118 7417 process.cpp:1912] Failed to >>>>>>>>>>>>>>> shutdown socket with fd 16: Transport endpoint is not connected >>>>>>>>>>>>>>> I1002 10:01:12.754132 7413 master.cpp:2553] Deactivating >>>>>>>>>>>>>>> slave 6a11063e-b8ff-43bd-86cf-e6eef0de06fd-S62 at slave(1)@ >>>>>>>>>>>>>>> 127.0.1.1:5051 (192.168.0.116) >>>>>>>>>>>>>>> I1002 10:01:12.754237 7416 hierarchical.hpp:768] Slave >>>>>>>>>>>>>>> 6a11063e-b8ff-43bd-86cf-e6eef0de06fd-S62 deactivated >>>>>>>>>>>>>>> I1002 10:01:12.754240 7413 replica.cpp:658] Replica >>>>>>>>>>>>>>> received learned notice for position 384 >>>>>>>>>>>>>>> I1002 10:01:12.754360 7413 leveldb.cpp:343] Persisting >>>>>>>>>>>>>>> action (20 bytes) to leveldb took 95171ns >>>>>>>>>>>>>>> I1002 10:01:12.754395 7413 leveldb.cpp:401] Deleting ~2 >>>>>>>>>>>>>>> keys from leveldb took 20333ns >>>>>>>>>>>>>>> I1002 10:01:12.754406 7413 replica.cpp:679] Persisted >>>>>>>>>>>>>>> action at 384 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>> Pradeep >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On 2 October 2015 at 02:35, Guangya Liu <gyliu...@gmail.com> >>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Pradeep, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Please check some of my questions in line. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Guangya >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Fri, Oct 2, 2015 at 12:55 AM, Pradeep Kiruvale < >>>>>>>>>>>>>>>> pradeepkiruv...@gmail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi All, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I am new to Mesos. I have set up a Mesos cluster with 1 >>>>>>>>>>>>>>>>> Master and 3 Slaves. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> One slave runs on the Master Node itself and Other slaves >>>>>>>>>>>>>>>>> run on different nodes. Here node means the physical boxes. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I tried running the tasks by configuring one Node cluster. >>>>>>>>>>>>>>>>> Tested the task scheduling using mesos-execute, works fine. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> When I configure three Node cluster (1master and 3 slaves) >>>>>>>>>>>>>>>>> and try to see the resources on the master (in GUI) only the >>>>>>>>>>>>>>>>> Master node >>>>>>>>>>>>>>>>> resources are visible. >>>>>>>>>>>>>>>>> The other nodes resources are not visible. Some times >>>>>>>>>>>>>>>>> visible but in a de-actived state. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Can you please append some logs from mesos-slave and >>>>>>>>>>>>>>>> mesos-master? There should be some logs in either master or >>>>>>>>>>>>>>>> slave telling >>>>>>>>>>>>>>>> you what is wrong. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> *Please let me know what could be the reason. All the >>>>>>>>>>>>>>>>> nodes are in the same network. * >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> When I try to schedule a task using >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> /src/mesos-execute --master=192.168.0.102:5050 >>>>>>>>>>>>>>>>> --name="cluster-test" --command="/usr/bin/hackbench -s 4096 >>>>>>>>>>>>>>>>> -l 10845760 -g >>>>>>>>>>>>>>>>> 2 -f 2 -P" --resources="cpus(*):3;mem(*):2560" >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> The tasks always get scheduled on the same node. The >>>>>>>>>>>>>>>>> resources from the other nodes are not getting used to >>>>>>>>>>>>>>>>> schedule the tasks. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Based on your previous question, there is only one node in >>>>>>>>>>>>>>>> your cluster, that's why other nodes are not available. We >>>>>>>>>>>>>>>> need first >>>>>>>>>>>>>>>> identify what is wrong with other three nodes first. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I*s it required to register the frameworks from every >>>>>>>>>>>>>>>>> slave node on the Master?* >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> It is not required. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> *I have configured this cluster using the git-hub code.* >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks & Regards, >>>>>>>>>>>>>>>>> Pradeep >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >