Yes i have run those commands but when i restart the slave i get always the
same amount of resources, the only thing that i can notice is a wrong
hostname after the modifications. i don't know if this could cause the
problem

2016-04-08 20:32 GMT+02:00 Vinod Kone <[email protected]>:

> Have you tried the remedy steps included in the error message?
>
> ------------------------------------------------------------
>
> To remedy this do as follows:
>
> Step 1: rm -f /tmp/mesos/meta/slaves/latest
>
>         This ensures slave doesn't recover old live executors.
>
> Step 2: Restart the slave.
>
>
>
> On Fri, Apr 8, 2016 at 11:29 AM, Stefano Bianchi <[email protected]>
> wrote:
>
>> Hi i would like to enter in this mailing list.
>> i'm currently doing my Master Thesis on Mesos and Calico.
>> I'm working at INFN, institute of nuclear physics. The goal of the thesis
>> is to build a PaaS where mesos is the scheduler and Calico must allow the
>> interconnection between multiple datacenters linked to the CERN.
>>
>> I'm exploiting an IaaS based on Openstack, here i have created 6 Virtual
>> Machines, 3 Masters and 3 Slaves, on one slave is running Mesos-DNS from
>> Marathon.
>> All is perfectly working, since i am on another network i changed
>> correctly the hostnames such that on mesos are resolvable and i tried to
>> run from marathon a simple http server which is scalable on all my machine.
>> So all is fine and working.
>>
>> The only thing that i don't like is that each 3 slaves have 1 CPU 10 GB
>> of disk memory and 2GB of RAM, but mesos currently show for each one only 5
>> GB of disk memory and 900MB of RAM.
>> So checking in documentation i found the command to manage the resources.
>> I stopped Slave1, for instance, and i have run this command:
>>
>> mesos-slave --master=MASTER_ADDRESS:5050
>> --resources='cpu:1;mem:2000;disk:9000'
>>
>> where i want set 2000 GB of RAM and 9000GB of disk memory.
>>  The output is the following:
>>
>> I0408 15:11:00.915324  7892 main.cpp:215] Build: 2016-03-10 20:32:58 by root
>>
>> I0408 15:11:00.915436  7892 main.cpp:217] Version: 0.27.2
>>
>> I0408 15:11:00.915448  7892 main.cpp:220] Git tag: 0.27.2
>>
>> I0408 15:11:00.915459  7892 main.cpp:224] Git SHA: 
>> 3c9ec4a0f34420b7803848af597de00fedefe0e2
>>
>> I0408 15:11:00.923334  7892 systemd.cpp:236] systemd version `219` detected
>>
>> I0408 15:11:00.923384  7892 main.cpp:232] Inializing systemd state
>>
>> I0408 15:11:00.950050  7892 systemd.cpp:324] Started systemd slice 
>> `mesos_executors.slice`
>>
>> I0408 15:11:00.951529  7892 containerizer.cpp:143] Using isolation: 
>> posix/cpu,posix/mem,filesystem/posix
>>
>> I0408 15:11:00.963232  7892 linux_launcher.cpp:101] Using 
>> /sys/fs/cgroup/freezer as the freezer hierarchy for the Linux launcher
>>
>> I0408 15:11:00.965541  7892 main.cpp:320] Starting Mesos slave
>>
>> I0408 15:11:00.966008  7892 slave.cpp:192] Slave started on 
>> 1)@192.168.100.56:5051
>>
>> I0408 15:11:00.966023  7892 slave.cpp:193] Flags at startup: 
>> --appc_store_dir="/tmp/mesos/store/appc" --authenticatee="crammd5" 
>> --cgroups_cpu_enable_pids_and_tids_count="false" 
>> --cgroups_enable_cfs="false" --cgroups_hierarchy="/sys/fs/cgroup" 
>> --cgroups_limit_swap="false" --cgroups_root="mesos" 
>> --container_disk_watch_interval="15secs" --containerizers="mesos" 
>> --default_role="*" --disk_watch_interval="1mins" --docker="docker" 
>> --docker_auth_server="https://auth.docker.io"; --docker_kill_orphans="true" 
>> --docker_puller_timeout="60" 
>> --docker_registry="https://registry-1.docker.io"; 
>> --docker_remove_delay="6hrs" --docker_socket="/var/run/docker.sock" 
>> --docker_stop_timeout="0ns" --docker_store_dir="/tmp/mesos/store/docker" 
>> --enforce_container_disk_quota="false" 
>> --executor_registration_timeout="1mins" 
>> --executor_shutdown_grace_period="5secs" 
>> --fetcher_cache_dir="/tmp/mesos/fetch" --fetcher_cache_size="2GB" 
>> --frameworks_home="" --gc_delay="1weeks" --gc_disk_headroom="0.1" 
>> --hadoop_home="" --help="false" --hostname_lookup="true" 
>> --image_provisioner_backend="copy" --initialize_driver_logging="true" 
>> --isolation="posix/cpu,posix/mem" --launcher_dir="/usr/libexec/mesos" 
>> --logbufsecs="0" --logging_level="INFO" --master="192.168.100.55:5050" 
>> --oversubscribed_resources_interval="15secs" --perf_duration="10secs" 
>> --perf_interval="1mins" --port="5051" --qos_correction_interval_min="0ns" 
>> --quiet="false" --recover="reconnect" --recovery_timeout="15mins" 
>> --registration_backoff_factor="1secs" --resources="cpu:1;mem:2000;disk:9000" 
>> --revocable_cpu_low_priority="true" --sandbox_directory="/mnt/mesos/sandbox" 
>> --strict="true" --switch_user="true" --systemd_enable_support="true" 
>> --systemd_runtime_directory="/run/systemd/system" --version="false" 
>> --work_dir="/tmp/mesos"
>>
>> I0408 15:11:00.967485  7892 slave.cpp:463] Slave resources: cpu(*):1; 
>> mem(*):2000; disk(*):9000; cpus(*):1; ports(*):[31000-32000]
>>
>> I0408 15:11:00.967547  7892 slave.cpp:471] Slave attributes: [  ]
>>
>> I0408 15:11:00.967560  7892 slave.cpp:476] Slave hostname: 
>> slave1.openstacklocal
>>
>> I0408 15:11:00.971304  7893 state.cpp:58] Recovering state from 
>> '/tmp/mesos/meta'
>>
>> *Failed to perform recovery: Incompatible slave info detected*.
>>
>> ------------------------------------------------------------
>>
>> Old slave info:
>>
>> hostname: "*slave_public_IP_i_set*"
>>
>> resources {
>>
>>   name: "cpus"
>>
>>   type: SCALAR
>>
>>   scalar {
>>
>>     value: 1
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> resources {
>>
>>   name: "mem"
>>
>>   type: SCALAR
>>
>>   scalar {
>>
>>     value: 920
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> resources {
>>
>>   name: "disk"
>>
>>   type: SCALAR
>>
>>   scalar {
>>
>>     value: 5112
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> resources {
>>
>>   name: "ports"
>>
>>   type: RANGES
>>
>>   ranges {
>>
>>     range {
>>
>>       begin: 31000
>>
>>       end: 32000
>>
>>     }
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> id {
>>
>>   value: "ad490064-1a6e-415c-8536-daef0d8e3572-S7"
>>
>> }
>>
>> checkpoint: true
>>
>> port: 5051
>>
>> ------------------------------------------------------------
>>
>> New slave info:
>>
>> hostname: "
>>
>> slave1.openstacklocal
>>
>> "
>>
>> resources {
>>
>>   name: "cpu"
>>
>>   type: SCALAR
>>
>>   scalar {
>>
>>     value: 1
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> resources {
>>
>>   name: "mem"
>>
>>   type: SCALAR
>>
>>   scalar {
>>
>>     value: 2000
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> resources {
>>
>>   name: "disk"
>>
>>   type: SCALAR
>>
>>   scalar {
>>
>>     value: 9000
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> resources {
>>
>>   name: "cpus"
>>
>>   type: SCALAR
>>
>>   scalar {
>>
>>     value: 1
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> resources {
>>
>>   name: "ports"
>>
>>   type: RANGES
>>
>>   ranges {
>>
>>     range {
>>
>>       begin: 31000
>>
>>       end: 32000
>>
>>     }
>>
>>   }
>>
>>   role: "*"
>>
>> }
>>
>> id {
>>
>>   value: "ad490064-1a6e-415c-8536-daef0d8e3572-S7"
>>
>> }
>>
>> checkpoint: true
>>
>> port: 5051
>>
>> ------------------------------------------------------------
>>
>> To remedy this do as follows:
>>
>> Step 1: rm -f /tmp/mesos/meta/slaves/latest
>>
>>         This ensures slave doesn't recover old live executors.
>>
>> Step 2: Restart the slave.
>>
>>
>>
>> I can notice two things:
>>
>>
>> 1)the message of failure;
>>
>> 2)the hostname is changed; the right one is a public IP i have set in order 
>> to resolve the hostname for mesos.
>>
>> As a consequence, when i start the slave, the resources are exaclty the 
>> same, nothing is changed.
>>
>> Can you please help me?
>>
>>
>> Thanks!
>>
>>
>>
>

Reply via email to