Inspecting the log now.

Reverse DNS seems to work if I understand you correctly:

$ host 174.129.76.5
5.76.129.174.in-addr.arpa domain name pointer
ec2-174-129-76-5.compute-1.amazonaws.com.

I am about to head out the office, so will try from home over the weekend.

Cheers!
Tim


On Fri, Jan 27, 2012 at 4:49 PM, Andrei Savu <[email protected]> wrote:
> Anything else relevant in the log files? Can you try a reverse DNS
> query using the terminal? (maybe your ISP is filtering them) Can you
> try the same setup from a VM inside the Amazon Cloud?
>
> PS: I am starting a cluster now from my machine to check that
> everything works fine.
>
> -- Andrei Savu
>
> On Fri, Jan 27, 2012 at 5:41 PM, Tim Robertson
> <[email protected]> wrote:
>> Thanks Andrei for the quick reply!
>>
>> I followed the destructions on:
>> http://code.google.com/speed/public-dns/docs/using.html
>>
>> Same result though I'm afraid.
>> Tim
>>
>> On Fri, Jan 27, 2012 at 4:11 PM, Andrei Savu <[email protected]> wrote:
>>> Hi Tim,
>>>
>>> And welcome to Apache Whirr! Let me give you some advices that could
>>> help you get this to work for you.
>>>
>>> I think the boostrap fails for you because Whirr fails at doing
>>> reverse DNS resolution for Amazon public IPs. Can you try
>>> switching to Google public dns servers? (8.8.8.8 & 8.8.4.4).
>>>
>>> PS: That article is a bit old but you've figured out all the needed changes!
>>>
>>> PS: you can also find me on IRC @ #whirr for a more interactive
>>> discussion (asavu)
>>>
>>> -- Andrei Savu / andreisavu.ro
>>>
>>> On Fri, Jan 27, 2012 at 4:57 PM, Tim Robertson
>>> <[email protected]> wrote:
>>>> Hi all,
>>>>
>>>> I am trying to follow the instructions on:
>>>>  http://www.bigfastblog.com/run-the-latest-whirr-and-deploy-hbase-in-minutes
>>>>
>>>> I took Whirr from here today (note this is different from the
>>>> instructions which seem to point at a non existing incubator path):
>>>>  http://svn.apache.org/repos/asf/whirr/trunk/
>>>>
>>>> It almost works for me, but the NameNode, JobTracker and HBase master
>>>> don't seem to start (Zookeeper does).
>>>> On the slaves I correctly have DataNode, TaskTracker and RegionServer
>>>> services running.
>>>>
>>>> I suspect something is going on with my key file or permissions, but I
>>>> am too naive to work out what is happening, so hopeful of some
>>>> guidance.
>>>>
>>>> SSH'ing onto the master and trying to start Hadoop manually by running
>>>> a start-all.sh as my user I get the following:
>>>>
>>>> tim@domU-12-31-39-0C-90-D1:~$ /usr/local/hadoop-0.20.205.0/bin/start-all.sh
>>>> chown: changing ownership of `/var/log/hadoop/logs': Operation not 
>>>> permitted
>>>> starting namenode, logging to
>>>> /var/log/hadoop/logs/hadoop-tim-namenode-domU-12-31-39-0C-90-D1.out
>>>> /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line 136:
>>>> /var/run/hadoop/hadoop-tim-namenode.pid: Permission denied
>>>> /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line 135:
>>>> /var/log/hadoop/logs/hadoop-tim-namenode-domU-12-31-39-0C-90-D1.out:
>>>> Permission denied
>>>> head: cannot open
>>>> `/var/log/hadoop/logs/hadoop-tim-namenode-domU-12-31-39-0C-90-D1.out'
>>>> for reading: No such file or directory
>>>> localhost: Warning: Permanently added 'localhost' (RSA) to the list of
>>>> known hosts.
>>>> localhost: chown: changing ownership of `/var/log/hadoop/logs':
>>>> Operation not permitted
>>>> localhost: starting datanode, logging to
>>>> /var/log/hadoop/logs/hadoop-tim-datanode-domU-12-31-39-0C-90-D1.out
>>>> localhost: /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line
>>>> 136: /var/run/hadoop/hadoop-tim-datanode.pid: Permission denied
>>>> localhost: /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line
>>>> 135: /var/log/hadoop/logs/hadoop-tim-datanode-domU-12-31-39-0C-90-D1.out:
>>>> Permission denied
>>>> localhost: head: cannot open
>>>> `/var/log/hadoop/logs/hadoop-tim-datanode-domU-12-31-39-0C-90-D1.out'
>>>> for reading: No such file or directory
>>>> localhost: chown: changing ownership of `/var/log/hadoop/logs':
>>>> Operation not permitted
>>>> localhost: starting secondarynamenode, logging to
>>>> /var/log/hadoop/logs/hadoop-tim-secondarynamenode-domU-12-31-39-0C-90-D1.out
>>>> localhost: /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line
>>>> 136: /var/run/hadoop/hadoop-tim-secondarynamenode.pid: Permission
>>>> denied
>>>> localhost: /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line
>>>> 135: 
>>>> /var/log/hadoop/logs/hadoop-tim-secondarynamenode-domU-12-31-39-0C-90-D1.out:
>>>> Permission denied
>>>> localhost: head: cannot open
>>>> `/var/log/hadoop/logs/hadoop-tim-secondarynamenode-domU-12-31-39-0C-90-D1.out'
>>>> for reading: No such file or directory
>>>> chown: changing ownership of `/var/log/hadoop/logs': Operation not 
>>>> permitted
>>>> starting jobtracker, logging to
>>>> /var/log/hadoop/logs/hadoop-tim-jobtracker-domU-12-31-39-0C-90-D1.out
>>>> /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line 136:
>>>> /var/run/hadoop/hadoop-tim-jobtracker.pid: Permission denied
>>>> /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line 135:
>>>> /var/log/hadoop/logs/hadoop-tim-jobtracker-domU-12-31-39-0C-90-D1.out:
>>>> Permission denied
>>>> head: cannot open
>>>> `/var/log/hadoop/logs/hadoop-tim-jobtracker-domU-12-31-39-0C-90-D1.out'
>>>> for reading: No such file or directory
>>>> localhost: chown: changing ownership of `/var/log/hadoop/logs':
>>>> Operation not permitted
>>>> localhost: starting tasktracker, logging to
>>>> /var/log/hadoop/logs/hadoop-tim-tasktracker-domU-12-31-39-0C-90-D1.out
>>>> localhost: /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line
>>>> 136: /var/run/hadoop/hadoop-tim-tasktracker.pid: Permission denied
>>>> localhost: /usr/local/hadoop-0.20.205.0/bin/hadoop-daemon.sh: line
>>>> 135: 
>>>> /var/log/hadoop/logs/hadoop-tim-tasktracker-domU-12-31-39-0C-90-D1.out:
>>>> Permission denied
>>>> localhost: head: cannot open
>>>> `/var/log/hadoop/logs/hadoop-tim-tasktracker-domU-12-31-39-0C-90-D1.out'
>>>> for reading: No such file or directory
>>>>
>>>> Running the same as sudo I get:
>>>>
>>>> tim@domU-12-31-39-0C-90-D1:~$ sudo
>>>> /usr/local/hadoop-0.20.205.0/bin/start-all.sh
>>>> starting namenode, logging to
>>>> /var/log/hadoop/logs/hadoop-root-namenode-domU-12-31-39-0C-90-D1.out
>>>> Warning: $HADOOP_HOME is deprecated.
>>>>
>>>> Error: JAVA_HOME is not set.
>>>> localhost: Warning: Permanently added 'localhost' (RSA) to the list of
>>>> known hosts.
>>>> localhost: Permission denied (publickey).
>>>> localhost: Permission denied (publickey).
>>>> starting jobtracker, logging to
>>>> /var/log/hadoop/logs/hadoop-root-jobtracker-domU-12-31-39-0C-90-D1.out
>>>> Warning: $HADOOP_HOME is deprecated.
>>>>
>>>> Error: JAVA_HOME is not set.
>>>> localhost: Permission denied (publickey).
>>>>
>>>> Below is the full log of the startup.
>>>>
>>>> Any help greatly appreciated!
>>>> Tim
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> $ bin/whirr launch-cluster --config hbase-ec2.properties
>>>> Bootstrapping cluster
>>>> Configuring template
>>>> Configuring template
>>>> Starting 1 node(s) with roles [hadoop-datanode, hadoop-tasktracker,
>>>> hbase-regionserver]
>>>> Starting 1 node(s) with roles [zookeeper, hadoop-namenode,
>>>> hadoop-jobtracker, hbase-master]
>>>> Nodes started: [[id=us-east-1/i-45c80920, providerId=i-45c80920,
>>>> group=hbase, name=hbase-45c80920, location=[id=us-east-1b, scope=ZONE,
>>>> description=us-east-1b, parent=us-east-1, iso3166Codes=[US-VA],
>>>> metadata={}], uri=null, imageId=us-east-1/ami-da0cf8b3, os=[name=null,
>>>> family=ubuntu, version=10.04, arch=paravirtual, is64Bit=true,
>>>> description=ubuntu-images-us/ubuntu-lucid-10.04-amd64-server-20101020.manifest.xml],
>>>> state=RUNNING, loginPort=22, hostname=domU-12-31-39-0E-49-D1,
>>>> privateAddresses=[10.192.74.31], publicAddresses=[107.22.17.222],
>>>> hardware=[id=c1.xlarge, providerId=c1.xlarge, name=null,
>>>> processors=[[cores=8.0, speed=2.5]], ram=7168, volumes=[[id=null,
>>>> type=LOCAL, size=10.0, device=/dev/sda1, durable=false,
>>>> isBootDevice=true], [id=null, type=LOCAL, size=420.0, device=/dev/sdb,
>>>> durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sdc, durable=false, isBootDevice=false], [id=null,
>>>> type=LOCAL, size=420.0, device=/dev/sdd, durable=false,
>>>> isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sde, durable=false, isBootDevice=false]],
>>>> supportsImage=And(ALWAYS_TRUE,Or(isWindows(),requiresVirtualizationType(paravirtual)),ALWAYS_TRUE,is64Bit()),
>>>> tags=[]], loginUser=ubuntu, userMetadata={Name=hbase-45c80920},
>>>> tags=[]]]
>>>> Nodes started: [[id=us-east-1/i-7bc8091e, providerId=i-7bc8091e,
>>>> group=hbase, name=hbase-7bc8091e, location=[id=us-east-1b, scope=ZONE,
>>>> description=us-east-1b, parent=us-east-1, iso3166Codes=[US-VA],
>>>> metadata={}], uri=null, imageId=us-east-1/ami-da0cf8b3, os=[name=null,
>>>> family=ubuntu, version=10.04, arch=paravirtual, is64Bit=true,
>>>> description=ubuntu-images-us/ubuntu-lucid-10.04-amd64-server-20101020.manifest.xml],
>>>> state=RUNNING, loginPort=22, hostname=domU-12-31-39-0C-90-D1,
>>>> privateAddresses=[10.215.147.31], publicAddresses=[75.101.188.54],
>>>> hardware=[id=c1.xlarge, providerId=c1.xlarge, name=null,
>>>> processors=[[cores=8.0, speed=2.5]], ram=7168, volumes=[[id=null,
>>>> type=LOCAL, size=10.0, device=/dev/sda1, durable=false,
>>>> isBootDevice=true], [id=null, type=LOCAL, size=420.0, device=/dev/sdb,
>>>> durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sdc, durable=false, isBootDevice=false], [id=null,
>>>> type=LOCAL, size=420.0, device=/dev/sdd, durable=false,
>>>> isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sde, durable=false, isBootDevice=false]],
>>>> supportsImage=And(ALWAYS_TRUE,Or(isWindows(),requiresVirtualizationType(paravirtual)),ALWAYS_TRUE,is64Bit()),
>>>> tags=[]], loginUser=ubuntu, userMetadata={Name=hbase-7bc8091e},
>>>> tags=[]]]
>>>> Wrote instances file /Users/tim/.whirr/hbase/instances
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports [2181]
>>>> for [192.38.28.12/32]
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports
>>>> [50070] for [192.38.28.12/32]
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports [8020,
>>>> 8021] for [75.101.188.54/32]
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports
>>>> [50030] for [192.38.28.12/32]
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports [8021]
>>>> for [75.101.188.54/32]
>>>> The permission '75.101.188.54/32-1-8021-8021' has already been
>>>> authorized on the specified group
>>>> Authorizing firewall
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports
>>>> [60010, 60000] for [192.38.28.12/32]
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports
>>>> [50030] for [192.38.28.12/32]
>>>> The permission '192.38.28.12/32-1-50030-50030' has already been
>>>> authorized on the specified group
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports [8021]
>>>> for [75.101.188.54/32]
>>>> The permission '75.101.188.54/32-1-8021-8021' has already been
>>>> authorized on the specified group
>>>> Authorizing firewall ingress to [us-east-1/i-7bc8091e] on ports
>>>> [60030, 60020] for [192.38.28.12/32]
>>>> Starting to run scripts on cluster for phase configure on instances:
>>>> us-east-1/i-45c80920
>>>> Starting to run scripts on cluster for phase configure on instances:
>>>> us-east-1/i-7bc8091e
>>>> Running configure phase script on: us-east-1/i-45c80920
>>>> Running configure phase script on: us-east-1/i-7bc8091e
>>>> configure phase script run completed on: us-east-1/i-45c80920
>>>> Successfully executed configure script: [output=starting datanode,
>>>> logging to 
>>>> /var/log/hadoop/logs/hadoop-hadoop-datanode-domU-12-31-39-0E-49-D1.out
>>>> Warning: $HADOOP_HOME is deprecated.
>>>>
>>>> No directory, logging in with HOME=/
>>>> starting tasktracker, logging to
>>>> /var/log/hadoop/logs/hadoop-hadoop-tasktracker-domU-12-31-39-0E-49-D1.out
>>>> Warning: $HADOOP_HOME is deprecated.
>>>>
>>>> No directory, logging in with HOME=/
>>>> starting regionserver, logging to
>>>> /var/log/hbase/logs/hbase-hadoop-regionserver-domU-12-31-39-0E-49-D1.out
>>>> No directory, logging in with HOME=/
>>>> , error=, exitCode=0]
>>>> configure phase script run completed on: us-east-1/i-7bc8091e
>>>> Successfully executed configure script: [output=No directory, logging
>>>> in with HOME=/
>>>> No directory, logging in with HOME=/
>>>> No directory, logging in with HOME=/
>>>> No directory, logging in with HOME=/
>>>> starting jobtracker, logging to
>>>> /var/log/hadoop/logs/hadoop-hadoop-jobtracker-domU-12-31-39-0C-90-D1.out
>>>> Warning: $HADOOP_HOME is deprecated.
>>>>
>>>> No directory, logging in with HOME=/
>>>> starting master, logging to
>>>> /var/log/hbase/logs/hbase-hadoop-master-domU-12-31-39-0C-90-D1.out
>>>> No directory, logging in with HOME=/
>>>> , error=12/01/27 14:43:35 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 1 time(s).
>>>> 12/01/27 14:43:36 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 2 time(s).
>>>> 12/01/27 14:43:37 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 3 time(s).
>>>> 12/01/27 14:43:38 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 4 time(s).
>>>> 12/01/27 14:43:39 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 5 time(s).
>>>> 12/01/27 14:43:40 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 6 time(s).
>>>> 12/01/27 14:43:41 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 7 time(s).
>>>> 12/01/27 14:43:42 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 8 time(s).
>>>> 12/01/27 14:43:43 INFO ipc.Client: Retrying connect to server:
>>>> 75.101.188.54/75.101.188.54:8020. Already tried 9 time(s).
>>>> Bad connection to FS. command aborted. exception: Call to
>>>> 75.101.188.54/75.101.188.54:8020 failed on connection exception:
>>>> java.net.ConnectException: Connection refused
>>>> , exitCode=0]
>>>> Finished running configure phase scripts on all cluster instances
>>>> Completed configuration of hbase
>>>> Hosts: 75.101.188.54:2181
>>>> Completed configuration of hbase role hadoop-namenode
>>>> Namenode web UI available at http://75.101.188.54:50070
>>>> Wrote Hadoop site file /Users/tim/.whirr/hbase/hadoop-site.xml
>>>> Wrote Hadoop proxy script /Users/tim/.whirr/hbase/hadoop-proxy.sh
>>>> Completed configuration of hbase role hadoop-jobtracker
>>>> Jobtracker web UI available at http://75.101.188.54:50030
>>>> Completed configuration of hbase
>>>> Web UI available at http://75.101.188.54
>>>> Wrote HBase site file /Users/tim/.whirr/hbase/hbase-site.xml
>>>> Wrote HBase proxy script /Users/tim/.whirr/hbase/hbase-proxy.sh
>>>> Completed configuration of hbase role hadoop-datanode
>>>> Completed configuration of hbase role hadoop-tasktracker
>>>> Starting to run scripts on cluster for phase start on instances:
>>>> us-east-1/i-7bc8091e
>>>> Running start phase script on: us-east-1/i-7bc8091e
>>>> start phase script run completed on: us-east-1/i-7bc8091e
>>>> Successfully executed start script: [output=, error=, exitCode=0]
>>>> Finished running start phase scripts on all cluster instances
>>>> Started cluster of 2 instances
>>>> Cluster{instances=[Instance{roles=[hadoop-datanode,
>>>> hadoop-tasktracker, hbase-regionserver], publicIp=107.22.17.222,
>>>> privateIp=10.192.74.31, id=us-east-1/i-45c80920,
>>>> nodeMetadata=[id=us-east-1/i-45c80920, providerId=i-45c80920,
>>>> group=hbase, name=hbase-45c80920, location=[id=us-east-1b, scope=ZONE,
>>>> description=us-east-1b, parent=us-east-1, iso3166Codes=[US-VA],
>>>> metadata={}], uri=null, imageId=us-east-1/ami-da0cf8b3, os=[name=null,
>>>> family=ubuntu, version=10.04, arch=paravirtual, is64Bit=true,
>>>> description=ubuntu-images-us/ubuntu-lucid-10.04-amd64-server-20101020.manifest.xml],
>>>> state=RUNNING, loginPort=22, hostname=domU-12-31-39-0E-49-D1,
>>>> privateAddresses=[10.192.74.31], publicAddresses=[107.22.17.222],
>>>> hardware=[id=c1.xlarge, providerId=c1.xlarge, name=null,
>>>> processors=[[cores=8.0, speed=2.5]], ram=7168, volumes=[[id=null,
>>>> type=LOCAL, size=10.0, device=/dev/sda1, durable=false,
>>>> isBootDevice=true], [id=null, type=LOCAL, size=420.0, device=/dev/sdb,
>>>> durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sdc, durable=false, isBootDevice=false], [id=null,
>>>> type=LOCAL, size=420.0, device=/dev/sdd, durable=false,
>>>> isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sde, durable=false, isBootDevice=false]],
>>>> supportsImage=And(ALWAYS_TRUE,Or(isWindows(),requiresVirtualizationType(paravirtual)),ALWAYS_TRUE,is64Bit()),
>>>> tags=[]], loginUser=ubuntu, userMetadata={Name=hbase-45c80920},
>>>> tags=[]]}, Instance{roles=[zookeeper, hadoop-namenode,
>>>> hadoop-jobtracker, hbase-master], publicIp=75.101.188.54,
>>>> privateIp=10.215.147.31, id=us-east-1/i-7bc8091e,
>>>> nodeMetadata=[id=us-east-1/i-7bc8091e, providerId=i-7bc8091e,
>>>> group=hbase, name=hbase-7bc8091e, location=[id=us-east-1b, scope=ZONE,
>>>> description=us-east-1b, parent=us-east-1, iso3166Codes=[US-VA],
>>>> metadata={}], uri=null, imageId=us-east-1/ami-da0cf8b3, os=[name=null,
>>>> family=ubuntu, version=10.04, arch=paravirtual, is64Bit=true,
>>>> description=ubuntu-images-us/ubuntu-lucid-10.04-amd64-server-20101020.manifest.xml],
>>>> state=RUNNING, loginPort=22, hostname=domU-12-31-39-0C-90-D1,
>>>> privateAddresses=[10.215.147.31], publicAddresses=[75.101.188.54],
>>>> hardware=[id=c1.xlarge, providerId=c1.xlarge, name=null,
>>>> processors=[[cores=8.0, speed=2.5]], ram=7168, volumes=[[id=null,
>>>> type=LOCAL, size=10.0, device=/dev/sda1, durable=false,
>>>> isBootDevice=true], [id=null, type=LOCAL, size=420.0, device=/dev/sdb,
>>>> durable=false, isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sdc, durable=false, isBootDevice=false], [id=null,
>>>> type=LOCAL, size=420.0, device=/dev/sdd, durable=false,
>>>> isBootDevice=false], [id=null, type=LOCAL, size=420.0,
>>>> device=/dev/sde, durable=false, isBootDevice=false]],
>>>> supportsImage=And(ALWAYS_TRUE,Or(isWindows(),requiresVirtualizationType(paravirtual)),ALWAYS_TRUE,is64Bit()),
>>>> tags=[]], loginUser=ubuntu, userMetadata={Name=hbase-7bc8091e},
>>>> tags=[]]}], configuration={hbase.zookeeper.quorum=75.101.188.54:2181,
>>>> hadoop.rpc.socket.factory.class.default=org.apache.hadoop.net.SocksSocketFactory,
>>>> hadoop.socks.server=localhost:6666,
>>>> hbase.zookeeper.property.clientPort=2181}}
>>>> You can log into instances using the following ssh commands:
>>>> 'ssh -i /Users/tim/.ssh/id_rsa -o "UserKnownHostsFile /dev/null" -o
>>>> StrictHostKeyChecking=no [email protected]'
>>>> 'ssh -i /Users/tim/.ssh/id_rsa -o "UserKnownHostsFile /dev/null" -o
>>>> StrictHostKeyChecking=no [email protected]'

Reply via email to