On Tue, 13 Feb 2018 13:50:55 -0600
Andy Bach <afb...@gmail.com> wrote:

> $ wget http://scripts.sil.org/OFL
> --2018-02-13 13:42:50--  http://scripts.sil.org/OFL
> Resolving scripts.sil.org (scripts.sil.org)... 209.12.63.143
> Connecting to scripts.sil.org (scripts.sil.org)|209.12.63.143|:80...
> connected.
> HTTP request sent, awaiting response... 302 Found
> Location:
> http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL
> [following] --2018-02-13 13:42:52--
> http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL
> Reusing existing connection to scripts.sil.org:80.
> HTTP request sent, awaiting response... 302 Moved Temporarily
> Location: /cms/scripts/page.php?site_id=nrsi&id=OFL&_sc=1 [following]
> --2018-02-13 13:42:52--
> http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL&_sc=1
> Reusing existing connection to scripts.sil.org:80.
> HTTP request sent, awaiting response... 302 Moved Temporarily
> Location: /cms/scripts/page.php?site_id=nrsi&id=OFL [following]
> --2018-02-13 13:42:53--
> http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL
> Reusing existing connection to scripts.sil.org:80.
> HTTP request sent, awaiting response... 200 OK
> Length: unspecified [text/html]
> Saving to: ‘OFL’
> 
>     [
> <=>
> ] 37,439      59.6KB/s   in 0.6s
> 
> 2018-02-13 13:42:55 (59.6 KB/s) - ‘OFL’ saved [37439]
> 
> so it may not be following the 302s. I'm not sure you're using the
> correct tool here.  A little more straight forward
> 
> andy@wiwmb-md-afb-mint:~/spam$ wget http://scripts.sil.org/robots.txt
> --2018-02-13 13:47:27--  http://scripts.sil.org/robots.txt
> Resolving scripts.sil.org (scripts.sil.org)... 209.12.63.143
> Connecting to scripts.sil.org (scripts.sil.org)|209.12.63.143|:80...
> connected.
> HTTP request sent, awaiting response... 200 OK
> Length: 36 [text/plain]
> Saving to: ‘robots.txt’
> 
> 100%[========================================================================================>]
>   
> 36          --.-K/s   in 0s
> 
> 2018-02-13 13:47:27 (2.99 MB/s) - ‘robots.txt’ saved [36/36]
> 
> but
> $ is_it_up.pl
> http://scripts.sil.org/robots.txt is DOWN!!!!
> 
> You might look at more LWP tools:
> $ lwp-dump https://www.sil.org
> HTTP/1.1 403 Forbidden
> Cache-Control: max-age=10
> Connection: keep-alive
> Date: Tue, 13 Feb 2018 19:49:47 GMT
> Server: cloudflare
> Content-Type: text/html; charset=UTF-8
> Expires: Tue, 13 Feb 2018 19:49:57 GMT
> CF-RAY: 3eca501a5d569895-LAX
> Expect-CT: max-age=604800, report-uri="
> https://report-uri.cloudflare.com/cdn-cgi/beacon/expect-ct";
> Set-Cookie: __cfduid=dd8038f4f2c995fa4b4c7fa8beb2b42f31518551387;
> expires=Wed, 13-Feb-19 19:49:47 GMT; path=/; domain=.sil.org; HttpOnly
> X-Frame-Options: SAMEORIGIN
> 
> <!DOCTYPE html>
> <!--[if lt IE 7]> <html class="no-js ie6 oldie" lang="en-US">
> <![endif]--> <!--[if IE 7]>    <html class="no-js ie7 oldie"
> lang="en-US"> <![endif]--> <!--[if IE 8]>    <html class="no-js ie8
> oldie" lang="en-US"> <![endif]--> <!--[if gt IE 8]><!--> <html
> class="no-js" lang="en-US"> <!--<![endif]--> <head>
> <title>Access denied | www.sil.org used Cloudflare to restrict
> access</title>
> <meta charset="UTF-8" />
> <meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
> <meta http-equiv=...
> (+ 2770 more bytes not shown)
> 
> so it's up, but "forbidden" probably as the user agent isn't set or
> some other policy reason.
> 
> 

I tried WWW::Mechanize, and (of course) got also 403.

Really strange. 

Is there another tool I could use for checking? I mean some tool in the
Perl universe?

-- 
Manfred



> On Tue, Feb 13, 2018 at 11:33 AM, Manfred Lotz <ml_n...@posteo.de>
> wrote:
> 
> > On Tue, 13 Feb 2018 10:47:42 -0600
> > Andy Bach <afb...@gmail.com> wrote:
> >  
> > > The site doesn't like 'head' requests? get works
> > > #!/usr/bin/perl
> > >
> > > use strict;
> > > use warnings;
> > >
> > > use LWP::Simple;
> > > #  my $url="https://shlomif.github.io/";;
> > > my $url="http://www.notabug.org/";;
> > > print "$url is ", (
> > >                 (! get($url)) ?  "DOWN!!!!"
> > >                                 : "up"
> > >                 ), "\n";
> > >
> > > $ is_it_up.pl
> > > http://www.notabug.org/ is up
> > >  
> >
> > You are right.
> >
> > But am afraid this is not all of it. If I test
> > http://scripts.sil.org/OFL then I get an error but it is fine in
> > firefox.
> >
> > Very strange.
> >
> > --
> > Manfred
> >
> >
> >  
> > >
> > > On Tue, Feb 13, 2018 at 5:25 AM, Manfred Lotz <ml_n...@posteo.de>
> > > wrote:
> > >  
> > > > Hi there,
> > > > Somewhere I found an example how to check if a website is up.
> > > >
> > > > Here my sample:
> > > >
> > > > #! /usr/bin/perl
> > > >
> > > > use strict;
> > > >
> > > > use LWP::Simple;
> > > > my $url="https://notabug.org";;
> > > > if (! head($url)) {
> > > >     die "$url is DOWN!!!!"
> > > > }
> > > >
> > > > Running above code I get
> > > >   https://notabug.org is DOWN!!!! at ./check_url.pl line 8.
> > > >
> > > >
> > > > However, firefox shows the site works ok.
> > > >
> > > >
> > > > What am I doing wrong?
> > > >
> > > >
> > > > --
> > > > Thanks,
> > > > Manfred
> > > >
> > > > --
> > > > To unsubscribe, e-mail: beginners-unsubscr...@perl.org
> > > > For additional commands, e-mail: beginners-h...@perl.org
> > > > http://learn.perl.org/
> > > >
> > > >
> > > >  
> > >
> > >  
> >
> > --
> > To unsubscribe, e-mail: beginners-unsubscr...@perl.org
> > For additional commands, e-mail: beginners-h...@perl.org
> > http://learn.perl.org/
> >
> >
> >  
> 
> 

--
To unsubscribe, e-mail: beginners-unsubscr...@perl.org
For additional commands, e-mail: beginners-h...@perl.org
http://learn.perl.org/


Reply via email to