Sorry - the WordPress plugin is W3 Total Cache, not W3 Super Cache. I always get those names scrambled.
Jerry On Feb 8, 2011, at 10:40 AM, Jerry Seeger wrote: > I'm still fiddling with the scripts on muddledramblings.com after a redesign, > but I intend to move static resources to a cookieless domain to improve > performance. This is a petty common tactic - sort of a poor man's CDN. The > key is that I can decide to do this. (Yes, I could rearrange my site and use > www.muddledramblings.com and cookieless.muddledramblings.com, but you're > making me do things a different way to support one Web browser.) > > (On a side note, muddledramblings.com's biggest performance problem right now > is the host. Don't use iPage. </rant>) > > Keep in mind that scripts not executing when expected can totally break a > site, not just make it less pleasant. A script that generates content must be > executed in a predictable fashion no matter where it came from. Long ago I > had a moon phase widget that generated content, and raised hell on browsers > that did not block correctly when the script loaded. (I once had a widget > with a script that generated a script. The results were... inconsistent.) > These days all browsers block correctly and the Web is a better place for it. > > I can't see telling Web designers, "If your script uses document.write, it > must come from the same domain or a known whitelist." (And let's hope the > latency of the whitelist server is really low.) I can't see telling Joe > Blogger why the visitor counter in his sidebar now writes the number at the > bottom of the page. > > The WordPress plugin W3 Super Cache includes features to automate moving > static content (including scripts) to a separate, cookieless domain. A lot of > people use the plugin, but I can't speak to how many use the pseudo-cdn > feature. My guess is not that many, but the ones who do will expect their > scripts to execute where encountered, before the rest of the page loads, as > mandated by the standards. > > The Web designer can already cause javascripts to load after the rest of the > page (the above plugin automates this as well). Were I to run ads, you can > bet that those scripts would not be loaded in the header (well, if I weren't > lazy you could bet it). If I'm not already loading Google analytics late, > it's because I haven't finished getting my script strategy finalized. > > While I would certainly like to see an automated mechanism for setting > external resource priority, allowing me to continue in my lazy ways and not > pay a performance price, (and make the Web more responsive in general, since > most of us are lazy), simple domain check is not adequate when it comes to > scripts. I wish I could think of an automated way to augment using the > domain, but all my ideas require knowing what's in the script ahead of time > (scripts that only define event handlers, for instance). > > Jerry Seeger > > On Feb 8, 2011, at 9:24 AM, Silvio Ventres wrote: > >> Do you have any example of scripts or css that are externally sourced, >> and where developer cares to reasonably optimize the web page? >> The main use case of such external scripts currently is ads and >> statistics gatherers for analysis. This, arguably, is not critical >> content that the user is interested in. >> >> If your argument is indeed "Web developer should have control", then, >> when you have no choice but including external scripts (ads, f.e.), >> you would probably hate those to break the latency of your website. >> If you are talking about the http://muddledramblings.com/ website, for >> example, you can clearly see that most scripts there are >> domain-internal. >> Do you deem your user experience more or less important than Google >> Analytics capability ? If Google Analytics hangs, for example, for 4 >> seconds, would you like the user to wait, or start reading while it >> loads? >> >> A change to HTML standard might be a good idea, though the problem >> here is that there are millions of pages on the 'net already, and the >> developers won't suddenly start changing them. >> >> This heuristic will allow the users to view 90% of the current Web >> more interactively. >> Keep in mind that at least 38% of all statistics is taken out of thin >> air :), but, really, please, show at least two pages which this >> heuristic will NOT work on. >> >> -- >> silvio >> >> On Tue, Feb 8, 2011 at 6:52 PM, Jerry Seeger <vikin...@mac.com> wrote: >>> My argument is less "it's the Web developer's fault" than it is "the Web >>> developer should have control." I am hardly a sophisticated Web developer >>> but I have javascript from a different domain that must be loaded first >>> and I have Google analytics, which I should load after the rest of the page >>> (though to be honest I'm not sure I do after my redesign... hm). While I >>> would love it if there were standardized rules for which scripts would be >>> loaded synchronously and which wouldn't, I would hate it if one browser >>> required me to move my scripts to a different domain. >>> >>> Having said all that, I hate it when I have to wait for a resource out >>> outside of my control, so I'd love to see a solution to this. If there were >>> a more reliable way than simple domain checking to prioritize content, that >>> would be fantastic. I think ideally this is something for the standards >>> board - perhaps an extension of the script and link tags to specify a >>> priority, or something like that. >>> >>> Jerry >>> >>> >>> On Feb 8, 2011, at 2:23 AM, Silvio Ventres wrote: >>> >>>> This argument - "web developer is to blame for choosing a slow >>>> ad/tracking/etc server" - is incorrect. >>>> Web developers in general do not have any control over the ad provider >>>> or, frankly, any other type of external functionality provider. >>>> Google Analytics being a good point in case, you would not want most >>>> of the world's web pages to suddenly hang if something happens inside >>>> Google. >>>> >>>> The web browser should clearly prioritize developer-controllable >>>> resources over ones that are beyond web developer's control. >>>> Also, as an application run by the user and not by the developer, the >>>> browser should arguably prioritize actual content against >>>> pseudo-content which purpose is functionality that is not visibile to >>>> the actual user, such as ad/tracker scripts. This actual content has >>>> higher probability to be important when sourced from the >>>> domain/subdomain of the webpage itself, based on current trends. >>>> >>>> Domain check is a reasonable approximation that fits both purposes. >>>> >>>> -- >>>> silvio >>>> >>>> >>>> On Tue, Feb 8, 2011 at 5:13 AM, Jerry Seeger <vikin...@mac.com> wrote: >>>>> I'm reasonably sure that javascript in the header must be loaded >>>>> synchronously, as it might affect the rest of the load. This is why tools >>>>> like YSlow advise Web designers to move javascript loads that are not >>>>> needed for rendering until after the rest of the page loads. >>>>> >>>>> Blocking on loading the css is less clear-cut, as in some cases it could >>>>> mean several seconds of ugly page. I don't know if it's right or wrong, >>>>> but a lot of pages out there rely on the CSS being loaded before the page >>>>> starts to render to avoid terrible layout and the appearance of items >>>>> meant to be hidden for the seconds it takes the css to load. >>>>> >>>>> In general, while things could certainly be improved, it's up to the >>>>> owner of the page to not rely on a a slow ad server, or build the page so >>>>> the ads load after the primary content. >>>>> >>>>> Jerry Seeger >>>>> >>>>> >>>>> On Feb 7, 2011, at 5:47 PM, Silvio Ventres wrote: >>>>> >>>>>> IE/Opera are delaying only for 4 seconds, same as Mobile Safari >>>>>> The reason looks to be the url for the script/css. >>>>>> If the url is the same twice, Chrome/Firefox serializes the requests, >>>>>> while IE/Opera/MobileSafari launches both requests simultaneously. >>>>>> >>>>>> Of course, requesting simultaneously doesn't fix anything, as you can >>>>>> see by trying a link-stuffed version at >>>>>> http://solid.eqoppa.com/testlag2.html >>>>>> >>>>>> This one has 45 css and 38 javascript links. It hangs all browsers >>>>>> nicely. >>>>>> The main point here is that it might be acceptable if it's coming from >>>>>> the webpage domain itself. >>>>>> But the links are coming from a completely different place. >>>>>> >>>>>> This is exactly what makes browsing pages with any third-party >>>>>> analytics, tracking or ad addons so slow and frustrating. >>>>>> Fixing priorities in subresource download should make experience >>>>>> considerably more interactive and fun. >>>>>> >>>>>> -- >>>>>> silvio >>>>> >>>>> >>> >>> >> _______________________________________________ >> webkit-dev mailing list >> webkit-dev@lists.webkit.org >> http://lists.webkit.org/mailman/listinfo.cgi/webkit-dev > > _______________________________________________ > webkit-dev mailing list > webkit-dev@lists.webkit.org > http://lists.webkit.org/mailman/listinfo.cgi/webkit-dev
_______________________________________________ webkit-dev mailing list webkit-dev@lists.webkit.org http://lists.webkit.org/mailman/listinfo.cgi/webkit-dev