Thanks alot Dmitriy ! I thinks its enough to get me started. will email if stuck some where .
Noman On Tue, Aug 3, 2010 at 7:59 PM, <[email protected]>wrote: > Send Wikitech-l mailing list submissions to > [email protected] > > To subscribe or unsubscribe via the World Wide Web, visit > https://lists.wikimedia.org/mailman/listinfo/wikitech-l > or, via email, send a message with subject or body 'help' to > [email protected] > > You can reach the person managing the list at > [email protected] > > When replying, please edit your Subject line so it is more specific > than "Re: Contents of Wikitech-l digest..." > > > Today's Topics: > > 1. Re: wikipedia is one of the slower sites on the web (Liangent) > 2. Re: wikipedia is one of the slower sites on the web (K. Peachey) > 3. Re: chanfing main page articles from drop down. help required > (Dmitriy Sintsov) > 4. Re: wikipedia is one of the slower sites on the web > (John Vandenberg) > 5. Re: wikipedia is one of the slower sites on the web (Platonides) > 6. Re: chanfing main page articles from drop down. help required > (Platonides) > 7. Re: Debian packages (was MediaWiki version statistics) > (David Gerard) > 8. Re: Showing bytes added/removed in each edit in "View > history" and "User contributions" (Aryeh Gregor) > 9. Re: Showing bytes added/removed in each edit in "View > history" and "User contributions" (soxred93) > > > ---------------------------------------------------------------------- > > Message: 1 > Date: Tue, 3 Aug 2010 18:54:38 +0800 > From: Liangent <[email protected]> > Subject: Re: [Wikitech-l] wikipedia is one of the slower sites on the > web > To: Wikimedia developers <[email protected]> > Message-ID: > > <[email protected]<aanlktikotqe0ptb1rct3j10m%[email protected]> > > > Content-Type: text/plain; charset=UTF-8 > > On 8/3/10, Lars Aronsson <[email protected]> wrote: > > Couldn't you just tag every internal link with > > a separate class for the length of the target article, > > and then use different personal CSS to set the > > threshold? The generated page would be the same > > for all users: > > So if a page is changed, all pages linking to it need to be parsed > again. Will this cost even more? > > > > ------------------------------ > > Message: 2 > Date: Tue, 3 Aug 2010 20:55:23 +1000 > From: "K. Peachey" <[email protected]> > Subject: Re: [Wikitech-l] wikipedia is one of the slower sites on the > web > To: Wikimedia developers <[email protected]> > Message-ID: > <[email protected]> > Content-Type: text/plain; charset=UTF-8 > > Would something like what is shown below get it even further down? > > a { color: blue } > a.1_byte_article, a.2_byte_article, a.3_byte_article, > a.4_byte_article, a.5_byte_article, a.6_byte_article, > a.7_byte_article, a.8_byte_article, a.9_byte_article, > a.10_byte_article,a.11_byte_article, a.12_byte_article, > a.13_byte_article, a.14_byte_article, a.15_byte_article, > a.16_byte_article, a.17_byte_article, a.18_byte_article, > a.19_byte_article, a.20_byte_article, a.21_byte_article, > a.22_byte_article, a.23_byte_article, a.24_byte_article, > a.25_byte_article, a.26_byte_article, a.27_byte_article, > a.28_byte_article, a.29_byte_article, a.30_byte_article, > a.31_byte_article, a.32_byte_article, a.33_byte_article, > a.34_byte_article, a.35_byte_article, a.36_byte_article, > a.37_byte_article, a.38_byte_article, a.39_byte_article, > a.40_byte_article, a.41_byte_article, a.42_byte_article, > a.43_byte_article, a.44_byte_article, a.45_byte_article, > a.46_byte_article, a.47_byte_article, a.48_byte_article, > a.49_byte_article, a.50_byte_article, a.51_byte_article, > a.52_byte_article, a.53_byte_article, a.54_byte_article, > a.55_byte_article, a.56_byte_article, a.57_byte_article, > a.58_byte_article, a.59_byte_article, a.60_byte_article, > a.61_byte_article, a.62_byte_article, a.63_byte_article, > a.64_byte_article, a.65_byte_article, a.66_byte_article, > a.67_byte_article, a.68_byte_article, a.69_byte_article, > a.70_byte_article, a.71_byte_article, a.72_byte_article, > a.73_byte_article, a.74_byte_article, a.75_byte_article, > a.76_byte_article, a.77_byte_article, a.78_byte_article, > a.79_byte_article, a.80_byte_article, a.81_byte_article, > a.82_byte_article, a.83_byte_article, a.84_byte_article, > a.85_byte_article, a.86_byte_article, a.87_byte_article, > a.88_byte_article, a.89_byte_article, a.90_byte_article, > a.91_byte_article, a.92_byte_article, a.93_byte_article, > a.94_byte_article, a.95_byte_article { color: red } > > > > ------------------------------ > > Message: 3 > Date: Tue, 03 Aug 2010 15:09:24 +0400 > From: Dmitriy Sintsov <[email protected]> > Subject: Re: [Wikitech-l] chanfing main page articles from drop down. > help required > To: Wikimedia developers <[email protected]> > Message-ID: > <[email protected]> > Content-Type: text/plain; charset="us-ascii"; format="flowed" > > * Noman <[email protected]> [Tue, 3 Aug 2010 12:04:31 +0500]: > > Thanks Dmitriy, > > i'm looking for div solution. as iframe will give scrolling if content > > are > > large. which is not required. > > now i was unable to find step by step approach to develop extension. > > > > if you have ne thing / example. i'll be waiting. > > > Maybe Extension:HTMLets will suits your needs. Otherwise, you have to > study MediaWiki developers site: > > 1. Perhaps one would setup a parser xml tag hook to generate proper > form/select/option and four corresponding div's code: > http://www.mediawiki.org/wiki/Manual:Tag_extensions > > from xml tag attributes one would generate full html which is required > to select titles and to place content of these into div's. > > 2. Perhaps one would use API to retrieve the pages whose title are taken > from option.value via javascript, then place these into div.innerHTML, > again via the Javascript: > http://www.mediawiki.org/wiki/API:Expanding_templates_and_rendering > Another possibility is to use Title and Article classes and do your own > AJAX handler: > http://www.mediawiki.org/wiki/Manual:Ajax > http://www.mediawiki.org/wiki/Manual:Title.php > http://www.mediawiki.org/wiki/Manual:Article.php > > However, that's probably a "reinventing of wheel". > > Sorry for not being able to provide full example - I am not a rich guy > and busy with projects to feed my family. Also, I am not the fastest > coder out there. > Dmitriy > > > > ------------------------------ > > Message: 4 > Date: Tue, 3 Aug 2010 21:24:03 +1000 > From: John Vandenberg <[email protected]> > Subject: Re: [Wikitech-l] wikipedia is one of the slower sites on the > web > To: Wikimedia developers <[email protected]> > Message-ID: > > <[email protected]<aj13u8urrck%2br%[email protected]> > > > Content-Type: text/plain; charset=ISO-8859-1 > > On Tue, Aug 3, 2010 at 8:55 PM, K. Peachey <[email protected]> wrote: > > Would something like what is shown below get it even further down? > > > > a { color: blue } > > a.1_byte_article, a.2_byte_article, a.3_byte_article, > > ... > > using an abbreviation like <x>ba would also help. > > Limiting the user pref to intervals of 10 bytes would also help. > > Also, as this piece of CSS is being dynamically generated,it only > needs to include the variations that occur in the body of the > article's HTML. > > Or the CSS can be generated by JS on the client side, which is what > Aryeh has been suggesting all along (I think). > > btw, I thought Domas was kidding. I got a chuckle out of it, at least. > > -- > John Vandenberg > > > > ------------------------------ > > Message: 5 > Date: Tue, 03 Aug 2010 13:55:13 +0200 > From: Platonides <[email protected]> > Subject: Re: [Wikitech-l] wikipedia is one of the slower sites on the > web > To: [email protected] > Message-ID: <[email protected]> > Content-Type: text/plain; charset=ISO-8859-1 > > Lars Aronsson wrote: > > On 08/01/2010 10:55 PM, Aryeh Gregor wrote: > >> One easy hack to reduce this problem is just to only provide a few > >> options for stub threshold, as we do with thumbnail size. Although > >> this is only useful if we cache pages with nonzero stub threshold . . > >> . why don't we do that? Too much fragmentation due to the excessive > >> range of options? > > > > Couldn't you just tag every internal link with > > a separate class for the length of the target article, > > and then use different personal CSS to set the > > threshold? The generated page would be the same > > for all users: > > > > <a href="My_Article" class="134_byte_article">My Article</a> > > That would be workable, eg. one class for articles smaller than 50 > bytes, other for 100, 200, 250, 300, 400, 500, 600, 700, 800, 1000, > 2000, 2500, 5000, 10000 if it weren't for having to update all those > classes whenever the page changes. > > It would work to add it as a separate stylesheet for stubs, though. > > > > > ------------------------------ > > Message: 6 > Date: Tue, 03 Aug 2010 13:59:06 +0200 > From: Platonides <[email protected]> > Subject: Re: [Wikitech-l] chanfing main page articles from drop down. > help required > To: [email protected] > Message-ID: <[email protected]> > Content-Type: text/plain; charset=ISO-8859-1 > > Noman wrote: > > Hi, > > i've installed mediawiki for a wiki project. > > now we have 4 sections on main page . like there are on wikipedia main > page. > > > > Now as its done in wikipedia these 4 boxes are tables and update on date > > criteria. > > > > Now i want to do is to give some kind a navigation bar like drop down or > > paging. > > so when user selects page from drop down the all 4 pages are updated with > > relevant articles. > > to clear more. how the for parts of table can be updated . and how to put > > combo box ( which is already filled with page numbers / article topics) > and > > which user selects any page from drop down all the 4 table rows are > updated. > > One way would be to have different pages. You seem to already have lists > of what four things should appear when selectiog X, so that would fit. > > Another approach that may fit you is the one of: > http://ca.wikipedia.org/wiki/Portada > > > > > ------------------------------ > > Message: 7 > Date: Tue, 3 Aug 2010 14:05:00 +0100 > From: David Gerard <[email protected]> > Subject: Re: [Wikitech-l] Debian packages (was MediaWiki version > statistics) > To: Wikimedia developers <[email protected]> > Message-ID: > <[email protected]> > Content-Type: text/plain; charset=UTF-8 > > On 3 August 2010 00:17, Edward Z. Yang <[email protected]> wrote: > > > > ? ?2. Distributors roll patches without telling upstream developers who > > ? ? ? would happily accept them into the mainline. > > > Has anyone reported the following as Debian bugs? > > * Package maintainer not sending patches back upstream > * Package maintainer not visible and active in MediaWiki development > * Package maintainer not visible and active in MediaWiki community > support, leaving supporting his packages to the upstream > > > - d. > > > > ------------------------------ > > Message: 8 > Date: Tue, 3 Aug 2010 10:53:57 -0400 > From: Aryeh Gregor > <[email protected]<simetrical%[email protected]> > > > Subject: Re: [Wikitech-l] Showing bytes added/removed in each edit in > "View history" and "User contributions" > To: Wikimedia developers <[email protected]> > Message-ID: > > <[email protected]<60dwcm3%2beq%[email protected]> > > > Content-Type: text/plain; charset=UTF-8 > > On Tue, Aug 3, 2010 at 1:14 AM, Liangent <[email protected]> wrote: > > Byte count is used. For example in Chinese Wikipedia, one of the > > criteria of "Did you know" articles is ">= 3000 bytes". > > I mean, is byte count used for anything where character count couldn't > be used just about as well? Like is there some code that uses rev_len > to figure out whether an article can fit into a field limited to X > bytes, or whatever? (That's probably unsafe anyway.) > > On Tue, Aug 3, 2010 at 3:48 AM, Robert Ullmann <[email protected]> > wrote: > > The revision size (and page size, meaning that of last revision) in > > bytes, is available in the API. If you change the definition there is > > no telling what you will break. > > The same could be said of practically any user-visible change. I > mean, maybe if we add a new special page we'll break some script that > was screen-scraping Special:SpecialPages. We can either freeze > MediaWiki and never change anything for fear that we'll break > something, or we can evaluate each potential change on the basis of > how likely it is to break anything. I can't see anything breaking too > badly if rev_len is reported in characters instead of bytes -- the > only place it's likely to be useful is in heuristics, and by their > nature, those won't break too badly if the numbers they're based on > change somewhat. > > > > ------------------------------ > > Message: 9 > Date: Tue, 3 Aug 2010 10:59:12 -0400 > From: soxred93 <[email protected]> > Subject: Re: [Wikitech-l] Showing bytes added/removed in each edit in > "View history" and "User contributions" > To: Wikimedia developers <[email protected]> > Message-ID: <[email protected]> > Content-Type: text/plain; charset=US-ASCII; delsp=yes; format=flowed > > Just butting in here, if I recall correctly, both the PHP-native > mb_strlen() and the MediaWiki fallback mb_strlen() functions are > considerably slower (1.5 to 5 times as slow). Unless there's another > way to count characters for multibyte UTF strings, this would not be > a feasible idea. > > -X! > > > > ------------------------------ > > _______________________________________________ > Wikitech-l mailing list > [email protected] > https://lists.wikimedia.org/mailman/listinfo/wikitech-l > > > End of Wikitech-l Digest, Vol 85, Issue 11 > ****************************************** > _______________________________________________ Wikitech-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/wikitech-l
