I dint think this is a function of a workaround. This is a function of Twitter having a good policy in place to prevent abuse.

You can do what you want by incrementally querrying the API. The API limits will make it take too long. Even with multiple accounts it will be months before you get a final list. Even then, I'm not sure you could keep on top of new user registrations.

Having acces to this data could only be used for nefarious efforts. What you want would be a spammers dream.

I think you would be better and faster to build a crawl farm and crawl all links on Twitter.com and parse the users out, bypassing the API.

Even with the API, as you add new records, those records you just added will expire, delete, get banned, blocked etc. There is no way you could ever have a reconciled system.

Consider if each username is an average 10 bytes. You have 520,000,000 bytes to download of just username data. Let's double that for http overhead and other misc data that will come over the wire. 1 billion bytes.

That's a strongly conservative terrabyte of data that you would have to download once a day and reconcile against the previous day. A terrabyte of just usernames.

Then you have all the CPU that you will need, network lag, time to insert into your data source.

This is not something that can be worked around. This is simply a limitation of scale, one that can not be overcome. You need a direct link to twitters data sources, ideally from within their data center to reduce network lag. This probably will not be approved :)
--
Scott
Iphone says hello.

On Jun 29, 2009, at 9:06 AM, Arunachalam <[email protected]> wrote:

Even if i have my account whitelisted, which have 20,000 request per hour, i need to run for many days which is not feasible. Any other workaround.

Any other way to get rid of these request limit.

Cheers,
Arunachalam


On Mon, Jun 29, 2009 at 7:01 PM, Abraham Williams <[email protected]> wrote:

There has been over 52000000 profiles created. You could just start at
1 and count up. Might take you a while though.

Abraham

On Mon, Jun 29, 2009 at 07:55, Arunachalam<[email protected]> wrote:
> Any idea how to implement the same using php / any other language.
> Im confused abt the implementation.
>
> Cheers,
> Arunachalam
>
>
> On Mon, Jun 29, 2009 at 5:57 PM, Cameron Kaiser <[email protected] >
> wrote:
>>
>> > I am looking to find the entire twitter user list ids.
>> >
>> > Social graph method provides the way to fetch the friends and followers
>> > id,
>> > thorough which we can access the profile of the person using user method
>> > -
>> > show. But this requires a code to be written to recursively crawl the
>> > list
>> > from any starting id and appending the followers and friends id of the
>> > person without duplicating.
>> >
>> > Do we have any other API to get entire list. If not, any other ways
>> > apart
>> > from crawling to get the entire list.
>>
>> No, and no, there are no other ways.
>>
>> --
>> ------------------------------------ personal:
>> http://www.cameronkaiser.com/ --
>>  Cameron Kaiser * Floodgap Systems * www.floodgap.com *
>> [email protected]
>> -- Careful with that Axe, Eugene. -- Pink Floyd
>> -------------------------------
>
>



--
Abraham Williams | Community Evangelist | http://web608.org
Hacker | http://abrah.am | http://twitter.com/abraham
Project | http://fireeagle.labs.poseurtech.com
This email is: [ ] blogable [x] ask first [ ] private.

Reply via email to