What's the size of your product feed as XML on disk?

On Thu, Jul 9, 2009 at 2:02 PM, JamesF<[email protected]> wrote:
>
> thats a good tip if the end user is downloading the xml...but in my
> case its a product feed going to vast.com which wants raw xml
>
> On Jul 9, 8:03 am, GravyFace <[email protected]> wrote:
>> Not really sure what the end-user/consumer of this XML looks like, but
>> I'd recommend also compressing it and providing the consumer of the
>> "feed" (more like a batch) with a .zip file they can download and
>> consume -- 20k records is much too bulky for any kind of near-time
>> consumption in my opinion, because the requestor's parser will most
>> likely be waiting for the request to complete (i.e. download the
>> entire XML file) before parsing it anyways -- might as well save some
>> bandwidth and transfer it reliably.
>>
>> Also, if you go this route, make sure you set the headers to
>> Content-disposition:attachment to force the browser to prompt
>> save/open, instead of trying to open in the browser.
>>
>> You can, however, provide a updated feed of "daily product updates"
>> for those who have already downloaded (what sounds like) your entire
>> product catalog.
>>
>> On Thu, Jul 9, 2009 at 2:48 AM, Bert Van den Brande<[email protected]> wrote:
>>
>>
>>
>> > Hi,
>>
>> > I've ran into the same problem as you did and ended up writing a Cake
>> > Shell script that periodically get's executed by a cronjob.
>> > The script script writes the .xml file to a location on the webserver,
>> > making it accessible through a Cake web page.
>>
>> > Don't have the code at hand here, but if you're interested I can dig it up.
>>
>> > Friendly greetings,
>> > Bert Van den Brande
>>
>> > On Wed, Jul 8, 2009 at 5:38 PM, JamesF<[email protected]> wrote:
>>
>> >> Hello all,
>>
>> >> I have been using cake's xmlhelper to render my xml output quite
>> >> successfully. but i need to create a rather large file (20000
>> >> records), for a product feed export. if i limit my results to 100 at a
>> >> time they render fine but obviously a brickwall is hit at higher
>> >> numbers php generates out of memory erros and the like.
>>
>> >> my idea was to render xmlHelpers output (the view) to a file instead
>> >> of live. then maybe the execution could be staggered and not timeout.
>> >> i dont want to run up against php's script execution time limit.
>>
>> >> maybe its time to bust out requestAction with a for loop that runs
>> >> through 100 records per iteration. i can set all that up, but am still
>> >> unsure about how to output the view into a file.
>>
>> >> anyone else run into a solution for this?
> >
>

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"CakePHP" group.
To post to this group, send email to [email protected]
To unsubscribe from this group, send email to 
[email protected]
For more options, visit this group at 
http://groups.google.com/group/cake-php?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to