I forgot to say after it reads all rows and writes the file, throws [1m [35m (600.1ms) [0m begin transaction [1m [36m (52.0ms) [0m [1mcommit transaction [0m failed to allocate memory Redirected to http://localhost:3000/upload_files/110 Completed 406 Not Acceptable in 1207471ms (ActiveRecord: 693.1ms)
On Friday, October 11, 2013 4:03:12 PM UTC-4:30, Monserrat Foster wrote: > > This is an everyday, initially maybe a couple people at the same time > uploading and parsing files to generate the new one, but eventually it will > extend to other people, so... > > I used a logger and It does retrieve and save the files using the > comparation. But it takes forever, like 30min or so in generating the file. > The process starts as soon as the files are uploaded but it seems to be > taking most of the time into opening the file, once it's opened it takes > maybe 5min at most to generate the new file. > > Do you know where can i find an example on how to read an xlsx file with > nokogiri? I can't seem to find one > > On Friday, October 11, 2013 11:12:20 AM UTC-4:30, Walter Lee Davis wrote: >> >> >> On Oct 11, 2013, at 11:30 AM, Monserrat Foster wrote: >> >> > One 30000+ row file and another with just over 200. How much memory >> should I need for this not to take forever parsing? (I'm currently using my >> computer as server and I can see ruby taking about 1GB in the task manager >> when processing this (and it takes forever). >> > >> > The 30000+ row file is about 7MB, which is not that much (I think) >> >> I have a collection of 1200 XML files, ranging in size from 3MB to 12MB >> each (they're books, in TEI encoding) that I parse with Nokogiri on a 2GB >> Joyent SmartMachine to convert them to XHTML and then on to Epub. This >> process takes 17 minutes for the first pass, and 24 minutes for the second >> pass. It does not crash, but the server is unable to do much of anything >> else while the loop is running. >> >> My question here was, is this something that is a self-serve web service, >> or an admin-level (one-privileged-user-once-in-a-while) type thing? In my >> case, there's one admin who adds maybe two or three books per month to the >> collection, and the 40-minute do-everything loop was used only for >> development purposes -- it was my test cycle as I checked all of the titles >> against a validator to ensure that my adjustments to the transcoding >> process didn't result in invalid code. I would not advise putting something >> like this live against the world, as the potential for DOS is extremely >> great. Anything that can pull the kinds of loads you get when you load a >> huge file into memory and start fiddling with it should not be public! >> >> Walter >> >> > >> > On Friday, October 11, 2013 8:44:22 AM UTC-4:30, Walter Lee Davis >> wrote: >> > >> > On Oct 10, 2013, at 4:50 PM, Monserrat Foster wrote: >> > >> > > A coworker suggested I should use just basic OOP for this, to create >> a class that reads files, and then another to load the files into memory. >> Could please point me in the right direction for this (where can I read >> about it)? I have no idea what's he talking about, as I've never done this >> before. >> > >> > How many of these files are you planning to parse at any one time? Do >> you have the memory on your server to deal with this load? I can see this >> approach working, but getting slow and process-bound very quickly. Lots of >> edge cases to deal with when parsing big uploaded files. >> > >> > Walter >> > >> > > >> > > I'll look up nokogiri and SAX >> > > >> > > On Thursday, October 10, 2013 4:12:33 PM UTC-4:30, Walter Lee Davis >> wrote: >> > > On Oct 10, 2013, at 4:36 PM, Monserrat Foster wrote: >> > > >> > > > Hello, I'm developing an app that basically, receives a 10MB or >> less XLSX files with +30000 rows or so, and another XLSX file with about >> 200rows, I have to read one row of the smallest file, look it up on the >> largest file and write data from both files to a new one. >> > > >> > > Wow. Do you have to do all this in a single request? >> > > >> > > You may want to look at Nokogiri and its SAX parser. SAX parsers >> don't care about the size of the document they operate on, because they >> work one node at a time, and don't load the whole thing into memory at >> once. There are some limitations on what kind of work a SAX parser can >> perform, because it isn't able to see the entire document and "know" where >> it is within the document at any point. But for certain kinds of problems, >> it can be the only way to go. Sounds like you may need something like this. >> > > >> > > Walter >> > > >> > > > >> > > > I just did a test reading a few rows from the largest file using >> ROO (Spreadsheet doesn't support XSLX and Creek look good but I can't find >> a way to read row by row) >> > > > and it basically made my computer crash, the server crashed, I >> tried rebooting it and it said It was already started, anyway, it was a >> disaster. >> > > > >> > > > So, my question was, is there gem that works best with large XLSX >> files or is there another way to approach this withouth crashing my >> computer? >> > > > >> > > > This is what I had (It's very possible I'm doing it wrong, help is >> welcome) >> > > > What i was trying to do here, was to process the files and create >> the new XLS file after both of the XLSX files were uploaded: >> > > > >> > > > >> > > > require 'roo' >> > > > require 'spreadsheet' >> > > > require 'creek' >> > > > class UploadFiles < ActiveRecord::Base >> > > > after_commit :process_files >> > > > attr_accessible :inventory, :material_list >> > > > has_one :inventory >> > > > has_one :material_list >> > > > has_attached_file :inventory, :url=>"/:current_user/inventory", >> :path=>":rails_root/tmp/users/uploaded_files/inventory/inventory.:extension" >> >> > > > has_attached_file :material_list, >> :url=>"/:current_user/material_list", >> :path=>":rails_root/tmp/users/uploaded_files/material_list/material_list.:extension" >> >> >> > > > validates_attachment_presence :material_list >> > > > accepts_nested_attributes_for :material_list, :allow_destroy => >> true >> > > > accepts_nested_attributes_for :inventory, :allow_destroy => true >> >> > > > validates_attachment_content_type :inventory, :content_type => >> ["application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"], >> :message => "Only .XSLX files are accepted as Inventory" >> > > > validates_attachment_content_type :material_list, :content_type >> => ["application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"], >> :message => "Only .XSLX files are accepted as Material List" >> > > > >> > > > >> > > > def process_files >> > > > inventory = Creek::Book.new(Rails.root.to_s + >> "/tmp/users/uploaded_files/inventory/inventory.xlsx") >> > > > material_list = Creek::Book.new(Rails.root.to_s + >> "/tmp/users/uploaded_files/material_list/material_list.xlsx") >> > > > inventory = inventory.sheets[0] >> > > > scl = Spreadsheet::Workbook.new >> > > > sheet1 = scl.create_worksheet >> > > > inventory.rows.each do |row| >> > > > row.inspect >> > > > sheet1.row(1).push(row) >> > > > end >> > > > >> > > > sheet1.name = "Site Configuration List" >> > > > scl.write(Rails.root.to_s + >> "/tmp/users/generated/siteconfigurationlist.xls") >> > > > end >> > > > end >> > > > >> > > > >> > > > -- >> > > > You received this message because you are subscribed to the Google >> Groups "Ruby on Rails: Talk" group. >> > > > To unsubscribe from this group and stop receiving emails from it, >> send an email to [email protected]. >> > > > To post to this group, send email to [email protected]. >> > > > To view this discussion on the web visit >> https://groups.google.com/d/msgid/rubyonrails-talk/bc470d4d-19c4-4969-8ba7-4ead7a35d40c%40googlegroups.com. >> >> >> > > > For more options, visit https://groups.google.com/groups/opt_out. >> > > >> > > >> > > -- >> > > You received this message because you are subscribed to the Google >> Groups "Ruby on Rails: Talk" group. >> > > To unsubscribe from this group and stop receiving emails from it, >> send an email to [email protected]. >> > > To post to this group, send email to [email protected]. >> > > To view this discussion on the web visit >> https://groups.google.com/d/msgid/rubyonrails-talk/0325dc87-0649-45fc-9d55-0fbcd8bed0a0%40googlegroups.com. >> >> >> > > For more options, visit https://groups.google.com/groups/opt_out. >> > >> > >> > -- >> > You received this message because you are subscribed to the Google >> Groups "Ruby on Rails: Talk" group. >> > To unsubscribe from this group and stop receiving emails from it, send >> an email to [email protected]. >> > To post to this group, send email to [email protected]. >> > To view this discussion on the web visit >> https://groups.google.com/d/msgid/rubyonrails-talk/ba633f69-5527-4dc1-8518-b6104e414e15%40googlegroups.com. >> >> >> > For more options, visit https://groups.google.com/groups/opt_out. >> >> -- You received this message because you are subscribed to the Google Groups "Ruby on Rails: Talk" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/rubyonrails-talk/87eace87-5b65-4265-8b5a-3fa804036629%40googlegroups.com. For more options, visit https://groups.google.com/groups/opt_out.

