Jim:

I took the opportunity to experiment with and without the '=50' while
setting up the file. Good one.

In managing environments from WinNT 4.0 through Win Serve 2003 I used a
command-line environment enhancer 'Take Command32' from JPSoft; the
inheritor of the 4DOS franchise. This batch file experience tempts me to
blow the dust off and see what it'll do in concert with R>.

Thanks again,

Bruce

> -------- Original Message --------
> Subject: [RBASE-L] - Re: Data File Relocation Routine
> From: James Bentley <[email protected]>
> Date: Wed, March 07, 2012 4:21 pm
> To: [email protected] (RBASE-L Mailing List)
> 
> 
> Bruce the "=50" was to prevent the data from wrapping to multiple lines. It 
> could have been any number that would 
> 
> keep the data from wrapping to multiple lines.
> 
> Some times we are seduced by the GUI and forget about the power of the 
> Windows command line.
> I use this technique to generate email mailings out of my database using a 
> command line program "BLAT.EXE".
> I started using the technique since my database was still in RBase for DOS. I 
> am in the process of converting
> it to  RBase for windows but will still use BLAT.
> 
>  
> Jim Bentley
> American Celiac Society
> [email protected]
> tel: 1-504-737-3293
> 
> 
> >________________________________
> > From: Bruce Chitiea <[email protected]>
> >To: RBASE-L Mailing List <[email protected]> 
> >Sent: Wednesday, March 7, 2012 3:39 PM
> >Subject: [RBASE-L] - Re: Data File Relocation Routine
> > 
> >Jim, thank you.
> >
> >I used your WRITE from the R> approach, using COPY instead of MOVE.
> >
> >Before I did, I set the source and target paths into variables:
> >
> >set vsource text = null
> >set vtarget text = null
> >set vsource = 'd:\0-source\'
> >set vtarget = 'd:\0-target\'
> >
> >After writing the REM line, the statement:
> >
> >('COPY' & .vsource + (ctxt(file_sn)) + 
> >+ '*.*' & .vtarget + file_set + '\')=50 +
> >from file_ident
> >
> >... produced the line-by-line batch file by serial number.
> >
> >Launched from R>.
> >
> >Files copied at rate of @1,000/min. Saved me at least a day of marking
> >and moving.
> >
> >I'd never once thought of setting up a batch file from within R:Base.
> >
> >The '=50' is new to me. Nice add to the tool bag.
> >
> >Thanks very much!
> >
> >Bruce
> >
> >
> >> -------- Original Message --------
> >> Subject: [RBASE-L] - Re: Data File Relocation Routine
> >> From: James Bentley <[email protected]>
> >> Date: Wed, March 07, 2012 10:01 am
> >> To: [email protected] (RBASE-L Mailing List)
> >> 
> >> 
> >> Bruce,
> >> 
> >> -- TO MOVE FILES IN THE FORM OF xxxxx.filename.ext
> >> -- create a batch file to accomplish this
> >> -- at the R> enter the following commands or better stil create a RMD file
> >> -- This assumes that the RBase working directory is where the common
> >> -- files are located and that the "a,b,c,d" directories are sub-directories
> >> -- of the common directory. If not use full path to "a,b,c,d" directories
> >> SET HEADINGS OFF
> >> OUTPUT MVFILES.BAT
> >> WRITE 'REM Batch file created',.#DATE,'at',.#TIME
> >> SELECT ('MOVE' & file_sn + '*.*' & file_set+'\')=50
> >> OUTPUT SCREEN
> >> -- This should write a series of 'MOVE 10001*.* a\
> >> -- if file_sn is INTEGER substitute (CTXT(file_sn)) in above
> >> -- if you are extra cautious use COPY instead of MOVE in above
> >> 
> >> -- You might want to enter RBEDIT MVFILES.BAT at R> to verify file
> >> -- after this open a WINDOWS command prompt and enter
> >> -- MVFILES.BAT
> >> 
> >> -- Alternatively you could try LAUNCH MVFILES.BAT at R>
> >> 
> >> 
> >>  
> >> Jim Bentley
> >> American Celiac Society
> >> [email protected]
> >> tel: 1-504-737-3293
> >> 
> >> 
> >> >________________________________
> >> > From: Jason Kramer <[email protected]>
> >> >To: RBASE-L Mailing List <[email protected]> 
> >> >Sent: Wednesday, March 7, 2012 10:27 AM
> >> >Subject: [RBASE-L] - Re: Data File Relocation Routine
> >> > 
> >> >
> >> >Bruce,
> >> >    If I understand correctly, you have a table that looks like
> >>     this:
> >> >file_sn    file_set
> >> >10001        a
> >> >10002        a
> >> >10003        b
> >> >10004        c
> >> >
> >> >    A single directory with all files in it, and directories named
> >>     a,b,c,d that you want to copy (or move) the files into, and an Excel
> >>     file with the same information?  Is that correct?
> >> >                                                               
> >>     Thanks,
> >> >                                                               
> >>     Jason
> >> >
> >> >Jason Kramer
> >> University Archives and Records Management
> >> 002 Pearson Hall
> >> (302) 831 - 3127 (voice)
> >> (302) 831 - 6903 (fax)
> >> >On 3/7/2012 11:20 AM, Bruce Chitiea wrote: 
> >> >All: RBASE 91_64
> >> 9.1.5.20214
> >> have PluginPowerPak 91_64
> >> Windows 7 GOAL
> >> Distribute classed sets of data files from a common folder to their
> >> respective set folders. BACKGROUND For a client with severe data-file 
> >> duplication and
> >> file-content-identification issues ranging across several thousand
> >> files, I've: - created a backup working set of those files - pre-pended a 
> >> 5-digit serial number to each file-copy into the format: 
> >> xxxxx.filename.ext - aggregated all files into a common folder
> >> - successfully parsed the folder's file list into text fields: xxxxx 
> >> filename ext filesize datestamp - processed these in Excel to identify 
> >> duplication and content-meaning,
> >> and to class files into four sets: [ a b c d ] - successfully Gateway'd 
> >> the file serial number and file set identifier
> >> into a two-column R:Base table: file_sn [ 10001,10002...2nnnn ] file_set [ 
> >> a b c d ] - created four empty file folders, each reserved for one file 
> >> class NEED I'm thinking the 'file_sn' and 'file_set' values are sufficient 
> >> input to
> >> a process that 'walks' the common file list and performs the
> >> distribution. I have no such routine and I'm SO FAR behind the 8-ball on
> >> this job ... What's the best way to handle this? MTIA Bruce Chitiea
> >> SafeSectors, Inc. 
> >> >
> >> >
> >
> >
> >
> >
> >
> >



Reply via email to