Re: APF authorization failure with DB2

2009-05-23 Thread Joel C Ewing

Vivian wrote:

On May 22, 2:31 pm, pauls2...@yahoo.com wrote:

When we run in straight batch PGM=MyPgm, it works.
When we run using IKJEFT01 with a call, it works.
When we run it as a DB2 job (IKJEFT01 with Run), it fails S047.  But,
all of our DB2 libs are also APF authorized and in the linklist.  By
all I mean SDSNLINK, SDSNEXIT, SDSNLOAD, SDXRRESL.  So, I'm
perplexed.

Is DSN in your TSO Authorized commands table?


Yes, it is.  We put it in the authorized commands and the authorized
programs, just in case...  No change.


Might try establishing DB2 connectivity by having your program call DB2 
CAF (Call Attachment Facility) from your program rather than using DB2 
DSN interface under batch TSO.  That way you can again invoke your 
program directly from JCL and at least get the added complication of 
IKJEFT01 batch TSO and DB2 DSN out of the picture.  If you need a 
different DB2 subsystem for testing, you can still have that passed to 
your program as a parameter to control the CAF call. Can't recall ever 
using CAF from an authorized program, but at least there shouldn't be 
any confusion about getting your program invoked as authorized and 
running as authorized up to the actual CAF call.


Converting batch DB2 to use CAF instead of the DSN interface has the 
added benefit that SMF job step stats will now reflect the real 
executed program and not be disguised as just another execution of IKJEFT01.

   JC Ewing

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: Enterprise COBOL for z/OS V4

2009-05-23 Thread R.S.

Richards, Robert B. pisze:

That is only true if you are paying full cap. Enterprise COBOL is a
Variable Workload License Charge (VWLC) product. You only pay for what
you use if you are doing sub-capacity pricing. Plus, if you are careful
to restrict the lpars that you allow compiler execution on, more money
can be saved. 


Not true. Previous version was also paid by LPAR. The only difference 
could be that v4 produces SMF89 and then do not need NOSMF89 entry.

However subcapacity pricing is available for COBOL V3.

--
Radoslaw Skorupka
Lodz, Poland


--
BRE Bank SA
ul. Senatorska 18
00-950 Warszawa
www.brebank.pl

Sd Rejonowy dla m. st. Warszawy 
XII Wydzia Gospodarczy Krajowego Rejestru Sdowego, 
nr rejestru przedsibiorców KRS 025237

NIP: 526-021-50-88
Wedug stanu na dzie 01.01.2009 r. kapita zakadowy BRE Banku SA (w caoci 
wpacony) wynosi 118.763.528 zotych. W zwizku z realizacj warunkowego 
podwyszenia kapitau zakadowego, na podstawie uchway XXI WZ z dnia 16 marca 
2008r., oraz uchway XVI NWZ z dnia 27 padziernika 2008r., moe ulec 
podwyszeniu do kwoty 123.763.528 z. Akcje w podwyszonym kapitale zakadowym 
BRE Banku SA bd w caoci opacone.

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: Submitting a Marketing REQUEST (was: BLOCK CONTAINS

2009-05-23 Thread Eric Bielefeld

Ed,

And yet, IBM is hiring 1,300 people in Dubuque Iowa.  I'm sure they will 
hire some who got laid off in other areas of the country, but it is a good 
sign.


Eric Bielefeld
Sr. Systems Programmer
Milwaukee, Wisconsin
414-475-7434


- Original Message - 
From: Ed Gould ps2...@yahoo.com

Excellent question. Although you might end up getting told to call the
1-800 IBM number. Chuckle they wouldn't know it if it cut their head in 
half.


IBM really screwed the customer over when they got rid of the people x 
many years ago. I am happy to be far far far away from that IBM mess. If 
people ask me if IBM is a good investment, I tell them to run run run away 
as fast as you can as sometime soon it is going to belly up. If they have 
any business left it will probably be in INDIA. If they are doing short 
term its a 50-50 gamble as far as I can see.


Ed 


--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: OAM 3995 objects

2009-05-23 Thread Joel C Ewing
	Over a year ago one of our DBA's converted us from OAM and the 3995 to 
remote DB2 objects on DASD on a non-mainframe server platform (as the 
whole point of the 3995 was long-term archival on random access media 
cheaper than mainframe DASD).  Our case might have been simpler than 
yours in that our only OAM archive and user of the 3995 was accessed 
through a single in-house subroutine and 3995 updates were only done 
weekly.  I don't know all the details, but I'm pretty sure he had to 
write his own program to pull objects from OAM and the 3995 and store 
them on the new platform; but by retaining some of the old naming 
conventions and modifying the common access subroutine, the move was 
almost transparent to the applications using the archive (except for 
elimination of the weekly 3995 drive and/or cherry picker failures and 
extended periods of unavailability!).
	I can remember no another hardware device that was the source of so 
many night calls, or that I was happier to see depart.  As this 
application was the only user of OAM, we were able to turn off that as 
well.
	Six months after the departure of the 3995, our snack vendor replaced 
our simple gravity-feed drink vending machines with improved vending 
machines with a cherry-picker extraction mechanism, a single point of 
failure like the 3995 that breaks at least once a month.  Go figure!

  Joel C Ewing, Data-Tronics Corp., Fort Smith, AR

Joel M Ivey wrote:

We're in need of migrating our transitioned objects to a better-performing
platform than the 3995.   Unfortunately OAM doesn't support transitioning
objects to (non-DB2) dasd.  Our other option is our VTS.  Are there any
shops that have migrated all their 3995 objects to another platform in order
to get rid of the 3995 altogether?   


Thanks,
Joel 
USC

Columiba, SC

...

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: Migrate at close

2009-05-23 Thread Ron Hawkins
David,

If you are using DFSMShsm duplex tapes for ML2 do you really need to have a
third copy as a DFSMShsm backup?

Ron

 -Original Message-
 From: IBM Mainframe Discussion List [mailto:ibm-m...@bama.ua.edu] On
Behalf Of
 O'Brien, David W. (NIH/CIT) [C]
 Sent: Wednesday, May 20, 2009 5:55 AM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: [IBM-MAIN] Migrate at close
 
 Lizzette,
 
   If the OP migrates to ML2 immediately at EOJ, then there is no HSM
Backup
 taken. The only option for a backup copy in this case is ABARs,
 
 Dave O'Brien
 NIH Contractor
 
 From: IBM Mainframe Discussion List [ibm-m...@bama.ua.edu] On Behalf Of
 Lizette Koehler [stars...@mindspring.com]
 Sent: Wednesday, May 20, 2009 8:39 AM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: Migrate at close
 
 Just so I have a clearer understanding -
   Is this to create a backup or DR file for offsite?
   Could you use a management class that would move it to ML2 a little
later?
 Does it need to be immediate?
 Could you have a second management class that you could use
occasionally
 that would migrate to ML2?
   Is this because you do not always want to migrate to ML2 but only
 occasionally?
 
 Thanks.
 
 Lizette
 
 
 
  In some cases ,  it would be nice to migrate a dataset to ML2,  as soon
  as it has been writen and closed.
  Can I achive this ?
 
  Search the archives at http://bama.ua.edu/archives/ibm-main.html
 
 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html
 
 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: ups batteries draining, can't switch to generators

2009-05-23 Thread Joel C Ewing
While YMMV, out experience has been that any utility power failure 
lasting more than 5-15 seconds is a solid failure and the outage will 
invariably be an hour or more while the utility company locates and 
fixes the problem.  This means that unless you have an extraordinary 
UPS, or functional generators able to recharge the UPS, you are going 
down --  The only issue is when or how.  Given the choice, a controlled 
shutdown from which restart is almost guaranteed is infinitely better 
than gambling the potential of adding hours of downtime and potential 
data loss from an abrupt termination for the questionable benefit of 
staying up for a few minutes longer.


One of issues sounds like a management problem.  Placing the power to 
make a shutdown decision solely in the hands of a duty manager who is 
not 100% available obviously doesn't work for decisions requiring a 10 
minute or better response time.  The other issue is that you must have 
automation support in place to minimize the z/OS shutdown time and 
documented emergency shutdown procedure that are required reading for 
whoever may have to effect the shutdown.


We have documented procedures for emergency system and hardware shutdown 
and z/OS automated procedures (using Netview automation and CBT freebie 
program NETINIT/NETSTOP) to take down online and batch systems and DB2 
as quickly as possible.  These are the same procedures used for normal 
IPL shutdown, so they are tested regularly.  Normally Operations would 
consult with whoever is on call in Technical Services (and someone is 
always available) and we would advise whether to initiate a system 
shutdown or do it ourselves if on site; but if communication is 
impossible within the allowed time frame, that decision must be made by 
the ranking Operator on site.


Our procedures also document a quick and dirty shutdown method if there 
is reason to believe the remaining UPS time is at best only one or two 
minutes instead of the typical 15+ minutes - namely, QUIESCE z/OS, 
SYSTEM RESET the production LPAR, and power down the processor and 
other hardware ASAP.  There is greater risk of logical damage - DB2 
threads in a questionable state and possibly a need to recover some 
specific tables from archive logs -- but doing a controlled hardware 
shutdown should at least eliminate any hardware issues on restart.

  Joel C Ewing

Kelly Bert Manning wrote:

Please don't laugh.

I work with applications on a non-sysplex and non-xrf, supported, z/OS
where there have been 3 cases of UPS batteries draining flat, 
followed by uncontrolled server crashes, in the past 17 years.


They all happened in October and November, gale season (Cue background
music with the Gales of November line by Gordon Lightfoot)

After the first one the data center operator said that they would consider
giving operators authority to shut down OS/390 if they were unable to
make immediate contact with the Duty Manager after discovering that
UPS batteries were draining during a power failure and that generator
power was not available or failed after starting.

Four weeks later a carbon copy crash occurred, inspriring a promise that
operators would start draining CICS and IMS message queues and stopping
and rolling back BMPs and DB2 online jobs, while there was still power
in batteries.

Roll forward to this decade, power off during gale season, generators
start, but one fails and goes offline, followed by other mayhem in the
power hardware. Back on batteries for 22 minutes, until they drain and
the z server crashes. Current operator says what promise to shut
everything down cleanly before the batteries drain?.

Is 22 minutes an unreasonable time figure for purging IMS messaqe
queues, bringing down CICS regions, draining initiators, and abending
and rolling back online IMS and DB2 jobs to the last checkpoint, swapping
logs, writing and dismounting log backups and turning off power before 
sudden power loss starts to play mayhem with disk and other hardware?


Oh did I mention, the 2 CPU single processor was only about 30% busy at the
time, the Sunday weekly low CPU use period.

We had a different sort of power outage after the first of the 2 crashes
last decade. Somebody working for one of the potential bidders used
a metal tape measure in an attempt to measure clearance around the
power cable entrance to the building. The resulting demonstration of
how much power moves through the space around a high voltage cable
destroyed several 3380 clone drives, in addition to crashing all
the OS/390 processors. I earned my DBA pay that day.

Bottom line, what should happen when UPS batteries start to drain and
there is no prospect of reliable, high quality, utility power being
restored quickly? Leave it up and roll the dice about losing work
in progress and log data (head crashes and cache controller microcode
bugs) or shut it down cleanly?


--
For IBM-MAIN subscribe 

Re: Migrate at close

2009-05-23 Thread R.S.

Ron Hawkins pisze:

David,

If you are using DFSMShsm duplex tapes for ML2 do you really need to have a
third copy as a DFSMShsm backup?


Human error. Even duplex copy can be deleted by mistake. Backup requires 
another action.


--
Radoslaw Skorupka
Lodz, Poland


--
BRE Bank SA
ul. Senatorska 18
00-950 Warszawa
www.brebank.pl

Sd Rejonowy dla m. st. Warszawy 
XII Wydzia Gospodarczy Krajowego Rejestru Sdowego, 
nr rejestru przedsibiorców KRS 025237

NIP: 526-021-50-88
Wedug stanu na dzie 01.01.2009 r. kapita zakadowy BRE Banku SA (w caoci 
wpacony) wynosi 118.763.528 zotych. W zwizku z realizacj warunkowego 
podwyszenia kapitau zakadowego, na podstawie uchway XXI WZ z dnia 16 marca 
2008r., oraz uchway XVI NWZ z dnia 27 padziernika 2008r., moe ulec 
podwyszeniu do kwoty 123.763.528 z. Akcje w podwyszonym kapitale zakadowym 
BRE Banku SA bd w caoci opacone.

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: Migrate at close

2009-05-23 Thread O'Brien, David W. (NIH/CIT) [C]
Hi Ron,

At DR I have the capability to Recall migrated datasets as well as Recover any 
that might be mistakenly deleted. To be completely safe, Yes, I believe both 
ML2 and Backups should be duplexed. Tape is cheap, recreating data if possible 
is not.

Dave O'Brien
NIH Contractor

From: IBM Mainframe Discussion List [ibm-m...@bama.ua.edu] On Behalf Of Ron 
Hawkins [ron.hawkins1...@sbcglobal.net]
Sent: Saturday, May 23, 2009 11:55 AM
To: IBM-MAIN@bama.ua.edu
Subject: Re: Migrate at close

David,

If you are using DFSMShsm duplex tapes for ML2 do you really need to have a
third copy as a DFSMShsm backup?

Ron

 -Original Message-
 From: IBM Mainframe Discussion List [mailto:ibm-m...@bama.ua.edu] On
Behalf Of
 O'Brien, David W. (NIH/CIT) [C]
 Sent: Wednesday, May 20, 2009 5:55 AM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: [IBM-MAIN] Migrate at close

 Lizzette,

   If the OP migrates to ML2 immediately at EOJ, then there is no HSM
Backup
 taken. The only option for a backup copy in this case is ABARs,

 Dave O'Brien
 NIH Contractor
 
 From: IBM Mainframe Discussion List [ibm-m...@bama.ua.edu] On Behalf Of
 Lizette Koehler [stars...@mindspring.com]
 Sent: Wednesday, May 20, 2009 8:39 AM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: Migrate at close

 Just so I have a clearer understanding -
   Is this to create a backup or DR file for offsite?
   Could you use a management class that would move it to ML2 a little
later?
 Does it need to be immediate?
 Could you have a second management class that you could use
occasionally
 that would migrate to ML2?
   Is this because you do not always want to migrate to ML2 but only
 occasionally?

 Thanks.

 Lizette


 
  In some cases ,  it would be nice to migrate a dataset to ML2,  as soon
  as it has been writen and closed.
  Can I achive this ?
 
  Search the archives at http://bama.ua.edu/archives/ibm-main.html

 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html

 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: BLOCK CONTAINS

2009-05-23 Thread Joel C Ewing

Paul Gilmartin wrote:

On Fri, 15 May 2009 09:27:42 -0400, Thompson, Steve wrote:

Welcome to the MVS world. In the MVS world, we are not device dependant,
nor are we data definition locked/blocked. We generally don't have to
recompile our programs, change the DTF contents (DCB in MVS), etc. just
because the file attributes change (xSAM to VSAM is the exception).


Huh???

If we are not device dependant, why is there such intense trepidation
and resistance to the mere suggestion of a device with a novel geometry
such as more bytes per track or more tracks per cylinder?  It doesn't
appear that you and I have been living in the same MVS world.

-- gil


It would be a rarity for application programs themselves to be device 
dependent in the z/OS world, but once you start dealing with instances 
of data used by those programs on DASD the dependency creeps back in in 
both obvious and subtle ways, in both the physical representation of the 
data and in the JCL which references or allocates that data set.  I 
think it would be fair to say that a systems programmer who has to 
design and implement a migration strategy and deal with effects that are 
more often than not performance and resource usage issues would see many 
more issues with DASD architecture migration than an applications 
programmer.


Just changing the number of cylinders per device, complicates DASD 
migration strategies and may require JCL adjustments to optimize 
allocations when the typical free space per volume or size of typical 
free space extents changes.  If you change the capacity of a track or 
cylinder, then all JCL, IDCAMS, or TSO dataset space allocations in 
those units are obviously suspect.


A change in track size raises issues in both block size and number of 
blocks per track that in general defy a completely automated solution 
and require some analysis to resolve.  Even with an ordinary sequential 
dataset, there is no guarantee that all access is via standard (QSAM) 
access methods with no use of pointers dependent in some way on track 
geometry.  And even if all access is QSAM, do you leave the block size 
alone and risk performance issues, or re-optimize block size and risk 
having to adjust JCL REGION sizes to allow for larger buffers?  Data set 
types that are known to be direct access, and which may contain internal 
relative track pointers cannot be copied without some internal modification.


Changes in track size or tracks per cylinder have especially subtle side 
effects when dealing with VSAM KSDS files.  VSAM Control Area (CA) size 
is tied to the physical cylinder architecture, optimum index CI-size is 
tied to the number of Data CIs per CA, which in turn is tied to both 
track size and cylinder size.  Failure to adjust a sub-optimum Index CI 
size when copying a KSDS to a new architecture could seriously degrade 
CA space utilization and increase total dataset space requirements, but 
automatically adjusting Index CI size for online files could have 
adverse side effects on manually defined LSR buffer pools in the online 
regions that access these KSDS files.  So again, any attempt to 
automatically migrate such files to a new architecture carries risks.


Having been through several architecture conversions in the past -- 3330 
 3350  3380(various models)  3390(various models) -- there is no 
doubt that (1)it can be done (but any old code with datasets that are 
tied to one architecture with no source or no support is a definite 
problem!), (2) a migration that changes the track/cylinder architecture 
takes more person hours to do correctly, (3) and having two different 
architectures in house during a transition is more complicated to 
manage.  All things being equal, I would much rather not use my scarce 
time dealing with DASD architecture migration.

Joel C Ewing

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: BLOCK CONTAINS

2009-05-23 Thread Ted MacNEIL
All things being equal, I would much rather not use my scarce time dealing 
with DASD architecture migration.

Having gone from 3330 -- 3350 -- 3380 -- 3390 (emulation mode) -- 3390 
(native), I agree 100%!

IBM promised, years ago, to not change the geometry again.

Better the devil you know; I have other things to waste my time on.
-
Too busy driving to stop for gas!

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: BLOCK CONTAINS

2009-05-23 Thread Paul Gilmartin
On Sat, 23 May 2009 19:12:19 +, Ted MacNEIL wrote:

All things being equal, I would much rather not use my scarce time dealing 
with DASD architecture migration.

Having gone from 3330 -- 3350 -- 3380 -- 3390 (emulation mode) -- 3390 
(native), I agree 100%!

IBM promised, years ago, to not change the geometry again.

Better the devil you know; I have other things to waste my time on.
-
Too busy driving to stop for gas!

You seem to be agreeing with Steve Thompson that In the MVS world, we
are not device dependant, only insofar as there is only one type of
device.  A weak assertion indeed.  Would you be willing to go so far
as to side with me and say that we fundamentally disagree with Mr.
Thompson?

-- gil

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: BLOCK CONTAINS

2009-05-23 Thread Ted MacNEIL
You seem to be agreeing with Steve Thompson that In the MVS world, we are not 
device dependant, only insofar as there is only one type of device.  A weak 
assertion indeed.

Not at all.
There are at least two device types -- tape and disk.
And, I can convert to either without re-compiling.
That is device independent.
-
Too busy driving to stop for gas!

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: Configuring APPN using Hipersockets

2009-05-23 Thread Chris Mason
John

I'm glad you managed to get it all working. You obviously don't need it but, 
for 
completeness for the archives, I may as well air the IP considerations for 
Enterprise Extender:

1. Clearly you need to define a static VIPA, the IP address given in the XCA 
GROUP statement.

DEVICE device_name VIRTUAL 0
LINK link_name VIRTUAL 0 device_name

HOME
 ...
 aaa.aaa.aaa.aaa link_name
 ...

2. Usual descriptions of the IP requirements for Enterprise Extender say that 
you need to establish a same host interface.
 
DEVICE IUTSAMEH device_name MPCPTP
LINK link_name MPCPTP IUTSAMEH

HOME
 ...
 bbb.bbb.bbb.bbb link_name
 ...

BSDROUTINGPARMS TRUE
 ...
 link_name mtu cost_metric subnet_mask 0
 ...
ENDBSDROUTINGPARMS

either as above or as a side-effect of specifying

IPCONFIG
 ...
 DYNAMICXCF bbb.bbb.bbb.bbb/num_mask_bits cost_metric 
 ...

However, some doubt has been thrown on this requirement by a recent 
thread Duplicate Home Addresses for VIPA and EE - Problem? running in 
February and March of this year in the IBMTCP-L list.[1]

Since you probably defined the DYNAMICXCF parameter of the IPCONFIG 
statement, you get your IUTSAMEH interface defined for you at no extra 
cost in terms of definition activity.

3. You will be using UDP ports 12000 to 12004. There have been recent 
discussions over whether or not you really need to specify entries in the PORT 
statement list when you do not actually need to protect your installation from 
people running socket programs without reference to the authority supposedly 
managing Communications Server (CS) IP.

Nevertheless it is good documentation practice - at least - to define your 
server ports in the PORT statement list. Of course if you need to specify 
additional parameters such as the BIND parameter, you will need to define an 
entry in the PORT statement list:

PORT
 ...
 12000 UDP NET
 12001 UDP NET
 12002 UDP NET
 12003 UDP NET
 12004 UDP NET
 ...

alternatively

PORTRANGE
 ...
 12000 5 UDP NET

where I have assumed that NET is the name of the started task procedure for 
VTAM.

Note that IBM has taken care to reserve port numbers 12000 to 12004 with 
the appropriate authority

http://www.iana.org/assignments/port-numbers

quote

entextxid 12000/tcp  IBM Enterprise Extender SNA XID Exchange
entextxid 12000/udp  IBM Enterprise Extender SNA XID Exchange
entextnetwk 12001/tcp  IBM Enterprise Extender SNA COS Network Priority
entextnetwk 12001/udp  IBM Enterprise Extender SNA COS Network Priority
entexthigh 12002/tcp  IBM Enterprise Extender SNA COS High Priority
entexthigh 12002/udp  IBM Enterprise Extender SNA COS High Priority
entextmed 12003/tcp  IBM Enterprise Extender SNA COS Medium Priority
entextmed 12003/udp  IBM Enterprise Extender SNA COS Medium Priority
entextlow 12004/tcp  IBM Enterprise Extender SNA COS Low Priority
entextlow 12004/udp  IBM Enterprise Extender SNA COS Low Priority

/quote

Just in case IBM might feel inspired in the future to have a flavour of 
Enterprise Extender which uses TCP as the transport protocol, the same ports 
are reserved with TCP.

However there are some products which have used at least port 12000 
without reference to the standards body. SHADOW DIRECT is one and there 
may be others lurking out there. If you are already unfortunately committed to 
one of these products - as well as marking them 0, assuming that's the lowest 
mark, the next chance you have to give them an evaluation - you should be 
able to avoid the unnecessary difficulty by means of the BIND parameter 
which would allow you to specify the Enterprise Extender static VIPA - 
assuming you have dedicated that static VIPA for use with Enterprise 
Extender.

4. The last IP topic is to establish routing to your partner. This can be 
either 
by static routing or dynamic routing using OMPROUTE. A route over 
HiperSockets interfaces should be quite straightforward. For more complicated 
configurations it is CS IP business-as-usual.

-

Chris Mason

[1] You can review posts in the IBMTCP-L list here:

http://www2.marist.edu/htbin/wlvindex?IBMTCP-L

On Wed, 20 May 2009 12:18:49 -0700, John Au john...@paccar.com 
wrote:

Chris,

Thanks a million.  Your examples helped.  We now see the APPN connection
being made between the LPARs.

Regards,

John

-Original Message-
From: IBM Mainframe Discussion List [mailto:ibm-m...@bama.ua.edu] On
Behalf Of Chris Mason
Sent: Tuesday, May 19, 2009 12:43 PM
To: IBM-MAIN@bama.ua.edu
Subject: Re: Configuring APPN using Hipersockets

John

Here's some samples to get you started taken from some definitions I set
up
for a customer last year:

Sample Enterprise Extender Major Node
-

 VBUILD TYPE=XCA
name PORT  MEDIUM=HPRIP,   enterprise extender *
   HPREELIV=YES, HPR liveness reduction for EE *
   IPPORT=12000,first of 5 IP port numbers *
   IPRESOLV=0, no resolver 

Re: BLOCK CONTAINS

2009-05-23 Thread R.S.

Ted MacNEIL pisze:

You seem to be agreeing with Steve Thompson that In the MVS world, we are not 
device dependant, only insofar as there is only one type of device.  A weak 
assertion indeed.


Not at all.
There are at least two device types -- tape and disk.
And, I can convert to either without re-compiling.
That is device independent.


IMHO more definitions of the independence can be formulated.
For me it is track size in SMS. As long you use track size as physical 
paramter of the disk you are device dependent. When you think about it 
as just some chunk of disk space (maybe even FBA) then you are independent.


BTW: When we compare MVS to other systems in this context it is good to 
know what they understand as device dependency. IMHO no need for 
recompilation is not independency.


My $0.02
--
Radoslaw Skorupka
Lodz, Poland


--
BRE Bank SA
ul. Senatorska 18
00-950 Warszawa
www.brebank.pl

Sąd Rejonowy dla m. st. Warszawy 
XII Wydział Gospodarczy Krajowego Rejestru Sądowego, 
nr rejestru przedsiębiorców KRS 025237

NIP: 526-021-50-88
Według stanu na dzień 01.01.2009 r. kapitał zakładowy BRE Banku SA (w całości 
wpłacony) wynosi 118.763.528 złotych. W związku z realizacją warunkowego 
podwyższenia kapitału zakładowego, na podstawie uchwały XXI WZ z dnia 16 marca 
2008r., oraz uchwały XVI NWZ z dnia 27 października 2008r., może ulec 
podwyższeniu do kwoty 123.763.528 zł. Akcje w podwyższonym kapitale zakładowym 
BRE Banku SA będą w całości opłacone.

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: ups batteries draining, can't switch to generators

2009-05-23 Thread Doug Fuerst
Why not simply fix the problems in the power systems, and test them 
regularly?
If this had been the case in the last 3 places I have worked, I would 
have been escorted off the premises, and my stuff thrown out after me.


Doug

snip


Kelly Bert Manning wrote:

Please don't laugh.

I work with applications on a non-sysplex and non-xrf, supported, z/OS
where there have been 3 cases of UPS batteries draining flat, 
followed by uncontrolled server crashes, in the past 17 years.


They all happened in October and November, gale season (Cue background
music with the Gales of November line by Gordon Lightfoot)

After the first one the data center operator said that they would 
consider

giving operators authority to shut down OS/390 if they were unable to
make immediate contact with the Duty Manager after discovering that
UPS batteries were draining during a power failure and that generator
power was not available or failed after starting.

Four weeks later a carbon copy crash occurred, inspriring a promise that
operators would start draining CICS and IMS message queues and stopping
and rolling back BMPs and DB2 online jobs, while there was still power
in batteries.

Roll forward to this decade, power off during gale season, generators
start, but one fails and goes offline, followed by other mayhem in the
power hardware. Back on batteries for 22 minutes, until they drain and
the z server crashes. Current operator says what promise to shut
everything down cleanly before the batteries drain?.

Is 22 minutes an unreasonable time figure for purging IMS messaqe
queues, bringing down CICS regions, draining initiators, and abending
and rolling back online IMS and DB2 jobs to the last checkpoint, 
swapping
logs, writing and dismounting log backups and turning off power 
before sudden power loss starts to play mayhem with disk and other 
hardware?


Oh did I mention, the 2 CPU single processor was only about 30% busy 
at the

time, the Sunday weekly low CPU use period.

We had a different sort of power outage after the first of the 2 crashes
last decade. Somebody working for one of the potential bidders used
a metal tape measure in an attempt to measure clearance around the
power cable entrance to the building. The resulting demonstration of
how much power moves through the space around a high voltage cable
destroyed several 3380 clone drives, in addition to crashing all
the OS/390 processors. I earned my DBA pay that day.

Bottom line, what should happen when UPS batteries start to drain and
there is no prospect of reliable, high quality, utility power being
restored quickly? Leave it up and roll the dice about losing work
in progress and log data (head crashes and cache controller microcode
bugs) or shut it down cleanly?


--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html




--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: ups batteries draining, can't switch to generators

2009-05-23 Thread Scott T. Harder
Hi All,

I have a couple of points to make on this topic:

1)  For all the well documented (and well taken) information about the
importance of shutting down a system orderly and cleanly, I find it
hard to remember when - in my experience - the system ever had
problems coming back up after a hard crash (I worked in OPS for 10+
yrs.).  Maybe back in the 308x days, but 3090+ ??  The hardware was
pretty resilient, as I remember.  I'm not saying that I recommend
anything other than a clean shutdown, but

2)  Kelly's post harken's me back to an old pet peev and that is;
Operations *used* to have good, knowledgable people who could make
decisions without calling 5 people to tell them what to do!  I saw,
firsthand, the dumbing down of OPS and it disturbed me greatly.  I had
mgmt come into Operations where I worked that *never* wanted to be at
fault... that was truly their #1 priority.  They achieved this through
never making a darn decision on their own... never sticking their neck
out no matter what the situation.  I remember one time where I
restarted the master catalog to resolve a problem; as called for by
the manual (ok... I think I could have gotten away with a lesser
evil), but my point is that my mgmt thought I was nuts (and just
lucky).  Maybe so, but as long as we put zombies who won't take action
based on knowledge and experience (and who are - most importantly -
empowered to do so), then more money must be spent on hardware,
systems and automation that will take the place of that.

Just my thoughts...

All the best,
Scott T. Harder

 Kelly Bert Manning wrote:
 Please don't laugh.

 I work with applications on a non-sysplex and non-xrf, supported, z/OS
 where there have been 3 cases of UPS batteries draining flat,
 followed by uncontrolled server crashes, in the past 17 years.

 They all happened in October and November, gale season (Cue background
 music with the Gales of November line by Gordon Lightfoot)

 After the first one the data center operator said that they would consider
 giving operators authority to shut down OS/390 if they were unable to
 make immediate contact with the Duty Manager after discovering that
 UPS batteries were draining during a power failure and that generator
 power was not available or failed after starting.

 Four weeks later a carbon copy crash occurred, inspriring a promise that
 operators would start draining CICS and IMS message queues and stopping
 and rolling back BMPs and DB2 online jobs, while there was still power
 in batteries.

 Roll forward to this decade, power off during gale season, generators
 start, but one fails and goes offline, followed by other mayhem in the
 power hardware. Back on batteries for 22 minutes, until they drain and
 the z server crashes. Current operator says what promise to shut
 everything down cleanly before the batteries drain?.

 Is 22 minutes an unreasonable time figure for purging IMS messaqe
 queues, bringing down CICS regions, draining initiators, and abending
 and rolling back online IMS and DB2 jobs to the last checkpoint, swapping
 logs, writing and dismounting log backups and turning off power before
 sudden power loss starts to play mayhem with disk and other hardware?

 Oh did I mention, the 2 CPU single processor was only about 30% busy at
 the
 time, the Sunday weekly low CPU use period.

 We had a different sort of power outage after the first of the 2 crashes
 last decade. Somebody working for one of the potential bidders used
 a metal tape measure in an attempt to measure clearance around the
 power cable entrance to the building. The resulting demonstration of
 how much power moves through the space around a high voltage cable
 destroyed several 3380 clone drives, in addition to crashing all
 the OS/390 processors. I earned my DBA pay that day.

 Bottom line, what should happen when UPS batteries start to drain and
 there is no prospect of reliable, high quality, utility power being
 restored quickly? Leave it up and roll the dice about losing work
 in progress and log data (head crashes and cache controller microcode
 bugs) or shut it down cleanly?

 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html



-- 
All the best,
Scott T. Harder

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: ups batteries draining, can't switch to generators

2009-05-23 Thread Scott T. Harder
Sorry should have said restarted the CATALOG address space.

On 5/23/09, Scott T. Harder scottyt.har...@gmail.com wrote:
 Hi All,

 I have a couple of points to make on this topic:

 1)  For all the well documented (and well taken) information about the
 importance of shutting down a system orderly and cleanly, I find it
 hard to remember when - in my experience - the system ever had
 problems coming back up after a hard crash (I worked in OPS for 10+
 yrs.).  Maybe back in the 308x days, but 3090+ ??  The hardware was
 pretty resilient, as I remember.  I'm not saying that I recommend
 anything other than a clean shutdown, but

 2)  Kelly's post harken's me back to an old pet peev and that is;
 Operations *used* to have good, knowledgable people who could make
 decisions without calling 5 people to tell them what to do!  I saw,
 firsthand, the dumbing down of OPS and it disturbed me greatly.  I had
 mgmt come into Operations where I worked that *never* wanted to be at
 fault... that was truly their #1 priority.  They achieved this through
 never making a darn decision on their own... never sticking their neck
 out no matter what the situation.  I remember one time where I
 restarted the master catalog to resolve a problem; as called for by
 the manual (ok... I think I could have gotten away with a lesser
 evil), but my point is that my mgmt thought I was nuts (and just
 lucky).  Maybe so, but as long as we put zombies who won't take action
 based on knowledge and experience (and who are - most importantly -
 empowered to do so), then more money must be spent on hardware,
 systems and automation that will take the place of that.

 Just my thoughts...

 All the best,
 Scott T. Harder

 Kelly Bert Manning wrote:
 Please don't laugh.

 I work with applications on a non-sysplex and non-xrf, supported, z/OS
 where there have been 3 cases of UPS batteries draining flat,
 followed by uncontrolled server crashes, in the past 17 years.

 They all happened in October and November, gale season (Cue background
 music with the Gales of November line by Gordon Lightfoot)

 After the first one the data center operator said that they would
 consider
 giving operators authority to shut down OS/390 if they were unable to
 make immediate contact with the Duty Manager after discovering that
 UPS batteries were draining during a power failure and that generator
 power was not available or failed after starting.

 Four weeks later a carbon copy crash occurred, inspriring a promise that
 operators would start draining CICS and IMS message queues and stopping
 and rolling back BMPs and DB2 online jobs, while there was still power
 in batteries.

 Roll forward to this decade, power off during gale season, generators
 start, but one fails and goes offline, followed by other mayhem in the
 power hardware. Back on batteries for 22 minutes, until they drain and
 the z server crashes. Current operator says what promise to shut
 everything down cleanly before the batteries drain?.

 Is 22 minutes an unreasonable time figure for purging IMS messaqe
 queues, bringing down CICS regions, draining initiators, and abending
 and rolling back online IMS and DB2 jobs to the last checkpoint,
 swapping
 logs, writing and dismounting log backups and turning off power before
 sudden power loss starts to play mayhem with disk and other hardware?

 Oh did I mention, the 2 CPU single processor was only about 30% busy at
 the
 time, the Sunday weekly low CPU use period.

 We had a different sort of power outage after the first of the 2 crashes
 last decade. Somebody working for one of the potential bidders used
 a metal tape measure in an attempt to measure clearance around the
 power cable entrance to the building. The resulting demonstration of
 how much power moves through the space around a high voltage cable
 destroyed several 3380 clone drives, in addition to crashing all
 the OS/390 processors. I earned my DBA pay that day.

 Bottom line, what should happen when UPS batteries start to drain and
 there is no prospect of reliable, high quality, utility power being
 restored quickly? Leave it up and roll the dice about losing work
 in progress and log data (head crashes and cache controller microcode
 bugs) or shut it down cleanly?

 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html



 --
 All the best,
 Scott T. Harder



-- 
All the best,
Scott T. Harder

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: BLOCK CONTAINS

2009-05-23 Thread Robert A. Rosenberg

At 20:24 -0300 on 05/18/2009, Clark Morris wrote about Re: BLOCK CONTAINS:


On 18 May 2009 13:35:40 -0700, in bit.listserv.ibm-main you wrote:

 Ah! NOBLOCK had F/80/80 whereas BLOCK1 had FB/80/80. BLOCK0 still 
had FB/80/27920. But it is still interesting, to me, that BLOCK 
CONTAINS 1 and no BLOCK CONTAINS at all can still create an 
optimally block dataset with little effort in the JCL and no program 
changes.


This is weird because the normal merge is DCB overrides JCL and JCL
overrides data set label (DSCB or tape label).  The code to
distinguish between BLKSIZE=0 in JCL and BLKSIZE not supplied must be
interesting to say the least.  If I get energetic, I'll check the
COBOL Programmers Guide.  Talk about ways to confuse the applications
programmers and subtle ways to screw up.


JCL BLKSIZE=0 gives SDB, no BLKSIZE gives 0. By the time the merge 
happens the FD has generates a DCB with F/80/80 for no Block 
Contains, FB/80/80 for Block Contains 1, and FB/80/0 for Block 
Contains 0 (thus only the last has a Blksize that can be filled in by 
the Merge).


--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: BLOCK CONTAINS

2009-05-23 Thread Robert A. Rosenberg

At 16:07 -0500 on 05/18/2009, Paul Gilmartin wrote about Re: BLOCK CONTAINS:


On Mon, 18 May 2009 18:18:20 +0200, Gilbert Saint-Flour wrote:


On Monday 18 May 2009 18:04, Paul Gilmartin wrote:


What a stupid necessity that programmers have to code BLOCK CONTAINS 0 !


 What happens if the programmer pre-allocates the data set?  
 It's still a stupid necessity, but it might help in dealing with
 situations where recompilation is impractical.


Well, pre-allocation is not very compatible with HSM, GDG, and a few other


I believe I understand the concern with GDG.  (Actually,
my understanding of GDG is so rudimentary I'm not qualified
to doubt the concern.)


GDG will get the blocksize from the DCB= clause, a DCB=DSN reference, 
or a DSCB with the DSNAME (without the .GVyy suffix) on the 
volume where the Catalog lives (or has this last last source become 
no longer relevant). Thus this information is always there for the 
predefined file.


--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: FTP and SMS (and DYNALLOC)

2009-05-23 Thread Lester, Bob
Hi Greg,

   No joy, I'm afraid.  Set it up as documented, but get the same error.
I think I'm getting bit by this statement in the doc.

---snip---

Volumes in dummy storage groups cannot be used when performing
volume allocations. For example, the following DD statement,  
where DUMMY1 is a volume in a dummy storage group, does not   
work: 
  
   //DD1  DD VOL=SER=DUMMY1,UNIT=SYSDA,DISP=SHR   

---snip---

   It appears that this applies to LOCSITE VOL= in the FTP client as
well.

Thanks!
BobL
 
 -Original Message-
 From: IBM Mainframe Discussion List [mailto:ibm-m...@bama.ua.edu] On
 Behalf Of Greg Shirey
 Sent: Friday, May 22, 2009 3:53 PM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: FTP and SMS (and DYNALLOC)
 
 I believe this is one the situations where a dummy storage group
 would
 be useful.  I couldn't find an example in the Implementing
 System-Managed Storage manual, but found the definition:
 
 A type of storage group that contains the serial numbers of volumes
 no
 longer connected to a system. Dummy storage groups allow existing
 JCL to
 function without having to be changed.
 
 HTH,
 Greg
 
 

--
This e-mail transmission may contain information that is proprietary, 
privileged and/or confidential and is intended exclusively for the person(s) to 
whom it is addressed. Any use, copying, retention or disclosure by any person 
other than the intended recipient or the intended recipient's designees is 
strictly prohibited. If you are not the intended recipient or their designee, 
please notify the sender immediately by return e-mail and delete all copies. 
OppenheimerFunds may, at its sole discretion, monitor, review, retain and/or 
disclose the content of all email communications. 
==

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: ups batteries draining, can't switch to generators

2009-05-23 Thread Ted MacNEIL
I saw, firsthand, the dumbing down of OPS and it disturbed me greatly.

When I started in this business, working as an operator was almost a 
requirement before becoming a SYSPROG.


I had mgmt come into Operations where I worked that *never* wanted to be at 
fault... that was truly their #1 priority.

BTDT. GTTS.

But, my management (at the time) still gave them the call of if/which changes 
would be implemented.
And, which changes to back out.
And, when I complained, they just said I didn't understand, having never been 
in the trenches.
I said, I have the scars to prove it.
Remember when you could do a $PQ and blow everything away (before $PQ,ALL was 
introduced and $PQ was an error).
I did that, once.


They achieved this through never making a darn decision on their own... never 
sticking their neck out no matter what the situation.

I've worked for many financial and government organisations.
That is the prevalent attitude in many departments, not just Computer Ops.
The problem becomes, eventually a decision will be made, due to the erosion of 
the situation, and you can only postpone for so long.

By the way, I think you have the wrong third letter in 'darn'.
-
Too busy driving to stop for gas!

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: FTP and SMS (and DYNALLOC)

2009-05-23 Thread Ted MacNEIL
No joy, I'm afraid.  Set it up as documented, but get the same error.
I think I'm getting bit by this statement in the doc.

I haven't really been following this thread, so please forgive me as I ask for 
clarification.

Are you saying that if an FTP outfile is allocated, under z/OS, that DFSMS does 
not intercede?
-
Too busy driving to stop for gas!

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: ups batteries draining, can't switch to generators

2009-05-23 Thread Scott T. Harder
On 5/23/09, Ted MacNEIL eamacn...@yahoo.ca wrote:
 When I started in this business, working as an operator was almost a
 requirement before becoming a SYSPROG.

Absolutely.


 BTDT. GTTS.

;-)


 But, my management (at the time) still gave them the call of if/which
 changes would be implemented.
 And, which changes to back out.

Yup.  Warrented, I think, though.

 And, when I complained, they just said I didn't understand, having never
 been in the trenches.

Wrong.

 I said, I have the scars to prove it.
 Remember when you could do a $PQ and blow everything away (before $PQ,ALL
 was introduced and $PQ was an error).
 I did that, once.

What about just a $P?  One day, everything came to a screaching halt
and nobody could figure it out.  Turned out, a Print Room Op had
indadvertently entered $P and the whole system drained.  ;-)



They achieved this through never making a darn decision on their own...
 never sticking their neck out no matter what the situation.

 I've worked for many financial and government organisations.
 That is the prevalent attitude in many departments, not just Computer Ops.
 The problem becomes, eventually a decision will be made, due to the erosion
 of the situation, and you can only postpone for so long.

Too much time wasted.  People on the front lines need to be empowered
to make decisions and they need to be well-paid, knowledgable
professionals.


 By the way, I think you have the wrong third letter in 'darn'.

You bet!  ;-)

 -
 Too busy driving to stop for gas!

 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html



-- 
All the best,
Scott T. Harder

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: FTP and SMS (and DYNALLOC)

2009-05-23 Thread Lester, Bob
 -Original Message-
 From: IBM Mainframe Discussion List [mailto:ibm-m...@bama.ua.edu] On
 Behalf Of Ted MacNEIL
 Sent: Saturday, May 23, 2009 7:34 PM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: FTP and SMS (and DYNALLOC)
 
 No joy, I'm afraid.  Set it up as documented, but get the same
 error.
 I think I'm getting bit by this statement in the doc.
 
 I haven't really been following this thread, so please forgive me as
 I ask for clarification.
 
 Are you saying that if an FTP outfile is allocated, under z/OS, that
 DFSMS does not intercede?
 -

Hi Ted,

   More or less, I guess.  FTP client on z/OS 1.9, pulling data from a
squatty box.  Relevant parms are:

   LOCSITE UNIT=SYSDA VOL=SYST01

   Get whatever.distributed.file 'PFDT.TSREL.OPPLZDCK.PULLOUT.SUN'

   SYST01 doesn't exist, but my SMS ACS routines point PFDT.** to a
particular SG.  The get fails with:

   PFDT.TSREL.OPPLZDCK.PULLOUT.SUN is on a direct access volume that is
not mounted and Noautomount is specified. 
Std Return Code = 16000, Error Code = 00018

   Putting SYST01 in a DUMMY SG doesn't work, same error.  Doc sort of
implies that it won't work.

   I clipped a volume to SYST01 and it works fine.  

   I was trying to get this to work without changing the ftp client
control cards.

Thanks!
BobL


--
This e-mail transmission may contain information that is proprietary, 
privileged and/or confidential and is intended exclusively for the person(s) to 
whom it is addressed. Any use, copying, retention or disclosure by any person 
other than the intended recipient or the intended recipient's designees is 
strictly prohibited. If you are not the intended recipient or their designee, 
please notify the sender immediately by return e-mail and delete all copies. 
OppenheimerFunds may, at its sole discretion, monitor, review, retain and/or 
disclose the content of all email communications. 
==

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html


Re: FTP and SMS (and DYNALLOC)

2009-05-23 Thread Gibney, Dave
  I just read that in detail again. The automount lead me to consider if
you're really in the HFS file system when you di the get. Try a
whatever the print local directory command is after the locsite and
before the get.

 -Original Message-
 From: IBM Mainframe Discussion List [mailto:ibm-m...@bama.ua.edu] On
 Behalf Of Lester, Bob
 Sent: Saturday, May 23, 2009 7:43 PM
 To: IBM-MAIN@bama.ua.edu
 Subject: Re: FTP and SMS (and DYNALLOC)
 
  -Original Message-
  From: IBM Mainframe Discussion List [mailto:ibm-m...@bama.ua.edu] On
  Behalf Of Ted MacNEIL
  Sent: Saturday, May 23, 2009 7:34 PM
  To: IBM-MAIN@bama.ua.edu
  Subject: Re: FTP and SMS (and DYNALLOC)
 
  No joy, I'm afraid.  Set it up as documented, but get the same
  error.
  I think I'm getting bit by this statement in the doc.
 
  I haven't really been following this thread, so please forgive me as
  I ask for clarification.
 
  Are you saying that if an FTP outfile is allocated, under z/OS, that
  DFSMS does not intercede?
  -
 
 Hi Ted,
 
More or less, I guess.  FTP client on z/OS 1.9, pulling data from a
 squatty box.  Relevant parms are:
 
LOCSITE UNIT=SYSDA VOL=SYST01
 
Get whatever.distributed.file 'PFDT.TSREL.OPPLZDCK.PULLOUT.SUN'
 
SYST01 doesn't exist, but my SMS ACS routines point PFDT.** to a
 particular SG.  The get fails with:
 
PFDT.TSREL.OPPLZDCK.PULLOUT.SUN is on a direct access volume that
is
 not mounted and Noautomount is specified.
 Std Return Code = 16000, Error Code = 00018
 
Putting SYST01 in a DUMMY SG doesn't work, same error.  Doc sort of
 implies that it won't work.
 
I clipped a volume to SYST01 and it works fine.
 
I was trying to get this to work without changing the ftp client
 control cards.
 
 Thanks!
 BobL
 
 


--
 
 This e-mail transmission may contain information that is proprietary,
 privileged and/or confidential and is intended exclusively for the
 person(s) to whom it is addressed. Any use, copying, retention or
 disclosure by any person other than the intended recipient or the
intended
 recipient's designees is strictly prohibited. If you are not the
intended
 recipient or their designee, please notify the sender immediately by
 return e-mail and delete all copies. OppenheimerFunds may, at its sole
 discretion, monitor, review, retain and/or disclose the content of all
 email communications.


==
 
 
 --
 For IBM-MAIN subscribe / signoff / archive access instructions,
 send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
 Search the archives at http://bama.ua.edu/archives/ibm-main.html

--
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to lists...@bama.ua.edu with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html