This document lists various MVS abend codes and provides brief explanations for each. Codes range from S001 to S013-88 and cover issues like incorrect DD statements, I/O errors, invalid record lengths, incorrect DCB parameters, and other errors opening and processing datasets. The document suggests checking DCBs, DD statements, block sizes, record formats, and other attributes when encountering these abend codes.
The document provides an overview of utilities used in the IBM Z/OS mainframe operating system. It discusses the objectives and agenda of a training course on IBM utilities. The first session covers the introduction and types of utilities, including dataset utilities, system utilities, and access method services. Common dataset utilities like IEFBR14, IEBGENER, IEBCOPY, and SORT are introduced. The document provides examples of using IEFBR14 to create and delete datasets, and examples of using IEBCOPY and IEBGENER to copy datasets and work with partitioned dataset members.
Here are a few ways to specify a dummy VSAM dataset in JCL:
1. Specify DISP=(NEW,PASS) to allocate a new dummy VSAM dataset that will be deleted at the end of the job.
2. Specify DISP=(MOD,PASS) to use an existing dummy VSAM dataset if it already exists, or create a new one.
3. Specify AMP='AMORG' to indicate to the system that this is a VSAM file even though it may not physically exist. The system will treat it like a VSAM file for authorization checking.
4. Omit the DD statement entirely for files that are only used for authorization checking and not actually accessed by the program.
1) To access the TSO service at the University of Georgia, users must identify themselves by entering their USERID and password during the LOGON process.
2) The Interactive System Productivity Facility/Program Development Facility (ISPF/PDF) is a component of TSO that facilitates interaction with the TSO service. This guide describes how to use ISPF/PDF.
3) Key conventions used in this guide include showing commands in UPPERCASE and information the user provides in lowercase. "Next to" indicates the character position immediately to the right of the indicating arrow.
Top jcl interview questions and answers job interview tipsjcltutorial
You'll likely be asked difficult questions during the interview. Preparing the list of likely questions in advance will help you easily transition from question to question.
Here are the key steps to reorganize an HDAM IMS database:
1. Back up the database before starting the reorganization process. This provides a recovery point in case of errors.
2. Run the DBDGEN utility to generate a new DBD for the database. This will incorporate any schema changes.
3. Run the HD Reorganization Unload (DFSURGU0) utility to unload all segments from the database to a sequential file.
4. Run the HD Reorganization Reload (DFSURGL0) utility to reload the segments from the sequential file into a new database with the new DBD. This rebuilds the database in a more efficient structure.
5. Run
Maintec Technologies provides affordable access to latest Mainframe infrastructure for development, training, additional short term capacity increases and outsourcing needs. Leverage our Mainframe infrastructure, without investing the time, expense and manpower it would take to build your own!.
DB2 runs on 5 address spaces that each perform essential functions:
- DSNMSTR controls connections to other systems and performs logging, recovery, and system management.
- DSNDBM1 supports data definition, manipulation, and retrieval.
- IRLMPROC controls concurrent data access and maintains integrity through locking.
- DSNDIST enables remote access to distributed databases.
- DSNSPAS provides an isolated environment to execute stored procedures.
The document provides an overview of the initial program load (IPL) process in z/OS. It describes how the bootstrap loads the operating system from disk into memory and passes control to the IPL control program IEAIPL00. This program initializes the system environment and loads initialization programs that initialize system areas, address spaces, and subsystems to start the operating system. The document also discusses system parameters and parmlib members that control the initialization process.
The document provides an overview of utilities used in the IBM Z/OS mainframe operating system. It discusses the objectives and agenda of a training course on IBM utilities. The first session covers the introduction and types of utilities, including dataset utilities, system utilities, and access method services. Common dataset utilities like IEFBR14, IEBGENER, IEBCOPY, and SORT are introduced. The document provides examples of using IEFBR14 to create and delete datasets, and examples of using IEBCOPY and IEBGENER to copy datasets and work with partitioned dataset members.
Here are a few ways to specify a dummy VSAM dataset in JCL:
1. Specify DISP=(NEW,PASS) to allocate a new dummy VSAM dataset that will be deleted at the end of the job.
2. Specify DISP=(MOD,PASS) to use an existing dummy VSAM dataset if it already exists, or create a new one.
3. Specify AMP='AMORG' to indicate to the system that this is a VSAM file even though it may not physically exist. The system will treat it like a VSAM file for authorization checking.
4. Omit the DD statement entirely for files that are only used for authorization checking and not actually accessed by the program.
1) To access the TSO service at the University of Georgia, users must identify themselves by entering their USERID and password during the LOGON process.
2) The Interactive System Productivity Facility/Program Development Facility (ISPF/PDF) is a component of TSO that facilitates interaction with the TSO service. This guide describes how to use ISPF/PDF.
3) Key conventions used in this guide include showing commands in UPPERCASE and information the user provides in lowercase. "Next to" indicates the character position immediately to the right of the indicating arrow.
Top jcl interview questions and answers job interview tipsjcltutorial
You'll likely be asked difficult questions during the interview. Preparing the list of likely questions in advance will help you easily transition from question to question.
Here are the key steps to reorganize an HDAM IMS database:
1. Back up the database before starting the reorganization process. This provides a recovery point in case of errors.
2. Run the DBDGEN utility to generate a new DBD for the database. This will incorporate any schema changes.
3. Run the HD Reorganization Unload (DFSURGU0) utility to unload all segments from the database to a sequential file.
4. Run the HD Reorganization Reload (DFSURGL0) utility to reload the segments from the sequential file into a new database with the new DBD. This rebuilds the database in a more efficient structure.
5. Run
Maintec Technologies provides affordable access to latest Mainframe infrastructure for development, training, additional short term capacity increases and outsourcing needs. Leverage our Mainframe infrastructure, without investing the time, expense and manpower it would take to build your own!.
DB2 runs on 5 address spaces that each perform essential functions:
- DSNMSTR controls connections to other systems and performs logging, recovery, and system management.
- DSNDBM1 supports data definition, manipulation, and retrieval.
- IRLMPROC controls concurrent data access and maintains integrity through locking.
- DSNDIST enables remote access to distributed databases.
- DSNSPAS provides an isolated environment to execute stored procedures.
The document provides an overview of the initial program load (IPL) process in z/OS. It describes how the bootstrap loads the operating system from disk into memory and passes control to the IPL control program IEAIPL00. This program initializes the system environment and loads initialization programs that initialize system areas, address spaces, and subsystems to start the operating system. The document also discusses system parameters and parmlib members that control the initialization process.
The document describes various DB2 online utilities including UNLOAD, LOAD, REBUILD INDEX, COPY, RECOVER, RUNSTATS, MODIFY RECOVERY, QUIESCE, and REORG. These utilities perform functions like unloading and loading data, rebuilding indexes, taking image copies of data, recovering data to a prior point in time, updating catalog statistics, and reorganizing tablespaces.
SMP/E is a tool used to install and maintain software on z/OS systems. It controls changes at the element level, selecting proper levels of elements to install from potential changes and calling system utilities to install the changes. SMP/E manages this process through zones like the Global, Target, and Distribution zones. It uses terminology like SYSMODs, elements, and FMIDs to track changes and prerequisites for installation. The basic process flow involves receiving updates, applying them to the Target zone, testing, then accepting them to the Distribution zone.
1. The document discusses various IBM mainframe utility programs that are used to perform tasks related to scheduling, datasets, and systems. Some examples provided include IEFBR14, IEBCOMPR, IEBCOPY, IEBGENER, and IDCAMS.
2. Many IBM utilities were originally developed by users and then modified by IBM. They typically use common JCL statements like SYSIN, SYSUT1, SYSUT2, and SYSPRINT.
3. The utilities covered in the document include scheduler utilities, dataset utilities, and system utilities that can be used for tasks like copying, compressing, comparing, updating, and managing datasets and catalogs.
This slide contains all the basic concepts of ISPF. It's giving the simple and easy step to get the knowledge of Interactive system productivity facility. If u like it then give me feedback on email anilbharti85@gmail.com Thanks v much.
A K Bharti
This document lists 27 interview questions and answers about VSAM files. It covers topics like the different types of VSAM files (ESDS, KSDS, RRDS), using IDCAMS commands, defining VSAM files including alternate indexes and GDGs, loading and initializing VSAM files, and troubleshooting errors.
The document provides an overview of Job Control Language (JCL) used to communicate with the IBM mainframe operating system. It describes the key components of JCL including JOB, EXEC and DD statements. JOB statements name a job and supply accounting/scheduling information. EXEC statements call programs for execution and can invoke cataloged procedures. DD statements define resources like input/output files used by the job. The document outlines the format, fields and common parameters used in each JCL statement type.
This document provides documentation for z/OS DFSORT PTFs UQ95214 and UQ95213, which were released in December 2004. The PTFs provide several enhancements to DFSORT and ICETOOL, including conditionally reformatting records; overlaying only selected parts of records; using larger numeric fields and constants; new date/time formats; restarting sequence numbers; improved join/match operations; and more. The document highlights and provides examples of the new features, and details changed messages associated with the PTFs.
This document provides an introduction to using Job Control Language (JCL) and the System Display and Search Facility (SDSF) on IBM mainframe systems. It explains the basic components of JCL, including the JOB, EXEC, and DD statements. It also describes how to create JCL procedures, override procedure statements, and use utilities and system libraries. The document concludes by explaining how SDSF allows users to monitor jobs, view outputs, and issue commands to the operating system.
This document discusses various DB2 database objects and utilities. It provides descriptions of storage groups, databases, tablespaces, tables, indexes, views, and the utilities for unload, load, reorganization, running statistics, and copy. It includes examples of creating and using these objects and utilities.
This document provides tips for improving productivity in TSO/ISPF. It discusses logging on using different 3278 terminal types, starting ISPF directly from the logon command, customizing ISPF settings such as log and list defaults, and using the ISPF editor more efficiently with features like masks, bounds, and macros. The overall goals are to describe ways to set up ISPF for maximum efficiency and learn techniques for getting more out of the editor.
The document provides an example JCL used to run a COBOL program. It includes:
1) A JOB statement that identifies the job name, programmer, class, and priority.
2) A STEP statement specifying the COBOL program "COBPROG" to execute.
3) DD statements defining the input and output files for the COBOL program.
The document then explains each part of the JCL and how it will execute the COBOL program on the MVS operating system.
This PPT File, helps with the Basic Interview Questions specially for DataBase Domain.. For more questions , please log in to www.rekruitin.com
By ReKruiTIn.com
The document contains interview questions for JCL (Job Control Language). It includes questions about primary and secondary allocations for datasets, dispositions, extents, DD statements, concatenating datasets, temporary datasets, overriding condition codes, restarting jobs, skipping steps, overriding DD statements in procedures, abend codes, JOBLIB, STEPLIB, JES2 vs JES3, job control statements, keywords vs positional parameters, JOB, EXEC and DD statements.
Practical Recipes for Daily DBA Activities using DB2 9 and 10 for z/OSCuneyt Goksu
This document discusses several practical DBA activities in DB2 9 and 10 for z/OS including recovering from accidentally dropping a table, defining a trusted context for security, including columns in indexes for performance, creating indexes on expressions, and using MAXTEMP_RID in version 10 for performance. Steps are provided for recovering a dropped table using log records, archive logs, and VSAM copy techniques. Trusted contexts are introduced for efficiently switching users without credentials. Including columns in indexes and new features in version 10 like MAXTEMP_RID are highlighted for potential performance improvements.
Mainframe jcl exec and dd statements part - 3janaki ram
EXEC STATEMENT(EXEC)
EXEC Statement is used to identify program name or procedure name.
Maximum we can code 255 EXEC statements in a JOB.
In EXEC statement has two kinds of parameters
Position parameter Keyword parameter
PGM TIME , COND
PROC REGION , PARM
If you not mention any thing default it takes PROC
PGM
This is a positional parameter which must be coded after EXEC with one blank. This parameter indicates the name of program (or) procedure name
Syntax
//STEPNAME EXEC PGM=REPORT (OR)
PROC = PROCEDURE NAME
PARM
It is keyword parameter and it is mainly used for passing the data to another program.
To pass input to Application program
To invoke complier option.
This is a keyword parameter at activity / step level must be coded with exec operand after PGM parameter.
This parameter allows MAX(100) characters.
To receive PARM parameter data, Cobol program must be coded with “Procedure Division Using Parameter ” option.
These parameters must be declared in the linkage section.
DATA DESCRIPTOR STATEMENT ( DD STMT)
It is used to identify files( input and output ) used in JCL
DD name act as a bridge b/w Cobol program and execution JCL
In DD statement has two kinds of parameters.
Position Parameter Keyword Parameter
* DSN,DISP
DATA SPACE,UNIT
DUMMY DCB,VOLUME
*
Position parameter at DD level used with SYSIN to pass data to Cobol programs this is known as In stream data any number of records can be passed to the program.
In stream data is used to pass values from JCL to Cobol dynamically.
To accept the values in Cobol program, we should have equivalent accept verbs.
Syntax
//SYSIN DD * ------- Entry of in stream data
100
200
/* ---------------------- End of in stream data
DATA
We can pass any thing to another program including special character
Syntax
//SYSIN DD DATA
100
/*
200
/*
DUMMY
All files are treated as End of the file
Syntax
//SYSIN DD DUMMY
NOTE
There is no input to the application all input files treated as End of the file.
DSN
Through DSN we can refer temporary / permanent file
Is a physical space or file where the records will be stored.
DISP
DISP parameter is used to identify the mode of the file.
DISP= ( STATUS, NORMAL , ABNORMAL )
TERMINATION TERMINATION
NEW DELETE DELETE
OLD KEEP KEEP
SHR CATLG CATLG
MOD UNCATLG UNCATLG
NEW Creating the data set first time
OLD Already created accessing the data set all resources are allocated exclusively . No other user can until is released by the current user. If dataset is not existing this creates.
SHR The data set can be accessed by multiple users at a time.
MOD Appending the records to the existing record
NOTE If the dataset is not existing , then MOD is assumed as NEW and writes records to the dataset.
Contains information about the DB2 DSNZPARM that forms the DB2 configuration parameters. All about the different types of zPARMs. A way to update them dynamically.
DB2 is a multi-platform database server that can scale from laptops to large systems handling terabytes of data. It provides tools for extending capabilities to support multimedia, is fully integrated for web access, and supports universal access and multiple platforms. The tutorial covered key DB2 concepts like instances, schemas, tables, and indexes. It demonstrated how to use Control Center and other GUIs to perform tasks like creating databases and tables, querying data, and setting user privileges. Java applications can also access DB2 data through JDBC.
The document discusses implementing a Parallel Sysplex which couples multiple z/OS systems together using hardware and software services. Key steps include defining coupling facility structures, configuring XCF signalling paths using CTCs or a coupling facility, formatting and configuring sysplex couple data sets, and defining CFRM policies to manage coupling facility resources.
DB2 is a relational database management system that runs on IBM mainframes. It uses SQL for data manipulation and definition. A COBOL program can use DB2 services by including host variables, a SQL communication area, and SQL statements. DB2 has major components including system services, locking services, database services, and distributed data facility. The database services component handles tasks like precompiling, binding, running SQL statements, data management, and buffer management.
IEBCOPY is a utility that can copy members between PDS datasets. It allows copying all members, selected members, or excluding specific members. It can also replace or rename members during the copy process. The utility can also compress a PDS dataset to free unused space. Example JCL is provided showing how to copy all members, select specific members, exclude members, replace and rename members, and compress a dataset.
The document discusses several methods for storing SAS data into Oracle tables, including PROC DBLOAD, the DATA step, PROC SQL, and PROC APPEND. It compares the performance of these different methods and describes how the INSERTBUFF option and BULKLOAD option can be used to optimize performance when loading large amounts of data. The BULKLOAD option utilizes Oracle's SQL*Loader utility to perform direct path loads for significantly faster loading compared to the other transactional methods.
The document describes various DB2 online utilities including UNLOAD, LOAD, REBUILD INDEX, COPY, RECOVER, RUNSTATS, MODIFY RECOVERY, QUIESCE, and REORG. These utilities perform functions like unloading and loading data, rebuilding indexes, taking image copies of data, recovering data to a prior point in time, updating catalog statistics, and reorganizing tablespaces.
SMP/E is a tool used to install and maintain software on z/OS systems. It controls changes at the element level, selecting proper levels of elements to install from potential changes and calling system utilities to install the changes. SMP/E manages this process through zones like the Global, Target, and Distribution zones. It uses terminology like SYSMODs, elements, and FMIDs to track changes and prerequisites for installation. The basic process flow involves receiving updates, applying them to the Target zone, testing, then accepting them to the Distribution zone.
1. The document discusses various IBM mainframe utility programs that are used to perform tasks related to scheduling, datasets, and systems. Some examples provided include IEFBR14, IEBCOMPR, IEBCOPY, IEBGENER, and IDCAMS.
2. Many IBM utilities were originally developed by users and then modified by IBM. They typically use common JCL statements like SYSIN, SYSUT1, SYSUT2, and SYSPRINT.
3. The utilities covered in the document include scheduler utilities, dataset utilities, and system utilities that can be used for tasks like copying, compressing, comparing, updating, and managing datasets and catalogs.
This slide contains all the basic concepts of ISPF. It's giving the simple and easy step to get the knowledge of Interactive system productivity facility. If u like it then give me feedback on email anilbharti85@gmail.com Thanks v much.
A K Bharti
This document lists 27 interview questions and answers about VSAM files. It covers topics like the different types of VSAM files (ESDS, KSDS, RRDS), using IDCAMS commands, defining VSAM files including alternate indexes and GDGs, loading and initializing VSAM files, and troubleshooting errors.
The document provides an overview of Job Control Language (JCL) used to communicate with the IBM mainframe operating system. It describes the key components of JCL including JOB, EXEC and DD statements. JOB statements name a job and supply accounting/scheduling information. EXEC statements call programs for execution and can invoke cataloged procedures. DD statements define resources like input/output files used by the job. The document outlines the format, fields and common parameters used in each JCL statement type.
This document provides documentation for z/OS DFSORT PTFs UQ95214 and UQ95213, which were released in December 2004. The PTFs provide several enhancements to DFSORT and ICETOOL, including conditionally reformatting records; overlaying only selected parts of records; using larger numeric fields and constants; new date/time formats; restarting sequence numbers; improved join/match operations; and more. The document highlights and provides examples of the new features, and details changed messages associated with the PTFs.
This document provides an introduction to using Job Control Language (JCL) and the System Display and Search Facility (SDSF) on IBM mainframe systems. It explains the basic components of JCL, including the JOB, EXEC, and DD statements. It also describes how to create JCL procedures, override procedure statements, and use utilities and system libraries. The document concludes by explaining how SDSF allows users to monitor jobs, view outputs, and issue commands to the operating system.
This document discusses various DB2 database objects and utilities. It provides descriptions of storage groups, databases, tablespaces, tables, indexes, views, and the utilities for unload, load, reorganization, running statistics, and copy. It includes examples of creating and using these objects and utilities.
This document provides tips for improving productivity in TSO/ISPF. It discusses logging on using different 3278 terminal types, starting ISPF directly from the logon command, customizing ISPF settings such as log and list defaults, and using the ISPF editor more efficiently with features like masks, bounds, and macros. The overall goals are to describe ways to set up ISPF for maximum efficiency and learn techniques for getting more out of the editor.
The document provides an example JCL used to run a COBOL program. It includes:
1) A JOB statement that identifies the job name, programmer, class, and priority.
2) A STEP statement specifying the COBOL program "COBPROG" to execute.
3) DD statements defining the input and output files for the COBOL program.
The document then explains each part of the JCL and how it will execute the COBOL program on the MVS operating system.
This PPT File, helps with the Basic Interview Questions specially for DataBase Domain.. For more questions , please log in to www.rekruitin.com
By ReKruiTIn.com
The document contains interview questions for JCL (Job Control Language). It includes questions about primary and secondary allocations for datasets, dispositions, extents, DD statements, concatenating datasets, temporary datasets, overriding condition codes, restarting jobs, skipping steps, overriding DD statements in procedures, abend codes, JOBLIB, STEPLIB, JES2 vs JES3, job control statements, keywords vs positional parameters, JOB, EXEC and DD statements.
Practical Recipes for Daily DBA Activities using DB2 9 and 10 for z/OSCuneyt Goksu
This document discusses several practical DBA activities in DB2 9 and 10 for z/OS including recovering from accidentally dropping a table, defining a trusted context for security, including columns in indexes for performance, creating indexes on expressions, and using MAXTEMP_RID in version 10 for performance. Steps are provided for recovering a dropped table using log records, archive logs, and VSAM copy techniques. Trusted contexts are introduced for efficiently switching users without credentials. Including columns in indexes and new features in version 10 like MAXTEMP_RID are highlighted for potential performance improvements.
Mainframe jcl exec and dd statements part - 3janaki ram
EXEC STATEMENT(EXEC)
EXEC Statement is used to identify program name or procedure name.
Maximum we can code 255 EXEC statements in a JOB.
In EXEC statement has two kinds of parameters
Position parameter Keyword parameter
PGM TIME , COND
PROC REGION , PARM
If you not mention any thing default it takes PROC
PGM
This is a positional parameter which must be coded after EXEC with one blank. This parameter indicates the name of program (or) procedure name
Syntax
//STEPNAME EXEC PGM=REPORT (OR)
PROC = PROCEDURE NAME
PARM
It is keyword parameter and it is mainly used for passing the data to another program.
To pass input to Application program
To invoke complier option.
This is a keyword parameter at activity / step level must be coded with exec operand after PGM parameter.
This parameter allows MAX(100) characters.
To receive PARM parameter data, Cobol program must be coded with “Procedure Division Using Parameter ” option.
These parameters must be declared in the linkage section.
DATA DESCRIPTOR STATEMENT ( DD STMT)
It is used to identify files( input and output ) used in JCL
DD name act as a bridge b/w Cobol program and execution JCL
In DD statement has two kinds of parameters.
Position Parameter Keyword Parameter
* DSN,DISP
DATA SPACE,UNIT
DUMMY DCB,VOLUME
*
Position parameter at DD level used with SYSIN to pass data to Cobol programs this is known as In stream data any number of records can be passed to the program.
In stream data is used to pass values from JCL to Cobol dynamically.
To accept the values in Cobol program, we should have equivalent accept verbs.
Syntax
//SYSIN DD * ------- Entry of in stream data
100
200
/* ---------------------- End of in stream data
DATA
We can pass any thing to another program including special character
Syntax
//SYSIN DD DATA
100
/*
200
/*
DUMMY
All files are treated as End of the file
Syntax
//SYSIN DD DUMMY
NOTE
There is no input to the application all input files treated as End of the file.
DSN
Through DSN we can refer temporary / permanent file
Is a physical space or file where the records will be stored.
DISP
DISP parameter is used to identify the mode of the file.
DISP= ( STATUS, NORMAL , ABNORMAL )
TERMINATION TERMINATION
NEW DELETE DELETE
OLD KEEP KEEP
SHR CATLG CATLG
MOD UNCATLG UNCATLG
NEW Creating the data set first time
OLD Already created accessing the data set all resources are allocated exclusively . No other user can until is released by the current user. If dataset is not existing this creates.
SHR The data set can be accessed by multiple users at a time.
MOD Appending the records to the existing record
NOTE If the dataset is not existing , then MOD is assumed as NEW and writes records to the dataset.
Contains information about the DB2 DSNZPARM that forms the DB2 configuration parameters. All about the different types of zPARMs. A way to update them dynamically.
DB2 is a multi-platform database server that can scale from laptops to large systems handling terabytes of data. It provides tools for extending capabilities to support multimedia, is fully integrated for web access, and supports universal access and multiple platforms. The tutorial covered key DB2 concepts like instances, schemas, tables, and indexes. It demonstrated how to use Control Center and other GUIs to perform tasks like creating databases and tables, querying data, and setting user privileges. Java applications can also access DB2 data through JDBC.
The document discusses implementing a Parallel Sysplex which couples multiple z/OS systems together using hardware and software services. Key steps include defining coupling facility structures, configuring XCF signalling paths using CTCs or a coupling facility, formatting and configuring sysplex couple data sets, and defining CFRM policies to manage coupling facility resources.
DB2 is a relational database management system that runs on IBM mainframes. It uses SQL for data manipulation and definition. A COBOL program can use DB2 services by including host variables, a SQL communication area, and SQL statements. DB2 has major components including system services, locking services, database services, and distributed data facility. The database services component handles tasks like precompiling, binding, running SQL statements, data management, and buffer management.
IEBCOPY is a utility that can copy members between PDS datasets. It allows copying all members, selected members, or excluding specific members. It can also replace or rename members during the copy process. The utility can also compress a PDS dataset to free unused space. Example JCL is provided showing how to copy all members, select specific members, exclude members, replace and rename members, and compress a dataset.
The document discusses several methods for storing SAS data into Oracle tables, including PROC DBLOAD, the DATA step, PROC SQL, and PROC APPEND. It compares the performance of these different methods and describes how the INSERTBUFF option and BULKLOAD option can be used to optimize performance when loading large amounts of data. The BULKLOAD option utilizes Oracle's SQL*Loader utility to perform direct path loads for significantly faster loading compared to the other transactional methods.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Various Open Source Cryptographic Libraries are being used these days to implement the
general purpose cryptographic functions and to provide a secure communication channel over
the internet. These libraries, that implement SSL/TLS, have been targeted by various side
channel attacks in the past that result in leakage of sensitive information flowing over the
network. Side channel attacks rely on inadvertent leakage of information from devices
through observable attributes of online communication. Some of the common side channel
attacks discovered so far rely on packet arrival and departure times (Timing Attacks), power
usage and packet sizes. Our research explores novel side channel attack that relies on CPU
architecture and instruction sets. In this research, we explored such side channel vectors
against popular SSL/TLS implementations which were previously believed to be patched
against padding oracle attacks, like the POODLE attack. We were able to successfully extract
the plaintext bits in the information exchanged using the APIs of two popular SSL/TLS
libraries.
Accelerate Reed-Solomon coding for Fault-Tolerance in RAID-like systemShuai Yuan
The document discusses accelerating Reed-Solomon erasure codes on GPUs. It aims to accelerate two main computation bottlenecks: arithmetic operations in Galois fields and matrix multiplication. For Galois field operations, it evaluates loop-based and table-based methods and chooses a log-exponential table approach. It also proposes tiling algorithms to optimize matrix multiplication on GPUs by reducing data transfers and improving memory access patterns. The goal is to make Reed-Solomon encoding and decoding faster for cloud storage systems using erasure codes.
Drivers connect applications to Cassandra clusters and maintain connections to nodes. They probe clusters to discover nodes, token ranges, and latency. Drivers are data-aware and can route queries to appropriate replicas or fail over if needed. Cassandra clusters can span multiple data centers for redundancy, workload separation, and geographic distribution of data and queries. Configuration files like cassandra.yaml and cassandra-env.sh are used to configure memory, data storage, caching, and other settings. Cassandra clusters should be provisioned on commodity servers using tools like cassandra-stress to test workloads and estimate needed nodes.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
International Journal of Engineering Research and Development is an international premier peer reviewed open access engineering and technology journal promoting the discovery, innovation, advancement and dissemination of basic and transitional knowledge in engineering, technology and related disciplines.
The document proposes adding memory channel testing to the OCP certification process to improve reliability and address issues like row hammer failures. It outlines conducting an electrical audit and protocol timing audit using a logic analyzer to qualitatively check signals and timing specifications are met. The goal is not comprehensive validation but a low-cost audit. Tests would check eyes, bursts, and for JEDEC spec compliance over an hour. Results would document for each server, slot, and channel. The testing also aims to investigate the row hammer failure mechanism and potential mitigation strategies.
1z0 034 exam-upgrade oracle9i10g oca to oracle database 11g ocpIsabella789
Our Practice exams ensures success on the 1Z0-034 Exam-Upgrade Oracle9i/10g OCA to Oracle Database 11g OCP visit@ https://www.troytec.com/1Z0-034-exams.html
The document provides an overview of the architecture and features of the TMS320C6713 digital signal processor. It describes the central processing unit, internal memory, general purpose register files, functional units, and peripheral options. The document is a user manual that contains multiple chapters, with each chapter providing examples and explanations of different aspects of the C6713 architecture and software development for the processor.
This document introduces ABAP programming concepts for reading database tables using SELECT statements in Open SQL. It discusses using SELECT to retrieve single records or multiple rows of data, filtering data with WHERE clauses, sorting results with ORDER BY, and working with system variables. Formal documentation methods are also introduced.
This document provides an agenda and introduction for a presentation on Apache Cassandra and DataStax Enterprise. The presentation covers an introduction to Cassandra and NoSQL, the CAP theorem, Apache Cassandra features and architecture including replication, consistency levels and failure handling. It also discusses the Cassandra Query Language, data modeling for time series data, and new features in DataStax Enterprise like Spark integration and secondary indexes on collections. The presentation concludes with recommendations for getting started with Cassandra in production environments.
This document introduces the theory behind geological database processes and provides detailed
examples using the geological database modelling functions in Surpac. By working through this
tutorial you will gain skills in the creation, use, and modification of geological databases.
The document discusses two options for achieving high availability for Oracle Database Standard Edition 2 (SE2):
1) Standard Edition High Availability (SEHA) provides an out-of-the-box failover cluster configuration using Oracle Grid Infrastructure that supports automatic failover between two nodes.
2) Using refreshable pluggable databases (PDBs) allows cloning a PDB from a primary database to a secondary database for read-only reporting or to refresh the secondary PDB periodically to propagate changes.
SecureFile LOBs provide efficient, secure storage of large objects like multimedia files in Oracle databases. They allow features like deduplication, encryption, and compression. A SecureFile LOB stores data in chunks within a LOB segment. The database manages read consistency using persistent undo instead of transaction rollback. DBAs can use the DBMS_SPACE and DBMS_LOBUTIL packages to view SecureFile LOB space usage and chunk mappings.
Level sensitive scan design(LSSD) and Boundry scan(BS)Praveen Kumar
This presentation contains,
Introduction,design for testability, scan chain, operation, scan structure, test vectors, Boundry scan, test logic, operation, BS cell, states of TAP controller, Boundry scan instructions.
Here are some useful GDB commands for debugging:
- break <function> - Set a breakpoint at a function
- break <file:line> - Set a breakpoint at a line in a file
- run - Start program execution
- next/n - Step over to next line, stepping over function calls
- step/s - Step into function calls
- finish - Step out of current function
- print/p <variable> - Print value of a variable
- backtrace/bt - Print the call stack
- info breakpoints/ib - List breakpoints
- delete <breakpoint#> - Delete a breakpoint
- layout src - Switch layout to source code view
- layout asm - Switch layout
Exam 1z0 062 Oracle Database 12c: Installation and AdministrationKylieJonathan
Training increase your chance of passing Oracle Database 12c: Installation and Administration exam. Examarea.com
provide you the updated study dumps for 1Z1-062 exam.
@ https://www.examarea.com/1Z1-062-exams.html
#IT #oracle #database #training #dumps #discount
DFSORT has many features beyond just sorting that can improve application performance and programmer productivity. These include ICEGENER for faster copying, OUTFIL for creating multiple outputs in one pass, and ICETOOL for performing complex tasks easily through a batch front-end. Examples show these features can reduce elapsed time by 85-89%, CPU time by 13-30%, and EXCP count by 89-92% compared to traditional approaches.
This document provides a mini-user guide for DFSORT's ICETOOL utility. It discusses the major features of ICETOOL for z/OS DFSORT V1R5 and DFSORT Release 14, including its JCL, control statements, and thirteen operators. Examples are provided to demonstrate how each operator works and how ICETOOL can be used to perform complex sorting, copying, reporting and analytical tasks using multiple data sets in a single job step.
DFSORT V1R5 and DFSORT R14 provide enhancements for larger fields and constants, new free form formats, and new features for the ICETOOL utility. The SPLICE operator allows joining fields from different input records based on a common key. New formats like SFF and UFF allow extracting numeric values from free form fields. COUNT can set a return code to control downstream job steps based on the number of records in a data set.
This document provides summaries of common IBM OS/VS system error messages and ABEND codes, including descriptions of what caused the error and suggestions for correcting it. Specific ABEND codes covered include 013, 0C1, 0C4, 0C5, 0C7, 106, 213, 222, 2F3, 322, 613, 722, 804 and more. Each error code section includes the associated error ID, description of the problem, and recommended corrective procedure.
This document provides an introduction to JCL (Job Control Language) including:
1. What JCL is and its basic syntax with JOB, EXEC, and DD statements.
2. The JOB statement identifies the start of a job. The EXEC statement identifies the program or procedure to run. The DD statement defines input/output resources.
3. Examples are provided for DD statements to define disk and tape datasets as well as instream data. Formats, dispositions, and other attributes can be specified.
This document provides an overview of the COBOL programming language. It discusses the history and structure of COBOL, including its divisions such as identification, environment, data, and procedure. It also describes COBOL coding conventions and basics such as literals, constants, variable declaration, and data types. Key points covered include COBOL being the first language for commercial applications, its English-like syntax, and use of four program divisions.
This document provides a user manual for COBOL. It begins with an introduction to COBOL including its language construction, character set, words, coding form, and divisions. It then discusses each division in detail: the Identification Division, Environment Division including its configuration and input-output sections, Data Division including data description and clauses, and Procedure Division including statements. It also covers topics like file handling, subprograms, and intrinsic functions. The document aims to help users understand the standard format of a COBOL program and write programs using features like VSAM files, modular design, and compiler options.
A beginner's guide to cobol programmingNirmal Pati
This document provides an introduction and overview of the COBOL programming language. It discusses the four main divisions of a COBOL program: identification, environment, data, and procedure. The identification division provides metadata about the program. The environment division defines any external files or devices. The data division allocates memory space for data. The procedure division contains the main logic of the program broken up into paragraphs. It also provides a simple "Hello World" example program to demonstrate the basic structure of a COBOL program.
The document lists many keyboard shortcuts for Windows and common Microsoft programs like Word, Excel, PowerPoint and Outlook. Some general shortcuts include CTRL+C to copy, CTRL+X to cut, and CTRL+V to paste. Navigation shortcuts include arrow keys to move the cursor and CTRL+TAB to switch between open programs. Dialog box shortcuts include TAB and SHIFT+TAB to move between options and ALT+underlined letter to select options.
The document describes the COBOL compilation process. The COBOL compiler prepares source programs for execution by producing an object module containing machine language instructions. This object module must then be processed by a linkage editor to be ready for execution. The compiler can compile single or multiple source programs at once, producing one or more object or load modules. It also provides debugging information, user exits, and options to control various aspects of compilation.
Main Java[All of the Base Concepts}.docxadhitya5119
This is part 1 of my Java Learning Journey. This Contains Custom methods, classes, constructors, packages, multithreading , try- catch block, finally block and more.
How to Make a Field Mandatory in Odoo 17Celine George
In Odoo, making a field required can be done through both Python code and XML views. When you set the required attribute to True in Python code, it makes the field required across all views where it's used. Conversely, when you set the required attribute in XML views, it makes the field required only in the context of that particular view.
How to Manage Your Lost Opportunities in Odoo 17 CRMCeline George
Odoo 17 CRM allows us to track why we lose sales opportunities with "Lost Reasons." This helps analyze our sales process and identify areas for improvement. Here's how to configure lost reasons in Odoo 17 CRM
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
বাংলাদেশের অর্থনৈতিক সমীক্ষা ২০২৪ [Bangladesh Economic Review 2024 Bangla.pdf] কম্পিউটার , ট্যাব ও স্মার্ট ফোন ভার্সন সহ সম্পূর্ণ বাংলা ই-বুক বা pdf বই " সুচিপত্র ...বুকমার্ক মেনু 🔖 ও হাইপার লিংক মেনু 📝👆 যুক্ত ..
আমাদের সবার জন্য খুব খুব গুরুত্বপূর্ণ একটি বই ..বিসিএস, ব্যাংক, ইউনিভার্সিটি ভর্তি ও যে কোন প্রতিযোগিতা মূলক পরীক্ষার জন্য এর খুব ইম্পরট্যান্ট একটি বিষয় ...তাছাড়া বাংলাদেশের সাম্প্রতিক যে কোন ডাটা বা তথ্য এই বইতে পাবেন ...
তাই একজন নাগরিক হিসাবে এই তথ্য গুলো আপনার জানা প্রয়োজন ...।
বিসিএস ও ব্যাংক এর লিখিত পরীক্ষা ...+এছাড়া মাধ্যমিক ও উচ্চমাধ্যমিকের স্টুডেন্টদের জন্য অনেক কাজে আসবে ...
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
This slide is special for master students (MIBS & MIFB) in UUM. Also useful for readers who are interested in the topic of contemporary Islamic banking.
This document provides an overview of wound healing, its functions, stages, mechanisms, factors affecting it, and complications.
A wound is a break in the integrity of the skin or tissues, which may be associated with disruption of the structure and function.
Healing is the body’s response to injury in an attempt to restore normal structure and functions.
Healing can occur in two ways: Regeneration and Repair
There are 4 phases of wound healing: hemostasis, inflammation, proliferation, and remodeling. This document also describes the mechanism of wound healing. Factors that affect healing include infection, uncontrolled diabetes, poor nutrition, age, anemia, the presence of foreign bodies, etc.
Complications of wound healing like infection, hyperpigmentation of scar, contractures, and keloid formation.
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
How to Fix the Import Error in the Odoo 17Celine George
An import error occurs when a program fails to import a module or library, disrupting its execution. In languages like Python, this issue arises when the specified module cannot be found or accessed, hindering the program's functionality. Resolving import errors is crucial for maintaining smooth software operation and uninterrupted development processes.
Leveraging Generative AI to Drive Nonprofit InnovationTechSoup
In this webinar, participants learned how to utilize Generative AI to streamline operations and elevate member engagement. Amazon Web Service experts provided a customer specific use cases and dived into low/no-code tools that are quick and easy to deploy through Amazon Web Service (AWS.)
Strategies for Effective Upskilling is a presentation by Chinwendu Peace in a Your Skill Boost Masterclass organisation by the Excellence Foundation for South Sudan on 08th and 09th June 2024 from 1 PM to 3 PM on each day.
हिंदी वर्णमाला पीपीटी, hindi alphabet PPT presentation, hindi varnamala PPT, Hindi Varnamala pdf, हिंदी स्वर, हिंदी व्यंजन, sikhiye hindi varnmala, dr. mulla adam ali, hindi language and literature, hindi alphabet with drawing, hindi alphabet pdf, hindi varnamala for childrens, hindi language, hindi varnamala practice for kids, https://www.drmullaadamali.com
1. MVS Abend Codes Page 1 of 13
MVS Abend Codes
1. QSAM FILE STATUS 90 Check the DD statement and the SELECT.
2. QSAM FILE STATUS 91 Attempting to process a closed file.
3. QSAM FILE STATUS 92 Probable user error. Attempting to process a closed file.
4. S001 Input record described incorrectly, length error, tried to read after encountering end-of-file, RECFM V is incompatible with
TRTCH ET, tried to write on an input file, read into a table whose subscript was zero, incorrect label parameter, tape being read in
wrong density, PDS directory is overwritten, NSL label had incorrect DSNAME, operator cancelled, true I/O failure, UCB show re-
dundnacy, harware error - bad tape or tape drive, secondary space allocation for DASD output file too small (see SP37), in COBOL:
incorrect use of BLOCK CONTAINS CLAUSE or had a RECORDING MODE F when generated, but not when read.
On DCB, add ',erropt=skp)' to skip a block with bad I/O. Check DCB in JCL against FD statement; In concatenation, largest blocksize
must be first. Verify BLOCK CONTAINS 0; verify FD. SUGGESTION:
5. S001-01 An I/O error occurred during BDAM create, BSAM, BISAM, QSAM, or BDAM and no SYNAD exit was specified by the
user.
6. S001-02 An error was encountered while attempting to close the dataset.
7. S001-03 For QSAM, an error was encountered that could not be accepted.
8. S001-04
For QSAM, ABE or an invalid value for EROPT parameter in the DCB and/or no error handling (SYNAD) exit was specified.
9. S001-05 For QSAM, a GET was issued after end-of-file.
10. S002 DCB had the wrong block size, IEHMOVE-attempt to rename dataset when new name already exists, wrong RECORD format
specified in JCL.
11. S002-04 An invalid record was encountered on a GET operation. The length field of a record descriptor word (RDW) for a variable
length record specifies a value less than 4.
12. S002-08 An invalid record was encountered on a PUT or WRITE operation. The record or block length plus the key length and
required overhead is greater than the DASD track capacity.
13. S002-10 An invalid record was encountered on a PUT or WRITE operation. The dataset uses the track overflow feature. The RDW for
a variable length record specifies a length greater than 32,752.
14. S002-14
An invalid record was encountered on a PUT or WRITE operation. The record length is greater than the blocksize specified in
the DCB.
15. S002-18 An invalid record was encountered on a PUT operation; the dataset uses the variable record format. The length value of the
RDW is either: less than 4, greater than 32,767, greater than the blocksize specified in the DCB, or less than 5 if ASA control
characters are being used.
16. S002-1C The error was issued for a record larger than the track capacity, and the record format does not use the track overflow feature.
17. S002-20 The error occurred during the creation of a direct dataset. A write macro was issued causing a secondary extent to be obtained.
The block will not fit on the amount of space allocated for the secondary extent.
18. S002-24 The error occurred during the creation of a direct dataset. A write macro was issued to write out a block larger than the
primary extent on the preallocated dataset. This could also occur if allocation of the primary extent as non-contiguous and any of the
secondary extents were smaller than the block.
19. S002-28 The error occurred during the creation of a direct dataset. During the execution of open it was detected that the blocksize was
larger than the primary extent.
20. S002-2C The error occurred while opening an ISAM dataset. Too many tracks were specified for cylinder overflow.
21. S013 Conflicting or incomplete parameters in DCB, such as BLKSIZE not a multiple of LRECL, or missing SYSIN DD; Tried to
create a PDS without allocating directory blocks; Member name specified in DD not found; no directory allocation subparameter in
DD; Open output dataset as input; Track overflow or updating attempted, but not supported by the OS. May be a record length or
OPEN statement error.
22. S013-08 ANSI standard labels were specified in the LABEL parameter but are not supported by the system.
23. S013-0C A buffer length of 0 was specified for a BDAM dataset for which dynamic buffering was requested.
24. S013-10 An OPEN macro was issued for a null dataset and blocksize and BUFL are both 0. Specify one or the other as non-zero.
25. S013-14 An OPEN macro was issued with output or outin specified. DCB specifies DSORG=PO, but the DSCB indicates the dataset is
not partitioned. Change DSORG to PS, or create dataset as PDS.
26. S013-18 An open was issued for a partitioned dataset. The specified member name was not found in the dataset.
27. S013-1C An OPEN macro was issued for a partitioned dataset, but an I/O error was encountered searching the directory.
28. S013-20 An OPEN macro was issued for a sequential dataset using the queued access technique with RECFM=FB, but blocksize is not
a multiple of LRECL, or for variable-length records, blocksize is not 4 bytes greater than the LRECL.
29. S013-24 An open was issued with input, inout, RDBACK, or updat specified, but the DCB MACRF did not specify EXCP, GET, or
READ.
30. S013-28 An OPEN macro was issued with output or outin specified, but the DCB MACRF did not specify EXCP, PUT, or WRITE.
31. S013-2C A sequential dataset using the queued access technique with exchange buffering was opened for input, but the buffer control
block address was 0.
08/07/20111
2. MVS Abend Codes Page 2 of 13
32. S013-30 A sequential dataset using the queued access technique with exchange buffering was opened for output, but the buffer control
block address was 0.
33. S013-34 An OPEN macro was issued for a dataset with blocksize and BUFL equal to 0. The system determined that it had to obtain
buffers but was unable to do so
34. S013-38 An OPEN macro was issued for a sequential dataset on a direct access device with track overflow, but the buffer control block
address was zero.
35. S013-3C A sequential dataset was opened for input or output, but the buffer control block address was 0. This type of error often
occurs is a DCB is shared by two or more tasks, or is opened and closed several times within one job step.
36. S013-40 A sequential or direct dataset was opened for input, but the buffer control block address was 0.
37. S013-44 An OPEN macro was issued for a dataset on a direct access device for which chained scheduling was specified, but the buffer
control block address was 0.
38. S013-48 An OPEN macro was issued for a sequential dataset using the queued access technique, but the buffer control block address
was zero.
39. S013-4C An OPEN macro was issued for a sequential dataset using the queued access technique. The system determined that a buffer
pool existed for this dataset and made the appropriate test shown below with unsatisfactory results: If the data was to be sent directly to
a unit record device (no spooling), the buffer length value in the buffer control block had to be equal or greater than the value specified
in the DCB for logical record length. Otherwise the buffer length in the buffer control block had to be equal or greater than the value
specified in the DCB for BLKSIZE.
40. S013-50 An OPEN macro issued for a dataset allocated to a printer did not have output specified as an open option.
41. S013-58 An OPEN macro was issued for a paper tape dataset and concatonation with unlike attributes was specified.
42. S013-5C An OPEN macro was issued for a sequential dataset using the queued access technique. The dataset contained spanned
variable length records larger than 32,756 but GET locate mode was not used.
43. S013-60 An OPEN macro was issued for a dataset with a DCB specifying RECFM=F, and blocksize was not equal to LRECL. Correct
the RECFM to FB, or make LRECL and blocksize equal.
44. S013-64 An OPEN macro was issued for a null dataset using an access method other than QSAM or BSAM. This is a no-no.
45. S013-68 An OPEN macro was issued for a dataset whose DCB specified a BLKSIZE greater than 32,767, which is the maximum
allowed.
46. S013-70 An OPEN macro was issued for a dataset on magnetic tape. A conflict exists among LABEL parameters on the DD statement
and DCBRECFM, DCBOPTCD, DCBBUFOF and DCBUSASI give the appearance of mixed USASI and EBCDIC attributes to the
dataset; or OPTCD=q was specified for a dataset on a devise other than magnetic tape.
47. S013-74 An OPEN macro was issued for an optical character reader dataset, but the open option did not specify INPUT.
48. S013-78 An OPEN macro was issued for an optical character reader dataset, but the BUFL parameter was specified as 0.
49. S013-7C An OPEN macro was issued for an optical character reader dataset, but the BUFL parameter in the DCB was specified as 0.
50. S013-80 An OPEN macro was issued for an optical character reader dataset, but the LRECL parameter was less than the BUFL.
51. S013-84 An OPEN macro was issued for an optical character reader dataset. The number of buffers specified in the buffer pool control
block is not the same as that specified in the DCBBUFNO.
52. S013-88 An OPEN macro was issued for a telecommunications device but the DCBDSORG did not specify so.
53. S013-8C An OPEN macro was issued for a direct organization dataset (BDAM) but the DCBRECFM parameter was not specific.
54. S013-A4 A DCB was opened for a SYSIN or SYSOUT dataset but the DSORG was not specified as PS.
55. S013-A8 An invalid record format was requested for a SYSIN or SYSOUT dataset. (RECFM=D, vs, or VBS are Invalid for SYSIN)
56. S013-B0 An OPEN macro was issued with the Reback option for a DCB specifying a record format of variable spanned records. These
are conflicting parameters.
57. S013-B4 An OPEN macro was issued with the INOUT/OUTIN option for a DCB specifying the QSAM MACRF values; these are
conflicting parameters.
58. S013-BC A SYSIN or SYSOUT DCB was opened with invalid options. Either open specified UPDAT or RDBACK, or the point
macro function was specified in MACRF=RP or WP. Repositioning or updating a spooled dataset is not permitted.
59. S013-C0 A SYSIN or SYSOUT Dataset could not be opened by a job entry subsystem. The failing DCB is not opened, however the
task is not terminated. Processing continues for other DCBS opened.
60. S013-C4 During the creation of a direct dataset, the open routine found that the DCB specified READ (R) or GET (G) in the MACRG
field. Only WRITE LOAD (WL) is allowed.
61. S013-C8 The open subsystem executor module was passed an error return code in register 15 after issuing the IEFSSREQ macro to
connect the user's ACB to the subsystem. This indicates the subsystem was not operating.
62. S028 A paging operation has not completed successfully due to one of the following:
-A permanent I/O error occurred while attempting a page-in or swap-in operation. The data being paged in or swapped in is lost.
-A real storage management routine or another system routine performing a service for RSM suffered an intermediate error. The
function performed is terminated.
-An auxiliary storage management routine suffered a translation error while using the control register of another address space to
update that address space's LSQA.
63. SO3B ISAM dataset to be processed, but not created or its DCB not closed after creation. - Possible cause: dataset was to be written on
but was opened as input. LRECL/BLKSIZE problem exists-invalid values or not multiples.
64. SO3D The error occurred during the execution of a QISAM or BISAM or BDAM macro instructions.
65. S03D-04 An OPEN macro was issued for an indexed sequential or direct dataset. The volume serial numbers on the DD statement were
not specified in the same order that the dataset was created.
66. S03D-08 An OPEN macro was issued for an indexed sequential dataset. The first volume of the dataset does not have a format2 DSCB.
67. S03D-0C An OPEN macro was issued for a direct dataset. The dataset has an indexed sequential organization.
68. S047 An unauthorized program requested a restricted SVC.
69. S071 The operator pressed the restart key to activate the system's recovery and termination process. The program running at the time
08/07/20111
3. MVS Abend Codes Page 3 of 13
the operator pressed the restart button was sent through abend processing because the operator determined it was a noncancelable loop
or wait state.
70. S07C Supervisor control recovery has detected an error which requires that the current task be terminated or that the current address
space be terminated.
71. S0B0 An uncorrectable error was detected by the SWA Manager.
72. S0B0-04 Invalid function requested.
73. S0B0-08 Invalid SVA (Does not point to the beginning of a SWA prefix or the SWA prefix has been destroyed.
74. S0B0-10 Invalid length (0 or negative for asign locate or attempting to read or write a record which is not 176 bytes, in move mode).
75. S0B0-14 Invalid coutn field (0 for read, write assign or 00 for write and assign).
76. S0B0-18 QMNGRIO macro was issued with both or neither of the read and write bits set.
77. S0B0-1C Invalid ID. (The caller attempted to write a block for the first time and has either passed an non-existing ID or has failed to
pass one and the block does not have an embedded ID.
78. S0B0-20 Invalid block pointer (write locate is attempting to write and is passing a block pointer which is not valied for the SVA
requested).
79. S0C1 Operation Exception - An Operation is code is not on the machine. Possible Causes:
subscript error - clobbered code
tried to read a file that was not open
misspelled DDName
error in parameters passed to subroutines
missing DD card
RECORDING MODE i wrong or density is incorrect
bad load module, possible bad object deck
FORTRAN - Missing DIMENSION statement, same name for array and subroutine
COBOL - subroutine PROGRAM ID is the same as the ENTRY name.
COBOL - tried to CALL within COBOL F SORT INPUT/OUTPUT PROCEDURE
COBOL - tried to CALL a subroutine which could not be found
COBOL - incomplete DCB for SORTIN
COBOL - using SORT verb, DDNAME was not SORTOUT when GIVING option used.
COBOL - executing sort-using after opening SORTIN
80. S0C2 Privileged operation exception. Unintentional branch to invalid instruction due to subscript error. COBOL -missing period at end
of paragraph or paragraph names missing GOBACK after sort verb. Logic fell into input procedure. ACCEPT verb was executed when
no SYSIN DD was available.
81. S0C3 Failure to close input/output file in source code.
82. S0C4 Protection exception - virtual address could not be translated into a real address; Possible subscript error or bad parms; Variable
length record larger than maximum target field size; missing or misspelled dd statement; I/O against unopened file; Using GOBACK in
an output procedure; Move to an area not passed in a CALL statement; COBOL Sort with STOP RUN in input or output procedure;
blocksize and logical record size equal for a variable length dataset.
83. S0C5 Index or subscript out of range; uninitialized index or subscript; invalid data address; Reference to a field in a record before
OPEN or after CLOSE; Closing the same file twice; improper exit of a PERFORM statement; 88 level cannot be last statement before
Report section; date not initialized; Problem with a GO TO, GO TO DEPENDING ON, or GO TO that has been ALTERED.
84. S0C6 Specifcation exception -an incorrect boundary was specified. Usually caused by unintentional branch to invalid instruction.
Ddname/Select statement conflict; ACCEPT statement with SYSIN DD and data missing; No GO TO statement in a paragraph named
in an ALTER statement; I/O attempted against a file not open.
85. S0C7 Data exception -decimal data is incorrect or improperly overlapped or not validly initialized. Subscript error, referenced beyond
table. COBOL -working storage not initialized. -bad data, should check data for errors. -Garbage in a filed being tested; subroutine
parm missing or in wrong order.
86. S0C8 Fixed point overflow exception note -masked by FORTRAN.
87. S0C9 Fixed point divide exception note-masked by FORTRAN.
88. S0CA Decimal overflow exception. The destination field is too small to contain the result field in a decimal operation.
89. S0CB Decimal divide exception. A quotient exceeds the specified data field size.
90. S0CC Exponent overflow exception. A floating point number exceeds the maximum size. Note- This error is detected and fixed in
FORTRAN.
91. S0CD Exponent underflow exception. A floating point number is smaller than the minimum. Note -FORTRAN will set the result to
zero and continue processing.
92. S0CE Significance exception a floating point addition or subtraction results in an all zero fraction note -this is masked in FORTRAN.
93. S0CF Floating point divide exception -divide by zero note- this error is detected and noted by FORTRAN.
94. S0D2 A program check was detected; an interruption code of 18 (translation specification exception) has occurred. There is invalid
data in either control registers 0 or 1, or a segment or page table. The error is the result of a hardware error or a program running in key
zero has caused data damage.
95. S0D3 A program check was detected; and interruption code of 19 (special operation error) has occurred. A privileged program issued a
set system mask. This instruction is not valid in OS/VS2 (since Release 2).
96. SOE3 The error occurred while processing a VIO dataset. The system was unable to assign, locate, fix, free, or access VIO pages for
the dataset. Please report this problem to the systems group.
97. SOF1 Program interruption in I/O interruption handler record described as wrong length.
08/07/20111
4. MVS Abend Codes Page 4 of 13
98. SOF2 Program interruption in Type 1 SVC routine clobbered IOB or other SVC paramters -see SOC1.
99. SOF3 Machine check interruption disk I/O failure or machine trouble.
100. SOF8 The issuer of an SVC was either in SRB mode, held a lock, or was disabled.
101. S100 A device to be used was not operational or a pseudo device previously allocated has been reallocated.
102. S102 The error occurred during execution of a post macro instruction. The control program found an invalid event control block
address, or the ECB was in a storage area with a protection key different from that of the task issuing the macro.
103. S106 Region size too small. An error was detected by the control program when it attempted to fetch the requested program into virtual
storage. Possible cause: Previous LKED step failed to output a SYSLMOD, needs larger SIZE=I/O error-machine trouble or disk
failure.
104. S106-0B A program check or loop occurred in fetch.
105. S106-0C Not enough storage was available for fetch to do a GETMAIN for the module or control blocks.
106. S106-0D The control program found an invalid record type in the load module
107. S106-0E The control program found an invalid address in the load module
108. S106-0F An uncorrectable input/output error occurred.
109. S122 Either the operator or the job entry subsystem cancelled the job, requesting a dump. Check the job log listing for more
information.
110. S12D Invalid segment table in an overlay program. See S0C1.
111. S130 The DEQ macro specified a resource not previously specified by an ENQ macro. That is, the program had not requested control
of a resource it was attempting to release.
112. S137 I/O error in end of volume label processing on magnetic tape.
113. S137-04 An I/O error occurred while writing an end of volume label or tape mark.
114. S137-08 An I/O error occurred while positioning the tape in preparation for label processing.
115. S137-0C An I/O error occurred reading a trailer label for a dataset opened with the input or inout option, or reading the header label for
a dataset opened with the RDBACK option.
116. S137-10 An I/O error occurred while positioning a magnetic tape at the end of the dataset.
117. S137-14 An I/O error occurred reading header labels for a dataset opened for input or inout, or reading the trailer label for a dataset
opened for RDBACK.
118. S137-18 An I/O error occurred while positioning a magnetic tape dataset at the first data record.
119. S137-1C An invalid trailer label was read during end of volume.
120. S137-20 An invalid header label was read during end of volume.
121. S137-24 A specific volume serial number was specified for the second or subsequent volume of an output dataset on magnetic tape.
During EOV processing, it was discovered that the expiration date (from the HDR1 label of the first dataset currently on the specified
volume) had not passed. When requested to specify whether the volume could be used in spit of the expiration date, the operator did
not reply 'U'. Ask the operator to reply 'U' or specify another volume serial.
122. S138 The error occurred during the execution of an ENQ macro. Two ENQ macros were issued for the same resource in the same task
without an intervening DEQ macro. The second ENQ macro did not specify test, use, or have in its RET operand.
123. S13E The task which created a subtask has issued a detach for that subtask, specifying STAE=NO, before that subtask has terminated.
124. S14F The routine attempted to execute the STATUS macro instruction for other than the STOP, STOP SYNCH, OR START function
and was not in supervisor key(0-7).
125. S16E The control program requested that a DEBCHK function be performed on a data extent block obtained from the DCB passed by
the program. That function could not be completed.
126. S16E-04 The indicated DEB pointer is not in the DEB table. A DEB whose address is not in the DEB table cannot be verified, deleted,
or purged.
127. S16E-08 Invalid type specified (macro not issued). Acceptable types are ADD, DELETE, VERIFY, and PURGE.
128. S16E-10 DEBDCBAD does not point to DCB. It is normally assumed that the DCBDEBAD field of the DCB points to the DEB, but
the DEBDCABAD field of the DEB must point to the DCB when TYPE=VERIFY, ADD, or DELETE.
129. S16E-14 AM value does not equal DEBAMTYP value. When a DEB pointer is added to the table, the access method pointer (AM)
value, if given is placed in the DEBAMTYP field of the DEB. If no AM value is coded, A O is inserted in the field, subsequent
DEBCHKS issued to verify or delete that DEB pointer must either specify the same AM value or omit the operand. When the operand
is omitted, no comparison is made.
130. S16E-18 DEB not on the TCB chain for type=ADD. Before the DEB pointer can be added to the table, the DEB itself must be queued
on the current TCB DEB chain.
131. S16E-1C DEBAMTYP or DEBTBLOF=0 for TYPE=ADD. Values other than 0 indicate a pointer to this DEB already exists in the
DEB table.
132. S16E-20 DEB table contains 32760 bytes and TYPE=ADD. The current DEB table does not have space for this new DEB pointer. To
increase the table size by the required increment of 24 would cause the table to exceed its maximum size.
133. S171 The real storage manager was invoked with a request for a PGFIX, PGFREE, PGLOAD or PGOUT service and the request was
illegal or invalid. The request is considered to be illegal if the storage range specified by the input parameters does not exist (a
GETMAIN was not issued for it). NOTE: The meanings of the contents of general registers 11, 12, 13 and 14 are provided for
diagnostic purposes in the full description of this abend in the SYSTEM CODES MANUAL. Possible causes:
input parameter error in virtual subarea list (VSL)
VSL not on a fullword boundary
VSL NOT IN FIXED STORAGE
undefined or conflicting option flags. End address of range not greater than beginning address.
an option was specified which is not supported by MVS. These are RSAOPT and ECBIND.
VSL is store protected from the caller
08/07/20111
5. MVS Abend Codes Page 5 of 13
ECB not on fullword boundary
ECB is store-protected from caller
ECB not specified for PGFIX.
ECB specified for PGOUT
134. S171-04 The error was detected by the page services routine. This generally indicates that the caller did not own the virtual storage
defined by the VSL list entry.
135. S171-16 The input VSL or ECB failed to pass the page services interface validity check.
136. S206 The address of the parameter list, or one of the parameters passed to a LINK, LOAD, XCTL, or DELETE macro invalid.
137. S213 DSCB not found; I/O error in reading or writing DSCB. Possible causes:
the data set is not the specified volume
DISP=MOD is not compatible with a volume reference
incorrect tape positioning.
138. S213-04 An I/O error occurred reading the FORMAT 1 DSCB, or the FORMAT 1 DSCB could not be found on the first volume
specified by the DD statement volume serial field.
139. S213-08 An OPEN macro was issued for a password protected dataset but the system was unable to locate the password dataset.
140. S213-0C An I/O error occurred reading a FORMAT 1 DSCB for a direct or indexed sequential dataset, or the FORMAT 1 DSCB could
not be found on the volume specified by the DD statement.
141. S213-18 An I/O error occurred writing back a FORMAT 1 DSCB.
142. S213-20 During an open, a volume contained more than 16 extends of the indicated dataset. Try browsing or editing the file via SPF
3.4. Check the upper right-most corner of the ISPF panel for an indicator message.
143. S213-28 An OPEN macro ws issued for a direct access dataset specifying UNIT=SYSDA, but the UNIT already contained 127 users,
whcih is the maximum allowed.
144. S214 I/O error in tape positioning or volume disposition.
145. S214-04 An I/O error occurred reading a user label on magnetic tape.
146. S214-08 An I/O error positioning a magnetic tape volume during execution of a CLOSE macro.
147. S222 Either the operator or the job enter subsystem cancelled the job. Check the job log listing for more information. If there is no
apparent explanation, contact operations before resubmitting. Possible causes:
Line or card estimate exceeded.
JCL error caused mount request
Excessive runtime.
148. S233 Invalid parameters have been passed to SVC dump.
149. S233-04 The address of the parameter list is zero.
150. S233-08 The parameter list is not a valid SVC dump or snap parameter list for MVS.
151. S233-0C The caller-supplied dataset is on an unsupported device.
152. S233-10 In a user-supplied storage range, the start address is greater than the end address.
153. S233-14 The user-supplied data (for HDR= or HDRAD=) is greater than 100 characters.
154. S233-18 The 4K SQA buffer has been requested (BUFFER=YES) but it is not serialized (by setting on the high order bit in
CVTSDBF)
155. S233-1C The parameter list or what it points to is in the 4K SQA buffer.
156. S233-20 The user supplied a DCB address and the DCB is not open, or the DCB address is invalid.
157. S233-24 The specified ASID parameter was syntactically invalid. The ASID was less than 0 or greater than the maximum value.
158. S233-28 An ASID specified in the ASID list pointed to the by the ASIDLST parameter was syntactically invalid. The ASID was less
than 0 or greater than the maximum value.
159. S233-2C The ASIDLST address is zero, or points to an area that the user cannot reference.
160. S233-38 The caller specifoed the 4K SQA buffer (BUFFER=YES) but and SVC dump was already in progress.
161. S237 The error occurred during end-of-volume label verification. Possible causes:
Incorrect volume serial
Incorrect volume mounted
Incorrect labels
162. S237-04 The block count in the DCB does not match that in the trailer label. A block of data has been misused or skipped.
163. S237-08 The DSname in a header label does not match that in the JFCB on the second or subsequent volume of a magnetic tape
dataset.
164. S23E The error was detected during execution of a DETACH macro instruction. Either:
The parameter passed to detach in register 1 was not a fullword address
The storage key of that address did not match that of the issuer of the detach
The parameter contained in the addressed fullword of the issuer was not the address of a subtask of the isser of the detach.
165. S240 An error occurred during execution of a RDJFCB macro.
166. S240-04 A RDJFCB macro instruction was issued, but the DCB did not contain a foundation extension block.
167. S240-08 A RDJFCB macro instruction was issued, but no EXLST address was found in the DCB.
168. S240-0C A RDJFCB macro instruction was issued, but no JFCB exit was specified in the DCB exit list.
169. S240-10 A RDJFCB macro instruction was issued, but the JFCB buffer is not within the user's storage.
170. S2F3 Job was being executed when system failure occurred- A system restart was performed.
171. S300 An error was detected when and I/O operatoin was requested; the storage protection key was not 0 or the DEB validity check
routine returned to EXCP wit a non-zero return code. In the abnormal termination dump, the TCB field TCBEXCPD (at offset C0)
08/07/20111
6. MVS Abend Codes Page 6 of 13
points to the EXCP problem determination area. The items in the problem determination area of greatest interest to you are: (all offsets
are in hex)
offset 10 contains a copy of the registers when EXCP determined the error condition.
offset 50 contains the contents of the Request Queue Element (RQE) if allocated, when the program check occurred.
172. S305 The error occurred during execution of a FREEMAIN macro:
The specified subpool could not be found.
The SP parameter was specified but the virtual strage area to be released was not within subpool zero.
The SP parameter was not specified but the virtual storage area to be released was not within subpool zero.
The SP parameter was specified correctly, but the boundaries of the storage areas to be freed were not completely described by
a descriptor queue element.
173. S306 The error occurred during the execution of the LINK, XCTL, ATTACH, or LOAD SERVICE ROUTINES. The authorized
routine requested a module, which could bot be found on an authorized library, but a copy of the module may have been found on a
non-authorized library.
174. S306-04 The requested program was not found in the indicated source.
175. S306-08 An uncorrectable I/O error occurred when the control program attempted to search the directory of the library containing the
module.
176. S306-0C The module could not be found in the LPA or in the LPA directory or an authorized library.
177. S30A The error occurred during execution of an R-form FREEMAIN macro for one of the following reasons:
A FREEMAIN for an entire subpool was requested (register 1 was zero) but the length in register 0 was NT set to zero.
The specified subpool could not be found.
The SP parameter was specified but the virtual storage area to be released was not within subpool zero.
The SP parameter was not specified, but the Virtual Storage Area to be released was not within subpool zero.
The SP parameter was specified correctly, but the boundaries of the storage areas to be freed were not completely described by
a descriptor queue element (DQE).
178. S313-04 (04 is the only possible return code associated with S313). An I/O error occurred reading a format 2 or format 3 DSCB during
the execution of an OPEN macro.
179. S314 I/O error in reading DSCB. SORTIN JCL gave wrong block size for a variable record.
180. S314-04 An I/O error occurred reading a DSCB for a dataset on a direct access device during execution of a CLOSE macro.
181. S314-08 An I/O error occurred reading a FORMAT 1 DSCB during execution of a CLOSE macro and standard user labels were
specified.
182. S322 Job or step time exceeded the specified limit. Program is in a loop. Insufficient time parameter on JOB or EXEC card.
183. S337 The error occurred when the end of a data set was reached. Possible causes:
No end of file routine was provided.
Tried to read past end of file.
Tried to write on a tape defined as input.
184. S337-04 The end of dataset was reached, but no end-of-dataset routine (EODAD) specified in the DCB for DD dummy dataset.
185. S337-08 No end-of-dataset routine (EODAD) specified in the DCB for DD dummy dataset.
186. S338 An unauthorized ptask attempted to use authorized options of the ENQ macro instruction.
187. S378 The error occurred during execution of an RC or RU form FREEMAIN macro. Possible causes:
A FREEMAIN for an entire subpool was requested (register 1 was zero) but the length in register 0 was not set to zero.
The SP parameter was specified, but the virtual storage area to be released was not within the subpool specified.
The SP parameter was not specified but the virtual storage areas to be released wit not within subpool zero.
The specified subpool could not be found.
The SP parameter was specified correctly, but the boundaries of the storage area to be freed were not completely described by
the descriptor queue element (DQE).
188. S400 The DCB in the DEB does not equal the DCB address in the IOB. In the abnormal termination dump, the TCB field TCBEXCPD
(at offset CO) points to the EXCP problem determination area. The items in the problem determination area of the greatest interest to
you are:
hex offset 10 contains a copy of the registers when EXCP determined the error condition;
hex offset 50 contains the contents of the request queue element (RQE) if allocated when the program check occurred.
189. S413 INPUT, INOUT, or RDBACK specified, but no vol ser in DD; or I/O error in reading volume label; or could not mount volume
on device; or volumes specified less than devices allocated. Possible causes:
Forgotten volume serial for input tape.
Had DISP new when should have had OLD.
Generated as NL, tried to read as SL.
Missing or inconsistent DCB information.
Volume could not be mounted so operated mounted scratch.
Tried to close a file without opening it.
Tried to open a second dataset on a tape without closing the first.
The requested drive was switched off.
SORT verb blew due to SD not matching FD with using or giving.
08/07/20111
7. MVS Abend Codes Page 7 of 13
190. S413-04 Possible causes:
No unit is available for mounting the volume for the dataset being opened.
The volume already on the allocated unit as specified by the VOL=SER is either permanently resident or reserved.
It could not be demounted in order to mount the required volume or
the volume cannot be demounted because another DCB is open on that device or
the device type is incompatible with the DSORG in the DCB.
This error may be due to a previous abnormal termination associated with the same unit inthe same step.
191. S413-08 An I/O error occurred positioning a magnetic tape volume.
192. S413-0C An I/O error occurred reading the volume label on a magnetic tape volume.
193. S413-10 An I/O error occurred writing a tape mark.
194. S413-18 The specified dataset was opened for input but no volume serial number was specified on the DD statement.
195. S413-1C An OPEN macro was issued for a dataset, but the volume sequence number on the associated DD statement was greater than
the number of volumes containing the dataset.
196. S413-20 An I/O error occurred reading the volume label on a direct access volume.
197. S413-24 An OPEN macro was issued for a dataset on magnetic tape. A density was specified in the DCB DEN parameter which was
incompatible with the recording density of the drive allocated to the dataset.
198. S413-34 LABEL=(N) was specified, where N is greater than 1, and VOL=SER was not specified for a tape dataset.
199. S513-04 (04 is the only possible return code associated with S513) an OPEN macro was issued for a magnetic tape dataset allocated to
a device that already has an open dataset on it.
200. S513-08 Label violates published STDS….is rejected.
201. S522 Job or TSO session exceeded maximum job wait time or operator did not mount the required tape within allowed time limit.
Waiting on datasets, may requeue later.
202. S604 Possible Causes:
GETMAIN error.
Missing SORTLIB card.
Subscript error.
Clobbered free queue element.
203. S605 Freemain error, see S604.
204. S60A Freemain queue element or address error. Possible Causes:
Needs more core.
JCL for input disks requested 2 units, gave 1 volume serial.
Missing DD card for SYS1.SORTLIB.
205. S613 I/O error in tape positioning or label processing. Possible Causes:
Tape label was not in requested density.
DCB missing.
DEN missing on a tape that requires it default DEN=3 (1600 BPI).
RECFM=F was specified for a file which was FB.
Incorrect header.
Tape positioning error.
Tape drive failure during OPEN.
206. S613-04 An I/O error occurred positioning a magnetic tape volume.
207. S613-08 An I/O error occurred reading a label on a magnetic tape.
208. S613-0C An invalid label was read from a magnetic tape volume. This error may be due to a previous abnormal termination associated
with the same tape since it was last mounted, possibly in a previous job or step, leaving the tape positioned improperly.
209. S613-10 An I/O error occurred writing a tape label.
210. S613-14 An I/O error occurred writing a tape mark after the header labels.
211. S614 Uncorrectable I/O error occurred in writing end-of-file during close processing a dataset on a direct access device or the operator
dismounted volume while program was processing.
212. S614-04 An I/O error occurred writing a file mark for a dataset on a direct access device during execution of a CLOSE macro.
213. S614-08 A file mark should have been written on an output dataset. The DCBFDAD field in the DCB indicated an extent number in
the DEB greater than the number of extents in the dataset. Consequently, it could not be determined where the file mark should have
been written.
214. S622 TSO session was cancelled. Possible Causes:
Operator stopped TSO or cancelled the user.
User signaled attn after the allocation process had completed.
The user disconnected his terminal from the system.
215. S637 I/O error in writing tape mark, tape positioning, reading label, sensing for file protection ring; DCB bit does not indicate
concatenation of unlike attributes.
216. S637-04 An I/O error occurred while reading a tape label, writing a tape mark, or positioning a magnetic tape volume.
217. S637-08 Following user trailer processing, an I/O error occurred positioning a magnetic tape.
218. S637-0C Concatenation of datasets with unlike attributes was detected, but not specified in the DCB.
219. S637-10 An I/O error occurred while positioning a magnetic tape dataset that was opened with the option input or inout to be read
08/07/20111
8. MVS Abend Codes Page 8 of 13
backward. If it is a tape with standard labels, the error occurred positioning at the labels. If it is a tape with no labels, the error occurred
positioning at the data.
220. S637-14 An I/O error in tape positioning occurred for a dataset with the leave option specified in the OPEN macro or in the FEOV
macro.
221. S637-18 An I/O error in tape positioning occurred for a dataset opened with the reread option.
222. S637-1C An I/O error in tape positioning occurred when FEOV was issued for a dataset with DISP=PASS and no open option 2
specified.
223. S637-24 An I/O error occurred rewinding a scratch magnetic tape volume. Either FEOV with a rewind option was issued, or no open
OPTION 2 was specified when the disposition was not PASS.
224. S637-2C An I/O error occurred while rewinding a magnetic tape volume prior to verifying the volume label.
225. S637-34 An I/O error occurred during end of volume processing while reading the volume label of a magnetic tape
226. S637-38 An I/O error occurred while positioning a tape without labels or with nonstandard labels.
227. S637-3C An I/O error occurred while positioning a concatonated magnetic tape dataset.
228. S637-40 An I/O error occurred positioning a magnetic tape dataset that was opened with the option input or input to be read forward.
229. S637-44 An I/O error occurred while checking sense bytes to determine if a file protect ring is on a magnetic tape containing a dataset
opened or inout.
230. S700 A program check occurred during EXCP processing of an I/O request. The program check occurred in a supervisor service
routine called by EXCP. This is likely not a programmer error. In the abnormal termination dump, the TCB field TCBEXCPD (at
offset C0) points to the EXCP problem determination area. The items of greatest interest to you are:
Hex offset 02 contains a flag byte indicating where the error occurred.
Bit 1 = 1 means error in SVC portion of the EXCP;
Bit 2 = 1 means error in SRB portion of the EXCP;
Bit 3 = 1 means error in PCI appendage;
Bit 4 = 1 error in CHE appendage;
Bit 5 = 1 means error in ABE appendage;
Bit 6 = 1 means error in EOE appendage;
Bit 7 = 1 means error in PGFX appendage;
Bit 8 = 1 means appendage is active;
No bits on means error in SIO appendage.
Hex offset 04 contains a copy of the program status word before RTM was entered.
Hex offset 0E contains the interruption code.
Hex offset 10 contains a copy of the registers when EXCP issued the abend macro.
Hex offset 54 contains the contents of the request queue element, if allocated, for the I/O request being processed.
231. S706 The requested load module was marked by the linkage editor as not executable. Bad prior LIKDEDIT -check the LKED
SYSPRINT.
232. S706-04 Improper program linkage (NON-IMS pgm linked used LNKTESTI Clist)
233. S713 Tried to write on a dataset whose data protection had not expired. Operator replied 'M' to IEC507D to honor exp date.
234. S713-04 A dataset on magnetic tape was opened for inout, outin, output, outinx, or extend, but the volume contained a dataset whose
expiration date had not been reached.
235. S713-08 An OPEN macro was issued with inout for a dataset on a direct access device with DISP=OLD specified on the DD statement.
The expiration date on the dataset had not been reached.
236. S714 I/O error in label processing -close for magnetic tape tape drive failure during close.
237. S714-04 An I/O error occurred writing trailer label 1 for a dataset on magnetic tape during execution of a CLOSE macro.
238. S714-08 An I/O error occurred writing trailer label 2 for a dataset on magnetic tape during execution of a CLOSE macro.
239. S714-0C An I/O error occurred writing a tape mark during execution of a CLOSE macro.
240. S722 The output limit specified by the OUTLIM keyword on the sysout DD statement or by the lines keyword on the jobparm DD
statement was exceeded.
241. S737 I/O error; DSCB not found for multi-volume or concatenated data set. Be sure all the data sets exist for concatenated data sets.
242. S737-04 An I/O error occurred reading or writing a DSCB during end of volume processing.
243. S737-08 An I/O error occurred reading a direct access volume label during end of volume processing.
244. S737-0C An error occurred reading the DSCB for a concatenated partitioned dataset.
245. S737-10 An I/O error occurred writing a file mark for a dataset on a direct access device.
246. S737-1C An I/O error occurred while reading a format 3 DSCB.
247. S737-24 A missing member name was detected by BLDL while searching for the TTR of a concatenated member.
248. S737-2C The error occurred when a FEOV macro was issued while attempting to write a file mark at the end of data. The DCBFDAD
field in the DCB indicated an extent number in the DEB greater than the number of extents in the dataset. Consequently it could not be
determined where the file mark should have been written.
249. S738 An unexpected error occurred during the execution of an ENQ macro. Related information is recorded in SYS1.LOGREC.
250. S800 An error occurred when EXCP attempted to fix a page for the EXCP request.
251. S804 Request for 0 bytes of main storage or not enough main storage available. Program exceeded region size.
252. S806 Possible causes:
Load module not found in loadlib.
BLDL detected error.
08/07/20111
9. MVS Abend Codes Page 9 of 13
Module not found or I/O error during directory search.
Missing JOBLIB or STEPLIB card.
Tried to execute a non-existent program.
Tried to execute in batch a program assembled with 'test' option.
SYSIN described incorrectly to LKED or name card missing.
Also, check if '**RACF** is missing or mispositioned on the jobcard.
253. S806-04 Module not found. The program entry point specified was not found in the indicated library (private library, job library, or
link library).
254. S806-08 An uncorrectable input/output error occurred when the BLDL control program routine attempted to search the directory of the
library that contained the program whose entry point was specified. This can occur if the indicated library is an uninitialized partitioned
dataset.
255. S806-0C A SVC routine required by the system could not be found in the link pack area.
256. S80A Request for 0 bytes of main storage or not enough main storage available. Need more core in region.
257. S813-04 (04 is the only possible return code associated with S813.) Open error. Possible causes:
Incorrect dataset name.
An OPEN macro was issued for a dataset on magnetic tape, but the dataset name on the header label did not match that in the
JFCB (Tape's been reused).
Wrong DSname or volumeserial -JCL disagrees with label.
Incorret record format or blocksize;
The requested drive is not switched to this machine.
258. S822 A region required to run the step could not be obtained. One of the following messages will be written to the programmer,
depending upon wheather the job was an ordinary job or a deferred checkpoint restart:
IEF085I Region Unavailable, Error Code=CDE.
IEF086I Region Unavailable For Restart, Error Code=CDE.
259. S822-08 IEFO85I - A V=V region was requested and a region size was specified which was larger than the private area, or a V=R
region was requested and a region size greater than the V=R area was specified. IEF1861 - The region parameter was increased so that
the region could not be obtained.
For ADDRESPC=REAL, the size of the real area was decreased.
For ADDRSPC=VIRT, the size of the private area decreased because the size of the nucleus increased or the size of the SQA or
the IPA increased.
260. S822-16 IEF085I - A V=R area was requested, but either long-fixed or damaged pages in the V=R area made it impossible to obtain
the requested region, or a V=R region was requested and there was not enough SQA available for the system to complete the
processing of the request; IEF186I - If a REAL region was requested, either long-fixed or damaged pages in the REAL area made it
impossible to obtain the required region.
261. S822-20 IEF085I and IEF186I - Either a V=V or V=R region was requested fragmentation by LSQA, SWA, or subpools 229 or 230
has made it impossible to obtain the requested region.
262. S822-24 IEF085I and IEF186I -A request for a V=R region could not be satisfied because the installation GETPART exit routine
rejected the request.
263. S837 Error during End-Of-Volume processing for a sequential dataset.
264. S837-04 A density conflict exists between the user-specified density and the unit's density capability.
265. S837-08 Specify more volume serial numbers or a larger volume count in the VOL parameter of the DD statement.
266. S837-0C Tape volume was requested and mounted but another dataset was processing the volume.
267. S878 GET MAIN error not enough virtual storage on region statement. Try a larger REGION parm on the JOB CARD.
268. S906 IMS use count limit error.
269. S90A The error occurred during the execution of an R-form FREEMAIN macro. The address of the storage area to be relesed was not
on a doubleword boundary (A multiple of 8).
270. S913 RACF security, Trying to access to a dataset not authorized to use. The error occurred during 1) The execution of an OPEN
macro or during end-of-volume processing for a password protected dataset, or 2) the execution of an OPEN macro involving a
checkpoint dataset.
Operator failed to supply correct password for a protected dataset.
Unauthorized attempt to use LABEL=BLP.
Previous dataset on this tape protected, new dataset is not protected.
271. S913-04 An OPEN macro was issued for a magnetic tape dataset with American National Standard labels. The volume accessability
byte (offset X'OA') in the volume label is not blank. This indicates the label was not written for use on an IBM system, or that it was
written by the user. The volume must be recreated.
272. S913-08 An OPEN macro was issued for a magnetic tape dataset with american national standard labels. The security byte in the
header label was not bland and not equal to X'F1'. This means the label either was not created on an IBM system or was created by the
user. The volume must be recreated for use on an IBM system.
273. S913-0C An OPEN macro was issued for a protected datset for which you are not allowed this type of access.
274. S913-10 An OPEN macro was issued to the VTOC for output processing by an unauthorized job step or job-step task.
275. S913-14 An OPEN macro was issued to concatenate checkpoint and noncheckpoint datasets.
276. S913-18 An open TYPE=J macro was issued for a magnetic tape volume. The JFCB was modified to indicate LABEL=BLP (Bypass
label processing) and the task was not authorized. BLP may be specified in the JCL (if the installation reader procedure allows it), but
08/07/20111
10. MVS Abend Codes Page 10 of 13
the JFCB may not be modified to indicate BLP unless the task is authorized.
277. S913-1C The error occurred during the execution of an open TYPE=J macro to a direct access device dataset. The dataset name
supplied to open was not available to the job. Either the dataset was bein opened for input and some other job had exclusive control of
it, or the dataset was being opened for an option other than input (thus requiring exclusive control) and some other job was using the
dataset.
278. S913-20 An OPEN macro using the EXCP access method was issued in which user-written appendages were required. The appendage
names were not included in the parmlib member IEAAPPOO, and the program issuing the open was not authorized (by APF or protect
key.).
279. S913-28 An OPEN macro was issued for a checkpoint dataset. The dataset organization was not BPAM or BSAM and the task was not
authorized via the authorized program facility (APF).
280. S913-2C An OPEN macro was issued to an ISAM dataset defined by two or three DD statements. Either the dataset names coded in
the DD staements were not all the same, or the JFCB passed to open (if an open TYPE=J is being used) has a dataset name different
from that coded in the DD statements.
281. S913-30 An OPEN macro was issued to write a daaset on a magnetic tape containing one or more previous datasets. The protection
mode of the dataset to be written was different than the protection mode of the previous dataset.
282. S913-38 An OPEN was issued for a RACF-protected dataset to which the user was not authorized.
283. S913-3C An OPEN was issued for a dataset with a Format-1 DSCB indicated RACF protection, but the dataset is not defined to
RACF.
284. S913-40 A VSAM data space being opened is RACF-defined.
285. SA00 A program check occurred in an appendage. In the abnormal termination dump, the TCB field TCBEXCPD (at offset CO) points
to the EXCP provided debugging area. The items in the debugging area of greatest interest to you are:
Hex offset 2 contains a flag byte indicating where the error occurred.
Bit 1 = 1 means error in SVC portion of the EXCP;
Bit 2 = 1 means error in SRB portion of the EXCP;
Bit 3 =1 means error in PCI appendage;
Bit 4 = 1 error in CHE appendage;
Bit 5 = 1 means error in ABE appendage;
Bit 6 = 1 means error in EOE appendage;
Bit 7 = 1 means error in PGFX appendage;
Bit 8 = 1 means appendage is active;
No bits on means error in SIO appendage.
Hex offset 04 contains a copy of the program status word before RTM was entered.
Hex offset 0E contains the interruption code.
Hex offset 10 contains a copy of the registers when EXCP issued the abend macro.
Hex offset 54 contains the contents of the request queue element, if allocated, for the I/O request being processed.
286. SA0A Inactive program overlaps overlaps free area; area to be released overlaps a free area.
287. SA13 The error occurred during execution of an OPEN macro on magnetic tape.
288. SA13-04 An unexpected load point was encountered whil positioning a tape. For NL tape this is probably a user error associated with
the use of a mutlivolume multfile NL tape.
289. SA13-08 The requested file sequence number is less than that of the first file on the SL tape during an open to the start of the file.
290. SA13-0C The requested file sequence number is less than that of the first file on the SL tape during an open to the end of a file.
291. SA13-10 A tape mark was read instead of a HDR1 label while forward spacing to the desired file on an SL tape. The multifile tape
ends before the desired file. When positioning to the end of file 1, this means the vol label is followed by a tape mark. Check the file
sequence number and the volume serial numbers, and that the job that wrote the tape wrote all the files.
292. SA13-14 A tape mark was read instead of a HDR1 label while opening for input to the start of the desired file on an SL tape. Thus, the
tape ends just before the diesired file. Check the file sequence number and the volume serial numbers and that the job that wrote the
tape wrote all the files.
293. SA13-18 An EOV1 label was read on the last SL tape volume while forward spacing to the desired file. If opening to the end of the
file, it could not be treated as the end of the dataset because it was for a previous file sequence number.
294. SA37 The error occurred during end of volume processing. The task is terminated unless the error is to be ignored as specified in the
DCB abend exit routine. Possible causes:
An open DCB may have been partially overlaid.
The DCB may have been closed in a SYNAD routine.
The DCB may have been automatically closed by a previous End Of Volume error where ignor was specified in the DBC abend
exit.
295. SA37-04 An SVC 55 (EOV) was issued, usually by a check, get, or put routine, against a DCB which was not open.
296. SA37-08 The data extent block (DEB) does not point to the DCB.
297. SA78 The error occurred during the execution of an RC or RU form FREEMAIN macro: Possible causes are:
The control program found that the address and length specifications in the release request defined an area to be freed that
overlapped a free area in virtual storage.
Part of the area being freed is still fixed in real storage. Fixed storage blocks cannot be freed.
298. SACC Not an error - FORTRAN accounting termination.
299. SB00 A program check occurred in an EXCP procedure. In the abnormal termination dump, the TCB field TCBEXCPD (at offset CO)
08/07/20111
11. MVS Abend Codes Page 11 of 13
points to the EXCP provided debugging area. The items in the debugging area of greatest interest to you are:
Hex offset 2 contains a flag byte indicating where the error occurred.
Bit 1 = 1 means error in SVC portion of the EXCP;
Bit 2 = 1 means error in SRB portion of the EXCP;
Bit 3 = 1 means error in PCI appendage;
Bit 4 = 1 error in CHE appendage;
Bit 5 = 1 means error in ABE appendage;
Bit 6 = 1 means error in EOE appendage;
Bit 7 = 1 means error in PGFX appendage;
Bit 8 = 1 means appendage is active;
No bits on means error in SIO appendage.
Hex offset 04 contains a copy of the program status word before RTM was entered.
Hex offset 0E contains the interruption code.
Hex offset 10 contains a copy of the registers when EXCP issued the abend macro.
Hex offset 54 contains the contents of the request queue element, if allocated, for the I/O request being processed.
Hex offset 80 is the start of the 160 byte blocks involved in the I/O request, if allocated.
300. SB0A The error occurred during execution of an R-form GETMAIN or FREEMAIN macro. Possible causes are:
A subpool number greater than 127 was specified by a problem program (a program not authorized to use valid subpool
numbers greater than 127).
An authorized program requested an invalid subpool.
Tried to change blocking factor while loading a dataset.
Control passed beyond end of program due to an invalid PERFORM.
301. SB14 At close a partitioned data set directory cannot be updated because:
the name is already in the directory or
no space is available in the directory or
an I/O error was encountered in the directory.
Possible cause: Allocated directory space in JCL, written as sequential dataset.
302. SB14-04 A duplicate name was found in the directory of the PDS.
303. SB14-0C The attempted update of the partitioned dataset found that there was no space left in the directory.
304. SB14-10 An I/O error occurred trying to update the directory of the partitioned dataset.
305. SB14-14 The close routine attempted to update the directory of the partitioned dataset, however the DCB of the dataset was not open.
306. SB14-18 Unsuccessful GETMAIN for stow workarea when close tried to update a partitioned dataset. Region size is too small.
307. SB37 Space problem, usually on input or output dataset. The data set on DASD output
already had 16 extents, but required more space or
secondary space was too small or
no more space was available on the volume or
at end of volume, the volume must be demounted, but the system is unable to dismount the volume.
308. SB37-04 During end-of-volume processing the system had to demount a volume in order to mount the next volume of the dataset. It
was unable to demount the volume for one of the following reasons:
The volume was permanently resident or was reserved.
Another job had datasets allocated on the volume.
The failing task had open datasets on the volume.
For an output dataset on a DASD no more space was available on the volume.
16 extents already obtained for the output dataset but more space is required.
VTOC is full.
For an output dataset on a tape volume, volume needed to be demounted because the reflective spot was encounted and more
records needed to be written, but the limit of output tape volumes had been reached.
309. SB37-08 During end-of-volume processing the system attempted to extend the dataset to a volume on which the DIRF bit was set. The
VTOC for the volume could not be converted to the standard format for one of the following reasons:
Two datasets were allocated to the same space on the volume.
A split cylinder dataset was located on the same cylinder as a non-split cylinder dataset.
The VTOC conversion routine is set to reject DIRF requests.
310. SB37-0C The mounted direct access volume was requested to continue processing the dataset, but the unit already contained the
maximum of 127 concurrent users.
311. SC03 A data set could not be closed by the control program, because the DCB had been erroneously modified. ISAM file -Bad block
size.
312. SC13 The error occurred during the execution of an OPEN macro for a concatenated partitioned or graphics dataset. Attempted to write
output to concatenated partitioned datasets missing or deleted joblib data set. Possible Cause: Graphics -system problem-wait for IPL.
313. SC13-04 The current task tried to open a graphics device that was previously opened and not closed.
314. SC13-10 An OPEN macro was issued specifying output or extend for a concatenated partitioned dataset, output datasets cannot be
08/07/20111
12. MVS Abend Codes Page 12 of 13
concatenated.
315. SD0D Invalid abend recursion during abnormal termination of subtask job step task terminated.
316. SD13 Open for graphics -DCB for other than graphics device.
317. SD23 The error ocurred during the execution of a WTO or WTOR macro for one of the following reasons:
The parameter list supplied to the WTOR macro does not begin on a fullword boundary.
A WTOR/MLWTO parameter list was specified.
A multiline WTO was specified and space was not available in subpool 229 for a workarea for SVC 35.
The parameter list passed by the user does not reside instorage that is accessible by the user.
Space was not available in subpool 231 for an ORE or WQE buffer.
318. SD2D Overlay supr. found invalid record type bad load module. Re-linkedit.
319. SD37-04 Space problem, usually in load library. (04 is the only possible return code associated with SD37) A dataset opened for output
used all the primary space, and no secondary space was requested. Either specify a larger primary quantity or add a secondary quantity
to the request. Try increasing the primary SPACE allocation.
320. SE37 Space problem, usually in a PDS. The error occurred during end of volume processing when an output operation was requested
for a dataset. Ran out of directory space in a PDS or output exceeded volume count.
321. SE37-04 A dataset opened for output used all space available on the current volume, and no more volumes were available:
Not enough volumes were specified for the dataset through the 'SER', volume count, or 'REF' subparameter of the volume
parameter.
When all the volumes were filled, the program attempted to write another record.
For a partitioned dataset, all space was filled when the program attempted to write another record
16 extents had been used when the program attempted to write another record, however no volume was available for
demounting.
322. SE37-08 On a dataset opened for output, end-of-volume had found a DSCB with a duplicate dataset name on the next volume, with a
volume sequence number less than that in the DEB.
A multi-volume physical sequential dataset was being written on a direct access device
all space was filled on a volume and an attempt was made to obtain space on the next specified volume but space either was not
available on that volume or the dataset already existed on that volume.
323. SF2D Overlay supervisor - Invalid Supervisor call. Bad Load Module. Re-LNKEDT entire module.
324. SF37 Hardware error.
325. U0004 GO step of FORTRAN - subrountine fileno - non sequential dataset.
326. U0008 GO step of FORTRAN - subrountine fileno - IDSRN out of range.
327. U0012 GO step of FORTRAN - subrountine fileno - dataset reference number has not been used.
328. U0016
Standard abend code for SYNC SORT failure. See diagnostic messages.
GO step of FORTRAN - Missing SD4060 DD card for SD4060 output.
GO step of FORTRAN - subroutine fileno - second argument not 'IN' or 'OUT'.
329. U0020
GO step of FORTRAN - subroutine fileno - previous file not closed.
ASMF could not open SYSIN. Check the SYSIN DD card.
I/O failure reading a bad track.
330. U0024
GO step of FORTRAN - subroutine fileno - device is not a tape.
RPG - alphanumeric data in numeric-specified field.
RPG - missing DD card.
331. U0030 GO step of FORTRAN - missing FT06F001 DD card.
332. U0040 RPG - input file out of sequence.
333. U0044 RPG - undefined record type.
334. U0064
Archived or inactive GDG.
RPG - QSAM I/O Error.
335. U0100 FDR - open error - unable to open disk or tape file.
336. U0101 FDR - disk I/O Error.
337. U0200 FDR - tape I/O Error.
338. U0201 FDR - tape End-Of-File before dataset or trailer record found.
339. U0204 PAN#8 - SYSPRINT DD statement missing.
340. U0240
Possible program looping. Program time out due to looping problem.
FORTRAN - SPIE.
341. U0261 Improper program linkage (IMS program linked used LNKTEST Clist); IMS program not linked using LNKTESTI (consider
this first, but other reasons are possible).
342. U0402 FDR - SYSPRINT or SYSPRIN-N error.
343. U0428 PSB not found. No IMS gen. U456 Program stopped.
08/07/20111
13. MVS Abend Codes Page 13 of 13
344. U0452 Transaction stopped. SUGGESTION: Restart using IMSRM.
345. U0456 Program/PSB stopped. SUGGESTION: Restart using IMSRM.
346. U0457 The same program is being run in two different jobs concurrently.
347. U0458 Database stopped.
348. U0474 Maximum time.; Message region stopped - contact data center.
349. U0476
Invalid PSB or no PSB. SUGGESTION: Check IMSGen and PSBDump.
Data Exception, Non numeric data in numeric field. See S0C7.
350. U0502 FDRDSF - Input control statement error.
351. U0519 Missing GO BACK, EXIT or STOP RUN.
352. U0688
IMS not active. SUGGESTION: Contact Tech Support.
JOB cancelled (JOB ran in the wrong JOB class). SUGGESTION: Rerun job with CLASS=J.
353. U0775
BMP program abended due to exceeding IMS calls between checkpoints.
Program has made too many calls without taking a checkpoint.
354. U0777 IMS step lock out.
355. U0778 Blocksize or LRECL problem; See S013
356. U0844 Production database out of space.
357. U0852
Another concurrent IMS job has the database locked for update. SUGGESTION: Rerun your job later.
Batch program is accessing a database that is being updated. SUGGESTION: Try rerunning the job later.
358. U0888 FDR - Diagnostic warning associated with text on SYSPRINT.
359. U0999 SYNCSORT - I/O error has occurred. See Diagnostic Messages U2000-.
360. U3303 Database stopped.
361. U3699 Invalid dump code used. Outside abend code range.
362. 4000 PL/1
08/07/20111