Mod_jk.conf file is NOT created

Hi Experts,
I am using Apache(2.0.40), JBOSS(3.0.3) and Tomcat(4.0.4) on IBM AIX 5.1 System.
After spending 1 week on Integrating Apache and Tomcat on IBM AIX System (and getting HELP from you guys from Mailing List), I finally builded "mod_jk.so" file for this platform.
I also configured Tomcat's "server.xml" and "workers.properties" file BUT when I start my JBOSS engine, "mod_jk.conf" file is NOT created automatically. I refereed the following document to configure that:
http://jakarta.apache.org/tomcat/tomcat-3.3-doc/mod_jk-howto.html
Do you experts have any suggestion/information as to why "mod_jk.conf" file is not created automatically? The above document says the file is CREATED AUTOMATICALLY once you start Tomcat!
Any related information on this would be appreciated.
THANKS!
Manoj G. Kithany
[email protected]

thanks, I have since figured this one out...
What I cant seem to do is get Apache to use Tomcat to serve any jsp's.
I have spent the last three days working with this and have gotten no where. I am rather frazzled and am about to abandon the thought of using anything Java for a long time.
Let me spell out what I have done for you:
#========================================================
C:\Java\Apache\Tomcat_4.1\conf\auto\mod_jk.conf
#========================================================
########## Auto generated on Fri Sep 13 18:46:55 EDT 2002##########
<VirtualHost localhost>
    ServerName localhost
    JkMount /manager ajp13
    JkMount /manager/* ajp13
    JkMount /examples ajp13
    JkMount /examples/* ajp13
    JkMount /tomcat-docs ajp13
    JkMount /tomcat-docs/* ajp13
    JkMount /webdav ajp13
    JkMount /webdav/* ajp13
    JkMount /admin ajp13
    JkMount /admin/* ajp13
</VirtualHost>#========================================================
C:\Java\Apache\Tomcat_4.1\conf\jk\workers.properties
#========================================================
# Start setup file
#==========================================================
workers.tomcat_home=C:/Java/Apache/Tomcat_4.1
workers.java_home=C:/Java/j2sdk1.4.0
ps=/
worker.list=MyWorker
#==========================================================
# Definition for MyWorker worker
#==========================================================
worker.MyWorker.port=8009
worker.MyWorker.host=192.168.0.2
worker.MyWorker.type=ajp13
worker.MyWorker.class_path=C:/Java/Apache/Tomcat_4.1/bin
worker.MyWorker.class_path=C:/Java/Apache/Tomcat_4.1/common/lib
worker.MyWorker.class_path=C:/Java/Apache/Tomcat_4.1/server/lib
worker.MyWorker.class_path=C:/Java/j2sdk1.4.0/lib/tools.jar
#==========================================================
# End setup file#==========================================================
C:\Java\Apache\Apache2\conf\httpd.conf
#==========================================================
about line# 212:
ServerName 192.168.0.2  # my internal ipat bottom of file:
LoadModule jk_module "C:/Java/Apache/Tomcat_4.1/module/mod_jk.dll"
Include C:/Java/Apache/Tomcat_4.1/conf/auto/mod_jk.conf
JkWorkersFile "C:/Java/Apache/Tomcat_4.1/conf/jk/workers.properties"
JkLogFile "C:/Java/Apache/Tomcat_4.1/logs/mod_jk.log"
JkLogLevel emerg
JkLogStampFormat "[%a %b %d %H:%M:%S %Y] "
JkMount /examples/*.jsp MyWorker
JkMount /examples/jsp/* MyWorkerWhen I put this URL in my browser:
http://192.168.0.2/examples/jsp/
I get a 404 error...
Apache keeps trying to serve the *.jsp's and it fails every time it trys to do so. I know this b/c the Tomcat logs are virtually empty while the Apache logs keep being added to like so:
192.168.0.2 - - [14/Sep/2002:19:04:57 - 0400] "GET /examples/jsp/snp/snoop.jsp HTTP/1.1" 404 294
Can anybody help me figure out how to configure Tomcat4.1 with Apache 2.0?

Similar Messages

  • RW-00023: Error:  DBC file was not created- When try to install 11i on OEL5

    I got an erorr in the appl-top log file as below, also my installation wizard showing post installation checks did not succeed. Please advice.....:-)
    DBC File Check
    RW-00023: Error: - DBC file was not created:
    File = /d01/oracle/visappl/fnd/11.5.0/secure/VIS_oracleapp1/vis.dbc
    HTTP Check
    checking URL = http://oracleapp1.srini.co.uk:8000
    RW-50015: Error: - Portal is not responding. The service might not have started on the port yet. Please check the service and use the retry button.
    RW-50015: Error: - Portal is not responding. The service might not have started on the port yet. Please check the service and use the retry button.
    JSP Check
    checking URL = http://oracleapp1.srini.co.uk:8000/OA_HTML/jsp/fnd/fndhelp.jsp?dbc=/d01/oracle/visappl/fnd/11.5.0/secure/VIS_oracleapp1/vis.dbc
    RW-50015: Error: - JSP is not responding. The service might not have started on the port yet. Please check the service and use the retry button.
    PHP Check
    checking URL = http://oracleapp1.srini.co.uk:8000/OA_HTML/US/ICXINDEX.htm
    RW-50015: Error: - Login Page is not responding. The service might not have started on the port yet. Please check the service and use the retry button.
    JSP Check
    checking URL = http://oracleapp1.srini.co.uk:8000/OA_HTML/jsp/fnd/fndhelp.jsp?dbc=/d01/oracle/visappl/fnd/11.5.0/secure/VIS_oracleapp1/vis.dbc
    RW-50015: Error: - JSP is not responding. The service might not have started on the port yet. Please check the service and use the retry button.
    PHP Check
    checking URL = http://oracleapp1.srini.co.uk:8000/OA_HTML/US/ICXINDEX.htm
    RW-50015: Error: - Login Page is not responding. The service might not have started on the port yet. Please check the service and use the retry button.

    Hi JD,
    I am able to fix this issue.....:-) Thanks for your help JD.
    As per the Metalink note 747424.1, i replaced all occurances of LD_ASSUME_KERNEL with XD_ASSUME_KERNEL only in DB-tier, but not in the Apps tier. So this was causing the
    above issue in the instalation of oracle 11i on OEL5
    --Once again i alter and replaced all occurances of LD_ASSUME_KERNEL with XD_ASSUME_KERNEL in both the files....
    (db tier)
    <DB_HOME>/appsutil/bin/adgetlnxver.sh (did it earlier)
    (apps tier)
    <APPL_TOP>/ad/11.5.0/bin/adgetlnxver.sh
    --run  the autoconfig in apps tier with completed successfully result (and also generated the  DBC files)
    --run the adstrtal.sh and retry the rapidwiz once more. ( Oracle Discoverer services VIS_oracleapp1 - addisctl.sh - this one is Failed)
    --This time everything checked OK in the instalation wizard.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • I have a problem. When exporting from PDF to PPTX error "unable to process the document in the module Save As file is not created." What to do?

    I have a problem. When exporting from PDF to PPTX error "unable to process the document in the module Save As file is not created." What to do?
    Windows 7 64
    PC

    everything works on a laptop (

  • PDF file is not created properly when I choose File- Print...- PDF-- Save as PDF

    PDF file is not created properly when I choose File->Print...In dialogue box if i select option Save as PDF under PDF button...But it works fine with Safari Browser..
    == This happened ==
    Every time Firefox opened
    == always

    My first thought is to make sure the fonts used are common. If there is a substitution, things will be off.

  • Log file is not created used util.logger

    I need to write a log into a file using java.util.logging.
    File name should read from properties file.
    Issue is file is not created in D:\TestLogging.log
    Below are the configuration:
    src\resources\logger.properties
    handlers=java.util.logging.FileHandler, java.util.logging.ConsoleHandler
    .level=INFO
    java.util.logging.ConsoleHandler.level=SEVERE
    java.util.logging.ConsoleHandler.formatter=java.util.logging.SimpleFormatter
    java.util.logging.FileHandler.level=ALL
    java.util.logging.FileHandler.pattern=D:\TestLogging.log
    # Number of output files to cycle through, by appending an
    # integer to the base file name:
    java.util.logging.FileHandler.count=1
    # Style of output (Simple or XML):
    java.util.logging.FileHandler.formatter=java.util.logging.SimpleFormatter
    SimpleLogger.java
    import java.util.logging.*;
    public final class SimpleLogger {
      public static void main(String argv[]) {
      SimpleLogger thing = new SimpleLogger();
      thing.doSomething();
      public void doSomething() {
      fLogger.finest("this is finest");
      fLogger.finer("this is finer");
      fLogger.fine("this is fine");
      fLogger.config("this is config");
      fLogger.info("this is info");
      fLogger.warning("this is a warning");
      fLogger.severe("this is severe");
      fLogger.logp(Level.INFO, this.getClass().toString(), "doSomething", "blah");
      Throwable ex = new IllegalArgumentException("Some exception text");
      fLogger.log(Level.SEVERE, "Some message", ex);
      fLogger.exiting(this.getClass().toString(), "doSomething");
      private static final Logger fLogger =
      Logger.getLogger(SimpleLogger.class.getName());
    Could you please help, any configuration is missed

    Malar,
    Check for the file in the server folder of your server installed directory.
    ie..<Drive where server installed>:\usr\sap\<name>\<instance>\j2ee\<server node>
    Regards
    Abu Bakar

  • Backups of the JVM.CONFIG file are not created

    Backups of the JVM.CONFIG file are not created.
    CF10 > Server Settings > Java and JVM   in part says:
    Backups of the jvm.config file are created when you hit the submit button. You can use this backup to restore from a critical change.
    CF10 Help page indicates jvm.bak will be created:
    http://help.adobe.com/en_US/ColdFusion/10.0/Admin/WSc3ff6d0ea77859461172e0811cbf3638e6-7ff c.html#WSc3ff6d0ea77859461172e0811cbf3638e6-7feb
    Note Help indicates incorrectly JVM.CONFIG and BAK are in cf_root\runtime\bin where as JVM.CONFIG is located cf_root\instance\bin .
    Thanks in advance, Carl.

    I must have not had enough coffee - JVM.BAK is created in where JVM.CONFIG is found. Pardon me.

  • Jar file: log file is not created!

    Hi, I create a log file with these instructions:
    //logger
              FileHandler fh = new FileHandler(LOG_FILE,true); //append mode
              fh.setFormatter(new SimpleFormatter());
              logger = Logger.getLogger(this.getClass().getName());
              logger.addHandler(fh);It works well as long I execute it normally, but when I create a .jar file and put this logger class inside logger file is not created..how come?

    Hi, I create a log file with these instructions:The code extract creates a Logger object, sure, but does that mean it creates a log file? The log file creation could be deferred until the first usage of the logger...
    It works well as long I execute it normally,Let's clarify: do you mean, when executing from your IDE, or when executing via a command line?
    but when I create a .jar file and put this logger class inside logger file is not created..how come?Let's clarify: do you mean, when you execute the same code packaged in a jar file?
    The most likely issue is configuration: apparently the various ways in which you run your program use different logging configurations.
    J.

  • Log file is not created using LOG4J

    Hi all,
    I want to use Log4j to log the details about my portal application. I would prefer using Log4j instead of SAP Logging. I have created properties file in my portal application which extends Abstract Portal Component. This my Log4j.properties file which i created under dist\PORTAL-INF\classses folder. I could read the Properties file. There is no problem on that. I am able to refer the API s available in Log4j. I am using Log4j-1[1].2.14.jar.
    log4j.rootLogger = INFO, R1
    log4j.appender.R1 = org.apache.log4j.RollingFileAppender
    log4j.appender.R1.File = LoggerForMyApplicationC.log
    log4j.appender.R1.MaxFileSize=100KB
    log4j.appender.R1.MaxBackupIndex=1
    log4j.appender.R1.layout=org.apache.log4j.PatternLayout
    log4j.appender.R1.layout.ConversionPattern=[%-5p] [%d] [%c] - [%m]%n
    But the problem is Log file is not created (LoggerForMyApplicationC.log) under \logs foler in SAP J2EE Server.
    Do i miss any other configuration? Please bring me the solution.
    Thanks,
    Malar

    Malar,
    Check for the file in the server folder of your server installed directory.
    ie..<Drive where server installed>:\usr\sap\<name>\<instance>\j2ee\<server node>
    Regards
    Abu Bakar

  • File to file secnario output file is not created.

    Hi,
    while creating the file to file secnario, file is picked by xi server and shown to sxmb_moni but output file is not created.
    Please help.

    HI
    Go to RWB->Message Monitoring->Message Display Tool(MDT)->Audit Log and check the status..
    Go to SMQ2 and check is there any queue is pending in the Queue..
    If so try to activate it. For this you can refer this blog-
    XI :  How to Re-Process failed XI Messages Automatically
    If not-then check the configuration again-
    1) Check the Receiver Communication channel for the FTP URL/File creation mode etc
    2) Are you able to see the mapped data in SXMB_MONI..
    3) Reciever Agreement/Reciever Determiation/Interface Determination
    Check if you have given a Write Access to the folder in which the file is to be placed
    Thanks
    Gaurav

  • OTC ssl.conf file does not have ssl wallet file location details

    Hi B2B Gurus,
    While doing the HTTPS configuration in OTC we observed that some of lines were missing in ssl.conf file located at <Oracle_Home>\Apache\Apache\conf. The below mentioned lines are missing in OTC’s ssl.conf but present in OracleB2B’s ssl.conf
    Even ssl wallet file location is also missing in OTC ssl.conf file as mentioned below.
    Can u please let us know as why these lines are missing or should we manually add these lines when we do HTTPS on OTC ?
    Listen 4444
    <VirtualHost default:4444>
    # General setup for the virtual host
    DocumentRoot "E:\Oracle_b2b\cachehome\Apache\Apache\htdocs"
    ServerName DSCP17506.TechMahindra.com
    ServerAdmin [email protected]
    ErrorLog "|E:\Oracle_b2b\cachehome\Apache\Apache\bin\rotatelogs logs/error_log 43200"
    TransferLog "|E:\Oracle_b2b\cachehome\Apache\Apache\bin\rotatelogs logs/access_log 43200"
    Port 443
    # SSL Engine Switch:
    # Enable/Disable SSL for this virtual host.
    SSLEngine on
    # SSL Cipher Suite:
    # List the ciphers that the client is permitted to negotiate.
    SSLCipherSuite ALL:!ADH:!EXPORT56:+HIGH:+MEDIUM:+LOW:+SSLv2:+EXP
    # Server Wallet:
    # The server wallet contains the server's certificate, private key
    # and trusted certificates. Set SSLWallet at the wallet directory
    # using the syntax: file:<path-to-wallet-directory>
    SSLWallet file:E:\Oracle_b2b\cachehome\Apache\Apache\conf\ssl.wlt\default
    # Certificate Revocation Lists (CRL):
    # Set the CA revocation path where to find CA CRLs for client
    # authentication or alternatively one huge file containing all
    # of them (file must be PEM encoded)
    # Note: Inside SSLCARevocationPath you need hash symlinks
    # to point to the certificate files. Use the provided
    # Makefile to update the hash symlinks after changes.
    #SSLCARevocationPath conf\ssl.crl
    #SSLCARevocationFile conf\ssl.crl\ca-bundle.crl
    # Client Authentication (Type):
    # Client certificate verification type and depth. Types are
    # none, optional and require
    #SSLVerifyClient require
    # Access Control:
    # With SSLRequire you can do per-directory access control based
    # on arbitrary complex boolean expressions containing server
    # variable checks and other lookup directives. The syntax is a
    # mixture between C and Perl. See the mod_ssl documentation
    # for more details.
    #<Location />
    #SSLRequire ( %{SSL_CIPHER} !~ m/^(EXP|NULL)-/ \
    # and %{SSL_CLIENT_S_DN_O} eq "Snake Oil, Ltd." \
    # and %{SSL_CLIENT_S_DN_OU} in {"Staff", "CA", "Dev"} \
    # and %{TIME_WDAY} >= 1 and %{TIME_WDAY} <= 5 \
    # and %{TIME_HOUR} >= 8 and %{TIME_HOUR} <= 20 ) \
    # or %{REMOTE_ADDR} =~ m/^192\.76\.162\.[0-9]+$/
    #</Location>
    # SSL Engine Options:
    # Set various options for the SSL engine.
    # o FakeBasicAuth:
    # Translate the client X.509 into a Basic Authorisation. This means that
    # the standard Auth/DBMAuth methods can be used for access control. The
    # user name is the `one line' version of the client's X.509 certificate.
    # Note that no password is obtained from the user. Every entry in the user
    # file needs this password: `xxj31ZMTZzkVA'.
    # o ExportCertData:
    # This exports two additional environment variables: SSL_CLIENT_CERT and
    # SSL_SERVER_CERT. These contain the PEM-encoded certificates of the
    # server (always existing) and the client (only existing when client
    # authentication is used). This can be used to import the certificates
    # into CGI scripts.
    # o StdEnvVars:
    # This exports the standard SSL/TLS related `SSL_*' environment variables.
    # Per default this exportation is switched off for performance reasons,
    # because the extraction step is an expensive operation and is usually
    # useless for serving static content. So one usually enables the
    # exportation for CGI and SSI requests only.
    # o CompatEnvVars:
    # This exports obsolete environment variables for backward compatibility
    # to Apache-SSL 1.x, mod_ssl 2.0.x, Sioux 1.0 and Stronghold 2.x. Use this
    # to provide compatibility to existing CGI scripts.
    # o StrictRequire:
    # This denies access when "SSLRequireSSL" or "SSLRequire" applied even
    # under a "Satisfy any" situation, i.e. when it applies access is denied
    # and no other module can change it.
    # o OptRenegotiate:
    # This enables optimized SSL connection renegotiation handling when SSL
    # directives are used in per-directory context.
    #SSLOptions FakeBasicAuth ExportCertData CompatEnvVars StrictRequire
    <Files ~ "\.(cgi|shtml)$">
    SSLOptions +StdEnvVars
    </Files>
    <Directory "E:\Oracle_b2b\cachehome\Apache\Apache\cgi-bin">
    SSLOptions +StdEnvVars
    </Directory>
    SetEnvIf User-Agent ".*MSIE.*" nokeepalive ssl-unclean-shutdown
    # Per-Server Logging:
    # The home of a custom SSL log file. Use this when you want a
    # compact non-error SSL logfile on a virtual host basis.
    CustomLog E:\Oracle_b2b\cachehome\Apache\Apache\logs\ssl_request_log "%t %h %{SSL_PROTOCOL}x %{SSL_CIPHER}x \"%r\" %b"
    </VirtualHost>
    Thanks for your help in Advance.
    Regards,
    Priyanka

    Hi Ramesh,
    We added the missing entries in ssl.conf file and tested it out. It is working fine. When we did slient install we are facing this error, we found that this is a bug in 10.1.2.3 AS patch (5983622) silent install. Raised an SR on this and Support is working to get this fixed.
    Regards,
    Priyanka

  • Bad file is not created during the external table creation.

    Hello Experts,
    I have created a script for external table in Oracle 10g DB. Everything is working fine except it does not create the bad file, But it creates the log file. I Cann't figure out what is the issue. Because my shell scripts is failing and the entire program is failing. I am attaching the table creation script and the shell script where it is refering and the error. Kindly let me know if something is missing. Thanks in advance
    Table Creation Scripts:_-------------------------------
    create table RGIS_TCA_DATA_EXT
    guid VARCHAR2(250),
    badge VARCHAR2(250),
    scheduled_store_id VARCHAR2(250),
    parent_event_id VARCHAR2(250),
    event_id VARCHAR2(250),
    organization_number VARCHAR2(250),
    customer_number VARCHAR2(250),
    store_number VARCHAR2(250),
    inventory_date VARCHAR2(250),
    full_name VARCHAR2(250),
    punch_type VARCHAR2(250),
    punch_start_date_time VARCHAR2(250),
    punch_end_date_time VARCHAR2(250),
    event_meet_site_id VARCHAR2(250),
    vehicle_number VARCHAR2(250),
    vehicle_description VARCHAR2(250),
    vehicle_type VARCHAR2(250),
    is_owner VARCHAR2(250),
    driver_passenger VARCHAR2(250),
    mileage VARCHAR2(250),
    adder_code VARCHAR2(250),
    bonus_qualifier_code VARCHAR2(250),
    store_accuracy VARCHAR2(250),
    store_length VARCHAR2(250),
    badge_input_type VARCHAR2(250),
    source VARCHAR2(250),
    created_by VARCHAR2(250),
    created_date_time VARCHAR2(250),
    updated_by VARCHAR2(250),
    updated_date_time VARCHAR2(250),
    approver_badge_id VARCHAR2(250),
    approver_name VARCHAR2(250),
    orig_guid VARCHAR2(250),
    edit_type VARCHAR2(250)
    organization external
    type ORACLE_LOADER
    default directory ETIME_LOAD_DIR
    access parameters
    RECORDS DELIMITED BY NEWLINE
    BADFILE ETIME_LOAD_DIR:'tstlms.bad'
    LOGFILE ETIME_LOAD_DIR:'tstlms.log'
    READSIZE 1048576
    FIELDS TERMINATED BY '|'
    MISSING FIELD VALUES ARE NULL(
    GUID
    ,BADGE
    ,SCHEDULED_STORE_ID
    ,PARENT_EVENT_ID
    ,EVENT_ID
    ,ORGANIZATION_NUMBER
    ,CUSTOMER_NUMBER
    ,STORE_NUMBER
    ,INVENTORY_DATE char date_format date mask "YYYYMMDD HH24:MI:SS"
    ,FULL_NAME
    ,PUNCH_TYPE
    ,PUNCH_START_DATE_TIME char date_format date mask "YYYYMMDD HH24:MI:SS"
    ,PUNCH_END_DATE_TIME char date_format date mask "YYYYMMDD HH24:MI:SS"
    ,EVENT_MEET_SITE_ID
    ,VEHICLE_NUMBER
    ,VEHICLE_DESCRIPTION
    ,VEHICLE_TYPE
    ,IS_OWNER
    ,DRIVER_PASSENGER
    ,MILEAGE
    ,ADDER_CODE
    ,BONUS_QUALIFIER_CODE
    ,STORE_ACCURACY
    ,STORE_LENGTH
    ,BADGE_INPUT_TYPE
    ,SOURCE
    ,CREATED_BY
    ,CREATED_DATE_TIME char date_format date mask "YYYYMMDD HH24:MI:SS"
    ,UPDATED_BY
    ,UPDATED_DATE_TIME char date_format date mask "YYYYMMDD HH24:MI:SS"
    ,APPROVER_BADGE_ID
    ,APPROVER_NAME
    ,ORIG_GUID
    ,EDIT_TYPE
    location (ETIME_LOAD_DIR:'tstlms.dat')
    reject limit UNLIMITED;
    _***Shell Script*:*----------------_*
    version=1.0
    umask 000
    DATE=`date +%Y%m%d%H%M%S`
    TIME=`date +"%H%M%S"`
    SOURCE=`hostname`
    fcp_login=`echo $1|awk '{print $3}'|sed 's/"//g'|awk -F= '{print $2}'`
    fcp_reqid=`echo $1|awk '{print $2}'|sed 's/"//g'|awk -F= '{print $2}'`
    TXT1_PATH=/home/ac1/oracle/in/tsdata
    TXT2_PATH=/home/ac2/oracle/in/tsdata
    ARCH1_PATH=/home/ac1/oracle/in/tsdata
    ARCH2_PATH=/home/ac2/oracle/in/tsdata
    DEST_PATH=/home/custom/sched/in
    PROGLOG=/home/custom/sched/logs/rgis_tca_to_tlms_create.sh.log
    PROGNAME=`basename $0`
    PROGPATH=/home/custom/sched/scripts
    cd $TXT2_PATH
    FILELIST2="`ls -lrt tstlmsedits*.dat |awk '{print $9}'`"
    NO_OF_FILES2="`ls -lrt tstlmsedits*.dat |awk '{print $9}'|wc -l`"
    $DEST_PATH/tstlmsedits.dat for i in $FILELIST2
    do
    cat $i >> $DEST_PATH/tstlmsedits.dat
    printf "\n" >> $DEST_PATH/tstlmsedits.dat
    mv $i $i.$DATE
    #mv $i $TXT2_PATH/test/.
    mv $i.$DATE $TXT2_PATH/test/.
    done
    if test $NO_OF_FILES2 -eq 0
    then
    echo " no tstlmsedits.dat file exists " >> $PROGLOG
    else
    echo "created dat file tstlmsedits.dat at $DATE" >> $PROGLOG
    echo "-------------------------------------------" >> $PROGLOG
    fi
    NO_OF_FILES1="`ls -lrt tstlms*.dat |awk '{print $9}'|wc -l`"
    FILELIST1="`ls -lrt tstlms*.dat |awk '{print $9}'`"
    $DEST_PATH/tstlms.datfor i in $FILELIST1
    do
    cat $i >> $DEST_PATH/tstlms.dat
    printf "\n" >> $DEST_PATH/tstlms.dat
    mv $i $i.$DATE
    # mv $i $TXT2_PATH/test/.
    mv $i.$DATE $TXT2_PATH/test/.
    done
    if test $NO_OF_FILES1 -eq 0
    then
    echo " no tstlms.dat file exists " >> $PROGLOG
    else
    echo "created dat file tstlms.dat at $DATE" >> $PROGLOG
    fi
    cd $TXT1_PATH
    FILELIST3="`ls -lrt tstlmsedits*.dat |awk '{print $9}'`"
    NO_OF_FILES3="`ls -lrt tstlmsedits*.dat |awk '{print $9}'|wc -l`"
    $DEST_PATH/tstlmsedits.datfor i in $FILELIST3
    do
    cat $i >> $DEST_PATH/tstlmsedits.dat
    printf "\n" >> $DEST_PATH/tstlmsedits.dat
    mv $i $i.$DATE
    #mv $i $TXT1_PATH/test/.
    mv $i.$DATE $TXT1_PATH/test/.
    done
    if test $NO_OF_FILES3 -eq 0
    then
    echo " no tstlmsedits.dat file exists " >> $PROGLOG
    else
    echo "created dat file tstlmsedits.dat at $DATE" >> $PROGLOG
    echo "-------------------------------------------" >> $PROGLOG
    fi
    NO_OF_FILES4="`ls -lrt tstlms*.dat |awk '{print $9}'|wc -l`"
    FILELIST4="`ls -lrt tstlms*.dat |awk '{print $9}'`"
    $DEST_PATH/tstlms.datfor i in $FILELIST4
    do
    cat $i >> $DEST_PATH/tstlms.dat
    printf "\n" >> $DEST_PATH/tstlms.dat
    mv $i $i.$DATE
    # mv $i $TXT1_PATH/test/.
    mv $i.$DATE $TXT1_PATH/test/.
    done
    if test $NO_OF_FILES4 -eq 0
    then
    echo " no tstlms.dat file exists " >> $PROGLOG
    else
    echo "created dat file tstlms.dat at $DATE" >> $PROGLOG
    fi
    #connecting to oracle to generate bad files
    sqlplus -s $fcp_login<<EOF
    select count(*) from rgis_tca_data_ext;
    select count(*) from rgis_tca_data_history_ext;
    exit;
    EOF
    #counting the records in files
    tot_rec_in_tstlms=`wc -l $DEST_PATH/tstlms.dat | awk ' { print $1 } '`
    tot_rec_in_tstlmsedits=`wc -l $DEST_PATH/tstlmsedits.dat | awk ' { print $1 } '`
    tot_rec_in_tstlms_bad=`wc -l $DEST_PATH/tstlms.bad | awk ' { print $1 } '`
    tot_rec_in_tstlmsedits_bad=`wc -l $DEST_PATH/tstlmsedits.bad | awk ' { print $1 } '`
    #updating log table
    echo "pl/sql block started"
    sqlplus -s $fcp_login<<EOF
    define tot_rec_in_tstlms     = '$tot_rec_in_tstlms';
    define tot_rec_in_tstlmsedits     = '$tot_rec_in_tstlmsedits';
    define tot_rec_in_tstlms_bad     = '$tot_rec_in_tstlms_bad';
    define tot_rec_in_tstlmsedits_bad='$tot_rec_in_tstlmsedits_bad';
    define fcp_reqid ='$fcp_reqid';
    declare
    l_tstlms_file_id number := null;
    l_tstlmsedits_file_id number := null;
    l_tot_rec_in_tstlms number := 0;
    l_tot_rec_in_tstlmsedits number := 0;
    l_tot_rec_in_tstlms_bad number := 0;
    l_tot_rec_in_tstlmsedits_bad number := 0;
    l_request_id fnd_concurrent_requests.request_id%type;
    l_start_date fnd_concurrent_requests.actual_start_date%type;
    l_end_date fnd_concurrent_requests.actual_completion_date%type;
    l_conc_prog_name fnd_concurrent_programs.concurrent_program_name%type;
    l_requested_by fnd_concurrent_requests.requested_by%type;
    l_requested_date fnd_concurrent_requests.request_date%type;
    begin
    --getting concurrent request details
    begin
    SELECT fcp.concurrent_program_name,
    fcr.request_id,
    fcr.actual_start_date,
    fcr.actual_completion_date,
    fcr.requested_by,
    fcr.request_date
    INTO l_conc_prog_name,
    l_request_id,
    l_start_date,
    l_end_date,
    l_requested_by,
    l_requested_date
    FROM fnd_concurrent_requests fcr, fnd_concurrent_programs fcp
    WHERE fcp.concurrent_program_id = fcr.concurrent_program_id
    AND fcr.request_id = &fcp_reqid; --fnd_global.conc_request_id();
    exception
    when no_data_found then
    fnd_file.put_line(fnd_file.log, 'Error:RGIS_TCA_TO_TLMS_CREATE.sh');
    fnd_file.put_line(fnd_file.log, 'No data found for request_id');
    fnd_file.put_line(fnd_file.log, sqlerrm);
    raise_application_error(-20001,
    'Error occured when executing RGIS_TCA_TO_TLMS_CREATE.sh ' ||
    sqlerrm);
    when others then
    fnd_file.put_line(fnd_file.log, 'Error:RGIS_TCA_TO_TLMS_CREATE.sh');
    fnd_file.put_line(fnd_file.log,
    'Error occured when retrieving request_id request_id');
    fnd_file.put_line(fnd_file.log, sqlerrm);
    raise_application_error(-20001,
    'Error occured when executing RGIS_TCA_TO_TLMS_CREATE.sh ' ||
    sqlerrm);
    end;
    --calling ins_or_upd_tca_process_log to update log table for tstlms.dat file
    begin
    rgis_tca_to_tlms_process.ins_or_upd_tca_process_log
                   (l_tstlms_file_id,
                   'tstlms.dat',
                   l_conc_prog_name,
                   l_request_id,
                   l_start_date,
                   l_end_date,
                   &tot_rec_in_tstlms,
                   &tot_rec_in_tstlms_bad,
                   null,
                   null,               
                   null,
                   null,
                   null,
                   null,
                   null,
                   l_requested_by,
                   l_requested_date,
                   null,
                   null,
                   null,
                   null,
                   null);
    exception
    when others then
    fnd_file.put_line(fnd_file.log, 'Error:RGIS_TCA_TO_TLMS_CREATE.sh');
    fnd_file.put_line(fnd_file.log,
    'Error occured when executing rgis_tca_to_tlms_process.ins_or_upd_tca_process_log for tstlms file');
    fnd_file.put_line(fnd_file.log, sqlerrm);
    end;
    --calling ins_or_upd_tca_process_log to update log table for tstlmsedits.dat file
    begin
    rgis_tca_to_tlms_process.ins_or_upd_tca_process_log
                   (l_tstlmsedits_file_id,
                   'tstlmsedits.dat',
                   l_conc_prog_name,
                   l_request_id,
                   l_start_date,
                   l_end_date,
                   &tot_rec_in_tstlmsedits,
                   &tot_rec_in_tstlmsedits_bad,
                   null,
                   null,               
                   null,
                   null,
                   null,
                   null,
                   null,
                   l_requested_by,
                   l_requested_date,
                   null,
                   null,
                   null,
                   null,
                   null);
    exception
    when others then
    fnd_file.put_line(fnd_file.log, 'Error:RGIS_TCA_TO_TLMS_CREATE.sh');
    fnd_file.put_line(fnd_file.log,
    'Error occured when executing rgis_tca_to_tlms_process.ins_or_upd_tca_process_log for tstlmsedits file');
    fnd_file.put_line(fnd_file.log, sqlerrm);
    end;
    end;
    exit;
    EOF
    echo "rgis_tca_to_tlms_process.sql started"
    sqlplus -s $fcp_login @$SCHED_TOP/sql/rgis_tca_to_tlms_process.sql $fcp_reqid
    exit;
    echo "rgis_tca_to_tlms_process.sql ended"
    _**Error:*----------------------------------*_
    RGIS Scheduling: Version : UNKNOWN
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    TCATLMS module: TCA To TLMS Import Process
    Current system time is 18-AUG-2011 06:13:27
    COUNT(*)
         16
    COUNT(*)
         25
    wc: cannot open /home/custom/sched/in/tstlms.bad
    wc: cannot open /home/custom/sched/in/tstlmsedits.bad
    pl/sql block started
    old 33:     AND fcr.request_id = &fcp_reqid; --fnd_global.conc_request_id();
    new 33:     AND fcr.request_id = 18661823; --fnd_global.conc_request_id();
    old 63:                &tot_rec_in_tstlms,
    new 63:                16,
    old 64:                &tot_rec_in_tstlms_bad,
    new 64:                ,
    old 97:                &tot_rec_in_tstlmsedits,
    new 97:                25,
    old 98:                &tot_rec_in_tstlmsedits_bad,
    new 98:                ,
    ERROR at line 64:
    ORA-06550: line 64, column 4:
    PLS-00103: Encountered the symbol "," when expecting one of the following:
    ( - + case mod new not null others <an identifier>
    <a double-quoted delimited-identifier> <a bind variable> avg
    count current exists max min prior sql stddev sum variance
    execute forall merge time timestamp interval date
    <a string literal with character set specification>
    <a number> <a single-quoted SQL string> pipe
    <an alternatively-quoted string literal with character set specification>
    <an alternatively-q
    ORA-06550: line 98, column 4:
    PLS-00103: Encountered the symbol "," when expecting one of the following:
    ( - + case mod new not null others <an identifier>
    <a double-quoted delimited-identifier> <a bind variable> avg
    count current exists max min prior sql st
    rgis_tca_to_tlms_process.sql started
    old 12: and concurrent_request_id = '&1';
    new 12: and concurrent_request_id = '18661823';
    old 18: and concurrent_request_id = '&1';
    new 18: and concurrent_request_id = '18661823';
    old 22: rgis_tca_to_tlms_process.run_tca_data(l_tstlms_file_id,&1);
    new 22: rgis_tca_to_tlms_process.run_tca_data(l_tstlms_file_id,18661823);
    old 33: rgis_tca_to_tlms_process.run_tca_data_history(l_tstlmsedits_file_id,&1);
    new 33: rgis_tca_to_tlms_process.run_tca_data_history(l_tstlmsedits_file_id,18661823);
    old 44: rgis_tca_to_tlms_process.send_tca_email('TCATLMS',&1);
    new 44: rgis_tca_to_tlms_process.send_tca_email('TCATLMS',18661823);
    declare
    ERROR at line 1:
    ORA-20001: Error occured when executing RGIS_TCA_TO_TLMS_PROCESS.sql ORA-01403:
    no data found
    ORA-06512: at line 59
    Executing request completion options...
    ------------- 1) PRINT   -------------
    Printing output file.
    Request ID : 18661823      
    Number of copies : 0      
    Printer : noprint
    Finished executing request completion options.
    Concurrent request completed successfully
    Current system time is 18-AUG-2011 06:13:29
    ---------------------------------------------------------------------------

    Hi,
    Check the status of the batch in SM35 transaction.
    if the batch is locked by mistake or any other error, now you can release it and aslo you can process again.
    To Release -Shift+F4.
    Also you can analyse the job status through F2 button.
    Bye

  • Planning File entry not created automatically

    Hi,
    What could be the reasons for not creating planning file entry automatically for a material..? Its MRP type is P3.
    How to check whether all the material in a particular plant is considered for MRP or not..?
    Does the transaction MDRE - Consistency Check have any role in this..?
    Pls. clarify.
    TIA
    Sabitha

    Dear Sabitha,
    In MD21, there are no entries for that material.
    It doesn't mean that planning File entry is not maintain for this material.
    If non of the tabs of NChge Pln,NChge P Hor, etc are activate then it means there is no req. for this material either with in planning horizon or out planning horizon.
    If planning file entry is not maintain for any material then you will get below msg in md21 at the time of execution,
    No planning file entry exists for this selection
    Message no. 61140
    Kindly check and revert back,
    Regards,
    Dhaval

  • Planning File Entry not created

    Hi,
    I have a list of materials with MRP type PD and P3 for which there is no planning file entries exists. I am getting the following msg in MD21,
    NO PLANNING FILE ENTRY EXISTS FOR THIS SELECTION
    I have the option of creating planning file entries manually, but i want to find out the reason for not creating planning file entry.
    What other parameters and configuration settings i need to check..? Should the BOM and routing mandatory for a material to create planning file entries?
    Pls. clarify.
    Thanks in advance,
    Sabitha

    Hi,
    When you create materials via LSMW, just before you start the LSMW, there is a check box for Create Planning File entries, i guess this checkbox was not ticked & hence no planning file entries.
    To double confirm, check if all the materials which you created via this LSMW have planning file entries, if it does not, it should answer your query.
    To create planning file entry, refer to my initial posting.
    Regards,
    Vivek

  • While compiling .pll file , plx  file is not creating .

    Hi ,
    While i try to compile .pll file in oracle forms 6i in my local machine , it is compiling without errors .
    but it is not generating .plx file. Can any one please tell me what could be the reason.
    Thanks in advance,
    Sanjeev.

    While i try to compile .pll file in oracle forms 6i in my local machine , it is compiling without errors .but it is not generating .plx file. Can any one please tell me what could be the reason.>
    Ctrl+Shift+K does not create executable , Ctrl+T does.
    Did you do Ctrl+T?
    Cheers,

  • Video Sidecar Files (.THM) Not Created When Video Exported As Original - They should be.

    When exporting video as in it's original format the .THM sidecar file (which contains all the relavent metadata for the video) should be included so that other programs (or even Lightroom itself if using the Add to This Catalog option during export) can have access to the metadata for the video.
    Here is my use case:
    After importing the original video clips I would like to do the following to keep my footage clean and conserve space (as we all know HD video eats a LOT of space)
    1. Trim excess / unneeded footage from the clip
    2. Export the clip in it's original format with a new name (such as MVI_1234_trimmed.mov) and have it added back to the catalog during the export process.
    3. Delete the original clip containing the unneeded footage.
    This workflow is almost perfect in the initial LR 4 beta release except that a copy of the .THM file is not included when the video is exported.  This means the new clip will not contain any metadata when added to LR or presumabley other applications.  I can manuall get around this by duplicating the original .THM file and renaming the duplicate to match the name of the new trimmed clip (example MVI_1234_trimmed.THM) then manually importing into Lightroom.  However this breaks the almost perfect workflow and increases time and room for error.  It would no longer be an automated process which could easily be applied to clips in bulk.
    Please have LR create the sidecar file with when exporting in original format or at least have an option for it!

    The proposed workflow may not be valid based on the discussion here: http://forums.adobe.com/thread/947245?tstart=0 but the sidecar file point still is.  Exporting as H.264 does not preserve metadata either.  LR shows no camera / exposure / date time information for the H.264 version.

Maybe you are looking for

  • SID Showing instead of friendly name after renaming the server

    I had to rename a windows 2012 R2 server and after I rename it, Sever is showing SID instead of user friendly name. Even when I try to add new domain group or user in any resourse of this server, the friendly name disappear, soon after I hit apply bu

  • Norwegian language for spell checking in pages

    Where can i donload norwegian language for spell checking in pages ? and how to install it? Thanks for any answers...

  • Question about intensity graphs

    Hi all Labview gurus and lovers, I have a question to all of you. Last time I encountered with the interesting problem connecting to the building of the intensity graphs from the measurements. Suppose I have a matrix of numbers actually the voltages

  • Plese go thought this

    Ex1.java:2: package javax.servlet does not exist import javax.servlet.*; ^ Ex1.java:3: package javax.servlet.Http does not exist import javax.servlet.Http.*; ^ Ex1.java:5: cannot find symbol symbol: class HttpServlet public class Ex1 extends HttpServ

  • CS4 and using Beyond Trust

    Has anyone had any experience/luck running CS4 PPro/AME/Encore without admin rights? If you got it to work... HOW? Our IT dept is taking away admin rights and functionality of these 3 is greatly diminished (like I can't get a finished product out of