Avoid creation of log file for external table

Hi
This script is creating log file in the ext directory. How to avoid it. Can you specify the syntex.
Thanks alot.
Bhaskar
CREATE TABLE datfiles_list
(file_name varchar2(255))
ORGANIZATION EXTERNAL
(TYPE ORACLE_LOADER
DEFAULT DIRECTORY ext_dir
ACCESS PARAMETERS (RECORDS DELIMITED BY NEWLINE)
LOCATION ('datfiles_list.txt')
);

Example
CREATE TABLE datfiles_list
(file_name varchar2(255))
ORGANIZATION EXTERNAL
(TYPE ORACLE_LOADER
DEFAULT DIRECTORY ext_dir
ACCESS PARAMETERS (RECORDS DELIMITED BY NEWLINE NOLOGFILE)
LOCATION ('datfiles_list.txt')
);

Similar Messages

  • Log files  in external table

    Hi,
    I have created an external table. And it has been run successfully. When i view the log file i am not able to find out the counts like
    no of records loaded,
    no of records rejected.
    Kindly advice me on this.
    Cheers
    Ramkannan.A

    Hi,
    ORGANIZATION EXTERNAL
    ( type oracle_loader
    default directory PARAGON_FEED_DIR
    access parameters
    records delimited by newline skip 1
    BADFILE PARAGON_FEED_LOG:'pgon_mapunmap.bad'
    LOGFILE PARAGON_FEED_LOG:'pgon_mapunmap.log'
    date_cache = 650000
    STRING SIZES ARE IN CHARACTERS
    CHARACTERSET = WE8ISO8859P1
    fields terminated by "," OPTIONALLY ENCLOSED BY '"' LDRTRIM
    missing field values are null
    In the Log file
    Field Definitions for table PGON_MAPUNMAP_LOAD
    Record format DELIMITED BY NEWLINE
    Data in file has same endianness as the platform
    Rows with all null fields are accepted
    Fields in Data Source:
    SOURCESYSTEMNAME CHAR (255)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    SOURCESYSTEMLOCATION CHAR (255)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    SOURCEID CHAR (255)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    SOURCESHORT CHAR (255)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    SOURCELEGAL CHAR (255)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    SOURCECOUNTRY CHAR (255)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    ORGID CHAR (255)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    ORGSHORT CHAR (255)
    I want to know how many records are loaded & how many records are rejected in Log file.
    But i am not able to see that

  • Avoid creation of .part files for not started downloads

    Hi,
    whenever I right click on an item and do a "save link as..." FF creates a .part file on the desktop (my regular download folder) even if I don't start and cancel the download at this point (the point where the app asks under which name I want to save the file).
    The .part file stays afterwards, even after I quit/restart FF.
    The same thing happens when I click on any download's link since I made FF ask where I want the files to be downloaded in the preferences.
    I noticed this behavior some time ago (around FF5) and I didn't bother since I clean my desktop regularly, but it's getting a bit annoying.
    Double question :
    - is this behavior normal ?
    - shouldn't it be more logical to create the .part file only if the download is actually started?
    Perhaps useful info : I use OsX10.5.8 and now on FF14

    https://bugzilla.mozilla.org/show_bug.cgi?id=420355

  • How to overwrite a log and bad file in external table in oracle 10g

    Hi,
    I have used external table in oracle 10g.whenever use select query in external table orace internally create one log file in specified directory,
    but this log file is growing.How can i overwrite the log file(old to replace with new).I need overwrite a log and bad file in external table.
    kindly give the solutions.
    By
    Siva

    I don't believe that is possible with the LOGFILE clause, but it may be with the BADFILE clause. Here is an excerpt from the documentation :
    The LOGFILE clause names the file that contains messages generated by the external tables utility while it was accessing data in the datafile. If a log file already exists by the same name, the access driver reopens that log file and appends new log information to the end. This is different from bad files and discard files, which overwrite any existing file. NOLOGFILE is used to prevent creation of a log file.
    If you specify LOGFILE, you must specify a filename or you will receive an error.
    If neither LOGFILE nor NOLOGFILE is specified, the default is to create a log file. The name of the file will be the table name followed by _%p and it will have an extension of .log.

  • How to avoid creation of _(file.*), coping from tiger to pc?

    how to avoid the creation of this file for ever on a pc environment??
    I explain: as i copy any image file, it goes to a spooler directory and that begins instantly to process the file to an OPI server, separating CMYK channels, converting it to high and low resolution...
    but as this _(file) is created (same name and extension, but 0Kb and with this _ underscore in the beggining of it name), it begins to process it, and as it really does not exists as an image itself, it crashes the OPI server, forcing us to restart it...
    and then, you can imagine the mess, because we're a newspaper, with more than 100 machines logged in two cities simultaneously...
    thanks for those who replied last topic I posted about some softwares to delete it! it really worked!!
    but now, can anyone have a clue how to DON'T create this file?
    thanks in advance!
    alex borba
    [email protected]

    I believe that command is supposed to stop the creation .DS_Store files on servers. Also, I think the plist file gets written to the user ~/Library/Pref, it probably ought to exist both there and in the /Library/Pref folder to really work.
    I didn't think the Save for Web option would work for your purposes. I don't believe anything will work except to strip the resource fork off the files before you send them. There is an Applescript that will do it, using a UNIX command in the script. You select a folder, run the Applescript, and it strips the resource fork off all files in the folder and creates a sub-folder called Stripped with resourceless copies of the files in the original:
    -- Strip resource fork and metadata in Tiger for one file.
    -- If a folder is selected creates an unresourced subfolder of the selected folder
    tell application "Finder"
         try
              set aFile to the selection as alias
              set aFolder to the container of aFile
         on error
              display dialog "Select a file"
              return
         end try
         set aFile to the quoted form of POSIX path of aFile
         do shell script "rsync -a " & aFile & " " & aFile & "Stripped; mv " & aFile & "Stripped " & aFile
         update aFolder
    end tell
    The "do shell script" line must be all a single line in Script Editor. It might wrap funny in the browser. Anyway, copy it, paste into Script Editor, hit compile, and save it as an application. I've only fooled with it a little tiny bit, so I'm not sure whether the rsync process would have any strange effects on complex Photoshop and EPS files or not.
    As far as I know the only way to send Mac files without resource forks to a server is to strip them from the files.
    Francine

  • Can we cleanse and transform data at flat file or external table level?

    Hi,
    I have some data that I want to cleanse and transform. I don't want to cleanse it after i populate the external table, I want to get done with it at flat file level or while populating the external table. Can we cleanse and transform data at flat file or external table level through Oracle or OWB 11.2? Is it possible to run a conditional load (i.e. having a where clause or if-else-then) for an external table? Can we call oracle functions for an external table at the time of creation?
    Thanks in advance.
    Regards,
    Ann.

    Hi Oleg,
    Thanks a lot for the clarification. :)
    So is there a way that I can clease the data within the text file through Oracle or OWB? I have datatype mismatches in the data and most of my data is getting rejected because of that. The way I can think of, for solving this problem, is to create the external table with all fields with datatype varchar and then cleansing the data. But it doesn't seem very effecient plus it will get very complicated because I have almost 80-90 fields.
    Any help?
    Thanks and regards,
    Ann.

  • Creation of rule file for adding aliases to Members

    Please help me in creation of rule file for adding of aliases to the members. I have created a new alias table along with default alias table.
    These are the members and alias names for the newly created alias table.
    0000002452     TA     
    0000002459     HJ
    0000002460     HK
    0000002462     HM
    0000004012     CB
    0000004015     EA
    0000004025     BA
    0000004070     DA
    0000004087     GC     
    0000004088     FD
    0000006231     CM
    0000009103     WP
    0000009106     WV
    0000009126     WW
    0000008458     NH
    0000008475     NY
    0000008481     NF
    0000008483     NG
    The below are the members and aliases for the default alias table.
    0000001209     P&S Storage Facilities
    0000001210     1210 US Procurement & Supply
    0000002452     Warren Truck Assembly
    0000002459     Toledo I Assembly
    0000002460     Toledo II Assembly
    0000002462     Toledo III Assembly
    0000004001     US Manufacturing Group Office
    0000004012     Jefferson North Assembly
    0000004015     Belvidere Assembly
    0000004025     Sterling Heights Assembly
    0000004070     Newark Assembly
    0000004087     St Louis North Assembly II
    0000004088     St Louis South Assembly II
    0000004199     Pilot Operations
    0000004501     Opm
    0000005111     Kokomo Casting
    0000005203     Trenton Engine
    0000005205     Kenosha Engine
    0000005209     New Mack Avenue Engine
    0000005304     Detroit Axle
    0000005305     Kokomo Complex
    0000005308     Indiana Transmission
    0000005403     New Castle Machining & Forge
    0000005404     McGraw Glass
    0000005406     Toledo Machining
    0000006203     Warren Stamping
    0000006204     Twinsburg Stamping
    0000006215     Sterling Heights Stamping
    0000006231     Conner Avenue Assembly
    0000002100     2100 Sales & Marketing
    0000005221     GEMA - Plant1 Ops
    0000009006     9006 Canadian Customs
    0000009008     Canadian Manufacturing Group Office
    0000009103     Windsor Assembly
    0000009104     Etobicoke Casting
    0000009106     Pillette Road Assembly
    0000009126     Bramalea Assembly
    0000008458     Atos
    0000008475     Saltillo Truck Plant
    0000008481     Toluca Assembly Plant
    0000008483     Lago Alberto Truck Plant
    Now I have to create rule files to update these aliases.
    Please elaborate, so that it will be clear for me. As I am new, please make it up with steps.

    You dont need to create a rule file. You can create a alias table like this in a notapad
    $ALT_NAME 'First'
    0000002452 TA
    0000002459 HJ
    0000002460 HK
    0000002462 HM
    0000004012 CB
    0000004015 EA
    0000004025 BA
    0000004070 DA
    0000004087 GC
    0000004088 FD
    0000006231 CM
    0000009103 WP
    0000009106 WV
    0000009126 WW
    0000008458 NH
    0000008475 NY
    0000008481 NF
    0000008483 NG
    $END
    and Save it as First.alt and You can import like this
    1)Open the outline in edit mode.
    2)Select Outline > Import alias table.
    3)In the Open dialog box, select the alias table import file.
    4)Click OK.
    go to outline properties and the new alias table must be made Set as active to view the alias names.
    you can create and import up to 10 alias tables

  • Is it possible to create materialized view log file for force refresh

    Is it possible to create materialized view log file for force refresh with join condition.
    Say for example:
    CREATE MATERIALIZED VIEW VU1
    REFRESH FORCE
    ON DEMAND
    AS
    SELECT e.employee_id, d.department_id from emp e and departments d
    where e.department_id = d.department_id;
    how can we create log file using 2 tables?
    Also am copying M.View result to new table. Is it possible to have the same values into the new table once the m.view get refreshed?

    You cannot create a record as a materialized view within the Application Designer.
    But there is workaround.
    Create the record as a table within the Application Designer. Don't build it.
    Inside your database, create the materialized with same name and columns as the record created previously.
    After that, you'll be able to work on that record as for all other within the Peoplesoft tools.
    But keep in mind do never build that object, that'll drop your materialized view and create a table instead.
    Same problem exists for partitioned tables, for function based-indexes and some other objects database vendor dependant. Same workaround is used.
    Nicolas.

  • How to use remote directory for external table

    Hi Folks,
    I have 2 Oracle 11GR2 64 bit database installed on Win 2008 server as prod1 and prod2.
    I have one directory created on prod1 server as EXT_TAB_DIR using the path as D:\OrsDWtest_dir .
    I want to use this directory in Prod2 server and use external table using this remote directory.
    I am able to access the Prod1 directory from Prod2 machine and also i have created Network map drive as Z drive pointing to that prod1 D:\OrsDWtest_dir directory. Also i checked read and Write permissions are there . I am able to create the external table but when i try to fetch the data i m getting below error ..
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04040: file IOMM_20121213_060736.csv in EXT_TAB_DIR not found
    now my doubt is this possible ? Can we use remote directory for External table ? or is there is there any alternative way to achieve same ?
    Thanks & Regards,
    Vikash Jain(DBA)

    could you confirm the name and the existence of this file "IOMM_20121213_060736.csv" ?
    same error like:
    http://www.oracle-base.com/articles/9i/external-tables-9i.php
    if the load files have not been saved in the appropriate directory the following result will be displayed.
    SQL> SELECT *
      2  FROM   countries_ext
      3  ORDER BY country_name;
    SELECT *
    ERROR at line 1:
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-29400: data cartridge error
    KUP-04040: file Countries1.txt in EXT_TABLES not found
    ORA-06512: at "SYS.ORACLE_LOADER", line 14
    ORA-06512: at line 1Edited by: Fran on 10-ene-2013 23:32

  • How to read any file using external tables.

    Hi folks,
    I have written an application that reads a series of csv files using external tables which works fine as long as I specify each file name in the directory i.e.......
    CREATE TABLE gb_test
    (file_name varchar2(10),
    rec_date date
    rec_name VARCHAR2(20),
    rec_age number,
    ORGANIZATION EXTERNAL
    TYPE ORACLE_LOADER
    DEFAULT DIRECTORY GB_TEST
    ACCESS PARAMETERS
    RECORDS DELIMITED BY NEWLINE
    FIELDS TERMINATED BY ','
    LOCATION ('data1.csv','data2.csv','data3.csv','data4.csv')
    PARALLEL 5
    REJECT LIMIT 20000;
    However I have discovered that I may not know the name of the files to be processed prior to the program being run so just want to read any file regardless of it's name (although it will always be a .csv file).
    Is there a way to ensure that you don't need to specify the files to be read in the LOCATION part of the syntax.
    Thanks in advance.
    Graham.

    Right, I have now completed this, however it's currently only working as SYS as opposed to any user, however here is a detail of the scenario and the steps required in case any of you guys need in the future ......
    The problem was I needed to search for csv files on my hard-drive. These files would be stored in a series of directories (a through to z), so I needed a way to read all 26 directories and process all files in these directories.
    The problem was, prior to running the program, the user would remove all the files in the directories and insert new ones, but it was never known how many he would decide to do each time.
    Solution: I created a table called stock_data_directories as follows ...
    create table stock_data_directories(sdd_rec_no number,
    sdd_table_name varchar2(50),
    sdd_directory_name varchar2(50),
    sdd_directory_path varchar2(100));
    Then inserted 26 records like ...
    insert into stock_data_directories(sdd_rec_no,sdd_table_name,sdd_directory_name,sdd_directory_path)
    values(1,'rawdata_a','KPOLLOCKA','C:\KPOLLOCK\A')
    insert into stock_data_directories(sdd_rec_no,sdd_table_name,sdd_directory_name,sdd_directory_path)
    values(2,'rawdata_b','KPOLLOCKB','C:\KPOLLOCK\B');
    etc...etc...
    Then created 26 DIRECTORIES E.G.
    CREATE OR REPLACE DIRECTORY KPOLLOCKA AS 'C:\KPOLLOCK\A';
    CREATE OR REPLACE DIRECTORY KPOLLOCKB AS 'C:\KPOLLOCK\B';
    Then created 26 external tables like the following ...
    CREATE TABLE rawdata_a
    (stock varchar2(1000),
    stock_date varchar2(10),
    stock_open VARCHAR2(20),
    stock_high varchar2(20),
    stock_low varchar2(20),
    stock_close VARCHAR2(30),
    stock_qty varchar2(20) )
    ORGANIZATION EXTERNAL
    TYPE ORACLE_LOADER
    DEFAULT DIRECTORY KPOLLOCKA
    ACCESS PARAMETERS
    RECORDS DELIMITED BY NEWLINE
    FIELDS TERMINATED BY ','
    LOCATION ('AA.csv')
    PARALLEL 5
    REJECT LIMIT 20000
    This basically says in directory rawdata_a it currently has 1 file called AA.csv.
    Then wrote a procedure as follows ...
    procedure p_process_files(pv_return_message OUT varchar2)is
    cursor c_get_stock_data_directories is
    select distinct sdd_directory_path,
    sdd_table_name
    from stock_data_directories
    order by sdd_table_name;
    vv_return_message varchar2(1000);
    begin
    -- here get the files for each directory
    for r_get_stock_directories in c_get_stock_data_directories loop
    p_build_external_table(r_get_stock_directories.sdd_directory_path,
         r_get_stock_directories.sdd_table_name,
         vv_return_message);
    end loop;
    end;
    then wrote a procedure called p_build_external_table as follows ...
    procedure p_build_external_table(pv_directory_path IN stock_data_directories.sdd_directory_path%type, -- e.g. 'C:\kpollock\A\
    pv_table_name IN stock_data_directories.sdd_table_name%type, -- e.g. rawdata_a
    pv_return_message OUT varchar2) is
    vv_pattern VARCHAR2(1024);
    ns VARCHAR2(1024);
    vv_file_name varchar2(4000);
    vv_start_string varchar2(1) := '''';
    vv_end_string varchar2(3) := ''',';
    vn_counter number := 0;
    vv_err varchar2(2000);
    BEGIN
    vv_pattern := pv_directory_path||'*';
    SYS.DBMS_BACKUP_RESTORE.searchFiles(vv_pattern, ns);
    FOR each_file IN (SELECT FNAME_KRBMSFT AS name FROM X$KRBMSFT) LOOP
    if each_file.name like '%.CSV' then
    vv_file_name := vv_file_name||vv_start_string||substr(each_file.name,instr(each_file.name,'\',1,3)+1)||vv_end_string;
         vn_counter := vn_counter + 1;
    end if;
    END LOOP;
    vv_file_name := substr(vv_file_name,1,length(vv_file_name)-1); -- remove final , from string
    execute immediate 'alter table '||pv_table_name||' location('||vv_file_name||')';
    pv_return_message := 'Successfully changed '||pv_table_name||' at '||pv_directory_path||' to now have '||to_char(vn_counter)||' directories';
    exception
    when others then
    vv_err := sqlerrm;
    pv_return_message := ' Error found updating directories. Error = '||vv_err;
    END;
    This reads every file in the directory and appends it to a list, so if it finds A.csv and ABC.csv, then using the dynamic sql, it alters the location to now read 'a.csv','abc.csv',
    It ignores all other file extentions.

  • The log file for database 'tempdb' is full

    Hi Experts,
    in my sync RFC to JDBC scenario I get the following error message which says:
    11.02.2009 11:23:20 Error Unable to execute statement for table or stored procedure.
    'Trns' (Structure 'XYZ')
    due to java.sql.SQLException:
    [Microsoft][SQLServer 2000 Driver for JDBC][SQLServer]
    The log file for database 'tempdb' is full. Back up the transaction
    log for the database to free up some log space.
    11.02.2009 11:23:20 Error JDBC message processing failed;
    reason Error processing request in sax parser: Error when executing
    statement for table/stored proc. 'Trns' (structure 'XYZ'):
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver for JDBC][SQLServer]
    The log file for database 'tempdb' is full. Back up the transaction log
    for the database to free up some log space.
    see here:
    http://img23.imageshack.us/img23/6541/unbenannthd9.jpg
    does anyone know how to solve this problem?
    thanks
    chris

    Christian,
    See this page for more information on the error you're getting: http://sqlserver2000.databases.aspfaq.com/why-is-tempdb-full-and-how-can-i-prevent-this-from-happening.html.
    Kind regards,
    Koen

  • History report error: | An Exceptional Error occurred. Application exiting. Check the log file for error 5022

    Hi all
    I've got a error msg when try to generate a report using Cisco history report tool:
    Error | An Exceptional Error occurred. Application exiting.  Check the log file for error 5022
    It only happens when choose report template: ICD_Contact_Service_Queue_Activity_by_CSQ_en_us.
    User tried samething on other PC, it working fine.
    only on user' own PC and only choose this report, error appears.
    user runing windows 7 and do not have crystal report installed
    tried reinstalled the software, doesn't work.
    also tried this: (https://cisco-support.hosted.jivesoftware.com/thread/2041254) - doesn't work
    then tried https://supportforums.cisco.com/docs/DOC-6209  - doesn't work
    attached the log file.
    thanks.

    wenqianyu wrote:From the log file:Looks like you get a Login Window.Error message showed up after username/password be enteredThere is an error in the log: Error happened in comparing UCCX version and HRC versionYou may need to do a clean uninstall, download the Historical report from the server, and install it again on the PC.Does this only happen to one PC or to every PC with this application?Wenqian 
    I have completely uninstalled the HRC, and download from server install again -- still doesn't work with exactly same error.
    this matter only happens on this PC, when user try same thing on other PC, it works.
    so i think it not relate to server or account.

  • (Cisco Historical Reporting / HRC ) All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054

    Hi All,
    I am getting an error message "All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054"  when trying to log into HRC (This user has the reporting capabilities) . I checked the log files this is what i found out 
    The log file stated that there were ongoing connections of HRC with the CCX  (I am sure there isn't any active login to HRC)
    || When you tried to login the following error was being displayed because the maximum number of connections were reached for the server .  We can see that a total number of 5 connections have been configured . ||
    1: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Current number of connections (5) from historical Clients/Scheduler to 'CRA_DATABASE' database exceeded the maximum number of possible connections (5).Check with your administrator about changing this limit on server (wfengine.properties), however this might impact server performance.
    || Below we can see all 5 connections being used up . ||
    2: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:[DB Connections From Clients (count=5)]|[(#1) 'username'='uccxhrc','hostname'='3SK5FS1.ucsfmedicalcenter.org']|[(#2) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#3) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#4) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#5) 'username'='uccxhrc','hostname'='47BMMM1.ucsfmedicalcenter.org']
    || Once the maximum number of connection was reached it threw an error . ||
    3: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Number of max connection to 'CRA_DATABASE' database was reached! Connection could not be established.
    4: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Database connection to 'CRA_DATABASE' failed due to (All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054.)
    Current exact UCCX Version 9.0.2.11001-24
    Current CUCM Version 8.6.2.23900-10
    Business impact  Not Critical
    Exact error message  All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054
    What is the OS version of the PC you are running  and is it physical machine or virtual machine that is running the HRC client ..
    OS Version Windows 7 Home Premium  64 bit and it’s a physical machine.
    . The Max DB Connections for Report Client Sessions is set to 5 for each servers (There are two servers). The no of HR Sessions is set to 10.
    I wanted to know if there is a way to find the HRC sessions active now and terminate the one or more or all of that sessions from the server end ? 

    We have had this "PRX5" problem with Exchange 2013 since the RTM version.  We recently applied CU3, and it did not correct the problem.  We have seen this problem on every Exchange 2013 we manage.  They are all installations where all roles
    are installed on the same Windows server, and in our case, they are all Windows virtual machines using Windows 2012 Hyper-V.
    We have tried all the "this fixed it for me" solutions regarding DNS, network cards, host file entries and so forth.  None of those "solutions" made any difference whatsoever.  The occurrence of the temporary error PRX5 seems totally random. 
    About 2 out of 20 incoming mail test by Microsoft Connectivity Analyzer fail with this PRX5 error.
    Most people don't ever notice the issue because remote mail servers retry the connection later.  However, telephone voice mail systems that forward voice message files to email, or other such applications such as your scanner, often don't retry and
    simply fail.  Our phone system actually disables all further attempts to send voice mail to a particular user if the PRX5 error is returned when the email is sent by the phone system.
    Is Microsoft totally oblivious to this problem?
    PRX5 is a serious issue that needs an Exchange team resolution, or at least an acknowledgement that the problem actually does exist and has negative consequences for proper mail flow.
    JSB

  • How to create different log files for each of web applications deployed in OC4J

    Hi All,
    I am using OC4J(from Oracle) v1.0.2.2 and Windows2000. Now I want to know
    1. how to create different log files for each of my deployed web applications ?
    2. what are the advantages in running multiple instances of oc4j and in what case we should run
    multiple instances of OC4J ?
    3. how to run OC4J as Windows2000 Service rather than Windows2000 Application ?
    Thanks and Regards,
    Kumar.

    Hi Avi,
    First of all I have given a first reading to log4j and I think there will some more easy way of logging debugging messages than log4j (If you could provide me a detailed explanation of a servlet,jsp,java bean that uses log4j and how to use log4j then it will be very helpful for me). The other easy ways (if I am not using log4j) to my problem i.e creating different log files for each of web applications deployed in oc4j are
    I have created multiple instances of OC4J that are configured to run on different ports and so on each instance I have deployed a single web application . And I started the 2 oc4j instances by transferring thier error/log messages to a file. And the other way is ..
    I have download from jakarta site a package called servhelper . This servhelper is a thread that is started in a startup servlet and stopped in the destroy method of that startup servlet. So this thread will automatically capture all the system.out.println's and will print those to a file. I believe that this thread program is synchronized. So in this method I need not run multiple instances of OC4J instead each deployed web application on single instance of oc4j uses the same thread program (ofcourse a copy of thread program is put in each of the deployed web applications directories) to log messages on to different log files.
    Can you comment on my above 2 approached to logging debugging messages and a compartive explanation to LOG4J and how to use LOG4J using a simple servlet, simple jsp is appreciated ...
    Thanks and Regards,
    Ravi.

  • LOG FILE for batch scripting in MAXL

    Hello,
    I just wanted to know how to create a LOG FILE for batch scripting.
    essmsh E:\Batch\Apps\TOG_DET\Scripts\unload_App.msh
    copy e:\batch\apps\tog_det\loadfile\gldetail.otl e:\hyperion\analyticservices\app\tog_det\gldetail /Y
    essmsh E:\Batch\Apps\TOG_DET\Scripts\Build_Hier_Data.msh
    REM
    ECHO OFF
    ECHO Loading GL actuals into WFS \ Combined......
    E:\HYPERION\common\Perl\5.8.3\bin\MSWin32-x86-multi-thread\PERL.EXE E:\Batch\Apps\WFS.COMBINED\AMLOAD\WFSUATAMLOAD.PLX
    E:\HYPERION\common\Perl\5.8.3\bin\MSWin32-x86-multi-thread\PERL.EXE E:\Batch\Apps\WFS.COMBINED\AMLOAD\WFSUATAMLOAD.PLX
    Drop object d:\NDM\Data\StampFiles\STAMPLOADBKUP.csv of type outline force;
    Alter object d:\NDM\Data\StampFiles\STAMPLOAD_cwoo.csv of type outline rename to d:\NDM\Data\StampFiles\STAMPLOADBKUP.CSV;
    SET LogFile=E:\Batch\Apps\TOG_DET\Logs.log
    This file does not generate log file can any help me what might be the problem? Even though some of the steps above are not correct it should generate me log file atleast. I need syntax or whatever it is to generate Log file.
    Regards
    Soma

    I wanted to have a logfile of the following batch script regardless of whether the script is running or not.
    essmsh E:\Batch\Apps\TOG_DET\Scripts\unload_App.msh
    copy e:\batch\apps\tog_det\loadfile\gldetail.otl e:\hyperion\analyticservices\app\tog_det\gldetail /Y
    essmsh E:\Batch\Apps\TOG_DET\Scripts\Build_Hier_Data.msh
    REM
    ECHO OFF
    ECHO Loading GL actuals into WFS \ Combined......
    E:\HYPERION\common\Perl\5.8.3\bin\MSWin32-x86-multi-thread\PERL.EXE E:\Batch\Apps\WFS.COMBINED\AMLOAD\WFSUATAMLOAD.PLX
    Drop object d:\NDM\Data\StampFiles\STAMPLOADBKUP.csv of type outline force;
    Alter object d:\NDM\Data\StampFiles\STAMPLOAD_cwoo.csv of type outline rename to d:\NDM\Data\StampFiles\STAMPLOADBKUP.CSV;
    What I really want is I need a log file of the above batch script, how the above scripts are running. I do not care whether they are giving me positive results but I need to know what is happening in logfile. HOw will the log file be generated.
    Regards
    SOma

Maybe you are looking for