PL/SQL procedure -- log files?

Say when i execute a PL/SQL procedure using SQL* Plus. Is there a place where these executions are stored/logged? Any trace files?
And, when a Java program calls my stored procedure, is there a place these transactions are stored, just to check what exactly is being passed to my stored procedure and what the procedure gave back as result set?
Any pointers will be appreciated. Thanks.

Hi
Use a system.out.println(parametername) to check what values you are passing to the procedure.. you can see the results in your application server console.
Thanks

Similar Messages

  • Summary SQL*LDR log file

    Hi,
    I need to get only a summary log file that show me records by error. for example if the log file has 5000 error for datatype and 300 for PK violation and 2000 for not null value, the problem that the log will show me this error row by row. can I get it summary like this
    -5000 rows rejected due to error for datatype
    -300 rows rejected due to PK violation
    -2000 rows rejected due to not null value
    Database: 9i
    O/S: Windows 2000 Server

    eng. Habeeli wrote:
    schavali wrote:
    Pl see your duplicate post here - big SQL*LDR log file
    Pl post what you have found in the documentation so far. AFAIK, there is no way to get a summary log file.
    Sriniits not duplicate ,I change to this form that specialized in sql loadre.
    I don't found any thing about summary log in the documnetPl post a link to the document you read. Did you see my comment above ?
    HTH
    Srini

  • How to Open Or read SQL Server log file .ldf

    Hi all,
    How to Open Or read SQL Server log file .ldf
    When ever we create database from sql server, it's create two file. (1) .mdf (2) .ldf.
    I want to see what's available inside the .ldf file.
    Thanks,
    Ashok

    I am not too sure but may be the below two undocumented commands might yield the desired result.
    DBCC Log
    Fn_dblog function
    Refer these links for more info,
    http://www.mssqlcity.com/Articles/Undoc/SQL2000UndocDBCC.htm
    http://blogs.sqlserver.org.au/blogs/greg_linwood/archive/2004/11/27/37.aspx
    http://searchsqlserver.techtarget.com/tip/0,289483,sid87_gci1173464,00.html
    Some 3rd party tools like Log Explorer can do the job for you.
    http://www.lumigent.com/products/le_sql.html
    - Deepak

  • Problem specifying SQL Loader Log file destination using EM

    Good evening,
    I am following the example given in the 2 Day DBA document chapter 8 section 16.
    In step 5 of 7, EM does not allow me to specify the destination of the SQL Loader log file to be on a mapped network drive.
    The question: Does SQL Loader have a limitation that I am not aware of, that prevents placing the log file on a network share or am I getting this error because of something else I am inadvertently doing wrong ?
    Note: I have placed the DDL, load file data and steps I follow in EM at the bottom of this post to facilitate reproducing the problem *(drive Z is a mapped drive)*.
    Thank you for your help,
    John.
    DDL (generated using SQL developer, you may want to change the space allocated to be less)
    CREATE TABLE "NICK"."PURCHASE_ORDERS"
        "PO_NUMBER"      NUMBER NOT NULL ENABLE,
        "PO_DESCRIPTION" VARCHAR2(200 BYTE),
        "PO_DATE" DATE NOT NULL ENABLE,
        "PO_VENDOR" NUMBER NOT NULL ENABLE,
        "PO_DATE_RECEIVED" DATE,
        PRIMARY KEY ("PO_NUMBER") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS NOCOMPRESS LOGGING TABLESPACE "USERS" ENABLE
      SEGMENT CREATION DEFERRED PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE
        INITIAL 67108864
      TABLESPACE "USERS" ;
    Load.dat file contents
    1, Office Equipment, 25-MAY-2006, 1201, 13-JUN-2006
    2, Computer System, 18-JUN-2006, 1201, 27-JUN-2006
    3, Travel Expense, 26-JUN-2006, 1340, 11-JUL-2006
    Steps I am carrying out in EM
    log in, select data movement -> Load Data from User Files
    Automatically generate control file
    (enter host credentials that work on your machine)
    continue
    Step 1 of 7 ->
      Data file is located on your browser machine
      "Z:\Documentation\Oracle\2DayDBA\Scripts\Load.dat"
       click next
    step 2 of 7 ->
      Table Name
      nick.purchase_orders
      click next
    step 3 of 7 ->
      click next
    step 4 of 7 ->
      click next
    step 5 of 7 ->
      Generate log file where logging information is to be stored
      Z:\Documentation\Oracle\2DayDBA\Scripts\Load.LOG
      Validation Error
      Examine and correct the following errors, then retry the operation:
      LogFile - The directory does not exist.

    Hi John,
    But, i did'nt found any error when i am going the same what you did.
    My Oracle Version is 10.2.0.1 and using Windows xp. See what i did and i got worked
    1.I created one table in scott schema :
    SCOTT@orcl> CREATE TABLE "PURCHASE_ORDERS"
      2  (
      3      "PO_NUMBER"      NUMBER NOT NULL ENABLE,
      4      "PO_DESCRIPTION" VARCHAR2(200 BYTE),
      5      "PO_DATE" DATE NOT NULL ENABLE,
      6      "PO_VENDOR" NUMBER NOT NULL ENABLE,
      7      "PO_DATE_RECEIVED" DATE,
      8      PRIMARY KEY ("PO_NUMBER") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS NOCOMPRESS LOGGING TABLESPACE "USERS" ENABLE
      9  )
    10  TABLESPACE "USERS";
    Table created.I logged into em Maintenance-->Data Movement-->Load Data from User Files-->My Host Credentials
    Here i total 3 text boxes :
    1.Server Data File : C:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\USERS01.DBF
    2.Data File is Located on Your Browser Machine : z:\load.dat <--- Here z:\ means other machine's shared doc folder; and i selected this option (as option button click) and i created the same load.dat as you mentioned.
    3.Temporary File Location : C:\ORACLE\PRODUCT\10.2.0\ORADATA\ORCL\ <--- I did'nt mentioned anything.
    Step 2 of 7 Table Name : scott.PURCHASE_ORDERS
    Step 3 of 7 I just clicked Next
    Step 4 of 7 I just clicked Next
    Step 5 of 7 I just clicked Next
    Step 6 of 7 I just clicked Next
    Step 7 of 7 Here it is Control File Contents:
    LOAD DATA
    APPEND
    INTO TABLE scott.PURCHASE_ORDERS
    FIELDS TERMINATED BY ',' OPTIONALLY ENCLOSED BY '"'
    PO_NUMBER INTEGER EXTERNAL,
    PO_DESCRIPTION CHAR,
    PO_DATE DATE,
    PO_VENDOR INTEGER EXTERNAL,
    PO_DATE_RECEIVED DATE
    And i just clicked on submit job.
    Now i got all 3 rows in purchase_orders :
    SCOTT@orcl> select count(*) from purchase_orders;
      COUNT(*)
             3So, there is no bug, it worked and please retry if you get any error/issue.
    HTH
    Girish Sharma

  • Logical sql in log file.

    Can someone please tell me how to see the complete sql query in the log file. If I run the same query the sql is not being produced I looked in the server log file and also manage sessions log file. It just says all columns from 'Subject Area'. I want to see all the joins and filters as well. Even for repeated queries how can I see complete sql. I set my logging level to 2.

    http://lmgtfy.com/?q=obiee+disable+query+caching
    http://catb.org/esr/faqs/smart-questions.html#homework

  • Repeated Errors in SQL Server log file

    I have hundreds of these errors saying 'Login failed for user 'Reporting' The user is not associated with a trusted SQL Server connection [CLIENT: ip address]
    The ip address is that of the server that sql server is installed on.
    Looking in my log file, all looks good until I get to Service Broker manager has started, then I get Error: 18452, Severity: 14 State: 1 then these two lines repeat about every minute, for the last 3 days!
    I think I must have just missed a tick box somewhere, but where?
    I have been into one of the databases, and input  and checked data, both via an application I wrote and SQL Server Management Studio.
    I am also having trouble connecting using my application to connect to the database, I can only connect if I use a Windows administrator account (this SQL Server 2005 running on a Windows 2003 Server, with the app on PC running Windows 2000)

    Hello Graham,
    did you find any solutions for that problem? I'm discovering a similar problem. The state of my error message is 5. According to the following source http://blogs.msdn.com/sql_protocols/archive/2006/02/21/536201.aspx this means the user is not known. This is correct. Neither the SQL-Server nor the Windows system has such a userid.
    The faulty user name is 'Reporting'. What is that user used for? I have set up other SQL Server 2005 servers but was never asked for such a name.
    Greetings,
    Frank

  • Big SQL*LDR log file

    Hi,
    I need to get only a summary log file that show me records by error. for example if the log file has 5000 error for datatype and 300 for PK violation and 2000 for not null value, the problem that the log will show me this error row by row. can I get it summary like this
    -5000 rows rejected due to error for datatype
    -300 rows rejected due to PK violation
    -2000 rows rejected due to not null value
    can u give me hand with this?

    Pl post details of OS and database versions. There is a dedicated SQL*Loader forum at Export/Import/SQL Loader & External Tables
    Have you reviewed the documentation for your database version at http://docs.oracle.com ? If so, what have you determined so far ?
    HTH
    Srini

  • SQL Server log file

    We are facing one issue can anyone please help me out.
    Logs are getting full in below path . Can we remove the logs ? is it archive logs ?
    Path : H:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\DATA

    If T-Log backups are configured then the log automatically get truncated once backup is compelted. If you dont have backups configured and you dont want to save the transactions you can manually truncate it
    Hi,
    From SQL Server 2008 onwards truncate_only is removed, it is replaced by
    backup log db_name to disk='Null'
    Also if your log is growing AND YOU DONT NEED POINT IN TIME RECOVERY switch recovery model to simple it will force checkpoint and logs will get truncated( If transaction does not requires the log) . Avoid using backup log to null or truncate only.
    Swapna,
    >>Also if you dont want the transactions to be saved then you can set the recovery model and Simple
    This is incorrect language used saving/committing transaction does not depends on recovery model. Recovery model only controls logging and recovery. if transaction is committed it will be present in database no matter what recovery model you use.
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • SQL query log File..

    Hi All,
    I am using a standalone BIP environment... Is there way we can see the actual query log the BIP generates? which location on the server..or how to get it activated.
    I want to see what is the actual sql query that is generated along with variables in it.
    Fyi : I want to have the query log n not the server error log..
    -dev

    Hi Eric,
    I would like you to re-check on the content level settings here as they are the primary causes of this kind of behavior. You could notice that the same information might have written down in the logical plan of the query too.
    Also, as per your description
    "In the SOURCES for this logical table, I've set the logical level of the content for E2 appropriately (detail level, same as E1)"
    I would like to check on this point again, as if you had mapped E2 to E1 in the same logical source with an inner join, you would get to set the content level at E1 levels themselves but not E2 (Now, that E2 would become a part of the E1 hierarchy too). This might be the reason, the BI Server is choosing to elimiate(null) the values from E2 too (even you could see them in the sql client)
    Hope this helps.
    Thank you,
    Dhar

  • Need PL/SQL procedure for file transfer local system to server location

    Hi Experts,
    The requirement is one file ex: text file,excel file is there in Local system suppose c:newfolder/test.txt.
    once run the concurrent program for this particular file in the local system should transfer into server location for the mentioned path.
    is it possible to do in PL/SQL or shell script.
    can you please share the code.
    Thanks
    Meher

    Meher Irk wrote:
    The requirement is one file ex: text file,excel file is there in Local system suppose c:newfolder/test.txt.
    once run the concurrent program for this particular file in the local system should transfer into server location for the mentioned path.Why do you want to copy a local file from the local client file system, via the Oracle database server, to a remote directory on the server?
    Why is the Oracle database server to be used? For what purpose?
    There are standard secure methods for a client to copy a file to a server. Such as sftp and scp. No complex Oracle database server layer needed in-between.
    So why do you want to use the Oracle database to act as the copy mechanism for you?
    It will only make sense if the local file is copied into the database (as a LOB or securefile) and managed by the database. I fail to see why Oracle should be acting as the go-between for the client file system copy process to the server file system. This introduces another moving part in the copy process. Adds more security issues. Adds more authentication issues. And what for?

  • SQL Loader Inserting Log File Statistics to a table

    Hello.
    I'm contemplating how to approach gathering the statistics from the SQL Loader log file to insert them into a table. I've approached this from a Korn Shell Script perspective previously, but now that I'm working in a Windows environment and my peers aren't keen about batch files and scripting I thought I'd attempt to use SQL Loader itself to read the log file and insert one or more records into a table that tracks data uploads. Has anyone created a control file that accomplishes this?
    My current environment:
    Windows 2003 Server
    SQL*Loader: Release 10.2.0.1.0
    Thanks,
    Luke

    Hello.
    Learned a little about inserting into multiple tables with delimited records. Here is my current tested control file:
    LOAD DATA
    APPEND
    INTO TABLE upload_log
    WHEN (1:12) = 'SQL*Loader: '
    FIELDS TERMINATED BY WHITESPACE
    TRAILING NULLCOLS
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , filler_field_3  FILLER
    , filler_field_4  FILLER
    , filler_field_5  FILLER
    , day_of_week
    , month
    , day_of_month
    , time_of_day
    , year
    , log_started_on          "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
    INTO TABLE upload_log
    WHEN (1:11) = 'Data File: '
    FIELDS TERMINATED BY ':'
    (  upload_log_id    RECNUM
    , filler_field_0   FILLER  POSITION(1)
    , input_file_name          "TRIM(:input_file_name)"
    INTO TABLE upload_log
    WHEN (1:6) = 'Table '
    FIELDS TERMINATED BY WHITESPACE
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , table_name              "RTRIM(:table_name, ',')"
    INTO TABLE upload_rejects
    WHEN (1:7) = 'Record '
    FIELDS TERMINATED BY ':'
    (  upload_rejects_id  RECNUM
    , record_number      POSITION(1)  "TO_NUMBER(SUBSTR(:record_number,8,20))"
    , reason
    INTO TABLE upload_rejects
    WHEN (1:4) = 'ORA-'
    FIELDS TERMINATED BY ':'
    (  upload_rejects_id  RECNUM
    , error_code         POSITION(1)
    , error_desc
    INTO TABLE upload_log
    WHEN (1:22) = 'Total logical records '
    FIELDS TERMINATED BY WHITESPACE
    (  upload_log_id      RECNUM
    , filler_field_0     FILLER  POSITION(1)
    , filler_field_1     FILLER
    , filler_field_2     FILLER
    , action                     "RTRIM(:action, ':')"
    , number_of_records
    INTO TABLE upload_log
    WHEN (1:13) = 'Run began on '
    FIELDS TERMINATED BY WHITESPACE
    TRAILING NULLCOLS
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , day_of_week
    , month
    , day_of_month
    , time_of_day
    , year
    , run_began_on            "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
    INTO TABLE upload_log
    WHEN (1:13) = 'Run ended on '
    FIELDS TERMINATED BY WHITESPACE
    TRAILING NULLCOLS
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , day_of_week
    , month
    , day_of_month
    , time_of_day
    , year
    , run_ended_on            "TO_DATE((:month ||' '|| :day_of_month ||' '|| :time_of_day ||' '|| :year), 'Mon DD HH24:MI:SS YYYY')"
    INTO TABLE upload_log
    WHEN (1:18) = 'Elapsed time was: '
    FIELDS TERMINATED BY ':'
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , elapsed_time
    INTO TABLE upload_log
    WHEN (1:14) = 'CPU time was: '
    FIELDS TERMINATED BY ':'
    (  upload_log_id   RECNUM
    , filler_field_0  FILLER  POSITION(1)
    , filler_field_1  FILLER
    , filler_field_2  FILLER
    , cpu_time
    )Here are the basic table create scripts:
    TRUNCATE TABLE upload_log;
    DROP TABLE upload_log;
    CREATE TABLE upload_log
    (  upload_log_id      INTEGER
    , day_of_week        VARCHAR2(  3)
    , month              VARCHAR2(  3)
    , day_of_month       INTEGER
    , time_of_day        VARCHAR2(  8)
    , year               INTEGER
    , log_started_on     DATE
    , input_file_name    VARCHAR2(255)
    , table_name         VARCHAR2( 30)
    , action             VARCHAR2( 10)
    , number_of_records  INTEGER
    , run_began_on       DATE
    , run_ended_on       DATE
    , elapsed_time       VARCHAR2(  8)
    , cpu_time           VARCHAR2(  8)
    TRUNCATE TABLE upload_rejects;
    DROP TABLE upload_rejects;
    CREATE TABLE upload_rejects
    (  upload_rejects_id  INTEGER
    , record_number      INTEGER
    , reason             VARCHAR2(255)
    , error_code         VARCHAR2(  9)
    , error_desc         VARCHAR2(255)
    );Now, if I could only insert a single record to the upload_log table (per table logged); adding separate columns for skipped, read, rejected, discarded quantities. Any advice on how to use SQL Loader to do this (writing a procedure would be fairly simple, but I'd like to perform all of the work in one place if at all possible)?
    Thanks,
    Luke
    Edited by: Luke Mackey on Nov 12, 2009 4:28 PM

  • Log files full of service broker errors, but it works OK

    We have a C# web application that is using SQL dependency to expire cached query data. Although everything is working okay we are seeing a lot of errors being generated particularly in our production environment.
    The first error messages is this:
    Service Broker needs to access the master key in the database 'SubscriberManager'. Error code:32. The master key has to exist and the service master key encryption is required.
    It shows up in both the server event log and the SQL server log files. I believe the actual content of the messages something of a red herring as I have created a database master key.
    I have also tried
    Recreating both the service master key and the database master key.
    Made sure the database owner is sa.
    Made sure the user account has permissions to create services, queues, procedures and subscribe query notifications
    Made sure the broker is enabled
    I have seen other people with similar errors whilst researching the issue but the error code almost always seems to be 25 or 26 not 32. I have been unable to find anything that tells me what these error codes mean so I'm not sure of the significance.
    Also I am seeing a lot of errors like this:
    The query notification dialog on conversation handle '{2FA2445B-1667-E311-943C-02C798B618C6}.' closed due to the following error: '-8490Cannot find the remote service 'SqlQueryNotificationService-7303d251-1eb2-4f3a-9e08-d5d17c28b6cf' because
    it does not exist.'.
    I understand that a certain number of these are normal due to the way that SqlDependency.Stop doesn't clean everything up in the database, but we are seeing thousands of these on one of our production servers.
    What is frustrating as we have been using SQL notifications for several years now without issue so something must have changed in either our application or the server setups to cause this but at this point I have no idea what.
    The applications are .net 4.0 MVC and WCF running on Windows 2012 servers calling SQL 2012 servers also running on Windows 2012.

    Hi Mark,
    1. for your question about possible memory pressure, if the used memory is below the Max Server Memory, then it's OK. If you have not set Max Server Memory, you should at least leave 4GB for your x64 system.
    2. for your original question, I suggest you can check my actions below:
    a. run this statement:
    Select name, is_master_key_encrypted_by_server
    from sys.databases
    Where name =
    'your_database_name'
    if the value of "is_master_key_encrypted_by_server" equals to "0", it means the database does not have a encrypted master key.
    b.if there is no encrypted master key, then the error may be hit by the "begin dialog conversation" statement (you can check your sql profiler trace to check".
    "Service Broker dialog security lets your application use authentication, authorization, or encryption for an individual dialog conversation (or dialog). By default,
    all dialog conversations use dialog security. When you begin a dialog, you can explicitly allow a dialog to proceed without dialog security by including the ENCRYPTION = OFF clause on the BEGIN DIALOG CONVERSATION statement. However, if a remote service binding
    exists for the service that the conversation targets, the dialog uses security even when ENCRYPTION = OFF."
    (http://msdn.microsoft.com/en-us/library/ms166036.aspx)
    Workarounds can be disabling the dialog security (using encryption = off) or create a master key. You can find more information in about URL.

  • Viewing SQL Audit Logs

    I am new at the SQL auditing feature. I have played around with it, and I have one problem that I hope someone can help me with. When I first turn on the SQL audit and the audit file is small I can view the log on my local machine with the log viewer in
    SSMS. However, as the log files continue to be created and grow, the log file viewer no longer works. It sits there for several minutes saying "initializing log file #1", then just comes back 0 records processed. Sometimes it comes back with an OutOfMemoryException.
    I have the SQL audit set to go to files, and the files are set to be 250 MB each.
    Is there some better way to look at the SQL audit log files generated, or is there some way to make the log file viewer actually work once the log files begin to grow? I know I can use Transact-SQL statements, but I was hoping to use the log file viewer
    or some other GUI viewer. Another issue I'm concerned about is being able to archive the log files to another location and having the ability to look at them.
    Thank you.

    Its looks like bug.
    http://connect.microsoft.com/SQLServer/feedback/details/709364/sql-server-audit-logs-do-not-display
    http://thomaslarock.com/2012/10/viewing-sql-server-2008-r2-audit-logs-using-ssms-2012/
    Use sys.fn_get_audit_file() system function to see the data that you already doing.
     http://msdn.microsoft.com/en-us/library/cc280765.aspx
    -Prashanth

  • Read sql error log, skip lines that are in a exception list

    Could someone help, I am creating a nice powershell script to read a sql server log file but to skip lines that are normal in a sql server. the "normal lines are held in a SQL table.
    To expand.
    I run a query to get the list of exception lines using invoke-sqlcmd  this creates $TextEXP
    this will contain things like "Microsoft Corporation", "All rights reserved", "Starting up Database"
    I then connect to a sql server using SMO and want to read in the error log where the text is not matched the values in $textExpI want to avoid reading extra data and process it. I have a work round but its not a nice clean as its hardcoded the match.
    $ENV =$srv.ReadErrorLog()
    |? {  $_.text
    -notmatch'This
    is an informational message only'-and$_.text
    -notmatch'No
    user action is required'-and$_.text
    -notmatch'found
    0 errors'-and$_.text
    -notmatch'Microsoft
    Corporation.'-and$_.text
    -notmatch'All
    rights reserved.'-and$_.text
    -notmatch'Server
    process ID is'-and$_.text
    -notmatch'System
    Manufacturer: '-and$_.text
    -notmatch'Starting
    up database'-and$_.text
    -notmatch'Using
    ''dbghelp.dll'' version'-and$_.text
    -notmatch'Authentication
    mode is'-and$_.text
    -notmatch'Logging
    SQL Server messages in file '-and$_.text
    -notmatch'Setting
    database option'-and$_.text
    -notmatch'The
    error log has been reinitialized. See the previous log for older entries'-and$_.text
    -notmatch'Server
    is listening on '-and$_.text
    -notmatch'Registry
    startup parameters:'-and$_.text
    -notmatch'Clearing
    tempdb database'-and$_.text
    -notmatch'Service
    Broker manager has started'-and$_.text
    -notmatch'The
    Service Broker protocol transport is disabled or not configured'`
    -and$_.ProcessInfo
    -notmatch"Logon"-and$_.logdate
    -ge$Sdate}

    So after some looking about on the web I found that you can use the | in a string
    the following will give an idea of how to use this (this is not a clean bit of code but will give you a starting point)
    $TextEXP  this is a data table from sql server with the list of values I want to skip
    The field name (col name) is extext
    Set the string to be empty
    $exclusions = ""
    #Create a string with the values in $TextExp
    Foreach($value in $TextExp){
    $exclusions = $exclusions + "$($value.extext)|"
    #remove the last pipe from the string
    $exclusions = $exclusions.substring(0,$exclusions.length-1)
    ##This will create a long string value|value|value###
    $err = $srv.readerrorLog() | ?{$_.text - notmatch $exclusions}
    ###end
    May need bit of a clean up and may be a better way but seems to do what I need for now.
    Thanks all for the help

  • Display data in log file using PL/SQL procedure

    Just as srw.message is used in Oracle RDF Reports to display data in log file in Oracle Apps, similarly how it is possible to display data in log file using PL/SQL procedure?
    Please also mention the syntax too.

    Pl post details of OS, database and EBS versions.
    You will need to invoke the seeded FND_LOG procedure - see previous discussions on this topic
    Enable debug for pl/sql
    https://forums.oracle.com/forums/search.jspa?threadID=&q=FND_LOG&objID=c3&dateRange=all&userID=&numResults=15&rankBy=10001
    HTH
    Srini

Maybe you are looking for

  • Issue to add t-codes & functions in ERM role request GRC 10.0 - SP 5

    Hi We are on SAP GRC 10.0 - SP5. We are facing a strange issue while creating a Role request. After Defining the role, we are unable to add actions and there is no option of add/delete. Similarly while adding Functions, we are getting an error of Fun

  • I just upgraded to 6.0.4.2 and now my purchased list is all messed up

    Any one know how to fix this? I just up graded to 6.0.4.2 Now when I look at my purchased list of songs it shows a bunch of old songs that I never purchased through ITunes? and many of the songs that I did purchase are showing up in the list 4 and 5

  • Online number in Thailand

    Is Online number available for use in Thailand? If not, when will it be available? Thanks in advance for your response. Solved! Go to Solution.

  • IMac G5 iSight Model. NEW HDD.

    Hi. I just replaced the old HDD with my New 500 GB Seagate Barracuda HDD. So, I'm trying to install new OS X Tiger too. When I put the disc in and turn it on with holding down the C key... It goes to the grey screen with apple sign on it and then sta

  • Creative Cloud LR Mobile

    Does anyone know if there is an additional charge to use the LR Mobile app on your iPad if you are a CC Photography member?  Most things seem to suggest it is part of the package but on the CC LR web site the following note is included: Note: If you