Getting error while data archiving in BWP.

Dear All,
I am getting following error while data archiving in BWP .
    Value 'NP' in field COMPOP not permitted.
Can anybody help me to solve this error.
Thanks in advance.
Regards,
Vaibhav

Hi,
May i know the field description? COMPOP?
Rgds,
Ravi

Similar Messages

  • Error while data loading

    Hi Gurus,
    I am getting error while data loading. At BI side when I  check the error it says Background Job Cancelled and when I check in R3 side I got the following error
    Job started
    Step 001 started (program SBIE0001, variant &0000000065503, user ID R3REMOTE)
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    DATASOURCE = 2LIS_11_V_ITM
             Current Values for Selected Profile Parameters               *
    abap/heap_area_nondia......... 2000683008                              *
    abap/heap_area_total.......... 4000317440                              *
    abap/heaplimit................ 40894464                                *
    zcsa/installed_languages...... ED                                      *
    zcsa/system_language.......... E                                       *
    ztta/max_memreq_MB............ 2047                                    *
    ztta/roll_area................ 6500352                                 *
    ztta/roll_extension........... 3001024512                              *
    4 LUWs confirmed and 4 LUWs to be deleted with function module RSC2_QOUT_CONFIRM_DATA
    ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    Job cancelled
    Please help me out what should I do.
    Regards,
    Mayank

    Hi Mayank,
    The log says it went to short dump due to temp space issue.as its the source system job ,check in the source system side for temp table space and also check at the BI side as well.
    Check with your basis regarding the TEMP PSA table space - if its out of space ask them to increase the table space and try to repeat the load.
    Check the below note
    Note 796422 - DBIF_RSQL_SQL_ERROR during deletion of table BWFI_AEDAT
    Regards
    KP
    Edited by: prashanthk on Jul 19, 2010 10:42 AM

  • Getting error while tring to Check in any document in Sharepoint 2007. Error - "Unable to validate data. System.Web.Configuration.MachineKey.GetDecodedData(Byte[] buf, Byte[] modifier, Int32 start, Int32 length, Int32& dataLength) ...

    Hi Team,
    I am getting error while tring to Check in any document in Sharepoint 2007. Error - "Unable to validate data. System.Web.Configuration.MachineKey.GetDecodedData(Byte[] buf, Byte[] modifier, Int32 start, Int32 length, Int32& dataLength).
    Anybody please let me know how to resolve the issue. It has affected our 9 site collection with almost 2000+ sites. Please email me your your suggestion or you can share your ideas here also.
    Tried to reset the IIS, checked the Internet connection speed (8MBps). Cleared the cache and all but no luck.
    Thanks,
    Pronob

    Hello,
    I have found this support window for you:
    http://support.microsoft.com/kb/556031
    http://technet.microsoft.com/en-in/library/cc298582%28v=office.12%29.aspx
    This may help you
    Hemendra:Yesterday is just a memory,Tomorrow we may never see
    Please remember to mark the replies as answers if they help and unmark them if they provide no help

  • Getting error while uploading gl master data using LSMW batch input data

    Hi Experts,
    I am getting error while uploading the GL master data through lsmw using batch input recording.
    After completion of all steps at the time of running batch input session error i am getting is "maintain EN language discription"
    and Pls tell me is it possible to upload transactional data using lsmw? if yes pls explaing briefly.

    Hello,
    Kindly post in the FI forum to get a better response.
    Regarding the issue you might check whether the correct column is being picked up where you have maintained the description.
    Kind Regards // Shaubhik

  • Getting error while taking MAX DB trans log backup.

    Hi,
    I am getting error while taking trans log backup of Maxdb database for archived log through data protector as below,
    [Critical] From: OB2BAR_SAPDBBAR@ttcmaxdb "MAX" Time: 08/19/10 02:10:41
    Unable to back up archive logs: no autolog medium found in media list
    But i am able to take complete data and incremental backup through data protector.
    I have already enabled the autolog for MAX DB database and it is writing that log file directly to HP-UX file system. Now i want to take backup of this archived log backup through data protector i.e. through trans log backup. So that the archived log which is on the file system after trans log backup completed will delete the archived logs in filesystem.  So that i don;t have to manually delete the archived logs from file system.
    Thanks,
    Subba

    Hi Lars,
    Thanks for the reply...
    Now i am able to take archive log backup but the problem is i can take only one archive file backup. Not multiple arhive log files generated by autolog at filesystem i.e /sapdb/MAX/saparch.
    I have enabled autolog and it is putting auto log file at unix directory i.e. /sapdb/MAX/saparch
    And then i am using the DataProtector 6.11 with trans log backup to backup the archived files in /sapdb/MAX/saparch. When i start the trans backup session through data protector it uses the archive stage command as "archive_stage BACKDP-Archive LOGBackup NOVERIFY REMOVE" If /sapdb/MAX/saparch has only one archive file it will backup and remove the file successfully. But if /sapdb/MAX/saparch has multiple archive files it gives an error as below,
      Preparing backup.
                Setting environment variable 'BI_CALLER' to value 'DBMSRV'.
                Setting environment variable 'BI_REQUEST' to value 'OLD'.
                Setting environment variable 'BI_BACKUP' to value 'ARCHIVE'.
                Constructed Backint for MaxDB call '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.
    bsi_in -c'.
                Created temporary file '/var/opt/omni/tmp/MAX.bsi_out' as output for Backint for MaxDB.
                Created temporary file '/var/opt/omni/tmp/MAX.bsi_err' as error output for Backint for MaxDB.
                Writing '/sapdb/data/wrk/MAX/dbm.ebf' to the input file.
                Writing '/sapdb/data/wrk/MAX/dbm.knl' to the input file.
            Prepare passed successfully.
            Starting Backint for MaxDB.
                Starting Backint for MaxDB process '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.
    bsi_in -c >>/var/opt/omni/tmp/MAX.bsi_out 2>>/var/opt/omni/tmp/MAX.bsi_err'.
                Process was started successfully.
            Backint for MaxDB has been started successfully.
            Waiting for the end of Backint for MaxDB.
                2010-09-06 03:15:21 The backup tool is running.
                2010-09-06 03:15:24 The backup tool process has finished work with return code 0.
            Ended the waiting.
            Checking output of Backint for MaxDB.
            Have found all BID's as expected.
        Have saved the Backup History files successfully.
        Cleaning up.
            Removing data transfer pipes.
                Removing data transfer pipe /var/opt/omni/tmp/MAX.BACKDP-Archive.1 ... Done.
            Removed data transfer pipes successfully.
            Copying output of Backint for MaxDB to this file.
    Begin of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
                #SAVED 1009067:1 /sapdb/data/wrk/MAX/dbm.ebf
                #SAVED 1009067:1 /sapdb/data/wrk/MAX/dbm.knl
    End of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
            Removed Backint for MaxDB's temporary output file '/var/opt/omni/tmp/MAX.bsi_out'.
            Copying error output of Backint for MaxDB to this file.
    Begin of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
    End of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
            Removed Backint for MaxDB's temporary error output file '/var/opt/omni/tmp/MAX.bsi_err'.
            Removed the Backint for MaxDB input file '/var/opt/omni/tmp/MAX.bsi_in'.
        Have finished clean up successfully.
    The backup of stage file '/export/sapdb/arch/MAX_LOG.040' was successful.
    2010-09-06 03:15:24
    Backing up stage file '/export/sapdb/arch/MAX_LOG.041'.
        Creating pipes for data transfer.
            Creating pipe '/var/opt/omni/tmp/MAX.BACKDP-Archive.1' ... Done.
        All data transfer pipes have been created.
        Preparing backup tool.
            Setting environment variable 'BI_CALLER' to value 'DBMSRV'.
            Setting environment variable 'BI_REQUEST' to value 'OLD'.
            Setting environment variable 'BI_BACKUP' to value 'ARCHIVE'.
            Constructed Backint for MaxDB call '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.bsi_
    in -c'.
            Created temporary file '/var/opt/omni/tmp/MAX.bsi_out' as output for Backint for MaxDB.
            Created temporary file '/var/opt/omni/tmp/MAX.bsi_err' as error output for Backint for MaxDB.
            Writing '/var/opt/omni/tmp/MAX.BACKDP-Archive.1 #PIPE' to the input file.
        Prepare passed successfully.
        Constructed pipe2file call 'pipe2file -d file2pipe -f /export/sapdb/arch/MAX_LOG.041 -p /var/opt/omni/tmp/MAX.BACKDP-Archive.1 -nowait'.
        Starting pipe2file for stage file '/export/sapdb/arch/MAX_LOG.041'.
            Starting pipe2file process 'pipe2file -d file2pipe -f /export/sapdb/arch/MAX_LOG.041 -p /var/opt/omni/tmp/MAX.BACKDP-Archive.1 -nowait >>/var/tmp/tem
    p1283767880-0 2>>/var/tmp/temp1283767880-1'.
            Process was started successfully.
        Pipe2file has been started successfully.
        Starting Backint for MaxDB.
            Starting Backint for MaxDB process '/opt/omni/lbin/sapdb_backint -u MAX -f backup -t file -p SAPDB.13576.1283767878.par -i /var/opt/omni/tmp/MAX.bsi_
    in -c >>/var/opt/omni/tmp/MAX.bsi_out 2>>/var/opt/omni/tmp/MAX.bsi_err'.
            Process was started successfully.
        Backint for MaxDB has been started successfully.
        Waiting for end of the backup operation.
            2010-09-06 03:15:25 The backup tool process has finished work with return code 2.
            2010-09-06 03:15:25 The backup tool is not running.
            2010-09-06 03:15:25 Pipe2file is running.
            2010-09-06 03:15:25 Pipe2file is running.
            2010-09-06 03:15:30 Pipe2file is running.
            2010-09-06 03:15:40 Pipe2file is running.
            2010-09-06 03:15:55 Pipe2file is running.
            2010-09-06 03:16:15 Pipe2file is running.
            Killing not reacting pipe2file process.
            Pipe2file killed successfully.
            2010-09-06 03:16:26 The pipe2file process has finished work with return code -1.
        The backup operation has ended.
        Filling reply buffer.
            Have encountered error -24920:
                The backup tool failed with 2 as sum of exit codes and pipe2file was killed.
            Constructed the following reply:
                ERR
                -24920,ERR_BACKUPOP: backup operation was unsuccessful
                The backup tool failed with 2 as sum of exit codes and pipe2file was killed.
        Reply buffer filled.
        Cleaning up.
            Removing data transfer pipes.
                Removing data transfer pipe /var/opt/omni/tmp/MAX.BACKDP-Archive.1 ... Done.
            Removed data transfer pipes successfully.
            Copying output of Backint for MaxDB to this file.
    Begin of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
    End of output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_out)----
            Removed Backint for MaxDB's temporary output file '/var/opt/omni/tmp/MAX.bsi_out'.
            Copying error output of Backint for MaxDB to this file.
    Begin of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
    End of error output of Backint for MaxDB (/var/opt/omni/tmp/MAX.bsi_err)----
            Removed Backint for MaxDB's temporary error output file '/var/opt/omni/tmp/MAX.bsi_err'.
            Removed the Backint for MaxDB input file '/var/opt/omni/tmp/MAX.bsi_in'.
            Copying pipe2file output to this file.
    Begin of pipe2file output (/var/tmp/temp1283767880-0)----
    End of pipe2file output (/var/tmp/temp1283767880-0)----
            Removed pipe2file output '/var/tmp/temp1283767880-0'.
            Copying pipe2file error output to this file.
    Begin of pipe2file error output (/var/tmp/temp1283767880-1)----
    End of pipe2file error output (/var/tmp/temp1283767880-1)----
            Removed pipe2file error output '/var/tmp/temp1283767880-1'.
        Have finished clean up successfully.
    The backup of stage file '/export/sapdb/arch/MAX_LOG.041' was unsuccessful.
    2010-09-06 03:16:26
    Cleaning up.
        Have encountered error -24919:
            Can not remove file '/var/tmp/temp1283767880-0'.
            (System error 2; No such file or directory)
        Could not remove temporary output file of pipe2file ('/var/tmp/temp1283767880-0' ).
        Have encountered error -24919:
            Can not remove file '/var/tmp/temp1283767880-1'.
            (System error 2; No such file or directory)
        Could not remove temporary output file of pipe2file ('/var/tmp/temp1283767880-1' ).
    Have finished clean up successfully.
    Thanks,
    Subba

  • Getting error while creating the Shopping cart.

    Hi ,
    I am getting error while creating the Shopping cart.
    a.Error in account assignment for item 1  (Item  Testing SC ) 
    b.Duplicates of Cost Centre T10063 are defined in SRM  (Item  Testing SC ) 
    Kindly provide the solutions.
    Thanks,
    Dev

    Try the following in the ERP backend system:
    Standard Hierarchy Inconsistencies
    Issue: one Cost Center is repeating in more than one node in Cost Center Standard Hierarchy.
    Update from SAP Global support, the following was the email received:
    in transaction KSH3 please run both the ambiguity and completeness check(Menu -> Extras -> Check and Help functions).
    If you think that your standard hierarchy is inconsistent you can check that as following:
    Run transaction Extras -> Hierarchy - Master data -> Test. The result shows you if there are in consistencies. If that is the case run also Extras -> Hierarchy - Master data -> Comparison.
    Alternatively, you can run the report 'RKCORRH1' (TN SE38).
    Run both the Hierarchy->Master data->test and the
    Hierarchy->Master data->comparison.
    As stated above inconsistency message showed after Test. Run the comparison and you get a similar message.
    Once the above two are run, again when you go to test, the inconsistency disappears.

  • Getting error while running Foreign currency valuation

    Hi,
    Thanks in advance
    User getting error while running Foreign Currncy valutaion i.e. Internal error with background job sheduling sapf100.
    Do not have good exposure in Foreign Currecy valutaion please help me out..
    Edited by: kudalamallikarjuna on Oct 12, 2011 4:00 PM

    Dear,
    Once you get the error, please go to SM37, put user ID and appropriate dates and Execute
    System will show all jobs submitted by the user id, check if there is any job which is cancelled, select the line and click on Job log,
    In job log please let us know what is shown there..
    Br,Vivek

  • Getting error while running Perl adcfgclone.pl appsTier

    hi ,
    i m getting error while running Perl adcfgclone.pl appsTier
    error:ksh: Perl: not found
    i check for per and ksh
    which perl
    /usr/bin/perl
    which ksh
    /usr/bin/ksh
    (my os out put for perl ,my os is SunOS testserver 5.10 Generic_137137-09 sun4u sparc SUNW,Ultra-250)
    perl -V
    /appl2/applmgr2/prodora/iAS/Apache/perl/lib/5.00503
    /appl2/applmgr2/prodora/iAS/Apache/perl/lib/site_perl/5.005
    /appl2/applmgr2/prodappl/au/11.5.0/perl
    /usr/perl5/5.8.4/lib/sun4-solaris-64int
    /usr/perl5/5.8.4/lib
    /usr/perl5/site_perl/5.8.4/sun4-solaris-64int
    /usr/perl5/site_perl/5.8.4
    /usr/perl5/site_perl
    /usr/perl5/vendor_perl/5.8.4/sun4-solaris-64int
    /usr/perl5/vendor_perl/5.8.4
    /usr/perl5/vendor_perl
    which i need to set in path
    PATH=$PATH:/usr/local/bin:/usr/ccs/bin:/usr/sfw/bin (like this )
    export PATH
    please help me out

    hi ,
    here is the out put
    # pwd
    /appl2/applmgr2/prodora/iAS/Apache/perl
    # perl -V
    Summary of my perl5 (revision 5 version 8 subversion 4) configuration:
    Platform:
    osname=solaris, osvers=2.10, archname=sun4-solaris-64int
    uname='sunos localhost 5.10 sun4u sparc SUNW,Ultra-2'
    config_args=''
    hint=recommended, useposix=true, d_sigaction=define
    usethreads=undef use5005threads=undef useithreads=undef usemultiplicity=undef
    useperlio=define d_sfio=undef uselargefiles=define usesocks=undef
    use64bitint=define use64bitall=undef uselongdouble=undef
    usemymalloc=n, bincompat5005=undef
    Compiler:
    cc='cc', ccflags ='-D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -xarch=v8 -D_TS_ERRNO',
    optimize='-xO3 -xspace -xildoff',
    cppflags=''
    ccversion='Sun WorkShop', gccversion='', gccosandvers=''
    intsize=4, longsize=4, ptrsize=4, doublesize=8, byteorder=87654321
    d_longlong=define, longlongsize=8, d_longdbl=define, longdblsize=16
    ivtype='long long', ivsize=8, nvtype='double', nvsize=8, Off_t='off_t', lseeksize=8
    alignbytes=8, prototype=define
    Linker and Libraries:
    ld='cc', ldflags =''
    libpth=/lib /usr/lib /usr/ccs/lib
    libs=-lsocket -lnsl -ldl -lm -lc
    perllibs=-lsocket -lnsl -ldl -lm -lc
    libc=/lib/libc.so, so=so, useshrplib=true, libperl=libperl.so
    gnulibc_version=''
    Dynamic Linking:
    dlsrc=dl_dlopen.xs, dlext=so, d_dlsymun=undef, ccdlflags='-R /usr/perl5/5.8.4/lib/sun4-solaris-64int/CORE'
    cccdlflags='-KPIC', lddlflags='-G'
    Characteristics of this binary (from libperl):
    Compile-time options: USE_64_BIT_INT USE_LARGE_FILES
    Locally applied patches:
    22667 The optree builder was looping when constructing the ops ...
    22715 Upgrade to FileCache 1.04
    22733 Missing copyright in the README.
    22746 fix a coredump caused by rv2gv not fully converting a PV ...
    22755 Fix 29149 - another UTF8 cache bug hit by substr.
    22774 [perl #28938] split could leave an array without ...
    22775 [perl #29127] scalar delete of empty slice returned garbage
    22776 [perl #28986] perl -e "open m" crashes Perl
    22777 add test for change #22776 ("open m" crashes Perl)
    22778 add test for change #22746 ([perl #29102] Crash on assign ...
    22781 [perl #29340] Bizarre copy of ARRAY make sure a pad op's ...
    22796 [perl #29346] Double warning for int(undef) and abs(undef) ...
    22818 BOM-marked and (BOMless) UTF-16 scripts not working
    22823 [perl #29581] glob() misses a lot of matches
    22827 Smoke [5.9.2] 22818 FAIL(F) MSWin32 WinXP/.Net SP1 (x86/1 cpu)
    22830 [perl #29637] Thread creation time is hypersensitive
    22831 improve hashing algorithm for ptr tables in perl_clone: ...
    22839 [perl #29790] Optimization busted: '@a = "b", sort @a' ...
    22850 [PATCH] 'perl -v' fails if local_patches contains code snippets
    22852 TEST needs to ignore SCM files
    22886 Pod::Find should ignore SCM files and dirs
    22888 Remove redundant %SIG assignments from FileCache
    23006 [perl #30509] use encoding and "eq" cause memory leak
    23074 Segfault using HTML::Entities
    23106 Numeric comparison operators mustn't compare addresses of ...
    23320 [perl #30066] Memory leak in nested shared data structures ...
    23321 [perl #31459] Bug in read()
    27722 perlio.c breaks on Solaris/gcc when > 256 FDs are available
    SPRINTF0 - fixes for sprintf formatting issues - CVE-2005-3962
    6663288 Upgrade to CGI.pm 3.33
    REGEXP0 - fix for UTF-8 recoding in regexps - CVE-2007-5116
    Built under solaris
    Compiled at Jul 31 2008 12:07:52
    @INC:
    /usr/perl5/5.8.4/lib/sun4-solaris-64int
    /usr/perl5/5.8.4/lib
    /usr/perl5/site_perl/5.8.4/sun4-solaris-64int
    /usr/perl5/site_perl/5.8.4
    /usr/perl5/site_perl
    /usr/perl5/vendor_perl/5.8.4/sun4-solaris-64int
    /usr/perl5/vendor_perl/5.8.4
    /usr/perl5/vendor_perl
    Thanks &Regard

  • Getting error while running SSRS report in SharePoint 2010

    Hi,
    I am getting below error while running a SSRS reports in SharePoint. Your help will be appreciated
    Webpage error details
    User Agent: Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.1; WOW64; Trident/4.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; .NET4.0C; .NET4.0E)
    Timestamp: Wed, 5 Nov 2014 11:58:42 UTC
    Message: Sys.WebForms.PageRequestManagerServerErrorException: An unknown error occurred while processing the request on the server. The status code returned from the server was: 503
    Line: 5
    Char: 62123
    Code: 0
    Thanks,
    Manjunath

    Hi,
    According to your post, my understanding is that you get error while running SSRS reports.
    Basically, If the Request Management service is enabled, it manages all incoming requests by evaluating logic rules against the user requests in order to determine what\which action to take and which machine or machines (targets) in the farm should handle
    the request.
    By default the Request time out value for Request Management service are as follows,
    RequestTimeout                    : 00:01:40
    RequestExecutionTimeout       : 01:00:00
    So the error throws from RM service for those reports who take more than 1 minute to generate the report.
    To solve this issue, we can stop the Request Management service.
    Or increase the time out
    value by making use of following PowerShell  script.
    $waUrl = "URLofWEbApp"
    $wa = Get-SPWebApplication $waUrl
    $rmSettings = $wa | Get-SPRequestManagementSettings
    $req=$wa.RequestManagementSettings
    $timeout2= New-TimeSpan -minutes 5
    Sreq.Requesttimeout=$timeout2
    $req.update()
    For more information:
    http://blogs.technet.com/b/sajiths/archive/2014/08/19/ssrs-reports-timed-out-with-39-503-39-error-sharepoint-2013.aspx
    Thanks & Regards,
    Jason
    Jason Guo
    TechNet Community Support

  • Getting error while running OIM in built purging script OIM_ReconArch.sh

    Getting error while running OIM in built purging script OIM_ReconArch.sh
    bash-3.00$ cat Err_Arch_Recon_2012_11_01_15_36_06.log
    -- Reconciliation Archival Execution Status
    BEGIN OIM_SP_ReconArchival('N','N',200000,'ARCHIVE_SELECT_WITH_CONS_DROP_REC','Both Events','20110101','20111231'); END;
    ERROR at line 1:
    ORA-00942: table or view does not exist
    ORA-06512: at "PREPRD.OIM_SP_RECONARCHIVAL", line 722
    ORA-00942: table or view does not exist
    ORA-01031: insufficient privileges
    ORA-06512: at line 1
    Please Help

    ORA-06512: at "PREPRD.OIM_SP_RECONARCHIVAL", line 722
    ORA-00942: table or view does not exist
    ORA-01031: insufficient privileges
    (1) Check whether the OIM_SP_RECONARCHIVAL table or view exists or not...
    (2) If it exists, check whether the DEV_OIM user has enough privilege to view this table or not...
    If not, give it all privilege by login as sys as sysdba
    GRANT ALL PRIVILEGES TO DEV_OIM;
    COMMIT;

  • Getting error while opening Enterprise project type (EPT)

    Hello All,
    In Project Server 2013, I am getting error while opening any Enterprise project type (EPT) from EPT page or when I am trying ot create any new EPT. How can I troublshoot this, please help.
    Thanks in advance

    I searched ULS logs and got the below error details.
    Exception occured in method 'Workflow.ReadWorkflows' System.Data.NoNullAllowedException: Column 'WORKFLOW_ASSOCIATION_NAME' does not allow nulls.     at System.Data.DataColumn.CheckNullable(DataRow row)     at System.Data.DataColumn.CheckColumnConstraint(DataRow
    row, DataRowAction action)     at System.Data.DataTable.RaiseRowChanging(DataRowChangeEventArgs args, DataRow eRow, DataRowAction eAction, Boolean fireEvent)     at System.Data.DataTable.SetNewRecordWorker(DataRow row,
    Int32 proposedRecord, DataRowAction action, Boolean isInMerge, Boolean suppressEnsurePropertyChanged, Int32 position, Boolean fireEvent, Exception& deferredException)     at System.Data.DataTable.InsertRow(DataRow row, Int64 proposedID,
    Int32 pos, Boolean fireEvent)     at System.Data.DataTable....
    Let me know your advise on this.

  • Getting error while importing metadata using View Objects

    Hi All,
    I am trying to create a repository using View Object in OBIEE 11.1.5.1 but getting error while viewing the data, after importing the metadata in the repository "[nQSError: 77031] Error occurs while calling remote service ADFService11G. Details: Runtime error for service -- ADFService11G - oracle/apps/fnd/applcore/common/ApplSession".
    I am also getting error "žADFException-2015: The BI Server is incompatible with the BI-ADF Broker Servlet: BI Server protocol version = null, BI-ADF Broker Servlet protocol version = 1" during testing my sample which is deployed to Admin server. I followed BI Adminstrator help file guide in order to create the sample for creating repository using view object.
    Admin server log says
    [2011-09-27T02:59:03.646-05:00] [AdminServer] [NOTIFICATION] [] [oracle.bi.integration.adf] [tid: [ACTIVE].ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: <anonymous>] [ecid: b260b57746aa92d3:-1f9ca26:1328fcfd3e6:-8000-0000000000006744,0] [APP: BIEEOrders] [[
    QUERY:
    <?xml version="1.0" encoding="UTF-8" ?><ADFQuery><Parameters></Parameters><Projection><Attribute><Name><![CDATA[Deptno]]></Name><ViewObject><![CDATA[AppModule.DeptViewObj1]]></ViewObject></Attribute><Attribute><Name><![CDATA[Dname]]></Name><ViewObject><![CDATA[AppModule.DeptViewObj1]]></ViewObject></Attribute><Attribute><Name><![CDATA[Loc]]></Name><ViewObject><![CDATA[AppModule.DeptViewObj1]]></ViewObject></Attribute></Projection><JoinSpec><ViewObject><Name><![CDATA[AppModule.DeptViewObj1]]></Name></ViewObject></JoinSpec></ADFQuery>
    [2011-09-27T02:59:04.199-05:00] [AdminServer] [ERROR] [] [oracle.bi.integration.adf.v11g.obieebroker] [tid: [ACTIVE].ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: <anonymous>] [ecid: b260b57746aa92d3:-1f9ca26:1328fcfd3e6:-8000-0000000000006744,0] [APP: BIEEOrders] java.lang.NoClassDefFoundError: oracle/apps/fnd/applcore/common/ApplSession[[
         at oracle.bi.integration.adf.ADFDataQuery.makeQueryBuilder(ADFDataQuery.java:81)
         at oracle.bi.integration.adf.ADFDataQuery.<init>(ADFDataQuery.java:70)
         at oracle.bi.integration.adf.ADFReadQuery.<init>(ADFReadQuery.java:15)
         at oracle.bi.integration.adf.ADFService.makeADFQuery(ADFService.java:227)
         at oracle.bi.integration.adf.ADFService.execute(ADFService.java:136)
         at oracle.bi.integration.adf.v11g.obieebroker.ADFServiceExecutor.execute(ADFServiceExecutor.java:185)
         at oracle.bi.integration.adf.v11g.obieebroker.OBIEEBroker.doGet(OBIEEBroker.java:89)
         at oracle.bi.integration.adf.v11g.obieebroker.OBIEEBroker.doPost(OBIEEBroker.java:106)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
         at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
         at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:300)
         at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at oracle.adf.share.http.ServletADFFilter.doFilter(ServletADFFilter.java:62)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:111)
         at java.security.AccessController.doPrivileged(Native Method)
         at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:313)
         at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:413)
         at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:94)
         at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:161)
         at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:136)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:27)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3715)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3681)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
         at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2277)
         at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2183)
         at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1454)
         at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
         at weblogic.work.ExecuteThread.run(ExecuteThread.java:178)
    Caused by: java.lang.ClassNotFoundException: oracle.apps.fnd.applcore.common.ApplSession
         at weblogic.utils.classloaders.GenericClassLoader.findLocalClass(GenericClassLoader.java:297)
         at weblogic.utils.classloaders.GenericClassLoader.findClass(GenericClassLoader.java:270)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:305)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:246)
         at weblogic.utils.classloaders.GenericClassLoader.loadClass(GenericClassLoader.java:179)
         ... 38 more
    Please suggest how to make it work.

    Hi,
    Thanks for providing the online help URL, i have already completed the steps mentioned in URL
    I am able to import metadata using view object but getting above error("[nQSError: 77031] Error occurs while calling remote service ADFService11G. Details: Runtime error for service -- ADFService11G - oracle/apps/fnd/applcore/common/ApplSession".") while validating the data.
    It fails at step 5 of "To import metadata from an ADF Business Component data source:" section of above URL.

  • Getting error while loading Flat File

    Hello All,
    I am getting error while loading flat file. Flat file is a CSV file.
    Value ',,,,,,,,' of characteristic 0DATE is not a number with 000008 spaces
    Data seprator  |
    Escape sign    ;
    It has 23708 entries , it s loading successfully till 23 665 entries
    Besides when checked in PSA
    for record having entries >23667 has calender day as ,,,,,, where as rest entries are  having date
    Besides when i checked in Flat file ,the total number of rows is 23,667 is there but i wonder why it has got 23,708 in
    RSMO
    Could you please let me know how to correct.
    regards
    path

    Hi,
    For date column you should maintain YYYYMMDD formate Eg: 20090601, kepp cursor on date column and right click and Formate >Custome>make it 00000000 then save teh file as .CSV . First type values on column and do like this formate and save it and without opening it load it. Once you open it you losw 00000000 formate you need to give again the same formate.
    Settings in Infopackage:
    Data Format  = CSV
    Data Separator = ,
    Escape Sign = ;
    http://help.sap.com/saphelp_nw2004s/helpdata/en/43/01ed2fe3811a77e10000000a422035/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/80/1a6567e07211d2acb80000e829fbfe/content.htm
    Thanks
    Reddy

  • Getting error while installing reporting servcies on existing sql server 2012.

    Hi Expaerts,
     sql 2012 sp1:
    getting error while installing Reporting services on existing sql server. at final step
    Error on scren: install_rsharepoint_cpu64_action : roll back , rolling back action.
    error on popup windows:  ' one or more files required to restore your computer to its previous state colud  not be found. restoration will not be found.
    Thanks in advance.

    Hello,
    Please help to collect the useful log information while install SQL Server Reporting Services. Here is the article for your reference, please see:
    Navigating the Setup Logs when you install SQL Server 2012 Reporting Services:
    http://blogs.msdn.com/b/jtarquino/archive/2012/03/14/navigating-the-setup-logs-when-you-install-sql-server-2012-reporting-services.aspx
    Regards,
    Elvis Long
    TechNet Community Support

  • Getting error while starting managed recovery

    Hi,
    I am getting error while starting managed recovery.Please suggest.
    Please find below error log
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION
    Attempt to start background Managed Standby Recovery process (OTP3)
    Wed Jun 23 19:24:10 2010
    MRP0 started with pid=50, OS id=3350
    MRP0: Background Managed Standby Recovery process started (OTP3)
    Fast Parallel Media Recovery enabled
    Managed Standby Recovery not using Real Time Apply
    Warning: Recovery target destination is in a sibling branch
    of the controlfile checkpoint. Recovery will only recover
    changes to datafiles.
    Datafile 1 (ckpscn 296765005228) is orphaned on incarnation#=1
    MRP0: Background Media Recovery terminated with error 19909
    Errors in file /ora/app/oracle/admin/OTP3/diag/rdbms/otp3/OTP3/trace/OTP3_mrp0_3350.trc:
    ORA-19909: datafile 1 belongs to an orphan incarnation
    ORA-01110: data file 1: '/ora_ots_gdb1/oradata/OTP3/datafile/o1_mf_SYSTEM_REORG1_.dbf'
    Errors in file /ora/app/oracle/admin/OTP3/diag/rdbms/otp3/OTP3/trace/OTP3_mrp0_3350.trc:
    ORA-19909: datafile 1 belongs to an orphan incarnation
    ORA-01110: data file 1: '/ora_ots_gdb1/oradata/OTP3/datafile/o1_mf_SYSTEM_REORG1_.dbf'
    MRP0: Background Media Recovery process shutdown (OTP3)
    Completed: ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION

    Hi,
    The standby database has been activated previously but since then reverted back into its standby role. Prior to the
    RMAN duplicate operation the "alter system reset control_files scope=spfile sid='*';" has not been executed (i.e. the
    control files have not been overwritten).
    check the incarnation on both the sides i.e primary & standby
    list incarnation of database;
    RMAN> reset database to incarnation 2;
    database reset to incarnation 2
    RMAN> exit
    hope you can solve by restting the incarnation.
    Thanks

Maybe you are looking for

  • Warning message in Lion

    Hi , I just installed Mac OSX Lion. When I download a set of images (like a manga for example) in Preview , a warning message saying that this is an image taken from the internet (and asking me if I'm sure I want to open it) keep bugging me. How to g

  • HT204406 iTunes Match ISSUE: we could not complete your itunes store request.  an unknown error occured (-50)

    I cannot turn on my iTunes Match.  When I try to add my computer it just says: "we could not complete your itunes store request.  an unknown error occured (-50)" and then I press ok and it says: "We could not complete your iTunes Store request.  The

  • My iPad screen seems to have frozen

    My iPad screen seems to have frozen and I cannot even power off any suggestions?

  • Missing tab in icon editor

    I'm pretty sure I'm missing a tab in my icon editor...  Can anybody tell me where my missing tab (forgot the name of it) has been gone to? Furthermore I can't synchronize with the NI icon library. The problems occured today for the first time. After

  • Help! Publish cancelled - and it AIN'T what you think!

    Hi all! For some reason when I publish with the same FTP settings I've used dozens of time, I get a message "Publishing has been cancelled.  Failed to create file: Account_Reminders.jpg."  Now - before you think you know the answer here, I have consu