Unable to fetch requisition lines data from package POR_CUSTOM_PKG.CUSTOM_VALIDATE_REQ_DIST

Problem Description
We need to validate the Distribution data and based on that we will throw Error with respective Line Number. But the line number is not in the parameter list of this procedure POR_CUSTOM_PKG.CUSTOM_VALIDATE_REQ_DIST. So we thought of deriving the Line Number by querying the table po_requisition_lines_all by passing the parent_req_line_id. But unfortunately the query is throwing no data found exception though the related line records are seen from the backend.
Request your help to fetch the requisition line number from this procedure POR_CUSTOM_PKG.CUSTOM_VALIDATE_REQ_DIST
Thanks & Regards,
Munish Mittal

Hi,
This is just for the information that the issue is resolved for me by using the below code.
begin
select parent_req_line_id into l_line_id_1 from dual;
select requisition_line_id, line_num into l_line_id, l_line_num
from po_requisition_lines_all
where requisition_line_id = l_line_id_1;
exception when others then
null;
  end;

Similar Messages

  • LMS 4.2.4 Error RICS0001:Internal Error,unable to process the collected data from the device.

    Hi
    LMS 4.2.4 inventory collection fails with Error RICS0001:Internal Error,unable to process the collected data from the device for few devices.
    Can anyone help me out.
    Regards
    Laxmi

    They are definately supported. However such issues are usually due to device packages, if unavailable or not installed.
    Please do following :
    1. Check Admin > software centre > device updates and check any devices packages available in Cisco.com and install it and recheck.
    2. Please do a snmpwalk on problematic device on sysObjectID and share result. Following is the way to do snmpwalk :
    GUI - Open the device in Device Centre (Inventory > Tools > Device Center) and open the device with IP and from tools select snmpwalk and do it for sysObjectID (1.3.6.1.2.1.1.2)
    CLI : CD to NMSROOT/objects/jt/bin from command prompt. run command as per following syntax:
    snmpwalk -v2c -c <snmp_community_string> 1.3.6.1.2.1.1.2
    Example :
    community string = test123
    IP = 10.1.1.1
    snmpwalk -v2c -c test123 10.1.1.1 1.3.6.1.2.1.1.2
    3. Please check your device package, by clicking on the number of packages in front of Inventory, conifg and image management from :
    Admin > System > Software Center > Device Update
    Also, check device identity under edit identity option from :
    Inventory > Device Administration > Add / Import / Manage Devices
    Check my screenshots for details :
    Please check and share results.
    -Thanks
    Vinod
    **Rating Encourages contributors, and its really free. **

  • RICS0001:Internal Error,unable to process the collected data from the device.

    Hi all,
    I've got the following error in Inventory Collection: 'RICS0001:Internal Error,unable to process the collected data from the device.'
    System is CW LMS 2.6
    If I search the web I get the following Cisco document:
    http://www.cisco.com/en/US/products/sw/cscowork/ps2073/prod_troubleshooting_guide09186a008036dff2.html
    '...in the log directory look for IC_Server.log.'
    IC_Server.log:
    [ Di Nov 16  15:54:27 CET 2010 ],INFO ,[Thread-25],com.cisco.nm.rmeng.inventory.ics.core.ICSCore,173,Got Async Request, User Name :admin
    [ Di Nov 16  15:54:27 CET 2010 ],INFO ,[Thread-25],com.cisco.nm.rmeng.inventory.ics.core.ICSCore,179,Request ID is : 1289919235488
    [ Di Nov 16  15:54:27 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.inventory.ics.core.CollectionController,309,Started processing device ID: 3341
    [ Di Nov 16  15:54:27 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.RMELogger,724,com.cisco.nm.rmeng.util.db.DatabaseConnectionPool,getConnection,59,Inside ICSDatabaseConnection, MAX_COUNT =20
    [ Di Nov 16  15:54:28 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.inventory.ics.core.CollectionController,387,Started processing device Name: 9.152.255.101
    [ Di Nov 16  15:54:28 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,g$eval,103,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:g$eval:populating ContainmentAG attributes, begins...
    [ Di Nov 16  15:54:28 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,populatingTheChassis,110,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:populatingTheChassis:ContainmentAG attributes,collection from the device begins...
    [ Di Nov 16  15:54:29 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,populatingTheChassis,147,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:populatingTheChassis:ContainmentAG attributes,collection from the device successful...
    [ Di Nov 16  15:54:29 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,populatingTheChassis,149,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:populatingTheChassis:ContainmentAG attributes,population begins...
    [ Di Nov 16  15:54:29 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,populatingTheChassis,216,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:populatingTheChassis:Before method getSlotsConfiguredStatistics
    [ Di Nov 16  15:54:29 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,populatingTheChassis,225,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:populatingTheChassis:After method getSlotsConfiguredStatistics
    [ Di Nov 16  15:54:29 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,populatingTheChassis,245,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:populatingTheChassis:ContainmentAG attributes,population completed...
    [ Di Nov 16  15:54:29 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.util.logger.XDILogger,77,com.cisco.nm.xms.xdi.pkgs.SharedInventoryCatIOS.ContainmentAGI_ENTITY_Mib_Non_Modular,g$eval,105,com.cisco.nm.xms.xdi.pkgs.SharedInventoryRouter:ContainmentAGI_ENTITY_Mib_Non_Modular:g$eval:populating ContainmentAG attributes, ends...
    [ Di Nov 16  15:54:43 CET 2010 ],INFO ,[Thread-14],com.cisco.nm.rmeng.inventory.ics.core.CollectionController,477,DP time is 15 seconds for 9.152.255.101
    [ Di Nov 16  15:54:44 CET 2010 ],ERROR,[Thread-14],com.cisco.nm.rmeng.util.logger.RMELogger,770,com.cisco.nm.rmeng.inventory.ics.invchange.AddInvChange,effect,33,Unexpected error :com.sybase.jdbc2.jdbc.SybSQLException: ASA Fehler -193: Primärschlüssel für Tabelle 'PhysicalElement' ist nicht eindeutig
        at com.sybase.jdbc2.tds.Tds.processEed(Tds.java:2834)
        at com.sybase.jdbc2.tds.Tds.nextResult(Tds.java:2156)
        at com.sybase.jdbc2.jdbc.ResultGetter.nextResult(ResultGetter.java:69)
        at com.sybase.jdbc2.jdbc.SybStatement.nextResult(SybStatement.java:220)
        at com.sybase.jdbc2.jdbc.SybStatement.nextResult(SybStatement.java:203)
        at com.sybase.jdbc2.jdbc.SybStatement.executeLoop(SybStatement.java:1766)
        at com.sybase.jdbc2.jdbc.SybStatement.execute(SybStatement.java:1758)
        at com.sybase.jdbc2.jdbc.SybPreparedStatement.execute(SybPreparedStatement.java:619)
        at com.cisco.nm.rmeng.inventory.ics.dbrep.DBRecord.insert(DBRecord.java:50)
        at com.cisco.nm.rmeng.inventory.ics.util.ICSDatabaseConnection.insert(ICSDatabaseConnection.java:91)
        at com.cisco.nm.rmeng.inventory.ics.invchange.AddInvChange.effect(AddInvChange.java:29)
        at com.cisco.nm.rmeng.inventory.ics.server.InvDataProcessor.processInvData(InvDataProcessor.java:394)
        at com.cisco.nm.rmeng.inventory.ics.core.CollectionController.run(CollectionController.java:849)
        at java.lang.Thread.run(Thread.java:534)
    [ Di Nov 16  15:54:44 CET 2010 ],ERROR,[Thread-14],com.cisco.nm.rmeng.inventory.ics.server.InvDataProcessor,448,ASA Fehler -193: Primärschlüssel für Tabelle 'PhysicalElement' ist nicht eindeutig
    com.sybase.jdbc2.jdbc.SybSQLException: ASA Fehler -193: Primärschlüssel für Tabelle 'PhysicalElement' ist nicht eindeutig
        at com.sybase.jdbc2.tds.Tds.processEed(Tds.java:2834)
        at com.sybase.jdbc2.tds.Tds.nextResult(Tds.java:2156)
        at com.sybase.jdbc2.jdbc.ResultGetter.nextResult(ResultGetter.java:69)
        at com.sybase.jdbc2.jdbc.SybStatement.nextResult(SybStatement.java:220)
        at com.sybase.jdbc2.jdbc.SybStatement.nextResult(SybStatement.java:203)
        at com.sybase.jdbc2.jdbc.SybStatement.executeLoop(SybStatement.java:1766)
        at com.sybase.jdbc2.jdbc.SybStatement.execute(SybStatement.java:1758)
        at com.sybase.jdbc2.jdbc.SybPreparedStatement.execute(SybPreparedStatement.java:619)
        at com.cisco.nm.rmeng.inventory.ics.dbrep.DBRecord.insert(DBRecord.java:50)
        at com.cisco.nm.rmeng.inventory.ics.util.ICSDatabaseConnection.insert(ICSDatabaseConnection.java:91)
        at com.cisco.nm.rmeng.inventory.ics.invchange.AddInvChange.effect(AddInvChange.java:29)
        at com.cisco.nm.rmeng.inventory.ics.server.InvDataProcessor.processInvData(InvDataProcessor.java:394)
        at com.cisco.nm.rmeng.inventory.ics.core.CollectionController.run(CollectionController.java:849)
        at java.lang.Thread.run(Thread.java:534)
    [ Di Nov 16  15:54:44 CET 2010 ],ERROR,[Thread-14],com.cisco.nm.rmeng.inventory.ics.core.CollectionController,861, Exception occured in process method while processing: 9.152.255.101 ASA Fehler -193: Primärschlüssel für Tabelle 'PhysicalElement' ist nicht eindeutig
    ICSException :: ASA Fehler -193: Primärschlüssel für Tabelle 'PhysicalElement' ist nicht eindeutig
        at com.cisco.nm.rmeng.inventory.ics.server.InvDataProcessor.processInvData(InvDataProcessor.java:463)
        at com.cisco.nm.rmeng.inventory.ics.core.CollectionController.run(CollectionController.java:849)
        at java.lang.Thread.run(Thread.java:534)
    I don't know where the German error message comes from. The whole system is English. Translation is: "ASA error -192: primary key for table 'PhysicalElement' isn't distinct'
    Thanks for your help!
    Alex

    Well known issue in LMS 2.6. Most likely you are hitting CSCsm97530. As a temporary solution you could remove and re-add problematic device from CS. To get a permanent fix you need a patch (provided by TAC).

  • LMS Prime 4.1 RICS0001:Internal Error,unable to process the collected data from the device.

    Hello,
    I have a problem with inventory collection from several devices getting
    "RICS0001:Internal Error,unable to process the collected data from the device."
    ERROR,[Thread-20],com.cisco.nm.rmeng.inventory.ics.server.InvDataProcessor,918,Device does not exist in RME, Please check if device is already deleted

    Did you schedule an inventory collection job?
    Delete schedule and recreate it using device group.
    Cheers,
    Michel

  • Standby database unable to fetch missing archive sequence from primary

    Hi,
    My primary oracle db version is 10.2.0.5 which is n Windows 2003 server 32 bit in file system
    My standby database version is 10.2.0.5 which in wondows 2008 server in ASM instance.
    some of the archivelogs are not fetching by standby database from primary db.because of that recovery is not happening and it waits for missing archive logs but its not transfering on satndby database.
    I tried many time by doing manuall..manuall its hapening but again sometimes its stucks and waits for next missing archivelog.
    standby alert log mesage
    Sun Jul 08 21:33:13 Pacific Daylight Time 2012
    FAL[client]: Failed to request gap sequence
    GAP - thread 1 sequence 128945-128945
    DBID 4176230406 branch 617106566
    FAL[client]: All defined FAL servers have been attempted.
    Check that the CONTROL_FILE_RECORD_KEEP_TIME initialization
    parameter is defined to a value that is sufficiently large
    enough to maintain adequate log switch information to resolve
    archivelog gaps.
    Sun Jul 08 21:44:03 Pacific Daylight Time 2012
    db_recovery_file_dest_size of 2048 MB is 0.00% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    trace file of standby
    Dump file d:\oracle\admin\mks01p\bdump\mks01p_mrp0_6628.trc
    Sun Jul 08 21:30:23 2012
    ORACLE V10.2.0.5.0 - 64bit Production vsnsta=0
    vsnsql=14 vsnxtr=3
    Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bit Production
    With the Oracle Label Security, OLAP, Data Mining and Real Application Testing options
    Windows NT Version V6.1 Service Pack 1
    CPU : 8 - type 8664, 4 Physical Cores
    Process Affinity : 0x0000000000000000
    Memory (Avail/Total): Ph:37019M/49108M, Ph+PgF:87277M/103214M
    Instance name: mks01p
    Redo thread mounted by this instance: 1
    Oracle process number: 19
    Windows thread id: 6628, image: ORACLE.EXE (MRP0)
    *** SERVICE NAME:() 2012-07-08 21:30:23.422
    *** SESSION ID:(95.1) 2012-07-08 21:30:23.422
    *** 2012-07-08 21:30:23.422 64949 kcrr.c
    MRP0: Background Managed Standby Recovery process started
    *** 2012-07-08 21:30:28.445 1170 krsm.c
    Managed Recovery: Initialization posted.
    *** 2012-07-08 21:30:28.445 64949 kcrr.c
    Managed Standby Recovery not using Real Time Apply
    *** 2012-07-08 21:30:39.833
    Recovery target incarnation = 1, activation ID = -118713082
    Influx buffer limit = 10000 min(50% x 191963, 10000)
    Successfully allocated 7 recovery slaves
    Using 158 overflow buffers per recovery slave
    Start recovery at thread 1 ckpt scn 7982577982885 logseq 128945 block 2
    *** 2012-07-08 21:30:40.784
    Media Recovery add redo thread 1
    *** 2012-07-08 21:30:40.784 1170 krsm.c
    Managed Recovery: Active posted.
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '+ORAREDO3/mks01p/mks01p_redo01b.rdo'
    ORA-17503: ksfdopn:2 Failed to open file +ORAREDO3/mks01p/mks01p_redo01b.rdo
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-00312: online log 1 thread 1: '+ORAREDO1/mks01p/mks01p_redo01a.rdo'
    ORA-17503: ksfdopn:2 Failed to open file +ORAREDO1/mks01p/mks01p_redo01a.rdo
    ORA-15001: diskgroup "ORAREDO1" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO1" does not exist or is not mounted
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '+ORAREDO3/mks01p/mks01p_redo01b.rdo'
    ORA-17503: ksfdopn:2 Failed to open file +ORAREDO3/mks01p/mks01p_redo01b.rdo
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-00312: online log 1 thread 1: '+ORAREDO1/mks01p/mks01p_redo01a.rdo'
    ORA-17503: ksfdopn:2 Failed to open file +ORAREDO1/mks01p/mks01p_redo01a.rdo
    ORA-15001: diskgroup "ORAREDO1" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO1" does not exist or is not mounted
    *** 2012-07-08 21:30:40.940 64949 kcrr.c
    Clearing online redo logfile 1 +ORAREDO1/mks01p/mks01p_redo01a.rdo
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '+ORAREDO3/mks01p/mks01p_redo01b.rdo'
    ORA-17503: ksfdopn:2 Failed to open file +ORAREDO3/mks01p/mks01p_redo01b.rdo
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-00312: online log 1 thread 1: '+ORAREDO1/mks01p/mks01p_redo01a.rdo'
    ORA-17503: ksfdopn:2 Failed to open file +ORAREDO1/mks01p/mks01p_redo01a.rdo
    ORA-15001: diskgroup "ORAREDO1" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO1" does not exist or is not mounted
    ORA-00349: failure obtaining block size for '+ORAREDO1/mks01p/mks01p_redo01a.rdo'
    ORA-15001: diskgroup "ORAREDO1" does not exist or is not mounted
    Error 349 creating/clearing online redo logfile 1
    *** 2012-07-08 21:30:40.987 64949 kcrr.c
    Clearing online redo logfile 1 complete
    *** 2012-07-08 21:30:41.034 64949 kcrr.c
    Media Recovery Waiting for thread 1 sequence 128945
    *** 2012-07-08 21:30:41.050 64949 kcrr.c
    Fetching gap sequence in thread 1, gap sequence 128945-128945
    Redo shipping client performing standby login
    OCISessionBegin failed -1
    .. Detailed OCI error val is 1031 and errmsg is 'ORA-01031: insufficient privileges
    *** 2012-07-08 21:30:41.533 64949 kcrr.c
    Error 1031 received logging on to the standby
    *** 2012-07-08 21:30:41.533 64949 kcrr.c
    FAL[client, MRP0]: Error 1031 connecting to MKS01P_PRD for fetching gap sequence
    ORA-01031: insufficient privileges
    *** 2012-07-08 21:31:11.610
    Redo shipping client performing standby login
    OCISessionBegin failed -1
    .. Detailed OCI error val is 1031 and errmsg is 'ORA-01031: insufficient privileges
    *** 2012-07-08 21:31:12.094 64949 kcrr.c
    Error 1031 received logging on to the standby
    *** 2012-07-08 21:31:12.094 64949 kcrr.c
    FAL[client, MRP0]: Error 1031 connecting to MKS01P_PRD for fetching gap sequence
    ORA-01031: insufficient privileges
    *** 2012-07-08 21:31:42.186
    Redo shipping client performing standby login
    OCISessionBegin failed -1
    .. Detailed OCI error val is 1031 and errmsg is 'ORA-01031: insufficient privileges
    *** 2012-07-08 21:31:42.654 64949 kcrr.c
    Error 1031 received logging on to the standby
    *** 2012-07-08 21:31:42.670 64949 kcrr.c
    FAL[client, MRP0]: Error 1031 connecting to MKS01P_PRD for fetching gap sequence
    ORA-01031: insufficient privileges
    *** 2012-07-08 21:32:12.747
    Redo shipping client performing standby login
    OCISessionBegin failed -1
    .. Detailed OCI error val is 1031 and errmsg is 'ORA-01031: insufficient privileges
    *** 2012-07-08 21:32:13.215 64949 kcrr.c
    Error 1031 received logging on to the standby
    *** 2012-07-08 21:32:13.230 64949 kcrr.c
    FAL[client, MRP0]: Error 1031 connecting to MKS01P_PRD for fetching gap sequence
    ORA-01031: insufficient privileges
    *** 2012-07-08 21:32:43.276
    Redo shipping client performing standby login
    OCISessionBegin failed -1
    .. Detailed OCI error val is 1031 and errmsg is 'ORA-01031: insufficient privileges
    *** 2012-07-08 21:32:43.759 64949 kcrr.c
    Error 1031 received logging on to the standby
    *** 2012-07-08 21:32:43.775 64949 kcrr.c
    FAL[client, MRP0]: Error 1031 connecting to MKS01P_PRD for fetching gap sequence
    ORA-01031: insufficient privileges
    *** 2012-07-08 21:33:13.821
    Check that the CONTROL_FILE_RECORD_KEEP_TIME initialization
    parameter is defined to a value that is sufficiently large
    enough to maintain adequate log switch information to resolve
    archivelog gaps.
    i created new password fle on bith primary and standby database also.
    Please help me to resolve this issue.

    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mounted
    ORA-15001: diskgroup "ORAREDO3" does not exist or is not mountedhave you mentioned parameter LOG_FILE_NAME_CONVERT in standby when online redo log locations are different?
    post from standby:-
    SQL> select name, state From v$asm_diskgroup;
    FAL[client, MRP0]: Error 1031 connecting to MKS01P_PRD for fetching gap sequence
    ORA-01031: insufficient privilegesPost from primary & standby
    SQL> select * from v$pwfile_users;
    User Profile for 919131
    919131     
    Handle:     919131  
    Status Level:     Newbie
    Registered:     Mar 6, 2012
    Total Posts:     16
    Total Questions:     8 (8 unresolved)
    OTN failed 100% to help you, then why you posted another question?
    First close all your old answered threads and then better continue your updates in your thread.
    Edited by: CKPT on Jul 9, 2012 11:45 AM

  • Unable to fetch 50,000 records from SQL using Orchestrator

    Hi Team,
    I have a table in MS SQL which is having more than 50,000 records. I am trying to fetch 50,000 records using orchestrator, but unable to fetch the records..
    There is no problem with the SQL query because I can able to get 40,000 records..
    I am using SCORCH DEV - SQL Integration Pack.
    Regards,
    Soundarajan.

    Hi,
    Thanks for your query.
    I have also used timeout parameter but it is not working.. As you said I also tried with Query database Activity which is out of the box...
    Now i can able to fetch more than 80,000 records but the output what i am getting is not in the format which we are looking for..
    Attached the output...
    How to edit this..?
    I tried to write the output in excel but all the data sits in the first column itself..
    Regards,
    Soundarajan.

  • Fetching complete PM data from SAP instance for reupload

    Hi,
    Is there any way out to fetch out the complete SAP PM master/ transaction data from existing SAP instance as ready reference or almost exact for uploading into another instance?
    Please suggest some effecient way with reduced effort
    Thanks
    -Lucky

    Lucky
    I'm assuming you don't want a full production copy..
    And there is not a standard tool to extract all PM data..
    But you should be able to use standard tools like the list-edit reports, LSMW, IBIP, etc.
    PeteA

  • Unable to extract Customer Master data from MDC

    Hello,
    We creating an MDM-XI-ERP2005 Customer scenario. We want to work with all bussiness scenarios (not customizing scenario).
    First I'm trying to extract Customer Master Data from EP 2005 using MDM_CLNT_EXTR.
    I have created a New Variant which properties are:
    Variant: CUSTOMER_TOTAL
    Description: Send all customer data to MDM
    Extraction Object: CUSTOMER_EXTRACT.
    Target System: PI7.
    But, the display jobs shows the following error: Could not determine recipients for message type MDMRECEIPT. Job cancelled after system exception.
    We need to focus the problem. We have done the following ALE configuration.
    SALE transaction:
    In Basic Settings -> Logical Systems -> Define Logical System we have add the following logical systems:
                  EUS100
                  MDM: Master Data Management
                  PI7: Process Integration 7.0
    WE21:
    Transactional RFC -> PI7, IDOC record types SAP Release 4.x
    RFC destination: SAPSLDAPI
    BD64
    New Model -> PFMC
    Add message type: Model iview: PFMC
                                 Sender: EUS100
                                  Receiver: PI7
                                 Message Type: DEBMDM
    WE20 transaction shows:
    Partner profiles -> Partner Type LS (logical system)
                        PI7: Outbound parmtrs. DEBMDM, SYNCH.
    Any idea will be very helpful,
    Thanks in advance
    Marta.
    Edited by: Marta Sánchez on Jul 9, 2008 12:34 PM

    Sorry because I made a mistake in my previous message. ALE seems to be correctly configured. I'll try to update correctly my current situation:
    In WE02 I can see al IDOCs that I'm trying to send to XI. All them have the IDoc Status = 3 (Data Passed to Port OK) instead of 12 Status Dispatch OK :(. So, It seems an tRFC error.
    SM58 shows me more trails: "Basic type 'DEBMDM06' is unknown"
    before that I have done the following steps at SAP ERP:
        1. WE21: I have created a PI7 port with 700PI70CLNT RFC.
        2. SM59: I have checked successfully 700PI70CLNT RFC connection.
        3. WE20: I have created a PI7 Partner Type LS with the following inbound and outbound parameters: DEBMDM and MDMRECEIPT.
        4. BD64: I have created a Distribution Model with DEBMDM and MDMRECEIPT message types.
    and at SAP PI:
         5.IDX: Port: SAPEUS. Client: 111. RFC Destination: PI7.
         6. technical and business system for sender SAP system(System Landscape Directory)...
    - Maybe some of the previous steps is not correct....
    - Is RFC Destination = PI7 is correct in step #5?
    - Is also necessary to configure more steps in PI system like RFC destinations or Transactional port?
    - I have read that is neccessary to check that the option "transfer idoc immediately" in we20. How?
    Please any idea will be very helpful.......
    Thank you!
    Best regards,
    Marta.

  • Fetching 7 recent dates from arraylist

    Hi experts,
    I have been stuck with an issue for which I seek your support/advice.
    I have an xml which returns numerous dates(say 150 or more) . I take all these dates from the xml date tag into an arraylist.
    The following getter method will return the arraylist with all the dates which I have taken from XML.
         public ArrayList<Date> getDates() {
              return dates;
         }My query is, I need to have the 7 recent dates to be taken as a collection again, from this arraylist.
    -3 dates before today, today's date, + 3 days dates after today. If today's date is not available, then the next immediate date will be the reference date for choosing -3 days and +3 days.
    I hope, I am clear with my question.
    I am going nuts thinking how to implement this for the past one day. If anyone has any idea regarding any simple methods which would help implementing this easily or any simple logic to implement this, kindly advice.
    Any help in this regard will be well appreciated.
    Looking Forward,
    Anees

    AneesAhamed wrote:
    Kayaman,
    I know this, but the real question is How to find
    * the one that matches the closest to the current date
    * if no today's date or 3 advanced dates are available, how to add, just the 3 recent dates (- 3 days dates)?
    This logic is bit tough and that is what I am stuck with!Well that depends on how you want it to work. If there is no date for today available, do you want the previous date or the next date, or do you want the one that is closer?
    I would basically divide it into parts like this:
    Case 1. All dates are available - Iterate through the List until you find current date, then get that and index-1, index-2, index-3 and index+1, index+2, index+3.
    Case 2. No current date or later dates are available - Iterate through the List until you get to the last one (and notice that it's not the current date), then get the last 3 dates.
    Case 3. Dates are available with gaps between them (like a date every 4 days or so, this is possible isn't it?) - Still iterate through the List looking for the current date, but as soon as you notice that the previous date is before current date and the next date is after the current date, choose one of them as the 'current' and get the 3 surrounding dates too.
    Basically there shouldn't be anything really difficult in this, if you know your requirements clearly.

  • Logic unable to handle midi controller data from my alesis andromeda

    I'm having problems recording knob movement (filter sweeps, lfo rate
    movement) into logic from my alesis andromeda. When logic is in record mode
    and I'm moving the filter from low to high very slowly, the data comes out
    all jumpy, with the midi in/out meter jumping all over the place.
    I found out that because the filters and knobs on the alesis andromeda are
    analog, it sends out NRPN data and not regular midi CC's. The NRPN data is
    in much larger increments than regular 0-127 midi data. I'm not sure why
    but logic was never made to handle this data directly. I've been told that
    there is a solution for filtering this NRPN data into data that logic can
    read/write/automate, but i'm unsure how.
    If anyone has any insight, it would be great appreciated.
    My setup is as follows:
    G5 Dual 2ghz
    Logic Pro 7.2.2
    Emagic Unitor 8 midi interface
    Alesis Andromeda Synth
    -Adam K

    brennon wrote:
    First, this data is recorded into a new MIDI region. Is there a way to record this directly into the existing region? If not, is there a way to copy the data from the new region into the existing region? If so, is there a way to copy ONLY the data from these selected MIDI parameters into the new region. For instance, say I've got my first region with everything but modulation data set, and I then record the modulation data into a new region. Can I only copy this modulation data into the existing region without blowing everything else away?
    *Either merge the regions by using the "glue" tool if you want to integrate all the controller data
    *To select only certain controller from an event list click on one event e.g. Modulation and then use "select similar" (my KC for this is shift+S) - all modulation data is selected and can be copied/pasted into another Midi region
    *You can also write data into existing regions using the Hyper Editor

  • Unable to find out remaining data from website or My Verizon - anyone else?

    For the last few days I have been unable to check data usage from either the Verizon website by logging in or from My Verizon on my phone - it gives a very useless "readout" but there's no way to determine how much data I have.  Guess this has been since the prepaid "system upgrade" a week ago.  Is this the case with everyone and does anyone have ANY idea when this will be resolved?  I just can't face calling customer service.
    A work-around is to call #3282 from your prepaid phone - it does tell you remaining data.  I have to assume that it's correct.

    I'm in the same boat.  I switched from postpaid to prepaid and it just shows my postpaid acct.  Myverizon app is usless I can't even log in or register with it...

  • How to  fetch the relational  data from the xml file registered in xdb

    Hi,
    I have to register the xml file into the  xdb repository and i have to fetch the data of the xml file as relational structure  through the select statement .
    i used the below query to register the xml file in xdb.
    DECLARE
    v_return BOOLEAN;
    BEGIN
    v_return := DBMS_XDB.CREATERESOURCE(
    abspath => '/public/demo/xml/db_objects.xml',
    data => BFILENAME('XML_DIR', 'db_objects.xml')
    COMMIT;
    END;
    Now i have to fetch the values in the xml file as relational data .
    whether it is possible ?
    can any one help me.
    Regards,
    suresh.

    When you transform your XMLdata to a xmltype you can do something like this for example:
    select
    extractvalue(value(p),'/XMLRecord/Session_Id') session_id,
    extractvalue(value(p),'/XMLRecord/StatementId') StatementId,
    extractvalue(value(p),'/XMLRecord/EntryId') EntryId
    from
    table(xmlsequence(extract(xmltype('
    <XMLdemo>
    <FormatModifiers><FormatModifier>UTFEncoding</FormatModifier></FormatModifiers>
    <XMLRecord>
    <Session_Id>117715</Session_Id>
    <StatementId>6</StatementId>
    <EntryId>1</EntryId>
    </XMLRecord>
    </XMLdemo>
    '),'/XMLdemo/*'))) p
    where extractvalue(value(p),'/XMLRecord/Session_Id') is not null;
    For this sample I've put a readable XML in plain text and convert it to xmltype so you can run it on your own database.

  • SPROXY unable to get Interface meta-data from XI Server

    Hi,
    In our SAP R/3 Enterprise system, the transaction SPROXY complains that "No connection to Integration Builder (only local data visible)".  It is unable to get the message interfaces defined in the SAP XI Server.
    Connection to SLD is configured correctly, and transaction SLDCHECK works correctly.
    Where can I configure the SAP R/3 system so that it can retrieve the message interfaces from SAP XI system ?
    thanks,
    Manish

    Hi Manish -
    Did you maintain the RFC destinations LCRSAPRFC and SAPSLDAPI?  And the corresponding J2EE connections?  These must be completed in addition to other possible configurations.
    Please check XI 3.0 Configuration Guide, section 8 (Connecting Business Systems with an Integration Engine to the Central Integration Server).  These have all the necessary steps.
    Regards,
    Jin

  • Unable to empty trash with data from Time Machine Drive

    Hi there,
    Cause of the current issue:
    I am hoping the experts can help me with this one. I have an external hard drive which is used for my Time Machine Backups. Recently there has been an error on the backup to this drive on occasion. This error only occurs when the system does the back up, and appears only periodically, never when I do the back up manually. One of the threads I read to solve this issue is to delete some older backup files from your time machine drive, and all is good. This worked for awhile.... but now a new problem:
    New Problem:
    I have been able to empty all of the trash save one particular set of folders. So the trash always appears with content.
    The structure of the trash content appears as such...
    Macintosh HD
    +-Users
    +--<username>
    +---Library
    +----FK6T0KGc9y4KBB <-- This changes each time I try to empty the trash.
    The library folder always has the time/date of the last time I tried to empty the trash.
    What I have tried:
    1. I have tried to empty the trash Securely. No good.
    2. I have tried the terminal command : sudo rm -rf ~/.Trash with no success
    3. I have tried this terminal command : sudo rm -rf ~/.Trash/* with no success
    4. I have tried just plain old Empty Trash. But I get the message "This operation cannot be completed because the item 'FK6T0KGc9y4KBB' is in use.
    Any thoughts, recommendations?

    Hey thanks for the help anyway VK.
    I tried that command and no go either.
    Last login: Fri Jul 25 08:13:39 on console
    iMac:~ tmlohnes$ sudo rm -rf /Volumes/Time\ Machine\ Backups/.Trashes/501/*
    Password:
    rm: /Volumes/Time Machine Backups/.Trashes/501/2008-03-28-160634/Macintosh HD/Users/tmlohnes/Library/8xMlmPGUHzilET: Directory not empty
    rm: /Volumes/Time Machine Backups/.Trashes/501/2008-03-28-160634/Macintosh HD/Users/tmlohnes/Library: Directory not empty
    rm: /Volumes/Time Machine Backups/.Trashes/501/2008-03-28-160634/Macintosh HD/Users/tmlohnes: Directory not empty
    rm: /Volumes/Time Machine Backups/.Trashes/501/2008-03-28-160634/Macintosh HD/Users: Directory not empty
    rm: /Volumes/Time Machine Backups/.Trashes/501/2008-03-28-160634/Macintosh HD: Directory not empty
    rm: /Volumes/Time Machine Backups/.Trashes/501/2008-03-28-160634: Directory not empty
    iMac:~ tmlohnes$
    I am wondering, would it make sense to try this same command if I were to do a safe boot?
    What about formatting the external hard drive and starting fresh?

  • How to fetch user specific data from View?

    Hi
    I have a requirement in which I need to display table data specific to Session_user only. Scenario is if user belongs to a specific region he should be able to see data related to that specific region only.
    I am thinking of passing session_user name in view query but do not know whether it's possible or not.If possible how to do this.
    What is the best way to achieve it?
    Thanks
    Vibzz

    Hello
    Vibzz,
    First thing is that you need to implement security in your application, if you configured security then, Take a session scope bean in your un bounded task flow
    inside that session bean , you can get the information from security context, then write a static block inside your session bean and modify the fallowing code according to your requirement
    ADFContext adfc=ADFContext.getCurrent();// get the context instance
    SecurityContext securityContext=adfc.getSecurityContext();// instantiate the SecurityContext
    String username=securityContext.getUserName(); // returns the username from the security Context who are currently logged in
    Once you written this code just assign this username to session scope variable
    Map sessionScope =
    (Map)ADFContext.getCurrent().getSessionScope().put("UserName", username);
    So that the username with the key UserName will be available throughout your session
    and Apply in where clause according to the user logged in, or else you may directly use
    adf.context.securityContext.userName at the bind variable what ever you used inside your view criteria.

Maybe you are looking for

  • Snooze problem since software update on N95

    I've had my N95 for about 18 months now. I don't use a lot of the things on it, but one I do use is the alarm clock to get me out of bed in time for work. I like to have a bit of "coming around" time so I usually have a few snoozes before getting up.

  • Window Controls Missing?

    Okay, back in my previous OSX, if I opened a folder I got a window with various control icons across the top.  Among them, on the far right, was an oblong button that toggled this control bar visible or invisible.  In Lion, this toggle button has dis

  • AE CS6 Crashes on Quit

    Every time I exit AE CS6 it crashes.  This is the prompt I get upon crash: Last log message was: <140735152419200> <ae.blitpipe> <2> HardwareBlitpipe Disengaged Please help.

  • How can a custom tag write multiple fragments in the response ?

    I have a custom tag that needs to write at least 2 html/xhtml fragments into the response in order to be html compliant. The custom tag outputs a script that must be located in the head section and other elements that refers to the script. Events han

  • What is syntax to include a file ??

    Hi, I am trying to include a JSP file using below syntax . <% String corFile="sample/cat/01.jsp"; corFile="/"+corFile; %> <%= corFile %> <%@ include file=corFile %> ==> It throws error. <%@ include file= "/sample/cat/o1.jsp " %> works. If I output th