Problem in loading the data into Infocube. 3.5 BW

Performed full load to Infocube but unable to get total records. For example from 1000 records got 990 records. In PSA it shows 1000 records. Request has performed no-errors.
Thanks in advance.

Hi Ahmed,
Infocube stores data in aggregated level, which means the key figures will be summed up instead of overwriting. The same property can be checked in your transformation rule types and make sure that the Aggregation Type is "Summation' and not "Overwrite". Please check for the summation of any one of your keyfigures and compare the same with that of PSA's Sum. If you find any difference then try to filter out records with smaller selections so that you can identify where the data is incorrect. Then probably you can take some corrective actions accordingly.
Regards,
Banu.

Similar Messages

  • How to write a procedure to load the data into a table using xml file as input to the procedure?

    Hi,
    Iam new to the xml,
    can u please anyone help me how to write procedure to load the data into a table using xml as input parameter to a procedure and xml file is as shown below which is input to me.
    <?xml version="1.0"?>
    <DiseaseCodes>
    <Entity><dcode>0</dcode><ddesc>(I87)Other disorders of veins - postphlebitic syndrome</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J04)Acute laryngitis and tracheitis</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J17*)Pneumonia in other diseases - whooping cough</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    </DiseaseCodes>.
    Regards,
    vikram.

    here is the your XML parse in 11g :
    select *
      from xmltable('//Entity' passing xmltype
    '<?xml version="1.0"?>
    <DiseaseCodes>
    <Entity><dcode>0</dcode><ddesc>(I87)Other disorders of veins - postphlebitic syndrome</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J04)Acute laryngitis and tracheitis</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J17*)Pneumonia in other diseases - whooping cough</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    </DiseaseCodes>
    ') columns
      "dcode" varchar2(4000) path '/Entity/dcode',
      "ddesc" varchar2(4000) path '/Entity/ddesc',
      "reauthflag" varchar2(4000) path '/Entity/reauthflag'
    dcode                                                                            ddesc                                                                            reauthflag
    0                                                                                (I87)Other disorders of veins - postphlebitic syndrome                           0
    0                                                                                (J04)Acute laryngitis and tracheitis                                             0
    0                                                                                (J17*)Pneumonia in other diseases - whooping cough                               0
    SQL>
    Using this parser you can create procedure as
    SQL> create or replace procedure myXMLParse(x clob) as
      2  begin
      3    insert into MyXmlTable
      4      select *
      5        from xmltable('//Entity' passing xmltype(x) columns "dcode"
      6                      varchar2(4000) path '/Entity/dcode',
      7                      "ddesc" varchar2(4000) path '/Entity/ddesc',
      8                      "reauthflag" varchar2(4000) path '/Entity/reauthflag');
      9    commit;
    10  end;
    11 
    12  /
    Procedure created
    SQL>
    SQL>
    SQL> exec myXMLParse('<?xml version="1.0"?><DiseaseCodes><Entity><dcode>0</dcode><ddesc>(I87)Other disorders of veins - postphlebitic syndrome</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity><Entity><dcode>0</dcode><ddesc>(J04)Acute laryngitis and tracheitis</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity><Entity><dcode>0</dcode><ddesc>(J17*)Pneumonia in other diseases - whooping cough</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity></DiseaseCodes>');
    PL/SQL procedure successfully completed
    SQL> select * from MYXMLTABLE;
    dcode                                                                            ddesc                                                                            reauthflag
    0                                                                                (I87)Other disorders of veins - postphlebitic syndrome                           0
    0                                                                                (J04)Acute laryngitis and tracheitis                                             0
    0                                                                                (J17*)Pneumonia in other diseases - whooping cough                               0
    SQL>
    SQL>
    Ramin Hashimzade

  • Error While loading the data into PSA

    Hi Experts,
           I have already loaded the data into my cube. But it doesn't have values for some fields. So again i modified the data in the flat file and tried to load them into PSA. But while starting the Info Package, I got an error saying that,
    "Check Load from InfoSource    
    Created       YOKYY  on   20.02.2008   12:44:33 
    Check Load from InfoSource , Packet IP_DS_C11
    Please execute the mail for additional information.
    Error message from the source system
    Diagnosis
    An error occurred in the source system.
    System Response
    Caller 09 contains an error message.
    Further analysis:
    The error occurred in Extractor .
    Refer to the error message.
    Procedure
    How you remove the error depends on the error message.
    Note
    If the source system is a Client Workstation, then it is possible that the file that you wanted to load was being edited at the time of the data request. Make sure that the file is in the specified directory, that it is not being processed at the moment, and restart the request.
    Pls help me in this......
    With Regards,
    Yokesh.

    Hi,
    After editing the file did you save the file and close it.
    This error may come if your file was open at the time of request.
    Also did you check the file path settings.
    If everything is correct try saving the infopackage once and loading again.
    Thanks,
    JituK

  • Does SSIS guarantee that it loads the data into SQL Server in the same order as it is in Excel

    Hi,
    We are trying to load several Excel files into SQL Server SSIS and we need to save the data in the database in the same order as it is in Excel..
    My question is, Does SSIS guarantee that it loads the data into SQL Server in the same order as it is in Excel. If so, we will add a sequence to ensure we have the order.
    Please advise.
    Thanks & Regards,
    Dhanumjay

    Thanks for your response.
    If it is one file then we can add an index column, but we have to load hundreds of files.
    How adding an index/key column to the table works unless SSIS picks and loads the data in the table in the same order as it is in Excel?
    Just to explain my question better,
    If excel has three rows  {a,b},{c,d},{e,f}, does SSIS ensure it loads the data in the same order in a table.
    Thanks.

  • Cannot load the data to infocube

    Hi,
    I am new to SAP BI.
    I am loading the data from my flat file to infocube. Flat file contains product, qty, amount and units of Qty is EA and Amount is INR and the units or fixed.  i need to have a Revenue field in the infocube and to populate that revenue field i am doing qty*amount in transformations using formula in transformations. My Revenue field unit is fixed currency INR. While loading data to infocube i can see error in DTP Monitor data cannot be loaded to cube. In DTP Monitor until Transformations it is showing in green and while loading to cube it is showing red color and making the whole request as red. But wthout this revenue field i can load the data to infocube correctly.
    Can any one please suggest me the things what i was doing wrong.

    Hi,
    Can you please check the error details by checking of the details for the failed step?
    This can help you to understand the exact casue of error.
    Regards,
    Geetanjali

  • Taking the data from interactive forms and load the data into SAP system?

    hi all,
    I want to know how to take the data from interactive forms and load the data into sap system?
    if u have any sample scenario, explain with that.
    thanks in advance
    Raja

    Hello,
    Check the program...
    SAPBC480_DEMO.
    Check the below threads
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/88e7ea34-0501-0010-95b0-ed14cfbeb85a
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/bfbcd790-0201-0010-679d-e36a3c6b89fa
    Thanks
    Seshu

  • AFTER LOADING THE DATA INTO THE INFOCUBE IT SHOULD NOT DISPLAY IN REPORTING

    Hi experts,
                     I have the data in infocube. When i selected the cube in reporting it should not display i.e; the data should not display in reporting.
    thanks nd regards,
    venkat

    Bit strange requirement. But if you have data in Cube and automatically its will be available for Reporting. unlike ODS, we do not have option of "Automatic request activation".
    However on repoeting side, you can restrict the values to #  and exclude everything from slection. But not good solution as such.
    Regards
    Pankaj

  • Stage tab delimited CSV file and load the data into a different table

    Hi,
    I pretty new to writing PL/SQL packages.
    We are using Application express for our development. We get CSV files which is stored as a BLOB content in a table. I need to write a trigger that would get executed once the user the uploads the file and parse thru the Blob content and upload or stage the data in a different table.
    I would like to see if there is any tutorial or article that could explain the above process with the example or sample code to do the same. Any help in this regard will be highly appreciated.

    Hi,
    This is slightly unusual but at the same time easy to solve. You can read through a blob using the dbms_lob package, which is one of the Oracle supplied packages. This is presumably the bit you are missing, as once you know how you read a lob the rest is programming 101.
    Alternatively, you could write the lob out to a file on the server using another built in package called utl_file. This file can be parsed using an appropriately defined external table. External tables are the easiest way of reading data from flat files, including csv.
    I say unusual because why are you loading a csv file into a blob? A clob is almost understandable but if you can load into a column in a table why not skip this bit and just load the data as it comes in straight into the right table?
    All of what I have described is documented functionality, assuming you are on 9i or greater. But you didn't provide a version so I can't provide a link to the documentation ;)
    HTH
    Chris

  • Problem while loading the data

    Hi Gurus,
    I am facing aproblem in PSA.Acctually the problem is that when we are  loading the data from R/3 to BI side sold to party errors were occured.I edited that error records and manual update done.But in cube manage i find that trasfore records are only 6 which are edited in PSA and added recordrs are 2651.But in that datapacket total records are 3061.
    My cube was rollup also.
    pls give my solution.I'll assigne points.
    Regards
    Roy

    Hi Raj,
    No need to match  datapacket total records to your target.Because depend on the design records are going to change.
    Thnx,
    Assign points if useful.

  • Idoc issue while loading the data into BI

    Hello Gurus,
    Initially i am having the Source System connection problem. After it was fixed by the basis i had followed the below process.
    I am loading the data using the Generic extractor specifying the selection and its a full load. When i load the data from R3 PP application  i found the below result in the monitor screen of BW.
    1. Job was completed in the R3 system and 1 million records are fetched by extractor.
    2. the records are not posted to the BW side Because of TRFC issue.
    3. i got the idoc number and process it in the BD87 tCode. But it was not process sucessfully it gives the following error " Error when saving BOM. Please see the application log."
    4. when i check the Application Log using the Tcode SLG1 with the help of time and date of that particular process it was in the yelow status.
    Kindly let me know i can i resolve this issue. i have already tried by repeating the infopackage job still i am facing the same issue. I have also check the connection its is ok.
    Regards

    hello veerendra,
    Thanks for your quick response. yes i am able to process it manually. after processing  it was ended with the status 51 application not posted.
    could you pls help me out with the same
    regard
    Edited by: KK on Nov 4, 2009 2:19 AM
    Edited by: KK on Nov 4, 2009 2:28 AM

  • Date problem while loading the data

    Hi Guys,
    When we are extraction the data from the DataSource  0CRM_SALES_ACT_1, it is giving an error for 0calday.
    If the date is greater than 2012, it is giving an error. i.e., it is taking date upto 31.12.2012. If the date is greater than that date is giving an error as "date is later than factory calendar definition".
    In that case we are modifying the date in the PSA and make date less than 2013 and loading the data.
    Even we have done Transfer Global setting at the source system and we checked the factory calendar in SCAL tcode it is setup to 2098.
    Can you please advice how to solve the problem when we are loading the date which is greater than 31.12.2012.
    Thanks in advance
    Jagan

    Hi Jagan,
    In the info package selections tab.
    for 0CALDAY field give to value as 31.12.9999 and try loading you will get all the records
    Regards,
    Venkatesh

  • Unable to load the data into HFM

    Hello,
    We created new HFM app configured that with FDM, generated output file through FDM and loaded that file through HFM directly 5-6 times, there was no issue till here.
    Then I loaded the file through FDM 4 times successfully, even for different months. But, after 4 loads I start getting Error. Attached is the error log .
    Please help us earliest..
    ** Begin fdmFM11XG6A Runtime Error Log Entry [2013-10-30-13:44:26] **
    Error:
    Code............-2147217873
    Description.....System.Runtime.InteropServices.COMException (0x80040E2F): Exception from HRESULT: 0x80040E2F
       at HSVCDATALOADLib.HsvcDataLoadClass.Load(String bstrClientFilename, String bstrClientLogFileName)
       at fdmFM11XG6A.clsFMAdapter.fDBLoad(String strLoadFile, String strErrFile, String& strDelimiter, Int16& intMethod, Boolean& blnAccumFile, Boolean& blnHasShare, Int16& intMode)
    Procedure.......clsHPDataManipulation.fDBLoad
    Component.......E:\Opt\Shared\Apps\Hyperion\Install\Oracle\Middleware\EPMSystem11R1\products\FinancialDataQuality\SharedComponents\FM11X-G6-A_1016\AdapterComponents\fdmFM11XG6A\fdmFM11XG6A.dll
    Version.........1116
    Identification:
    User............fdmadmin
    Computer Name...EMSHALGADHYFD02
    FINANCIAL MANAGEMENT Connection:
    App Name........
    Cluster Name....
    Domain............
    Connect Status.... Connection Open
    Thanks,'
    Raam

    We are working with the DB team but they have confirmed that they is no issue with the TB, the process we have followed
    As a standard process – while loading the data from FDM or manually to HFM – we don’t write any SQL query. Using the web interface – data would be loaded to HFM application. This data can we viewed by different reporting tools (smart view(excel)/HFR Report/etc.)
    There is no any official documents on oracle website which talk about Insert SQL query which is used to insert data to HFM tables. Even, Hyperion does not provide much details on its internal tables used. Hyperion does not provide much insight on internal structure of HFM system.
    As per Hyperion blogs/forums on internet –HFM stores the base level data in so called DCE tables (for example EMHFMFinal _DCE_1_2013 where EMHFMFinal  is application name, 1 identifies the Scenario and 2013 the Year).  Each row in the DCE table contains data for all periods of a given combination of dimensions (also called an intersection).
    We are trying to load same data file with a replace option( it should delete the existing data before loading the data file).

  • Unable to load the data into Cube Using DTP in the quality system

    Hi,
    I am unable to load the data from PSA to Cube using DTP in the quality system for the first time
    I am getting the error like" Data package processing terminated" and "Source TRCS 2LIS_17_NOTIF is not allowed".
    Please suggest .
    Thanks,
    Satyaprasad

    Hi,
    Some Infoobjects are missing while collecting the transport.
    I collected those objects and transported ,now its working fine.
    Many Thanks to all
    Regards,
    Satyaprasad

  • Error loading the data into ODS - Message no. BRAIN060?

    Hi,
    I am getting following error while load the data from flat file, data loaded successfully from flat file to PSA but I got following error while updating the data from PSA to data target:
    Value '010384 javablue' (hex. '30003100300033003800340020006A0061007600610062006C') of characteristic 0PO_NUMBER contains invalid characters
    Message no. BRAIN060
    Diagnosis
    The following standard characters are valid in characteristic values as default:
    !"%&''()*+,-./:;<=>?_0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ
    Characteristic values are not allowed if they only consist of the character "#" or begin with "!". If the characteristic is compounded, this also applies to each partial key.
    You are trying to load the invalid characteristic value 1. (hexidecimal representation 30003100300033003800340020006A0061007600610062006C).
    I am trying to load Value '010384 javablue' to "0PO_NUMBER" info object for ODS with in one row with some other data.
    Any idea or any input to resolve this issue?
    Thanks in advance for any input.
    Steve

    Thanks Soumya, I have maintained the upper case letters, but I am loading upper and lower case mixed to PO number? and it is not working. What is the solution to this? If I set lower case property to the PO number infoobject, it won't accept upper case. If I uncheck the lower case then it won't accept lower case letters. I cannot add upper and lower case letters in RSKC, because it accepts up to 72 characters, I have already have more than 60 characters (special char, numbers and 26 upper case letters).
    I have already tried transfer routine but you can either convert to lower or upper but it doesn't work. we need both uper and lower case for po number, and R/3 is accepting it. why BW doesn't accept both?
    Any idea what can be done?
    Thanks in advance for your help.
    Steve

  • Error loading the data into cube

    Hi All,
    wen i am loading the data in the ASO Cube its giving warning but i did every thing correct yester day it was loaded correctly buttoday its not loading its giving warning ASO application ignore update to derived cells even i clear the data wats this error
    so plz can any one help
    Reading Rule SQL Information For Database accdb
    Reading Rules From Rule Object For Database accdb
    Parallel dataload enabled: [1] block prepare threads, [1] block write threads.
    Aggregate storage applications ignore update to derived cells. 29 cells skipped
    Data Load Elapsed Time with http://dataldcu.rul : http://0.094 seconds
    Database import completed
    Output columns prepared: [0]
    Thanks
    Ram

    In ASO cubes you can only load data at level 0 (=leaf level). Every upload to aggregated dimension members are ignored. The warning message "ignore update to derived cells" is meaning that you are trying to load data at node level. As this is not possible with ASO cubes, essbase ignores that data.

Maybe you are looking for

  • Problem in performing multiple Point-In-Time Database Recovery using RMAN

    Hello Experts, I am getting an error while performing database point in time recovery multiple times using RMAN. Details are as follows :- Environment: Oracle 11g, ASM, Database DiskGroups : DG_DATA (Data files), DG_ARCH(Archive logs), DG_REDO(Redo l

  • PDF Files in Reader Saving As Instead of Saving

    Hello, I am running Adobe Reader XI. My colleagues are all running other PDF editing software (I don't get the PDF editing software since "I don't need it.") When I open a PDF document that is on our shared server, I will sign the document on Reader

  • Color management with HP officejet 6500 E709 serie

    Can you pls tell me what my color management settings should be set to on mty HP officejet 6500 E 709 series? It is connected to an HP Elitebook 6930p.  All cartridges are new but none of the "red/pink" colors print.  Thank you.

  • How to attach files from "Office" to E Mail

    Hi, I am using a Nokia Lumia 720. and I have an outlook account, which I use with regularity. I have faced a problem which is still not resolved - how to attach files in the mail. 1. You open Outlook Mail or any mailbox that you are using and then yo

  • FS-CD Dunning

    Hi FS-CD gurus, After running VPVA transaction the dunning levels are created in the FKKMAKO table. But the print dates are not updated. Due to this VYM10(Dunning history) transaction is not showing this dunning level in the output. Please advise on