Data Not coming From DSO to Cube

Hi All,
I have data in DSO. When i run the DTP , request status is green.
Now when i go to cube and see the data the request has came but data has not transferred .....
The Request Status is green in cube but data is not cuming....
No transferred data.
How can bring data in cube..??
Regards,
Rohit

Hi,
When you loaded the DSO......after that are you loading the cube for the first time?If so .....and still data is not coming..then check if the last loaded request somehow got deleted from the DSO....
If after loading the DSO....you have already loaded the cube...and this is the second time you are loading the cube.........then obviously it will pick 0 records...........since delta DTP always picks the fresh new request..........if there is no new request in the source .......then it will pick 0 records.......
Also both the cased are not working.........then check the End routine or Start routine.if any.......
Thanks and Regards,
Debjani.......

Similar Messages

  • Delta records not updating from DSO to CUBE in BI 7

    Hi Experts,
    Delta records not updating from DSO to CUBE
    in DSO keyfigure value showing '0' but in CUBE same record showing '-I '
    I cheked in Change log table in DSO its have 5 records
    ODSR_4LKIX7QHZX0VQB9IDR9MVQ65M -  -1
    ODSR_4LKIX7QHZX0VQB9IDR9MVQ65M -   0
    ODSR_4LIF02ZV32F1M85DXHUCSH0DL -   0
    ODSR_4LIF02ZV32F1M85DXHUCSH0DL -   1
    ODSR_4LH8CXKUJPW2JDS0LC775N4MH -   0
    but active data table have one record - 0
    how to corrcct the delta load??
    Regards,
    Jai

    Hi,
    I think initially the value was 0 (ODSR_4LH8CXKUJPW2JDS0LC775N4MH - 0, new image in changelog) and this got loaded to the cube.
    Then the value got changed to 1 (ODSR_4LIF02ZV32F1M85DXHUCSH0DL - 0, before image & ODSR_4LIF02ZV32F1M85DXHUCSH0DL - 1, after image). Now this record updates the cube with value 1. The cube has 2 records, one with 0 value and the other with 1.
    The value got changed again to 0 (ODSR_4LKIX7QHZX0VQB9IDR9MVQ65M - (-1), before image &
    ODSR_4LKIX7QHZX0VQB9IDR9MVQ65M - 0, after image). Now these records get aggregated and update the cube with (-1).
    The cube has 3 records, with 0, 1 and -1 values....the effective total is 0 which is correct.
    Is this not what you see in the cube? were the earlier req deleted from the cube?

  • Data is not updating from DSO to CUBE

    Hi Friends,
    Can any one please Guide me.
    Data is updating upto ODS but not from this lower level ODS to top of the  cube from last  week and displaying error as " Exception condition "DATE_INVALID" raised" .
    How i can proceed to find which date field is getting wrong date or not getting date ..Can any body tell that how we can analize the possibilities in this case to find the error and solution .I need to solve this today,Can any body guide me please.
    BR,
    Venkat

    hi
    check the transformations between dso and cube, and check any inconsistence. if not then follow the link
    http://www.consolut.com/en/s/sap-ides-access/d/s/doc/F-DATE_CONVERT_TO_FACTORYDATE
    this may help you.

  • Data is not uploaded from dso to cube

    Dear Experts,
    In one of my process chains uploading of the data is taking place from dso to the cube through DTP.
    But only data in data package 1 are transfeered .
    The data gets stuck up in data package 2 and data package 3 so the request remains in yellow state only.
    Please suggest.

    Dear All,
    As per ur issue, try to check the following,
    1) Check the short dump is there are not in SM12.
    2)If found then try to load manually  the packet2 and 3 , (we have got an option on the right most corner to load them manually).
    3) if 1 & 2 are unsuccessful, try to keep an option of PSA and load the data again.
    kindly let me know if you have any issues.
    Thanks and Kind Regards,
    Saisandeep.

  • Data Extraction failure from DSO to Cube

    Hello Experts,
    The Data load to Cube( zfiar_c03) from DSO (zfiar_03) using a full
    update is failing throwing the error message"Incomplete update due to
    errors in single records ".We had deleted the failed request from
    infocube and loaded the the same req from psa by correcting or removing
    the # values but it failed again resulting in the same error.
    When we checked the previous successful requests which got loaded to
    cube from dso it also contains the # values , but the loading was
    successful.
    our BW reporting is seriously impacted due to the failure coz our
    collection,recievables..etc all business critical report is based on
    this cube.
    Request you all to help..
    thanks,
    Sapta

    Hi Sapta Jyothi,
                     I am not sure what the actual problem is. But, from my prior experience, I fell this is related to the Process chain configuration provided you are using process chains.
                     If  you are using a process chain , please check if there is any other failed step in the Process chain. If that is the case, even if the failure due to invalid entries is fixed the QM status of the load to the cube will not turn Green till the process chain status turns green or yellow. You can check from the details tab of the info-package load run and see if the process chain log step is red. If this is the case, then you will have to take action to see that other steps are completed succesfully. And also the Process chain in this case needs to be modelled in a way, such that, when a load to a cube fails the successive steps like creation of cube indeces should not be triggered.
    Thanks ,
    Nithin.

  • Data not coming from DOE to Mobile After defining Rule for device attribute

    Hi All,
    I have created a DO and rule for it.In case of Bulk Rule for all definition when i triggere extract from Portal then all the data comes to outbound queue but when i define rule for Device attribute then no data comes to my Outboun queue.Here is the scenario what i am doing :
    1. I have order header in my backend which has a field named "Work_Center" and this will be criteria field.
    2. In CDS table i have all the records for all the work center.
    3. Now in RMM under customized , i have added an attribute named "Work_center".
    4. Now i defined a rule with Device attribute mapping and activated the rule.
    5. Now on Portal i assigned this data object and in the device attribute tab i assigned the value(this value exist in CDS table for few orders) of a   Work center to the attribute "Work_Center" .
    6. Then i triggrere extract but its Outbound queue is empty, what could be the reason.
    Is my approach is correct
    Regards,
    Abhishek

    Hi Abhishek,
    You can check one ore thing, after you have performed all the steps till step 5, i.e. just before triggering
    extract. Check if the AT table for ur DO has entries based on the criteria specified by you...
    1. In the workbench click on the Data Object, and then right click and select "View Metadata".
    2. Select Distribution Model tab.
    3. Now select your DO's Association table.
    4. For the input field DEVICE ID specify your corresponing device id,and also for status field specify it 
        as "I"  and execute
    If there are any entries now in the AT table, and on triggering extract if they are not coming to the
    outbound Q there is some EXTRACT Q blocked. And is there were no entries in the AT then the rule
    specified is not  the satifying.
    Thanks,
    Swarna
    Now if you have entries w

  • Header data not coming from vendor master when creating SA using BAPI

    Hello,
    I am creating SA using BAPI_SA_CREATE. I am passing the vendor code and other header related data in the header table of the BAPI.
    SA is created successfully with all the data that is passed to the BAPI. Now if the payment term and inco terms data is passed initial to BAPI then this data is not taken from vendor master. This data appears blank in SA.
    Shouldn't this data be picked up from vendor master as they are maintained properly in vendor master?
    I just need confirmation whether BAPI picks or not from vendor master if the data (payment term and inco terms) passed is initial.
    Thanks and Regards,
    Saba

    Hi,
    Are u passing X to corresponding fields of HEADER structure what ever you have populated in HEADER structure .

  • Added new field to cube but data not passed from DSO

    Hope someone can help.
    (BI 7.0) We added new fields into a cube. The fields already existed in the DSO. When we ran the process chain in development for the first time after making this change, we notice that the 'historical' data for these fields is populated in the cube. When we do this in our quality system only new data passed to the cube is updated. In development in the sub-chain DTP request we see all previous requests listed under selections. In quality it is only the latest request. The only difference is that the DeltaInit flag in the DTP request in development is ticked (extraction mode) - but not in quality. Does anyone know whyb this is?

    hi peter,
    Adding fields to cube doesn't affect delta status...The delta DTP should be able to handle delta requests automatically.
    I guess in you quality system, the cube already got all requests updated from ODS before you importing the change request. And in develop system none of the requests in ODS were updated before the change.
    Regards,
    Frank

  • Change Log Data from DSO to CUBE

    Hi Experts,
      Is it possible to load/transfer change log data as well from DSO to Cube
    Thanks

    Hi,
    Yes it is possible.
    Create a DTP from DSO to cube.
    In the extraction tab of DTP choose the option
    Datasource Change log
    Hope it Helps!!
    Reward if useful
    Manish

  • Delta records are not loading from DSO to info cube

    My query is about delta loading from DSO to info cube. (Filter used in selection)
    Delta records are not loading from DSO to Info cube. I have tried all options available in DTP but no luck.
    Selected "Change log" and "Get one request only" and run the DTP, but 0 records got updated in info cube
    Selected "Change log" and "Get all new data request by request", but again 0 records got updated
    Selected "Change log" and "Only get the delta once", in that case all delta records loaded to info cube as it was in DSO and  gave error message "Lock Table Overflow" .
    When I run full load using same filter, data is loading from DSO to info cube.
    Can anyone please help me on this to get delta records from DSO to info cube?
    Thanks,
    Shamma

    Data is loading in case of full load with the same filter, so I don't think filter is an issue.
    When I follow below sequence, I get lock table overflow error;
    1. Full load with active table with or without archive
    2. Then with the same setting if I run init, the final status remains yellow and when I change the status to green manually, it gives lock table overflow error.
    When I chnage the settings of DTP to init run;
    1. Select change log and get only one request, and run the init, It is successfully completed with green status
    2. But when I run the same DTP for delta records, it does not load any data.
    Please help me to resolve this issue.

  • Data load from DSO to Cube in BI7?

    Hi All,
    We just migrated a dataflow from 3.5 to 7 in development and moved to production. So till now in production, the dataloads happend using the infopackages.
    a. Infopackage1 from datasource to ODS and
    b. Infopackage2 from ODS to the CUBE.
    Now after we transported the migrated dataflow to production, to load the same infoproviders I use
    1. Infopackage to load PSA.
    2. DTP1 to load from PSA to DSO.
    3. DTP2 to load from DSO to CUBE.
    step1 and step2 works fine but when I run the DTP2 it is getting terminated. But now when I tried the step b (above), it loads the CUBE fine using the infopackage. So I am unable to understand why the DTP failed and why the infopackage load is successful. In order to use the DTP do we need to do any cleanup when using it for first time? Please let me know if you have any suggestions.
    Please note that the DSO already has data loaded using infopackage.  (Is this causing the problem?)
    Thanks,
    Sirish.

    Hi Naveen,
    Thanks for the Reply. The creation of DTP is not possible without a transformation.
    The transformation has been moved to production successfully.

  • Problem while  data processing TRANSACTION data from DSO to CUBE

    Hi Guru's,
    we are facing problem while  data processing TRANSACTION data from DSO to CUBE. data packets  processing very slowly  and updating .Please help me regarding this.
    Thanks and regards,
    Sridhar

    Hi,
    I will suggest you to check a few places where you can see the status
    1) SM37 job log (give BIrequest name) and it should give you the details about the request. If its active make sure that the job log is getting updated at frequent intervals.
    2) SM66 get the job details (server name PID etc from SM37) and see in SM66 if the job is running or not. See if its accessing/updating some tables or is not doing anything at all.
    If its running and if you are able to see it active in SM66 you can wait for some time to let it finish.
    3) RSMO see what is available in details tab. It may be in update rules.
    4) ST22 check if any short dump has occured.
    You can also try SM50 / SM51 to see what is happening in the system level like reading/inserting tables etc.
    If you feel its active and running you can verify by checking if the number of records has increased in the cube.
    Thanks,
    JituK

  • Error in loading data from DSO to Cube

    Hi
    Daily process chains are executed and the data is loaded from DSO to several cubes.
    On 7.11.2008, there was an error in one of the cube. When checked in the short dump, the following is the dump msg:
    DBIF_REPO_PART_NOT_FOUND. I looked into the note 854261 and had a chat with our basis guy, he said our kernal and patch levels are more then that mentioned in the note.
    Strange thing is, from 8.11.2008 to till date, the loads are running fine but the problem was only on that date.
    So now i have several request in the cube with loads of data but "WITHOUT" reporting. Because of error on 7.11.2008.
    Can anyone help in solving the issue. I want to re-execute the load again and see if the problem occurs again or not.
    Regards
    Annie

    Hi,
       Check the reconstruction tab. If the request is green, you could try deleteing and reconstructing the incorrect request.
    Regards.

  • Error while loading data from DSO to Cube

    Dear Experts,
    I have change update routines of the update rule for DSO & update rule between DSO & cube. when i uploading data in DSO , working fine. But when uploading from DSO to cube , it giving error.
    Error Message :
    DataSource 8ZDSO_ABB does not have the same status as the source system in the Business Information Warehouse.
    The time stamp in the source system is 30.09.2010 14:05:17.
    The time stamp in the BW system is 09.05.2008 16:41:10.
    I executed program RS_TRANSTRU_ACTIVATE_ALL , still giving same error.
    Regards,
    Anand Mehrotra.

    Dear Experts,
    Thanks for help.Problem solved.
    Regards,
    Anand Mehrotra.

  • Semantic Partitioning Delta issue while load data from DSO to Cube -BW 7.3

    Hi All,
    We have created the semantic partion with the help of BADI to perform Partitions. Everthing looks good,
    first time i have loaded the data it is full update.
    Second time i init the data load. I pulled the delta  from DSO to Cube . DSO is standard where as Cube is partition with the help of semantic partition. What i can see is the records which are updated in Latest delta are shown into the report rest all are ignored by the system.
    I tried compression but still it did not worked.
    Can some has face this kind
    Thanks

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

Maybe you are looking for