Data Load with 0 records

Hi,
How a system should react under following cases:
1. Full load bringing in 0 records
2. Init load bringing in 0 records
3. Delta load bringing in 0 records
Note: here by 0 records I mean actually the load has no records.
For each of the above case will the load turn green or remain yellow and then times out.
I always have different reactions from the system for these cases.  Appreciate view from experts…
Thank you,
sam

Jr roberto setting which you said exists true,
but i have that setting marked as green.
i did an init load which pulled in 0 records. this is correct. now eventhough the green is checked for 0 records in rsmo settings the load errored out after the time out setting in infopack
and the main traffic light is still running...with  No errors could be found. The current process has probably not finished yet.
any tips..

Similar Messages

  • Tracking Data Changes with Record History (WHO)

    Hi all ,
    I've a custom form & i need to Tracking Data Changes with Record History (WHO) will anyone show me example of how it's done because i tried to go through the application developer guide but unfortunatelly this part isnot tht clear with me
    waiting your reply
    thanks

    Hi,
    Good question, a colleague asked me the same yesterday - i will find out officially for you.
    Regards,
    H

  • Master Data Load - Number records not match

    I am loading the 0PLANT table.  In R3, it shows 966 records.   After running the InfoPackage, there are 966 records in the PSA, none in error.  But I get a message - Number of records  requested does not match number transferred (packet 1).
    This is a first use of a new source system to this BW.  All master data loads give me the same error.
    What is setup wrong?

    Hello RoberT,
    & How r u ?
    For the same 0PLANT we load data on daily basis. Not only for this but also for more Master Data Objects. For some of them we used to get this message but not as an Error. On pressing the Update Tree (Refresh) it will change the status.
    Anyway u got the records in the PSA so there is no problem with the data load. What signal u find in the Monitor ? Yellow or Red ? If yellow wait for some time, press Update Tree.
    Best Regards....
    Sankar Kumar
    +91 98403 47141

  • Material data load with two unit (0MAT_UNIT) failed?

    Hi,
    I am loading the 0MAT_UNIT data from R/3 QAS to BW QA, I am getting below error.
    0MAT_UNIT : Data record 34 ('000000000000010249 ') : Duplicate data record RSDMD 191
    There two records for same material number with two different units (EA and PAK), Data loaded to BW dev successfully even if duplicate records for material with two units but it failed in BW QA.
    Any setting I am missing?
    Please help.
    Thanks,
    Steve

    If you look at the definition of the P table (/BI0/PMAT_UNIT) you will find that both material and unit are keys. So the system recognizes a record as duplicate only when a given material with a given unit exists twice in the data. Based on what you are saying, this is not the case. You have two separate units for a given material. Can you please check to see if there is another record for the same material in a different datapacket ??

  • Data loading with 1:N & M:N relationship

    Hi,
    Which data (master or transaction) we can load with 1 : N relationship and M : N relationship.
    Can anyone give me the idea.
    Thx

    In case of master data,the char infoobjects with master data attributes take care of 1:N relation.For e.g for material master data,the material number would be the key based on which the data would be loaded for each new material.In case of transaction data for 1:N relation,you may take up DSO in which the primary keys can be made the key fields of DSO based on which the repeating combinations would be overwritten.You may use an infoset when there is a requirement of inner join(intersection) or left outer join between master data/DSO objects. The infocube can be used for M:N relation.For union operation betwen two infoproviders,multiprovider may be used.
    Hope this helps you out.
    Regards,
    Rinal

  • If data loading fails records should ROLLBACK and only error records .....

    Hi All,
    In my process data loading from source to target,for example iam loading 100 records ,in that 50th record having error,once the error record occured total loading process want to fail, records needs to rollback and error record need to move error table....any one can you guide on this process...
    Thanks in advance....

    30021986 wrote:
    Hi All,
    In my process data loading from source to target,for example iam loading 100 records ,in that 50th record having error,once the error record occured total loading process want to fail, records needs to rollback and error record need to move error table....any one can you guide on this process...
    Thanks in advance....Hi ,
    Test this ..
    Set flow control to YES
    Set Maximum number of errors allowed to 0 (zero)
    Run the interface

  • Full Data Loads with a DSO as source

    I am using a DSO as a source object running a DTP full load. When I look at the SQL running it is performing a select statement grouping all of the characteristics and summing the key figures.
    Is there any way of changing the extract so that it extracts the source data on a record by record basis (as it does in 3.x)?

    50M records is quite large.  Can you break down the data into smaller chunks ... I personally would make the chunks as close to about 3-5M as I could. You may be able to increase this amount if you are not looking up data or have complex update rules -- if everything is mapped 1:1 (no transformations) between the two DSO's, then you can increase the size a lot (maybe 10M).
    Unless you have a large amount of memory, you will not be able to load large amounts of data like this in one shot.
    Back in the 2.x days, it wasn't recommend to load more than a Million records into a data target, so feel lucky that they increased this limit a lot!
    Brian

  • In EIS at data loading Few Records are getting rejected

    Hello,
    I m loading the data in Essbase through EIS.I am using user defined SQL for data loading.
    In log this this is showing that 22000 records are loaded and 564 records rejected.
    Can any Expert let me know why these records are getting rejected.
    Thx
    Shivendra

    If you look on your EIS server, under hyperion\eis there is a logfiles directory. Under it are folders for every load. In these folders are the error files. If you look at the file, the first column shows you the member that got rejected. You can then look to see if it is in your outline. My guess is your custom sql is bringing back rows that are not in the cube

  • Data Load with warnings, but dataload.err empty

    Hi guys,
    laoding data on EAS Console for a Essbase database gives below message:
    Database has been cleared before loading.
    dataload.err is EMPTY.
    Any ideas??
    Some data is loaded but how can I find out where it stopped or is all my data loaded?
    Message:
    Parallel dataload enabled: [1] block prepare threads, [1] block write threads.
    Transaction [ 0x190004( 0x4fd89600.0x64578 ) ] aborted due to status [1003050].
    There were errors, look in H:\My Documents\temp\dataload.err
    Database import completed ['AXN_Sim'.'BewMod']
    Output columns prepared: [0]
    Thanks Bernd

    Hi,
    There is nothing much you can say whether the data is completely loaded or not. You have to check by yourself on that.
    Since the loading was aborted with a warning, it could be a network issue..
    Try to load a small file and then check whether you are facing the problem or you coulfd also post your finding on that..
    Thanks,
    CM

  • Data Load with HAL Problem with Decimal-delimiter/Language Setting

    Hi,
    we use HAL to load Data into Essbase. The Data is with a . as Decimal Point like in
    1252.25 When i upload these Data with HAL the number becomes 125225.00
    The Locale of my OS is German
    The Locale i specified for the file in HAL is English
    If i change the Locale of my BS to English the Problem disappears but thats annoying.
    Has anybody else such a problem ?
    Is There a Solution for this ?
    Thanks.
    Kevin

    Reading over John's blog, we created a rule file.
    But, it looks the load process using the rule file seems to be hanging? are we missing any thing? We are using comman seperator and having number of lines to skip as zero.

  • Master Data load duplicated records - 0ART_SALES_ATTR

    Hi All,
    I'm trying to load Master Data in the Test System and I'm getting an error that says that There are duplicates of the data record 70114 with the key '10 30 000000000001034170 ' for characteristic 0MAT_SALES .
    This is wrong, no data is duplicated, I already check the PSA and it's fine.
    The dataflow is the 3.5.
    The Infopackage has been set up with Update Initialization with Data Transfer.
    The PSA and the Infoobjects are empty, they have never been loaded.
    The Processing in the Infopackage has been set as PSA and Infoobject (Package by package).
    I can't modify the Infopackage as I am in a Test System.
    Any Idea on how to solve this?

    The problem was that the person who designed the dataflow made a mistake and swapeed two fields.

  • Time Data loading with missing format

    Hi expert,
    after we load time data, find the following missing information:
    0CALDAY:
    20.080.112 (Error)
    2008.01.12 (correct one)
    0CALMONTH:
    200.801(Error)
    2008.01(correct one)
    0CALQUARTE
    20.081(Error)
    20081(correct one)
    Could anyone let us know how to correct data with missing format to correct one and how to do this step by step? Many thank!

    What is the source for this data ?
    Have it corrected in the source
    Or
    Correct in PSA and load it from there.
    Does it happen only for a few records ?

  • Data load with ABAP control

    Hi:
    I need to load the data to a Cube with ABAP code in info-Package.... but am a Novice in ABAP..!
    Need your help.
    I need to load the data to the Cube only the records thta have "NETDUEDATE" gretaed than System Date
    (SY-DATUM) and drop the records that do not meet the condition. 
    Not sure what would be the code / syntax for putting this condition!
    Appreciate your help.
    Thanks....Pbs

    Hi PBS,
    You can write a ABAP logic for your requirement. In the data selction tab of the Info-package, give the type as abap routine and the variable in the type. You can implement the logic with IF-ENDIF condition.
    Reffer:
    ABAP routine in Infopackage
    ABAP Routine in InfoPackage.
    Regards,
    Satyam

  • Data loading with routine displays zero for key fig values

    Hi all,
    my source field is amount...........and target field is research amount...........
    if i restrict amount(source field) with cost element(coaeom7) and value type(010 then it is equal to target field research amount.
    for this my code is
    IF COMM_STRUCTURE-VTYPE = '010' AND
       COMM_STRUCTURE-COSTELMNT = 'COAEOM12'.
       RESULT = COMM_STRUCTURE-AMOUNT.
    ENDIF.
    but when i load the data it displays only zeros.........
    Pleas suggest
    Regards,
    Raj.

    Hi Raj,
    Do u need costelement values other than 'COAEOM12' into the target.
    Are you writing this routine in Start Routine/End Routine. if its BI.7
    If you are loading data from soucre cube/DSO to target structure, in the transformation you need to write start routine, before which you need to map amount field, cost element,and value type from source to Research amount in target,
    The you need to write the code
    If source_packake cost element = COAEOM12'  and source_packake value type =010
    then research amount = amount
    End If.
    i hope this will solve your problem.

  • Error in Data loading with ODS

    Hi Guru's
                 I'm loading the data from the Flat file  to the ODS
    the key figure net masss volume  with the data type quantity show s the value multiply thousand but it in the psa  it shows correct value what is there in the source system
    for eg:   value for this key figure  48.984 in the flat file means 48.(point) 984
    but it is showing in the ods 48,984(comma) so some where it is multiplying by 1000 like it is applying  coumn level
    and there is no transfer rules  it is direct mapping
    there is no routines
    please let me know where is the problem
    Points will be assign

    Hi Sanjay,
    It is not multiplied by 1000. The settings are like that. The decimal seperator used here is ',' instead of '.' .
    Steps to change the settings:
    Goto Menu-> system->user Profile->own data->in the following screen of user profile maintanance, goto tab Defaults. There you can find a drop down for Decimal notations
    From there you can change the settings. You can also use the T.code SU01.
      After changing the settings logoff and then login for the changes to take place.
      Hope this helps you.
    Regards,
    Yokesh.

Maybe you are looking for

  • Services in Ozark MO?

    I'm wondering if there will be services in Ozark MO. We had it when we lived in Springfield but now live in Ozark and would really love to have UVerse in Ozark.

  • How do I sign documents from my E-mail

    I would like to sign documents using electronic signiture, How would I go about doing that ?

  • How can I get multirate scanning to work?

    I'm working with a PCI-6025E DAQ. I need to read a channel connect to a tension meter for a certain amout of time. I'm trying to use multi-rate scanning as described in the documentation. I am recieving a -10003 code when calling SCAN_Start and I don

  • Due Date for Emails

    Hi Experts, System: SAP CRM 2007 (WEB UI) I am trying to set-up ERMS Due date for incoming emails. I am trying out the Standard funtionality by using the two services FG_ESCALTIME UT_ESCALDETERM Can any one help me with the Properties that I should s

  • Weird error with AFP shares

    Hi everyone, I am using an OS X 10.8.5 Server with mostly 10.6.8 Clients. The clients usually open, save, copy to and from AFP shares provided by the server. Sometimes it happens that a client gets thrown out of a folder it is currently viewing on th