Error in Data loading with ODS

Hi Guru's
             I'm loading the data from the Flat file  to the ODS
the key figure net masss volume  with the data type quantity show s the value multiply thousand but it in the psa  it shows correct value what is there in the source system
for eg:   value for this key figure  48.984 in the flat file means 48.(point) 984
but it is showing in the ods 48,984(comma) so some where it is multiplying by 1000 like it is applying  coumn level
and there is no transfer rules  it is direct mapping
there is no routines
please let me know where is the problem
Points will be assign

Hi Sanjay,
It is not multiplied by 1000. The settings are like that. The decimal seperator used here is ',' instead of '.' .
Steps to change the settings:
Goto Menu-> system->user Profile->own data->in the following screen of user profile maintanance, goto tab Defaults. There you can find a drop down for Decimal notations
From there you can change the settings. You can also use the T.code SU01.
  After changing the settings logoff and then login for the changes to take place.
  Hope this helps you.
Regards,
Yokesh.

Similar Messages

  • Essbase Error(1003050): Data Load Transaction Aborted With Error (1220000)

    Hi
    We are using 10.1.3.6, and got the following error for just one of the thousands of transactions last night.
    cannot end dataload. Essbase Error(1003050): Data Load Transaction Aborted With Error (1220000)
    The data seems to have loaded, regardless of the error. Should we be concerned, and does this suggest something is not right somewhere?
    Your assistance is appreciated.
    Cheers

    Hi John
    Not using a load rule.
    There were two other records which rejected based on absentee members. The error message was different for them, and easily fixed.
    But this error doesn't tell us much. I will monitor the next run to see if the problem persists.
    Thanks

  • Common errors in Data Loading

    Can anybody tell me what are the common error we come across in BW/BI for following:-
    1] Data Loading in Infocube/ODS/DSO
    2] Aggregates
    3] PSA
    4] DTP
    5] Transformations
    Thanks in advance

    Hi,
    Here are the list of common issues we face while data loading in to BW:
    Data Loading in Infocube/ODS/DSO:
    1) Missing SID issue or missing master data
    2) Data load cancelled becasue of various exceptions. Eg: Message Type 'x', DBIF SQL Error,
    3) Data records found in duplicate
    4) Alpha Confirming Vaue error
    5) Invalid time interval
    6) Time overlap error
    7) Job Cancelled in source system
    8) RFC Connection Error
    9) Data Source Replication error
    10) Data load locked by active change run
    11) Attributes to target by user 'xyz'
    12) Job cancelled as the previous load is still running
    13) Target locked by a change run
    (Some times data loads dont fail but run for long time then usual with out any progess, because of struck tRFCs in source sytem, source sytem performance is poor and the job is in release state for long time, etc...)
    Aggregate Rollups:
    1) Nofilled aaggregates available, rollup not possible
    2) Rollup locked by active change run
    3) Job cancelled because of various exceptions
    PSA Updations:
    1) Job cancelled because of various exception
    2) Missing SID value
    ODS Activaitons:
    1) Request to be activate shoud be green - (Loading of data in to ODS in still going on)
    2) Key exists in duplicate
    3) Job cancelled because of various exceptions (Short Dumps)
    Attribue Change Run:
    1) Locked by a rollup
    2) Job cancelled because of various exceptions
    3) Locked by another change run
    Hope it helps....
    Cheers,
    Habeeb

  • Error while data loading

    Hi Gurus,
    I am getting error while data loading. At BI side when I  check the error it says Background Job Cancelled and when I check in R3 side I got the following error
    Job started
    Step 001 started (program SBIE0001, variant &0000000065503, user ID R3REMOTE)
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    DATASOURCE = 2LIS_11_V_ITM
             Current Values for Selected Profile Parameters               *
    abap/heap_area_nondia......... 2000683008                              *
    abap/heap_area_total.......... 4000317440                              *
    abap/heaplimit................ 40894464                                *
    zcsa/installed_languages...... ED                                      *
    zcsa/system_language.......... E                                       *
    ztta/max_memreq_MB............ 2047                                    *
    ztta/roll_area................ 6500352                                 *
    ztta/roll_extension........... 3001024512                              *
    4 LUWs confirmed and 4 LUWs to be deleted with function module RSC2_QOUT_CONFIRM_DATA
    ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    Job cancelled
    Please help me out what should I do.
    Regards,
    Mayank

    Hi Mayank,
    The log says it went to short dump due to temp space issue.as its the source system job ,check in the source system side for temp table space and also check at the BI side as well.
    Check with your basis regarding the TEMP PSA table space - if its out of space ask them to increase the table space and try to repeat the load.
    Check the below note
    Note 796422 - DBIF_RSQL_SQL_ERROR during deletion of table BWFI_AEDAT
    Regards
    KP
    Edited by: prashanthk on Jul 19, 2010 10:42 AM

  • Summing up key figure in Cube - data load from ODS

    Hello Gurus,
    I am doing a data load from ODS to cube.
    The records in ODS are at line-item level, and all needs to be summed up at header level. There is only one key-figure. ODS has header and line-item fields.
    I am loading only header field data ( and not the item-field data) from ODS to Cube.
    I am expecting only-one record in cube ( for all the item-level records) with all the key-figures summed up. But couldn't see that.
    Can anyone please explain how to achieve it.
    I promise to reward points.
    =====
    Example to elaborate my point.
    In ODS
    Header-field  item-field   quantity
    123                301          10
    123                302           20
    123                303           30
    Expected record in Cube
    Header-field       Quantity
       123                    60  
    ====================
    Regards,
    Pramod.

    Hello Oscar and Paolo.
    Thanks for the reply. I am using BW 7.0
    Paolo suggested:
    >>If you don't add item number to cube and put quantity as adition in update rules >>from ODS to cube it works.
    I did that still I get 3 records in cube.
    Oscar Suggested:
    >>What kind of aggregate do you have for your key figure in update rules (update >>or no change)?
    It is "summation". And it cannot be changed. (Or at least I do not know how to change it.)
    There are other dimensions in the cube - which corresponds to the field(s) in ODS.
    But, I just mentioned these two (i.e. header and item-level) for simplicity.
    Can you please help?
    Thank you.
    Pramod.

  • Error is data loading from 3rd party source system with DBCONNECT

    Hi,
    We have just finished an upgrade of SAP BW 3.10 to SAP NW 7.0 EHP1.
    After the upgrade, we are facing a problem with data loads from a third party Oracle source system using DBConnect.
    The connection is working OK and we can see the tables in the source system. But we cannot load the data.
    The error in the monitor is as follows:
    'Error message from the source system
    Diagnosis
    An error occurred in the source system.
    System Response
    Caller 09 contains an error message.
    Further analysis:
    The error occurred in Extractor .
    Refer to the error message.'
    But, unfortunately, the error message has no further information.
    If we look at the job log in sm37, the job finished with the following log -                                                                               
    27.10.2009 12:14:19 Job started                                                                                00           516          S 
    27.10.2009 12:14:19 Step 001 started (program RSBATCH1, variant &0000000000119, user ID RXSAHA)                    00           550          S 
    27.10.2009 12:14:23 Start InfoPackage ZPAK_4FMNJ2ZHNNXC6HT3A2TYAAFXG                                              RSM1          797          S 
    27.10.2009 12:14:24 Element NOAUTHORITYCHECK is not available in the container                                     OL           356          S 
    27.10.2009 12:14:24 InfoPackage ZPAK_4FMNJ2ZHNNXC6HT3A2TYAAFXG created request REQU_4FMXSQ6TLSK5CYLXPBOGKF31G     RSM1          796          S 
    27.10.2009 12:14:24 Job finished                                                                                00           517          S 
    In a BW 3.10 system, there is no  message related to element NOAUTHORITYCHECK. So, I am wondering if this is something new in NW 7.0.
    Thanks in advance,
    Rajib

    There will be three things to get the errors like this
    1.RFC CONNECTION FAILED
    2.CHECK THE SOURCE SYSTEM
    3.CHECK IT OUT WITH Oracle Consultants WEATHER THEY ARE FILLING UP THE LOADS.TELL THEM TO STOP
    4.CHECK I DOC PROCESSING
    5.FINALLY MEMORY ISSUES.
    6.CATCH THE DATA SOURCE FIRST CHANGE IT AND THEN ACTIVATE AND RUN THE LOAD
    7.LAST IS MEMORY ISSUE.
    and also Check the RFC connection in SM59 If  it is ok then
    check the SAP note : 692195 for authorization
    Santosh

  • Need help with Rollback data if there is error in Data load

    Hi All,
    We are trying to load data to Oracle 11g database. We want a trigger, procedure or something like that to rollback the data if there are errors in load. Is it possible to do rollback after all the records has been parsed ? So if we try to load 100 records and if there are 30 records with error, we want to rollback after all the 100 records are parsed.
    Please advice.

    >
    Thanks for the suggestion. I'll try that option. So currently we are only loading data that is validated and erroneous records are rejected using trigger. So we don't get any invalid data in table. But Now users are saying that all the records should be rejected if there is even one error in the data load.
    >
    I generally use a much simpler solution for such multi-stage ETL processes.
    Each table has a IS_VALID column that defaults to 'N'. Each step of the process only pulls data with the flag set to 'Y'.
    That allows me to leave data in the table but guarantee that it won't be processed by subsequent stages. Since most queries that move data from one stage to another ultimately have to read table rows (i.e. they can't just use indexes) it is not a performance issue to add a predicate such as "'AND IS_VALID = 'Y'" to the query that accesses data.
    1. add new unvalidated data - automatically flagged an invalid by default of 'N" on IS_VALID column
    2. run audit step #1 - capture the primary key/rowid of any row failing the audit.
    3. run audit steps #2 through #n - capture error row ids
    4. Final step - update the data setting IS_VALID to 'Y' only if a row passes ALL audits: that is, only if there are NO fatal errors for that row captured in the error table.
    That process also allows me to capture every single problem that any row has so that I can produce reports for the business users that show everything that is wrong with the data. There are some problems that the business wan't to ignore, others that can be fixed in the staging tables then reprocessed and others that are rejected since they must be fixed in the source system and that can take several days.
    For data that can be fixed in the staging tables the data is fixed and then the audit is rerun which will set the IS_VALID flag to 'Y' allowing those 'fixed' rows to be included in the data that feeds the next processing stage.

  • Segmentation fault error during data load in parallel with multiple rules

    Hi,
    I'm trying to do sql data load in parallel with multiple rules (4 or 5 rules, maybe), i'm getting a "segmentation fault" error. I tested 3 rules file and it worked fine. we're using Essbase system 9.3.2., with UDB (v8) as the sql data source. ODBC driver is DataDirect 5.2 DB2 Wire Protocol Driver (ARdb222). Please let me know if you have any information on this.
    thx.
    Y

    Hi Thad,
    I was wondering, if system is unicode or non unicode that should not matter the amount and currency field . As currencies are defined by SAP and it is in pure English at least a currency code part of 3 Chars. 
    Could this because of some incosistency of data ??
    I would like to know for Which currency  had some special characters it it in that particular record ??
    Hope that helps.
    Regards
    Mr Kapadia

  • No data load into ODS although sourcesystem connection is ok

    Hello experts,
    although everthing is ok with my sourcesystem connection, SAP BW loads no data into ODS and Cube. In the ODS-administration-view the requests are registrated, QM-Satus is green but ODS-id is zero.
    Everything worked fine when running the process chain, but suddenly it stops when activating the data in the ods because there are no data in it.
    Any suggestions?
    thanx
    Axel

    Hi Axel,
    well the error message you specified is only poosible if the datasource is not installed.I m a bit confused.
    Check if you are able to see this datasource in rsa6 in the source system.
    Also,try replicating the source system on the BW side,then check your respective infosource,whether they are available..?
    hope it helps,
    regards,
    Parth.

  • Error Message when loading an ODS from an ODS

    Within a process chain we are loading an ODS from another ODS.  This fails with the following message:
    Job REQU_49HIKNN81WXVPRI80DC2TVCUF removed from periodical scheduling and terminated
    Message no. RSM1170
    Diagnosis
    An end-date for the job is specified when the job is scheduled (no longer "start after date/time").
    When the job is started for the first time, this end-date is applied to ALL subsequent periodical jobs.
    This is different to the standard R/3 batch schedulings.
    The end-date is checked each time the job is restarted, and if the end-date has passed, the job is removed from the schedule.
    This is what has happened in this case.
    Procedure
    Reschedule the job if you want to keep it running.
    Does anyone know where I can see/change the end date for this job or how I reschedule it?

    Dear Barry,
    1) Just go and make your  req Red both n Data target and PC
    2)Delete the error request
    3)Get the Infopackage from PC header.
    4)reshedule it again
    Hope it solves your problem
    Assign points if helpful
    Regards
    Bala

  • Short Dump Error While Data load

    Hi Experts,
    Data load to an ODS in production server has been failed because of Short dump error. The error message shows " OBJECTS_OBJREF_NOT_ASSIGNED ".
    Request you to please help me out with a solution ..
    Regards,
    Vijay

    Hi Vijay,
    follow the steps below (May Help you)
    Goto Monitor screen>Status Tab> using the wizard or the menu path >Environment -> Short dump> In the warehouse
    Select the Error and double click
    Analyse the error from the message.
    1.-->Go to Monitor
    -->Transactional RFC
    -->In the Warehouse
    -->Execute
    -->EDIT
    -->Execute LUW
    Refresh the Transactional RFC screen and go to the Data Target(s) which have failed and see the status of the Request-->Now it should be green.
    2.In some cases the above process will not work (where the Bad requests still exists after the above procedure).
    In that case we need to go to the Data Targets and need to delete the bad requests and do the update again
    Regards,
    BH

  • Error in Data updation From ODS TO CUBE

    hello Experts
    We are working on BW 3.5
    for FI GL line items we are loading the data to ODS and than to Cube.
    earlier there was a process chain failure 2 months back so delta were stuck for 2 months now about 10 million records have been uploaded in ODS but when i am tryint to upload these records to CUBE via delta upload  its giving Error
    Value 'Ã
     ABC100711396 ' (hex. 'C30DC2A04C4344313030373131333936 ') of characteristic 0REF_DOC_NO c
    i would like to know how to correct this and from where this error is coming as the same data is loaded to ODS and there its OK and secondly as there is no PSA bw ODS and Cube Do i need to delete the request and upload again
    what if same error appears again should i write a start routine in update rule Between ODS and CUBE for special characters
    thanks for replies

    Hi,
    Goto All Elementary Tests in RSRV
    Goto Master data
    Select the required test
    Select your Infoobject
    Execute the test
    You will get the result with errors or no errors
    If you get errors, then you click "Correct Error" button.
    Regrds,
    Suman

  • PERFORM_CONFLICT_TAB_TYPE  shortdump during data loading into ODS

    HAI
    Im trying to load the data into ODS in Production and QA system . But im getting the shortdump says that <b>PERFORM_CONFLICT_TAB_TYPE</b> .
    But in development system , data is loading into ODS.
    So please tell me wht i have to do
    i will assing the points
    rizwan

    here it is..
    Note 707986 - Writing in trans. InfoCubes: PERFORM_CONFLICT_TAB_TYPE
    Summary
    Symptom
    When data is written to a transactional InfoCube, the termination PERFORM_CONFLICT_TAB_TYPE occurs. The short dump lists the following reasons for the termination:
               ("X") The row types of the two tables are incompatible.
               ("X") The table keys of the two tables do not correspond.
    Other terms
    transactional InfoCube, SEM, BPS, BPS0, APO
    Reason and Prerequisites
    The error is caused by an intensified type check in the ABAP runtime environment.
    Solution
    Workaround for BW 3.0B (SP16-19), BW 3.1 (SP10-13)
               Apply the attached correction instructions.
    BW 3.0B
               Import Support Package 20 for 3.0B (BW3.0B Patch20 or SAPKW30B20) into your BW system. The Support Package is available oncenote 0647752 with the short text "SAPBWNews BW3.0B Support Package 20", which describes this Support Package in more detail, has been released for customers.
    BW 3.10 Content
               Import Support Package 14 for 3.10 (BW3. 10 Patch14 or SAPKW31014) into your BW system. The Support Package is available once note 0601051  with the short text "SAPBWNews BW 3.1 Content Support Package 14" has been released for customers.
    BW3.50
               Import Support Package 03 for 3.5 (BW3.50 Patch03 or SAPKW35003) into your BW system. The Support Package is available once note 0693363 with the short text "SAPBWNews BW 3.5 Support Package 03", which describes this Support Package in more detail, has been released for customers.
    The notes specified may already be available to provide advance information before the Support Package is released. However, in this case, the short text still contains the term "Preliminary version" in this case.
    Header Data
    Release Status: Released for Customer
    Released on: 18.02.2004  08:11:39
    Priority: Correction with medium priority
    Category: Program error
    Primary Component: BW-BEX-OT-DBIF Interface to Database
    Secondary Components: FIN-SEM-BPS Business Planning and Simulation
    Releases
    Software
    Component Release From
    Release To
    Release And
    subsequent
    SAP_BW 30 30B 30B  
    SAP_BW 310 310 310  
    SAP_BW 35 350 350  
    Support Packages
    Support
    Packages Release Package
    Name
    SAP_BW_VIRTUAL_COMP 30B SAPK-30B20INVCBWTECH
    Related Notes
    693363 - SAPBWNews BW SP03 NW'04 Stack 03 RIN
    647752 - SAPBWNews BW 3.0B Support Package 20
    601051 - SAPBWNews BW 3.1 Content Support Package 14
    Corrections Instructions
    Correction
    Instruction Valid
    from Valid
    to Software
    Component Ref.
    Correction Last
    Modifcation
    301776 30B 350 SAP_BW J19K013852 18.02.2004  08:03:33
    Attributes
    Attribute Value
    weitere Komponenten 0000031199
    German
    English
    Vishvesh

  • No data loaded from ODS to Cube

    Hi,
    I am trying to do init with data transfer from 0FIAP_O03 to 0FIAP_C03...
    though there is data in the ODS, nothing is loaded to the cube. It just says
    No data available
    Diagnosis
    The data request was a full update.
    In this case, the corresponding table in the source system does not
    contain any data.
    System Response
    Info IDoc received with status 8.
    Procedure
    Check the data basis in the source system.
    Any suggestions?
    Thanks
    Shalini

    Hi
    "Info IDoc received with status 8" implies that there is no data in the source.
    But as you say, that data is there in  the source(ODS), then there is some inconsistency in the data source and you need to replicate the data source and reactivate that.
    IDoc Status 8 signifies that there is no data in the source and its not any error.
    For more info on idocs check the below link
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/0087780d-81b0-2a10-eb82-86e1062f1a2d
    Also goto sm58 and clear the idoc there.
    Hope this helps
    Regards
    Shilpa

  • Error in data load from application server

    Well, this problem again!
    In datasource for dataload by flatfile, tab Extraction i selected Load Text-Type File From Application Server.
    In tab Proposal:
    Converter: Separated with Separator (for Example, CSV).
    No. of Data Records: 9198
    Data load sucefull.
    I add 1 record:
    Converter: Separated with Separator (for Example, CSV).
    No. of Data Records: 9199
    So appear a  message error "Cannot convert character sets for one or more characters".
    I look in line 9199 in file and not have any special character. I use AL11 and debug to see this line.
    When i load file from local workstation, this error not occur.
    What's happen?

    Hi Rodrigo,
    What type of logical file path have yopu created in the application server for loading this file.
    Is it a UNIX file path or an ASCII file path.
    Did you check how the file looks in the application server AL11?
    Prathish

Maybe you are looking for