SAP HANA calculation field

Hello,
I have a query on the new calculated field in SAP HANA. What is the usage of this field, what are the pre-requisites and potential case when it is recommended to use such fields? Do we need to have a calculative view to use the calcluated field or can it be used directly in ABAP code?
Kindly let me know. We are in the process of code optimization for HANA.
Best Regards,
Mohit

Hey Mohit,
can you please give a bit more details, e.g. documentation about what you refer to as "the new calculated field in SAP HANA".
Concerning the details of usage within HANA, i.e. whether you need a calculation view etc. you might have a look at the SAP Help (Introduction - SAP HANA SQL and System Views Reference - SAP Library)) or maybe directly in the HANA SQL reference (http://help.sap.de/hana/SAP_HANA_SQL_Script_Reference_en.pdf).
From my answer I think you can directly guess, that there's no way in directly accessing the calculated field from ABAP - except of course if you access it using native SQL (ADBC or EXEC SQL).
For code optimization in SAP HANA (from an ABAP developers perspective), you might want to have a look at our Open SAP course (ABAP Development for SAP HANA - Dr. Jasmin Gruschke and Jens Weiler) or in the documentation given in our "Get Started" Section of ABAP for SAP HANA namely ABAP for SAP HANA Reference Scenario and ABAP for SAP HANA Reference Scenario - AS ABAP 7.4 Support Packages.
In case you need more HANA (less ABAP) related information for your question, you might additionally contact the HANA experts on SAP HANA Developer Center.
Cheers,
Jasmin

Similar Messages

  • How to make a BW Infoprovider for a SAP HANA Calculation-/Analytcial-View

    Hi
    I have got a BW 73 with a Hana.
    How can I  make a BW Infoprovider for a SAP HANA Calculation-/Analytcial-View in SAP BW ?
    Thank You
    Best Martin

    Hi Rekha,
    No need to do anything, just we need do migrate standard info cube to In-memory Optimized Info Cube. Once it is turned to In-Memory optimized Info Cube(We will have F table surrounded by master data tables) and no dimension tables in HANA.
    F-tables is nothing but Analytic View in HANA and Master data tables are nothing but Attributes in HANA.
    once you migrate to In-Memory Info Cube, your tables are ready in HANA and you can create different view.
    have a look at the below videos.
    https://www.experiencesaphana.com/blogs/bwonhana/2011/10/18/prerequisites-for-bw-on-hana
    http://www.sdn.sap.com/irj/scn/data-warehousing-elearning
    Regards,
    Venkatesh

  • Error while creating a BO Universe (using IDT) on top of SAP HANA Calculation View with Input Parameters

    Hi All..
          We are trying to create a Universe (using IDT-Version4.0) on top of the HANA Calculation view. The Calculation view has 4 input parameters, So
    in Universe side we have created the respective prompts. For the creation of derived table we have used the following code
    SELECT *
    FROM "_SYS_BIC"."<schema_name>.<CV_Calculation_View_test>"
    'PLACEHOLDER'=('$$IP_A$$','@Prompt(A)'),
    'PLACEHOLDER'=('$$IP_B$$','@Prompt(B)'),
    'PLACEHOLDER'=('$$IP_C$$','@Prompt(C)'),
    'PLACEHOLDER'=('$$IP_D$$','@Prompt(D)')
    While validating the above code we are getting an error.
    I have attached the snapshot of that error for your reference. Please find the attachment and help me in resolving it.
    Thanks in advance.

    Hello George,
    I don't have any personalization set on the info space. Also I am using mapped Account in BI to connect to HANA. The confusing part is that it is able to validate the infospace with the input parameters and index it. However query does not return any results. I even tried running the same query which explorer sends to HANA in the SQL editor and there too the same results,the query does not return anything. The model does return data when I do a data preview and if accessed from other tools like AAO.
    Also when I use SSO connection to HANA, indexing of the infospace fails. Where can I see the error log?
    Thanks,

  • Hi,How to add field to sap Liquidity calculation module tables?

    Hi Experts,
            How to add field to sap Liquidity calculation module tables?
            and how to add a field to a particular transaction code using a customer exits.
    please tell me in detail.
    thanks inadvance,
    Regards,
    Rekha

    Hi Pranab,
    Please follow the below steps to create an extra field and write code for that field through Infoset.
    1-->Change in Infoset
    u2022Go to SQ02 , enter Infoset name and click on Change Button
    u2022Go to Extras (F5) button displayed on application bar.
    u2022In Extras tab, click on 'Create' icon to create additional field E_NAME1, give as type C (character) and give desired length and. Enter header description  as 'ShipToName'
    u2022Select this field i.e E_NAME1 in one of the field group of Infoset.
    u2022Go to Code section  (Shift+F8) of infoset, Select Record Processing Event and write your logic code (condition) in this code section:
    if vbpa-adrnr = space.
    E_NAME1 = kna1-name1.
    else.
    E_NAME1 = adrc-name1.
    endif.
    2-->Generate the Infoset.
    3-->Change in Query
    u2022Go to SQ01(in a new session), give your query name (by selecting your user group) and click on change button.
    u2022Check the field group in which you have added E_NAME1 field, then check E_NAME1 from Fields screen
    u2022Click on 'Basic List' button; give line (row no.) and sequence (column no.) for extra fields.
    u2022Execute/Test the Query, you will get desired result.
    Please let me know, if you need more information.
    Regards,
    Dinesh
    Edited by: Dinesh Tiwari on Oct 29, 2009 7:13 AM

  • SAP HANA XSODATA service (Service exception: column store error.)

    Hi all,
    i have a problem with my calculation view using xsodata service on it. (There's an input parameter called P
    _SWERK)
    In my calculation view, the data origin are two analytic views (on which the input parameter P_SWERK should be filter data at beginning of the sql script code).
    First i read the analytic views with function CE_OLAP_VIEW and after i do a CE_PROJECTION function on them using the input parameter P_SWERK like a filter on field SWERK.
    But when i run my application on browser the following error occurs :
    <message xml:lang="en-US">Service exception: column store error.</message>
    The link is this :
    http://host:port/Project_DM/services/Test/TEST_ZIIG_PDM_CALC_VIEW_FINAL_service.xsodata/PianiDiManutenzioneParameters(P_SWERK='CO05')/Results
    The service definition is :
    service {
    "EricssonItalgas/TEST_ZIIG_PDM_VIEW_FINAL.calculationview" as "PianiDiManutenzione" keys generate local "ID"
    parameters via entity;
    The SAP HANA AWS revision is 60.
    Someone could you help me,please?
    Thanks in advance.
    Dario.

    Hi Dario,
    Does the calculation view work without xsodata service? From the URL, your XS project name should be Project_DM, but from the xsodata source, the project name is EricssonItalgas. I'm confused with this. Did you use rewrite_rules or?
    Best regards,
    Wenjun

  • SAP HANA One and Predictive Analysis Desktop - Time Series Algorithms

    I have been working on a Proof-of-Concept project linking the SAP Predictive Analysis Desktop application to the SAP HANA One environment.
    I have modeled that data using SAP HANA Studio -- created Analytic views, Hierarchies, etc. -- following the HANA Academy videos.  This has worked very well in order to perform the historical analysis and reporting through the Desktop Application. 
    However, I cannot get the Predictive Analysis algorithms -- specifically the Time Series algorithms -- to work appropriately using the Desktop tool. It always errors out and points to the IndexTrace for more information, but it is difficult to pinpoint the exact cause of the issue.  The HANA Academy only has videos on Time Series Algorithms using SQL statements which will not work for my user community since they will have to constantly tweak the data and algorithm configuration. 
    In my experience so far with Predictive Analysis desktop and the Predictive Algorithms, there is a drastic difference between working with Local .CSV / Excel files and connecting to a HANA instance.  The configuration options for using the Time Series Algorithms are different depending upon the data source, which seems to be causing the issue.  For instance, when working with a local file, the Triple Exponential Smoothing configuration allows for the specification of which Date field to use for the calculation.  Once the data source is switched to HANA, it no longer allows for the Date field to be specified.  Using the exact same data set, the Algorithm using the local file works but the HANA one fails. 
    From my research thus far, everyone seems to be using PA for local files or running the Predictive Algorithms directly in HANA using SQL.  I can not find much of anything useful related to combing PA Desktop to HANA. 
    Does anyone have any experience utilizing the Time Series Algorithms in PA Desktop with a HANA instance?   Is there any documentation of how to structure the data in HANA so that it can be properly utilized in PA desktop? 
    HANA Info:
    HANA One Version: Rev 52.1
    HANA Version: 1.00.66.382664
    Predictive Analysis Desktop Info:
    Version: 1.0.11
    Build: 708
    Thanks in advance --
    Brian

    Hi,
    If you use CSV or XLS data source you will be using Native Algorithm or R
    Algorithm in SAP Predictive Analysis.
    When you connect HANA, SAP Predictive Analysis uses PAL Algorithm which runs
    on HANA server.
    Coming to your question regarding difference,
    In SAP PA Native Algorithm, we could provide the Data variable, Algorithm
    picks the seasonal information from the Data column. Both R and SAP HANA PAL
    does not support Date Column. We need configure seasonal information in
    Algorithm properties.
    R Properties
    1) Period : you need to mention the periodicity of the Data
    Monthly : (12)
    Quarter : (4)
    Custom : you can use it for week or Daily or hourly.
    2) Start Year: need to mention Start year.
    Start year is not used by algorithm for calculating Time series, but it helps
    PA to generate Visualization ( Time series chart) by simulating year and
    periodicity information.
    3) Starting Period:
    if your data is Quarterly and you have data recordings from Q2, mention 2 in
    start period.
    Example.
    If the data periodicity is Monthy and my data starts from Feb 1979, we need to provide following information,
    Period: 12
    Start year: 1979
    start Period: 2
    PAL Properties. : Same as properties defined in R.
    Thanks
    Ashok
    [email protected]

  • Open HUB ( SAP BW ) to SAP HANA through DB Connection data loading , Delete data from table option is not working Please help any one from this forum

    Issue:
    I have SAP BW system and SAP HANA System
    SAP BW to SAP HANA connecting through a DB Connection (named HANA)
    Whenever I created any Open Hub as Destination like DB Table with the help of DB Connection, table will be created at HANA Schema level ( L_F50800_D )
    Executed the Open Hub service without checking DELETING Data from table option
    Data loaded with 16 Records from BW to HANA same
    Second time again executed from BW to HANA now 32 records came ( it is going to append )
    Executed the Open Hub service with checking DELETING Data from table option
    Now am getting short Dump DBIF_RSQL_TABLE_KNOWN getting
    If checking in SAP BW system tio SAP BW system it is working fine ..
    will this option supports through DB Connection or not ?
    Please follow the attachemnet along with this discussion and help me to resolve how ?
    From
    Santhosh Kumar

    Hi Ramanjaneyulu ,
    First of all thanks for the reply ,
    Here the issue is At OH level ( Definition Level - DESTINATION TAB and FIELD DEFINITION )
    in that there is check box i have selected already that is what my issue even though selected also
    not performing the deletion from target level .
    SAP BW - to SAP HANA via DBC connection
    1. first time from BW suppose 16 records - Dtp Executed -loaded up to HANA - 16 same
    2. second time again executed from BW - now hana side appaended means 16+16 = 32
    3. so that i used to select the check box at OH level like Deleting data from table
    4. Now excuted the DTP it throws an Short Dump - DBIF_RSQL_TABLE_KNOWN
    Now please tell me how to resolve this ? will this option is applicable for HANA mean to say like , deleting data from table option ...
    Thanks
    Santhosh Kumar

  • Currency conversion error in SAP HANA

    Hi,
    I am new to SAP HANA and learning to create information views in HANA studio (SAP HANA SP6 on Cloudshare, HANA studio 1.0.68). I am trying to create a simple analytic view (on purchaseOrderItem table in SAP_HANA_EPM_DEMO sample database) to have GrossAmount converted to EUR.
    I added a calculated column as follows:
    When i click on "OK", i get error -
    The check box “Calculate before aggregation” has been unchecked, because the definition of the calculated column contains measures with currency conversion, restricted measures or operands with input parameters. For such a calculated column the calculation is always done after the aggregation."
    and checkbox "calculate before aggregation" get unchecked. See screenshot below:
    Please suggest what could be reason? Thanks in advance.
    Regards,
    Amit

    Hi Amit,
    If you uncheck the "Calculate before aggregation" checkbox and activate the view, you will see in the generated log that a Calc scenario is created. (a view with /olap wrapper). Due to the calc scenario, the aggregation is defined as the default behavior for the KFs and hence the calculation cannot be done before aggregation.
    By the way, I did not understand why do you need calculate before aggregation for a KF which is just a copy of another KF. If you need Gross amount in Local currency and EUR, then just perform the currency conversion without "Calculate before aggregation" checkbox. It will work.
    Regards,
    Ravi

  • SAP HANA modelling Standalone

    Hello Experts,
    We are in the process of HANA Standalone implementation and design studio as reporting tool. When I am modeling, I did not figure out answers to some of the below questions .Below are the questions. Experts, please help.
    Best way of modeling: The SAP HANA LIVE is completely built on calculation view; there are no Attribute and Analytical views. I have got different answer why there is only Calculation view and there are no Alaytic view and Attribute views. We are in SP7 latest version. This is a brand new HANA in top of non-SAP (DB2 source).  What is the best way to model this scenario, meaning, can we model everything in the Calculation view’s like SAP HANA live or do you suggest using the standard attribute, analytical and calculation views to do the data model. Is SAP moving away from AV & AT to only calculation Views to simply the modeling approach?
    Reporting: We are using the design studio as front end tool. Just for example, if we assume that we are
    Using the BW, we bring all the data in to BW from different sources, build the cubes and use the bex query. Here in bex query we will be using the restricted key figures, calculated key figures calculations etc. From the reporting wise, we have the same requirements, calculations, RKF, CKF,Sum, Avg etc. if we are Using the design studio on top of standalone HANA, where do I need to implement all these calculations? Is it in different views?  (From reporting perspective, if it’s BW system, I would have done all the calculations in BEx.)
    Universe: If we are doing all the calculations in SAP HANA like RKF. CKF and other calculations , what is the point in having additional layer of universe , because the reporting compnets cam access the queries directly on views .In one of our POC , we found that the using universe affect performance.
    Real time reporting: Our overall objective is to give a real time or close to real time reporting requirements, how data services can help, meaning I can schedule the data loads every 3 or 5 min to pull the data from source. If I am using the Data services, how soon I can get the data in HANA, I know it depends on the no of records and the transformations in between the systems & network speed. Assuming that I will schele the job every 2 min and it will take another 5 min to process the Data services job , is it fair to say the my information will be available on the BOBJ tools with in 10 min from the creation of the records.
    Are there any new ETL capabilities included in SP7, I see some additional features included in SP7. Is some of the concepts discussed are still valid, because in SP7 we have star join concept.
    Thanks
    Magge

    magge kris wrote:
    Hello Experts,
    We are in the process of HANA Standalone implementation and design studio as reporting tool. When I am modeling, I did not figure out answers to some of the below questions .Below are the questions. Experts, please help.
    Best way of modeling: The SAP HANA LIVE is completely built on calculation view; there are no Attribute and Analytical views. I have got different answer why there is only Calculation view and there are no Alaytic view and Attribute views. We are in SP7 latest version. This is a brand new HANA in top of non-SAP (DB2 source).  What is the best way to model this scenario, meaning, can we model everything in the Calculation view’s like SAP HANA live or do you suggest using the standard attribute, analytical and calculation views to do the data model. Is SAP moving away from AV & AT to only calculation Views to simply the modeling approach?
    >> I haven't read any "official" guidance to move away from typical modeling approach, so I'd say stick with the usual approach- AT, then AV, then CA views. I was told that the reason for different approach with HANA Live was to simplify development for mass production of solutions.
    Reporting: We are using the design studio as front end tool. Just for example, if we assume that we are
    Using the BW, we bring all the data in to BW from different sources, build the cubes and use the bex query. Here in bex query we will be using the restricted key figures, calculated key figures calculations etc. From the reporting wise, we have the same requirements, calculations, RKF, CKF,Sum, Avg etc. if we are Using the design studio on top of standalone HANA, where do I need to implement all these calculations? Is it in different views?  (From reporting perspective, if it’s BW system, I would have done all the calculations in BEx.)
    >> I'm not a BW guy, but from a HANA perspective - implement them where they make the most sense. In some cases, this is obvious - restricted columns are only available in Analytic Views. Hard to provide more complex advice here - it depends on your scenario(s). Review your training materials, review SCN posts and you should start to develop a better idea of where to model particular requirements. (Most of the time in typical BI scenarios, requirements map nicely to straightforward modeling approaches such as Attribute/Analytic/Calculations Views. However, some situations such as slowly-changing dimensions, certain kinds of calculations (i.e. calc before aggregation with BODS as source - where calculation should be done in ETL logic) etc can be more complex. If you have specific scenarios that you're unsure about, post them here on SCN.
    Universe: If we are doing all the calculations in SAP HANA like RKF. CKF and other calculations , what is the point in having additional layer of universe , because the reporting compnets cam access the queries directly on views .In one of our POC , we found that the using universe affect performance.
    >>> Depends on what you're doing. Universe generates SQL just like front-end tools, so bad performance implies bad modeling. Generally speaking - universes *can* create more autonomous reporting architecture. But if your scenario doesn't require it - then by all means, avoid the additional layer if there's no added value.
    Real time reporting: Our overall objective is to give a real time or close to real time reporting requirements, how data services can help, meaning I can schedule the data loads every 3 or 5 min to pull the data from source. If I am using the Data services, how soon I can get the data in HANA, I know it depends on the no of records and the transformations in between the systems & network speed. Assuming that I will schele the job every 2 min and it will take another 5 min to process the Data services job , is it fair to say the my information will be available on the BOBJ tools with in 10 min from the creation of the records.
    Are there any new ETL capabilities included in SP7, I see some additional features included in SP7. Is some of the concepts discussed are still valid, because in SP7 we have star join concept.
    >>> Not exactly sure what your question here is. Your limits on BODS are the same as with any other target system - doesn't depend on HANA. The second the record(s) are committed to HANA, they are available. They may be in delta storage, but they're available. You just need to work out how often to schedule BODS - and if your jobs are taking 5 minutes to run, but you're scheduling executions every 2 minutes, you're going to run into problems...
    Thanks
    Magge

  • Universe based on SAP HANA OLAP connection

    Can I create universe based on SAP HANA OLAP connection ?
    How can i access SAP HANA view which was created using OLAP connection in WEBI and Dashboard? It does not allow creating Universe on top of SAP HANA OLAP connection and give an error.
    SAP BusinessObjects query and reporting applications can directly connect to OLAP SAP HANA connections. No universe is required, only a published OLAP SAP HANA connection.
    Thanks

    OLAP connection are only meant for tools that make use of OLAP connectivity only like BEx ..and HANA analytic views and calculated views are also treated as OLAP cubes for them..
    So these connection is only available with analysis product like Analysis Office (which is meant to be used with Bex or OLAP cubes ) and Design studio which is again mainly built around OLAP paradigm..
    You can not create a universe or Webi on Olap connection on HANA as of now.. However it could be like exposed in a future release of BO which is not public yet...

  • Issues while creating new user in SAP HANA

    Hello Team
                       When i am trying to create a new user in SAP HANA studio i can see that there has been a new field added of DATA validity where there are two options  a) Valid From and b)Valid Unitl . No matter whatever dates i give in this i got this error which creating the user . Status :- inactive
    Reason :- outside validity period . PFA screen shot attached below . Please suggest what dates should be given in  this field with some sample example .
    Regards

    Prag,
    Try this. Execute the following in a SQL window started by a userid that has been granted the USER ADMIN system privilege:
    ALTER USER BODS1 VALID FROM NOW    UNTIL FOREVER;
    You can use a date instead of FOREVER --- '2016-12-31 23:59'.
    Good luck,
    Robert

  • Column Value change in BO report vs HANA calculation view

    Hi,
    We are having BO  reports on  Standard hana views.
    We have some columns  in calculation view as below in Hana studio
    Column Name     Column Purpose
    1.Cost Center---(Gives cost center number)
    2.Cost center Name--(Gives description of cost center )
    3.Cost Element-which --(Gives cost center number)
    4.Cost Element Name-(Gives description of cost center)
    when we preview the data with these coulumns in hana studio there are giving the result as specified above.
    But if I run the report in BO using the above 4 columns ,I could see only descriptions for all the four columns as below
    1.Cost Center---(Gives description of cost center)
    2.Cost center Name--(Gives description of cost center )
    3.Cost Element-which --(Gives description of cost center)
    4.Cost Element Name-(Gives description of cost center)
    As per our requirement we should get numbers for cost center and cost element.
    I have observed in semantics that cost center and cost element have label columns mapped to them as below
    cost center has labelcolumn as costcenter name
    cost element has labelcolumn
    Please see the attached file . Is this happend because of the labelcolumn mapping
    we want the cost center and cost element to be displayed as numbers only.  Please suggest me on this.
    Thanku

    Hi Chandra,
    TEXT is coming because you have defined as "Label Column" . In HANA, Data preview the column names will only come but not "Descriptions".
    What is your reporting tool?  If it is AO, have a look on the below blog:
    Using Text Joins to enrich Attributes with "Texts" in SAP HANA with SAP BO Analysis Office
    Regards,
    Krishna Tangudu

  • SAP HANA Input parameters in SAP BO Analysis

    Hi, dear experts.
    Please help me with one problem in SAP BO Analysis. Problem is: in sap hana studio i made calculation view that has some input parameters (Param. type - Direct, Sematic Type - Date, Data type - Date). Then i made filter with this parameters in one of projection of my calculation view. When i push data preview in Hana Studio all input parameters are ready to input and they also have "Value help dialog". All is great! SAP HANA Studio version is 1.0.7000 (Build id: 386119). HANA DB version is 1.00.70.00.386119.
    On workstation i installed SAP BO Analysis edition for Microsoft Office 32-bit and driver for SAP HANA 70 Client Win32, made ODBC connection to HANA server. From SAP BO Analysis I found my view and started it. I see my input parameters, and see "value help dialog", but i can't choose value: it's not available for input. Here is the problem: what i have to do to enter the parameters?
    I also try to use variable but value help dialog is empty. Please help me with this issue.
    Message was edited by: Tom Flanagan

    Hello Andrei,
    Tested on:
    HANA Rev 70
    Have tested for variables and the value help is coming.
    Regards,
    Krishna Tangudu

  • SAP Hana implementation project - beginning tasks

    Hello,
    We want to implement a new SAP landscape running with SAP Hana database. As I understand, the SAP Hana database is a database that runs on a specific server appliance, am I right?
    For we can know which SAP Hana server appliance to buy we need first to estimate the sizing about it! We want to implement SAP ERP product (EHP6 or 7) on Hana DB, but we don´t know how to start... this is, we don´t know how to measure this, in order to buy the best solution (server appliance) for our own scenario!
    I checked some information in the following areas:
    service.sap.com/sizing
    and also the Quick Sizer tool (service.sap.com/quicksizer) but this is so much confused tool... I don´t understand it at all.
    Can you help me with your experience which SAP Hana server appliance we need to buy for this kind of project?!
    Thank you,
    samid raif

    Hi Dario,
    Does the calculation view work without xsodata service? From the URL, your XS project name should be Project_DM, but from the xsodata source, the project name is EricssonItalgas. I'm confused with this. Did you use rewrite_rules or?
    Best regards,
    Wenjun

  • R/3 archiving - wrong calculated field on 2LIS_02_ITM extractor

    Hello.
    An info and a question for who developed an archiving with r/3..
    We have found an issue related with the R/3 archiving and LIS_02_ITM.
    A calculed field (Delivery Quantity Variance in %); (field name = MABW) (you can see it at http://help.sap.com/saphelp_nw04/helpdata/en/58/f6383fdb800804e10000000a114084/frameset.htm )gives wrong data.
    The MABW is calculated using MSEG infos.
    MSEG infos were archived (and deleted), so, the calculation goes bad and returns wrong value.
    In front of the archiving process, we potentially will find problems due to the standard extractor behaviour.
    Have you ever noticed the same behaviour because of the wrong calculation onto the fields?
    thank you in advance for your cooperation...!
    Dieguz
    The details are the following :
    We have a R/3 PRODUCTION system (called MP0), connected with three BW
    ones (MP2,
    MP4, AP4).
    Someone performed an R/3 data archiving, using an archiving tool
    (IXOS).They archived the data with a one-to-one object approach :
    FI_DOCUMNT archived 18 months from clearing date.
    MM_MATBEL archived 24 mesi from posting date
    SD_VBRK archived 24 mesi from posting date
    RV_LIKP archived 24 mesi from posting date
    SD_VBAK archived 24 mesi from posting date
    MM_ACCTIT 18 months from posting date
    OIG_SHIPMNT 24 months from posting date
    MM_INV_BEL 24 months from posting date
    RL_TA 12 months from posting date
    RL_TB 12 months from posting date
    In the R/3 system, after the archiving, those documents were deleted.
    Our BW systems extract data using custom extractors and standard ones.
    - 2LIS_02_*
    - 2LIS_03_BF
    - 2LIS-11_*
    - 2LIS-17_*
    - 2LIS_13_*
    - 0FI_AR_4
    - 0FI_GL_4
    - 0FI_AP_4
    - 0EC_PCA_3
    - 0CO_OM_CCA_9
    We have found problems with 2LIS_02_ITM;
    During the delta extraction, we found records related with Purchase
    Order, with GR archived, that have MBAW field (Delivery Quantity
    Variance in %
    ) with "-100" value; the correct value should be "0" (because the
    Purchase Order was completed).
    MP0 is a 40B R/3 SYSTEM SP 83, PLUGIN 2004.1
    Thank you in advance for you cooperation
    best regards
    Diego

    Hi Waldemar,
    LOEKZ  (Deletion indicator in purchasing document) is not available in standard extract structure we have append that field with append structure in RSA 6.
    BUKRS is available in standard extract structure :
    In  lo-cockpit open extract structure and select BUKRS(Company code) field transfer to left side .
    2. click  data source maintenance check BUKRS field .
    you will get BUKRS field in 2lis_o2_ITM
    Thanks
    Sree

Maybe you are looking for