SAP HANA Performance Optimization

Hi Experts.
Is there any documentation out there that relates to optimizing SAP HANA components such as optimizing the SAP Hana models, data loading using SAP LT, etc. I would like to fine-tune what I have already built.
Thanks,
Regards,
Thatoyame
Email: [email protected]

Hello Thatoyame,
since HANA is pretty fast in so many cases, there isn't much available about general performance tuning. In cases where performance isn't great, a detailed root cause analysis (i.e. trace) is the way to go. See SAP HANA Appliance Software – SAP Help Portal Page (Admin section) for some details on tracing.
For data modeling, check Werner Steyn's webinar at Building Advanced Data Models with SAP HANA - Webinar Details.
Regards,
Marc
SAP Customer Solution Adoption (CSA)

Similar Messages

  • SAP Hana FAQ Frequently Asked Questions

    Hi All,
    SAP have published a fantastic Knowledge Base Article containing an extremely detailed
    SAP Hana FAQ - Frequently Asked Questions
    Go and download it now, it's here:
         2000003 - FAQ: SAP HANA
    And then come back to it from time to time, because not everything in there is released yet.
    Best regards,
    Andy.

    For the sake of completeness, there are now a number of very useful Hana FAQ OSS Notes, here's the list so far:
    . OSS 2000003 - FAQ: SAP HANA
    . OSS 2000000 - FAQ: SAP HANA Performance Optimization
    . OSS 1640741 - FAQ: "DB users for the DBA Cockpit for SAP HANA"
    . OSS 2039883 - FAQ: SAP HANA database and storage snapshots
    . OSS 2044468 - FAQ: SAP HANA Partitioning
    . OSS 2057595 - FAQ: SAP HANA High Availability
    . OSS 1999997 - FAQ: SAP HANA Memory
    . OSS 1999880 - FAQ: SAP HANA System Replication
    . OSS 2044468 - FAQ: SAP HANA Partitioning
    . OSS 2000002 - FAQ: SAP HANA SQL Optimization
    . OSS 2073112 - FAQ: SAP HANA Studio
    . OSS 1999998 - FAQ: SAP HANA Lock Analysis
    . OSS 1905137 - FAQ: SAP HANA - Obsolete tables
    . OSS 1999930 - FAQ: SAP HANA I/O Analysis
    . OSS 1914584 - SAP HANA Live Browser FAQ
    . OSS 2057046 - FAQ: SAP HANA Delta Merges
    . OSS 1642148 - FAQ: SAP HANA Database Backup & Recovery
    . OSS 2104291 - FAQ - SAP HANA multitenant database containers
    . OSS 2100009 - FAQ SAP HANA Savepoints
    . OSS 2082286 - FAQ: SAP HANA Graph
    . OSS 2081591 - FAQ: SAP HANA Table Distribution
    . OSS  2053330 - FAQ: Operations Recommendation on SAP HANA Alerts
    If you are interested in Hana Basis related OSS Notes, then the longest list of Hana OSS Notes publicly available on the internet is over here.
    Best regards,
    Andy

  • Performance Optimization Self Service- SAP help requirement

    Hi,
    I want to know whether for SAP's help is required for performing the self service of Performance Optimization.
    If we collect ST12 trace and use it to perform the self service then is the report which is generated from the self service sufficient to take further action or will I need some SAP expertise to implement / take corrective actions?
    In short, whether I can do the Performance Optimization by myself or I need help from SAP?
    Regards,
    Vishal

    hi,
    1) Is this service available to all the customer? (by all the customers I mean "Max Attention", "Enterprise Support" etc)
    i answer this above is it, from mz above reply, have you checked
    enterprise support customers can get five EGI sessions as free per year. please check
    http://service.sap.com/esacademy
    - click browse egis
    for your second question also I answered above
    Does the report itself gives suggestions or we need to provide the report to SAP
    here my reply above
    because Guided procedure itself the proven methodlogy from SAP, the report provides the lots of suggestions against the SAP best practices.
    you can use it yourself most of the time. if still you need expert guidance from SAP, book for EGI sessions. they called as expert guided implementations, remote support. duration might vary based on the session.
    again, service report is the source, you have to review yourself, if you are in EGI, sap use that report for guiding. Please review
    Thanks
    Jansi

  • CPU Work load optimization in SAP HANA

    Hi All,
    In virtual SAP HANA , there is an option to assign/optimize the vCPU to corresponds to Physical CPU as mentioned in the document below,
    .http://www.vmware.com/files/pdf/SAP_HANA_on_vmware_vSphere_best_practices_guide.pdf
    But,in Physical SAP HANA Systems, are the below mentioned features available,
    Can we assign dedicated CPU cores manually to a particular user/users ?
    Or, Is there a way to reserve  certain CPU cores for particular Application/Schema Threads/Sessions?
    Thanks for your help!!
    -Gayathri
    Message was edited by: Julie Blaufuss

    Nope, there is no such cpu-core based allocation option available.
    You can limit how many worker threads will be created, in case you need to balance the CPU usage on in a multi instance setup.
    However, you don't have any control on the actual amount of CPU resource that any SAP HANA instance, let alone DB user or query has.
    Comparing this with what you can do in vSphere is not suitable, as we are looking at a different level of abstraction here.
    To SAP HANA the machine that vSphere emulates will have x cores and SAP HANA will use these x cores - all of them.
    It's important not to forget that you have an additional layer of indirection here and things like CPU core binding can easily have negative side effects.
    For the SAP HANA users it would be more interesting to have a workload management within SAP HANA that would allow to manage different requirements on responsiveness and resource usage (it's not just CPU...). And that is what the SAP HANA development is working on.
    Maybe we're lucky and see some features in this direction later this year.

  • Open HUB ( SAP BW ) to SAP HANA through DB Connection data loading , Delete data from table option is not working Please help any one from this forum

    Issue:
    I have SAP BW system and SAP HANA System
    SAP BW to SAP HANA connecting through a DB Connection (named HANA)
    Whenever I created any Open Hub as Destination like DB Table with the help of DB Connection, table will be created at HANA Schema level ( L_F50800_D )
    Executed the Open Hub service without checking DELETING Data from table option
    Data loaded with 16 Records from BW to HANA same
    Second time again executed from BW to HANA now 32 records came ( it is going to append )
    Executed the Open Hub service with checking DELETING Data from table option
    Now am getting short Dump DBIF_RSQL_TABLE_KNOWN getting
    If checking in SAP BW system tio SAP BW system it is working fine ..
    will this option supports through DB Connection or not ?
    Please follow the attachemnet along with this discussion and help me to resolve how ?
    From
    Santhosh Kumar

    Hi Ramanjaneyulu ,
    First of all thanks for the reply ,
    Here the issue is At OH level ( Definition Level - DESTINATION TAB and FIELD DEFINITION )
    in that there is check box i have selected already that is what my issue even though selected also
    not performing the deletion from target level .
    SAP BW - to SAP HANA via DBC connection
    1. first time from BW suppose 16 records - Dtp Executed -loaded up to HANA - 16 same
    2. second time again executed from BW - now hana side appaended means 16+16 = 32
    3. so that i used to select the check box at OH level like Deleting data from table
    4. Now excuted the DTP it throws an Short Dump - DBIF_RSQL_TABLE_KNOWN
    Now please tell me how to resolve this ? will this option is applicable for HANA mean to say like , deleting data from table option ...
    Thanks
    Santhosh Kumar

  • Currency conversion error in SAP HANA

    Hi,
    I am new to SAP HANA and learning to create information views in HANA studio (SAP HANA SP6 on Cloudshare, HANA studio 1.0.68). I am trying to create a simple analytic view (on purchaseOrderItem table in SAP_HANA_EPM_DEMO sample database) to have GrossAmount converted to EUR.
    I added a calculated column as follows:
    When i click on "OK", i get error -
    The check box “Calculate before aggregation” has been unchecked, because the definition of the calculated column contains measures with currency conversion, restricted measures or operands with input parameters. For such a calculated column the calculation is always done after the aggregation."
    and checkbox "calculate before aggregation" get unchecked. See screenshot below:
    Please suggest what could be reason? Thanks in advance.
    Regards,
    Amit

    Hi Amit,
    If you uncheck the "Calculate before aggregation" checkbox and activate the view, you will see in the generated log that a Calc scenario is created. (a view with /olap wrapper). Due to the calc scenario, the aggregation is defined as the default behavior for the KFs and hence the calculation cannot be done before aggregation.
    By the way, I did not understand why do you need calculate before aggregation for a KF which is just a copy of another KF. If you need Gross amount in Local currency and EUR, then just perform the currency conversion without "Calculate before aggregation" checkbox. It will work.
    Regards,
    Ravi

  • SAP HANA modelling Standalone

    Hello Experts,
    We are in the process of HANA Standalone implementation and design studio as reporting tool. When I am modeling, I did not figure out answers to some of the below questions .Below are the questions. Experts, please help.
    Best way of modeling: The SAP HANA LIVE is completely built on calculation view; there are no Attribute and Analytical views. I have got different answer why there is only Calculation view and there are no Alaytic view and Attribute views. We are in SP7 latest version. This is a brand new HANA in top of non-SAP (DB2 source).  What is the best way to model this scenario, meaning, can we model everything in the Calculation view’s like SAP HANA live or do you suggest using the standard attribute, analytical and calculation views to do the data model. Is SAP moving away from AV & AT to only calculation Views to simply the modeling approach?
    Reporting: We are using the design studio as front end tool. Just for example, if we assume that we are
    Using the BW, we bring all the data in to BW from different sources, build the cubes and use the bex query. Here in bex query we will be using the restricted key figures, calculated key figures calculations etc. From the reporting wise, we have the same requirements, calculations, RKF, CKF,Sum, Avg etc. if we are Using the design studio on top of standalone HANA, where do I need to implement all these calculations? Is it in different views?  (From reporting perspective, if it’s BW system, I would have done all the calculations in BEx.)
    Universe: If we are doing all the calculations in SAP HANA like RKF. CKF and other calculations , what is the point in having additional layer of universe , because the reporting compnets cam access the queries directly on views .In one of our POC , we found that the using universe affect performance.
    Real time reporting: Our overall objective is to give a real time or close to real time reporting requirements, how data services can help, meaning I can schedule the data loads every 3 or 5 min to pull the data from source. If I am using the Data services, how soon I can get the data in HANA, I know it depends on the no of records and the transformations in between the systems & network speed. Assuming that I will schele the job every 2 min and it will take another 5 min to process the Data services job , is it fair to say the my information will be available on the BOBJ tools with in 10 min from the creation of the records.
    Are there any new ETL capabilities included in SP7, I see some additional features included in SP7. Is some of the concepts discussed are still valid, because in SP7 we have star join concept.
    Thanks
    Magge

    magge kris wrote:
    Hello Experts,
    We are in the process of HANA Standalone implementation and design studio as reporting tool. When I am modeling, I did not figure out answers to some of the below questions .Below are the questions. Experts, please help.
    Best way of modeling: The SAP HANA LIVE is completely built on calculation view; there are no Attribute and Analytical views. I have got different answer why there is only Calculation view and there are no Alaytic view and Attribute views. We are in SP7 latest version. This is a brand new HANA in top of non-SAP (DB2 source).  What is the best way to model this scenario, meaning, can we model everything in the Calculation view’s like SAP HANA live or do you suggest using the standard attribute, analytical and calculation views to do the data model. Is SAP moving away from AV & AT to only calculation Views to simply the modeling approach?
    >> I haven't read any "official" guidance to move away from typical modeling approach, so I'd say stick with the usual approach- AT, then AV, then CA views. I was told that the reason for different approach with HANA Live was to simplify development for mass production of solutions.
    Reporting: We are using the design studio as front end tool. Just for example, if we assume that we are
    Using the BW, we bring all the data in to BW from different sources, build the cubes and use the bex query. Here in bex query we will be using the restricted key figures, calculated key figures calculations etc. From the reporting wise, we have the same requirements, calculations, RKF, CKF,Sum, Avg etc. if we are Using the design studio on top of standalone HANA, where do I need to implement all these calculations? Is it in different views?  (From reporting perspective, if it’s BW system, I would have done all the calculations in BEx.)
    >> I'm not a BW guy, but from a HANA perspective - implement them where they make the most sense. In some cases, this is obvious - restricted columns are only available in Analytic Views. Hard to provide more complex advice here - it depends on your scenario(s). Review your training materials, review SCN posts and you should start to develop a better idea of where to model particular requirements. (Most of the time in typical BI scenarios, requirements map nicely to straightforward modeling approaches such as Attribute/Analytic/Calculations Views. However, some situations such as slowly-changing dimensions, certain kinds of calculations (i.e. calc before aggregation with BODS as source - where calculation should be done in ETL logic) etc can be more complex. If you have specific scenarios that you're unsure about, post them here on SCN.
    Universe: If we are doing all the calculations in SAP HANA like RKF. CKF and other calculations , what is the point in having additional layer of universe , because the reporting compnets cam access the queries directly on views .In one of our POC , we found that the using universe affect performance.
    >>> Depends on what you're doing. Universe generates SQL just like front-end tools, so bad performance implies bad modeling. Generally speaking - universes *can* create more autonomous reporting architecture. But if your scenario doesn't require it - then by all means, avoid the additional layer if there's no added value.
    Real time reporting: Our overall objective is to give a real time or close to real time reporting requirements, how data services can help, meaning I can schedule the data loads every 3 or 5 min to pull the data from source. If I am using the Data services, how soon I can get the data in HANA, I know it depends on the no of records and the transformations in between the systems & network speed. Assuming that I will schele the job every 2 min and it will take another 5 min to process the Data services job , is it fair to say the my information will be available on the BOBJ tools with in 10 min from the creation of the records.
    Are there any new ETL capabilities included in SP7, I see some additional features included in SP7. Is some of the concepts discussed are still valid, because in SP7 we have star join concept.
    >>> Not exactly sure what your question here is. Your limits on BODS are the same as with any other target system - doesn't depend on HANA. The second the record(s) are committed to HANA, they are available. They may be in delta storage, but they're available. You just need to work out how often to schedule BODS - and if your jobs are taking 5 minutes to run, but you're scheduling executions every 2 minutes, you're going to run into problems...
    Thanks
    Magge

  • Process to Upgrade SAP HANA from SPS06 to SPS07 on distributed systems

    Hi Experts,
    We have a requirement to upgrade our SAP HANA system system from SPS06 to SPS07, but I have below queries on the same:
    1) Do we need to consider any Notes before applying the upgrade for pre and post steps (as per my knowledge we have to consider post upgrade note: 1962472, are there any other too?)
    2) As per my knowledge, we just need to stop SAP app systems and then backup of HANA system
    3) Also we have a distributed environment in our landscape (three multiple hosts), what could be the upgrade approach, is it same as like single host environment (as /hana/shared will be the common to all distributed hosts)
    Could anyone clarify my questions please.
    Also for your information we have BW systems on top of HANA DB
    Thanks very much in advance !
    Kind Regards,
    Arun Reddy
    Message was edited by: Tom Flanagan

    Hi John,
    Thank you for the response, our source and target levels are (Source - 1.00.69.00.385196 Target - 1.00.74.03.392810)
    I just wanted to ask you one more quick question here, can't we upgrade the above mentioned versions in one go? or we need to upgrade in two steps like, at first upgrading from 1.00.69.00.385196 to 1.00.73.00.389160  and then to target level 1.00.74.03.392810?
    Also as you said in distributed environment we can upgrade as same like single host environment, so it same like below approach?
    1) stopping the hana database
    2) navigate to media path and then execute hdblcm in gui or command line
    3) select option upgrading existing system
    4) provide the passwords and system poped-up inputs
    5) select the target levels of HANA components
    6) once upgrade done successfully, then need to upgrade hana client in all app servers
    if my above assumption is correct, then I have couple of questions like:
    1) in which host we need to login and perform the upgrade? (as we are having three hosts- HANADB1, HANADB2, HANADB3) , is it okay we can do from any of the three hosts?
    2) Also if upgrade stopped at middle due to errors, do we have option like to start the upgrade from the point where it stop? else we need to restore the lat successful backup and then again need to perform the upgrade from starting point?
    Sorry for asking many questions, i have done the upgrade in single host environment, but not on distributed. So wanted to be make sure with all points to action the activity.
    Thanks for your help and patience in advance !
    Kind Regards,
    Arun Reddy

  • Regarding performance optimization and tuning...

    hi all,
    <b>please provide me the performance tuning scenarios and parameters of an R/3 system with Oracle..</b>
    i heartly welcome all docs and pdf links or notes related to this issue..
    please provide ur suggestions at the earliest...
    expecting ur response..
    <i>Vineeth</i>

    Hello,
    there are many SAP Notes regarding performance issues. Here are just a couple of them:
    618868
    805934
    793113
    805934
    Please also have a look at the lists of the relating Notes at the end of each Note.
    But still much more effective would be to read the book of
    <a href="http://www.sap-press.de/katalog/buecher/titel/gp/titelID-1155?GalileoSession=66220888A2.lRCIISlE">T.Schneider Performance Optimization Guide</a>.
    It's the best performance tuning guide. The course ADM315 (or BC315?) ist also very helpful.
    Regards,
    Natalia

  • Use cases for the SAP HANA Appliance for SAP FI-Accts.Payable

    Hello SAP HANA friends:
    Are there any compelling use cases for the SAP HANA Appliance for SAP FI-AP Accounts Payable module?
    CO-PA and CRM have many immediate use cases, but I'm looking for the use case for Accounts Payable,
    either from a cash flow perspective or performance.
    Any and all comments are appreciated.
    Regards,
    Joe Gonzales
    856 912 1136

    Hi,
    In SQL console please execute below query .
    GRANT SELECT ON SCHEMA <schema Name> TO "_SYS_REPO";
    BR
    Sumeet

  • SAP Netweaver Performance Tuning

    Guys,
    I am good in SAP J2EE Engine Performance Tuning. I would like to learn SAP ABAP Based Server Performance tuning end to end. Please recommend me some links/docs to start with the performance tuning for different OS like HP-UX/AIX & for Databases : Oracle/MaxDB etc.
    Thank you in advance.
    Cheers, Karthick

    Hi Karthick, for example for SAP with Oracle on AIX -->
    white paper -->
    http://www-03.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/WP100377
    But SAP with Oracle tuning are usually not very much dependent on the platform, of course there are differences but they are not much.
    Personally, I bought the SAP Performance Optimization Guide -->
    http://www.sap-press.com/products/SAP-Performance-Optimization-Guide-(5th-Edition).html?
    and SAP Database Administration with Oracle
    http://www.sap-press.com/products/SAP-Database-Administration-with-Oracle.html
    Nice books. Regards.

  • SAP HANA calculation field

    Hello,
    I have a query on the new calculated field in SAP HANA. What is the usage of this field, what are the pre-requisites and potential case when it is recommended to use such fields? Do we need to have a calculative view to use the calcluated field or can it be used directly in ABAP code?
    Kindly let me know. We are in the process of code optimization for HANA.
    Best Regards,
    Mohit

    Hey Mohit,
    can you please give a bit more details, e.g. documentation about what you refer to as "the new calculated field in SAP HANA".
    Concerning the details of usage within HANA, i.e. whether you need a calculation view etc. you might have a look at the SAP Help (Introduction - SAP HANA SQL and System Views Reference - SAP Library)) or maybe directly in the HANA SQL reference (http://help.sap.de/hana/SAP_HANA_SQL_Script_Reference_en.pdf).
    From my answer I think you can directly guess, that there's no way in directly accessing the calculated field from ABAP - except of course if you access it using native SQL (ADBC or EXEC SQL).
    For code optimization in SAP HANA (from an ABAP developers perspective), you might want to have a look at our Open SAP course (ABAP Development for SAP HANA - Dr. Jasmin Gruschke and Jens Weiler) or in the documentation given in our "Get Started" Section of ABAP for SAP HANA namely ABAP for SAP HANA Reference Scenario and ABAP for SAP HANA Reference Scenario - AS ABAP 7.4 Support Packages.
    In case you need more HANA (less ABAP) related information for your question, you might additionally contact the HANA experts on SAP HANA Developer Center.
    Cheers,
    Jasmin

  • Performance optimization

    Hi All,
    i am trying to learn sap basis. i want to know that what is performance optimization in sap basis and how it will be done. kindly help me. thanks in advance
    Regards,
    Subhash

    Hi Subhash,
    an answer to that question would easily fill a book, so I'd like to suggest one:
    SAP Performance Optimization Guide
    Thomas Schneider
    SAP Press
    ISBN 1-59229-069-8
    It's <b>the</b> performance bible and covers every aspect of performance optimization.
    Kind regards
    Dirk

  • FICO for SAP HANA

    Hello,
    I've reading a lot of documentation from SAP about the new SAP HANA. But i can't yet find the difference between SAP FICO in no-HANA Systems and Smart Finance and Simple Finance in SAP HANA sytem's.
    As far as i understood, if i migrate my FICO from no-HANA to HANA system, i would have, obligatorily, the New GL functionality and the New Asset Accounting.
    But i have two question, if anyone could assist it would be of great help:
    1. By simply migrating data would i imediatly benefit of the new architectural design? (column oriented, in-memory, total tables and index tables transformed to views such as all BSI* and BSA*) or do I need to instal an add-on (Simple Finance or Smart Finance)??
    2. I recently entered a SAP HANA machine from one of our clients, and BSIS, for example, was not a view as was expected. Was there something missing, some sort of activation?
    Thank you very much.
    Best Regards,
    Nuno Anjos

    Please read the release notes in this link SAP Simple Finance Add-On for SAP Business Suite powered by SAP HANA – SAP Help Portal Page to understand what's new in SAP FICO as compared to the earlier versions.
    Regarding your questions above, here are my thoughts:
    1. You need to install the add-on first (a software update) and then perform the database migration.
    2. Once you install the financial add-on for Suite on HANA (SAP Simple Finance Add-On for SAP Business Suite powered by SAP HANA), BSIS should appear as a DDL SQL View in SE11.

  • SAP HANA One and Predictive Analysis Desktop - Time Series Algorithms

    I have been working on a Proof-of-Concept project linking the SAP Predictive Analysis Desktop application to the SAP HANA One environment.
    I have modeled that data using SAP HANA Studio -- created Analytic views, Hierarchies, etc. -- following the HANA Academy videos.  This has worked very well in order to perform the historical analysis and reporting through the Desktop Application. 
    However, I cannot get the Predictive Analysis algorithms -- specifically the Time Series algorithms -- to work appropriately using the Desktop tool. It always errors out and points to the IndexTrace for more information, but it is difficult to pinpoint the exact cause of the issue.  The HANA Academy only has videos on Time Series Algorithms using SQL statements which will not work for my user community since they will have to constantly tweak the data and algorithm configuration. 
    In my experience so far with Predictive Analysis desktop and the Predictive Algorithms, there is a drastic difference between working with Local .CSV / Excel files and connecting to a HANA instance.  The configuration options for using the Time Series Algorithms are different depending upon the data source, which seems to be causing the issue.  For instance, when working with a local file, the Triple Exponential Smoothing configuration allows for the specification of which Date field to use for the calculation.  Once the data source is switched to HANA, it no longer allows for the Date field to be specified.  Using the exact same data set, the Algorithm using the local file works but the HANA one fails. 
    From my research thus far, everyone seems to be using PA for local files or running the Predictive Algorithms directly in HANA using SQL.  I can not find much of anything useful related to combing PA Desktop to HANA. 
    Does anyone have any experience utilizing the Time Series Algorithms in PA Desktop with a HANA instance?   Is there any documentation of how to structure the data in HANA so that it can be properly utilized in PA desktop? 
    HANA Info:
    HANA One Version: Rev 52.1
    HANA Version: 1.00.66.382664
    Predictive Analysis Desktop Info:
    Version: 1.0.11
    Build: 708
    Thanks in advance --
    Brian

    Hi,
    If you use CSV or XLS data source you will be using Native Algorithm or R
    Algorithm in SAP Predictive Analysis.
    When you connect HANA, SAP Predictive Analysis uses PAL Algorithm which runs
    on HANA server.
    Coming to your question regarding difference,
    In SAP PA Native Algorithm, we could provide the Data variable, Algorithm
    picks the seasonal information from the Data column. Both R and SAP HANA PAL
    does not support Date Column. We need configure seasonal information in
    Algorithm properties.
    R Properties
    1) Period : you need to mention the periodicity of the Data
    Monthly : (12)
    Quarter : (4)
    Custom : you can use it for week or Daily or hourly.
    2) Start Year: need to mention Start year.
    Start year is not used by algorithm for calculating Time series, but it helps
    PA to generate Visualization ( Time series chart) by simulating year and
    periodicity information.
    3) Starting Period:
    if your data is Quarterly and you have data recordings from Q2, mention 2 in
    start period.
    Example.
    If the data periodicity is Monthy and my data starts from Feb 1979, we need to provide following information,
    Period: 12
    Start year: 1979
    start Period: 2
    PAL Properties. : Same as properties defined in R.
    Thanks
    Ashok
    [email protected]

Maybe you are looking for