Performance issue on a virtual cube

Hi BW gurus,
I am working on Consolidation virtual cube and the query performance thru that cube is very bad. I know that we cannot build aggregates or partition on a virtual cube ..what should be my approach then.....
Your suggestions will be appreciated with lots of points

Hi Nick,
If you can not move out of the virtualcube option, then I think you should try to improve the performance of the virtual cube. This mainly ABAP work. You can use SE30 to analyze what parts of the code is taking too much time. Follow this steps:
1) Create a breakpoint in the function module of your virtualcube
2) Go to listcube and initiate an extraction from your virtualcube.
3) On a separate session, run SE30 to start an analysis of the extraction process for your virtual cube.
You can use the report from SE30 as a starting point on your performance optimization work.
Note: Transaction ST05 can help you determine what database calls are taking a long time.
Hope this helps.

Similar Messages

  • Performance issue in browsing SSAS cube using Excel for first time after cube refresh

    Hello Group Members,
    This is a continuation of my earlier blog question -
    https://social.msdn.microsoft.com/Forums/en-US/a1e424a2-f102-4165-a597-f464cf03ebb5/cache-and-performance-issue-in-browsing-ssas-cube-using-excel-for-first-time?forum=sqlanalysisservices
    As that thread is marked as answer, but my issue is not resolved, I am creating a new thread.
    I am facing a cache and performance issue for the first time when I try to open a SSAS cube connection using Excel (using Data tab  -> From Other Sources --> From Analysis Services) after daily cube refresh. In end users system (8 GB RAM but around
    4GB available RAM), for the first time, it takes 10 minutes to open the cube. From next run onwards, its open up quickly within 10 secs.
    We have daily ETL process running in high end servers. The configuration of dedicated SSAS cube server is 8 core, 64GB RAM. In total we have 4 cube DB - out of which for 3 is full cube refresh and 1 is incremental refresh. We have seen after daily cube
    refresh, it takes 10 odd minutes to open the cube in end users system. From next time onwards, it opens up really fast with 10 secs. After cube refresh, in server systems (32 GB RAM, around 4GB available RAM), it takes 2 odd minutes to open the cube.
    Is there, any way we could reduce the time taken for first attempt ?
    As mentioned in my previous thread, we have already implemented a cube wraming cache. But, there is no improvement.
    Currently, the cumulative size of the all 4 cube DB are more than 9 GB in Production and each cube DB having 4 individual cubes in average with highest cube DB size is 3.5 GB. Now, the question is how excel works with SSAS cube after
    daily cube refresh?
    Is it Excel creates a cache of the schema and data after each time cube is refreshed and in doing so it need to download the cube schema in Excel's memory? Now to download the the schema and data of each cube database from server to client, it will take
    a significant time based on the bandwidth of the network and connection.
    Is it anyway dependent to client system RAM ? Today the bigest cube DB size is 3.5 GB, tomorrow it will be 5-6 GB. Now, though client system RAM is 8 GB, the available or free RAM would be around 4 GB. So, what will happen then ?
    Best Regards, Arka Mitra.

    Could you run the following two DMV queries filling in the name of the cube you're connecting to. Then please post back the row count returned from each of them (by copying them into Excel and counting the rows).
    I want to see if this is an issue I've run across before with thousands of dimension attributes and MDSCHEMA_CUBES performance.
    select [HIERARCHY_UNIQUE_NAME]
    from $system.mdschema_hierarchies
    where CUBE_NAME = 'YourCubeName'
    select [LEVEL_UNIQUE_NAME]
    from $system.mdschema_levels
    where CUBE_NAME = 'YourCubeName'
    Also, what version of Analysis Services is it? If you connect Object Explorer in Management Studio to SSAS, what's the exact version number it says on the top server node?
    http://artisconsulting.com/Blogs/GregGalloway

  • Cache and performance issue in browsing SSAS cube using Excel for first time

    Hello Group Members,
    I am facing a cache and performance issue for the first time, when I try to open a SSAS cube connection using Excel (using Data tab  -> From Other Sources --> From Analysis Services) after daily cube refresh. In end users
    system (8 GB RAM), for the first time, it takes 10 minutes to open the cube. From next run onwards, its open up quickly within 10 secs.
    We have daily ETL process running in high end servers. The configuration of dedicated SSAS cube server is 8 core, 64GB RAM. In total we have 4 cubes - out of which for 3 is full cube refresh and 1 is incremental refresh. We have seen after
    daily cube refresh, it takes 10 odd minutes to open the cube in end users system. From next time onwards, it opens up really fast with 10 secs. After cube refresh, in server systems (16 GB RAM), it takes 2 odd minutes to open the cube.
    Is there, any way we could reduce the time taken for first attempt ?
    Best Regards, Arka Mitra.

    Thanks Richard and Charlie,
    We have implemented the solution/suggestions in our DEV environment and we have seen a definite improvement. We are waiting this to be deployed in UAT environment to note down the actual performance and time improvement while browsing the cube for the
    first time after daily cube refresh.
    Guys,
    This is what we have done:
    We have 4 cube databases and each cube db has 1-8 cubes.
    1. We are doing daily cube refresh using SQL jobs as follows:
    <Batch xmlns="http://schemas.microsoft.com/analysisservices/2003/engine">
    <Parallel>
    <Process xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ddl2="http://schemas.microsoft.com/analysisservices/2003/engine/2" xmlns:ddl2_2="http://schemas.microsoft.com/analysisservices/2003/engine/2/2" xmlns:ddl100_100="http://schemas.microsoft.com/analysisservices/2008/engine/100/100" xmlns:ddl200="http://schemas.microsoft.com/analysisservices/2010/engine/200" xmlns:ddl200_200="http://schemas.microsoft.com/analysisservices/2010/engine/200/200">
    <Object>
    <DatabaseID>FINANCE CUBES</DatabaseID>
    </Object>
    <Type>ProcessFull</Type>
    <WriteBackTableCreation>UseExisting</WriteBackTableCreation>
    </Process>
    </Parallel>
    </Batch>
    2. Next we are creating a separate SQL job (Cache Warming - Profitability Analysis) for cube cache warming for each single cube in each cube db like:
    CREATE CACHE FOR [Profit Analysis] AS
    {[Measures].members}
    *[TIME].[FINANCIAL QUARTER].[FINANCIAL QUARTER]
    3. Finally after each cube refresh step, we are creating a new step of type T-SQL where we are calling these individual steps:
    EXEC dbo.sp_start_job N'Cache Warming - Profit Analysis';
    GO
    I will update the post after I receive the actual im[provement from UAT/ Production environment.
    Best Regards, Arka Mitra.

  • Issue populating Enhaced Virtual Cube Fields

    Hi Gurus,
    Please help me with the below Issue. I tried a lot on SDN before writing this Post.
    Architecture:
    I have a Standard Multi-provider - "0FIAR_M30". It has got a Standard Cube: 0FIAR_C30 & a virtual Cube: 0FIAR_R30.
    Cube: 0FIAR_C30 is getting Data from R/3 DS: 0FI_AR_3 and Cube: 0FIAR_R30 is getting data from DS: 0FI_AR_30.
    I enhaced the two DS for fields: Plant, Sales Org, Created By, Sales Dist.
    Data is populating properly in the enhaced fields for standard Cube, but in case of Virtual cube, data is getting loaded only till PSA and but when i am trying to display data for Virtual Cube. No data is getting Populated in the enhaced Fields.
    Options Tried:
    1. I used Z Info-objects with template of the standard Info-objects. I made all the Attributes of the Z-info-objects as Display wherever the attributes are navigational.
    I used the below posts while investigating on the issue:
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/a00ba5ee-9840-2d10-f385-933182d62b12?QuickLink=index&overridelayout=true
    Please provide your suggestions n Inputs.
    Thanks
    Sameer

    Hi Sameer,
    We are also facing the same issue.
    Could you please provide how you have corrected the issue
    Thanks,
    Jyothi.

  • BCS Virtual Cubes Performance Tuning

    Hi All,
    We are working on improving the performanc tuning of queries on BCS Virtual Cubes (with services).
    Any specific changes (from RAM to specific properties on queries, Virtual Cubes with services) that you have seen working in improving the performance in yuor environemnt is greately appreciated.
    Thanks,
    - Shashi

    Thanks a lot Marc,
    We are on NW2004, with the following support pack levels
    SAP_BW     350     0016
    FINBASIS     300     0012
    BI_CONT     353     0008
    SEM-BW     400     0012
    I have checked the service market place and the current SP's available are:
    SAP_BW     350     0019
    FINBASIS     300     0015
    BI_CONT     353     0013
    SEM-BW     400     0015
    Which Service Packs that you suggest us to go with for performance issues on BCS Virtual Infoprovider queries?
    Thanks,
    - Shashi

  • BW BCS cube(0bcs_vc10 ) Report huge performance issue

    Hi Masters,
    I am working out for a solution for BW report developed in 0bcs_vc10 virtual cube.
    Some of the querys is taking more 15 to 20 minutes to execute the report.
    This is huge performance issue. We are using BW 3.5, and report devloped in bex and published thru portal. Any one faced similar problem please advise how you tackle this issue. Please give the detail analysis approach how you resolved this issue.
    Current service pack we are using is
    SAP_BW 350 0016 SAPKW35016
    FINBASIS 300 0012 SAPK-30012INFINBASIS
    BI_CONT 353 0008 SAPKIBIFP8
    SEM-BW 400 0012 SAPKGS4012
    Best of Luck
    Chris
    BW BCS cube(0bcs_vc10 ) Report huge performance issue

    Ravi,
    I already did that, it is not helping me much for the performance. Reports are taking 15 t0 20 minutes. I wanted any body in this forum have the same issue how
    they resolved it.
    Regards,
    Chris

  • Report issue on Virtual cube with services

    Hi,
    I'm following a how-to document to report on a virtual cube using a custom function module and everything seems to be OK except it is dumping at the following line which is included in the How To document "How Tou2026Implement a virtual InfoProvider with Services".
    Both <l_s_data> and E_T_DATA  are of same data types as discussed in the how-to document.
    append <l_s_data> to E_T_DATA
    These are the declarations which are same as HowTo document
    E_T_DATA TYPE  STANDARD TABLE
    FIELD-SYMBOLS: <l_s_data> TYPE ANY.
    Did any one face this issue. Greatly appreciate any insight into this issue. We are on BW 3.0B.
    the dump i'm getting is ...
    Data objects in a Unicode program are not convertible.
    Error analysis                                                                               
    The statement    "MOVE src TO dst"   requires the operands "dst" and "src" to be comvertible.                                                                             
    Since this statement occurs in a Unicode program, the special  convertibility rules for Unicode programs apply.
    Thanks,
    Hari Immadi
    http://immadi.com
    SEM BW Analyst

    Hi,
    In the debugging this is what I found out.. Any further insight?
    Virtual cube has only 1 characteristic Z_CRNUM & only 1 keyfigure Z_SR
    Report has only Z_CRNUM & Z_SR
    <L_S_DATA> has following record & value
    component      type     length     contents
    K____504     P     9     1.000
    S____504     C     18     16EA4D-R
    K____504 is the technical name of Z_SR in virtual provider.
    S____504 is the technical name of Z_CRNUM in virtual provider.
    E_T_DATA (type standard table) has following structure with a total of 5 fields.
    S____504    K____504           &KEYEND 1ROWCOUNT   Z_SR
    Regards
    Hari Immadi
    http://immadi.com
    SEM BW Analyst

  • Inventory Cube performance Issue

    Hi All,
    This is not something new, but an old issue traditionally with this cube. I have customized 0IC_C03 for my requirement and having serious performance issues. It has 0CALDAY, 0MATERIAL, 0PLANT as non-cumulative value parameters. i hav eadded movement types (temproarily for validation purpose). But my query always timed out, unless I specify the material. There are close to 40K materials are being maintained. The values are all fine between ECC and BI afterdata loads. So we are thinking may be snap shot approach would hlp us resolve the performanc eissues.
    Anybody has implementeted snap-shot approach for inventory? I know it is a loading issue, but we think we could deal with that rather than performanc eissue when the users execute the query.
    if anybody has done it, could ou provide the steps?
    Thanks,
    Alex.

    Hi Jameson - Thanks for your response.
    We thought that would be the case. Have raised a SR with oracle and they are investigating on it. We have also sent an EIFF file to Oracle for investigation.
    Both the DBs are in the same environment (AIX 6.1) and DBAs have confirmed both the DBs have the same system parameters.
    Even if we keep aside comparing to 11.2.0.1, for some reason 11.2.0.3 seems to be very slow. Even a simple cube (2 Dim and 2 Measures) with 9K records takes around 15 min to get refreshed and it takes ages to view the data.
    Havent generated the AWR report, will see if we can do the same.
    rgds,
    Prakash S

  • Building a new Cube Vs Restricted Key figure in Query - Performance issue

    Hi,
    I have a requirement to create  a OPEX restricted key figure in Query. The problem is that the key figure should be restricted to about 30 GL Accounts and almost 300 Cost centers.
    I do not know if this might cause performance issue in the query. At the moment, I am thinking of creating a new OPEX cube and load only those 30 GL Accounts, 300 cost  centers and Amount. and  include OPEX  in multiprovider in order to get OPEX
    amount in the report.
    whats the best solution - creating OPEX restricted key figure or OPEX cube ?
    thanks,
    Bhat

    I think you should go for cube as all the restrcited key figure are calculated at OLAP runtime so it will definitely affect the query performance.There are a lot of costcenter for which you have to restrict it so definitely during the runtime of query it will take a lot of time to fetch tha data from infoprovider.Its better that you create a cube with the restrictions and include it in MP.It will definitely save a lot of time during query execution
    Edited by: shyam agarwal on Feb 29, 2012 10:26 AM

  • Cube Refresh Performance Issue

    We are facing a strange performance issue related to cube refresh. The cube which used to take 1 hr to refresh is taking around 3.5 to 4 hr without any change in the environment. Also, the data that it processes is almost the same before and now. Only these cube out of all the other cubes in the workspace is suffering the performance issue over a period of time.
    Details of the cube:
    This cube has 7 dimensions and 11 measures (mix of sum and avg as aggregation algorithm). No compression. Cube is partioned (48 partitions). Main source of the data is a materialized view which is partitioned in the same way as the cube.
    Data Volume: 2480261 records in the source to be processed daily (almost evenly distributed across the partition)
    Cube is refreshed with the below script
    DBMS_CUBE.BUILD(<<cube_name>>,'SS',true,5,false,true,false);
    Has anyone faced similar issue? Please can advise on what might be the cause for the performance degradation.
    Environment - Oracle Database 11g Enterprise Edition Release 11.2.0.3.0
    AWM - awm11.2.0.2.0A

    Take a look at DBMS_CUBE.BUILD documentation at http://download.oracle.com/docs/cd/E11882_01/appdev.112/e16760/d_cube.htm#ARPLS218 and DBMS_CUBE_LOG documentation at http://download.oracle.com/docs/cd/E11882_01/appdev.112/e16760/d_cube_log.htm#ARPLS72789
    You can also search this forum for more questions/examples about DBMS_CUBE.BUILD
    David Greenfield has covered many Cube loading topics in the past on this forum.
    Mapping to Relational tables
    Re: Enabling materialized view for fast refresh method
    DBMS CUBE BUILD
    CUBE_DFLT_PARTITION_LEVEL in 11g?
    Reclaiming space in OLAP 11.1.0.7
    Re: During a cube build how do I use an IN list for dimension hierarchy?
    .

  • ASO Cube Performance Issue

    We have been working on 2 ASO cubes, performance was great. No modification hasn't been done ever since, but we have been experiencing performance issue now. What could be the possible cause or how I could resolve this. Thank you.

    'Performance issue' isn't very descriptive - performance of what? Query? Data load? Aggregation? Restructure?
    As a start, has the volume of data (input data cells) been increasing significantly?

  • Query performance on remote/virtual cube

    All:
    We are on BW 3.5.
    Can anyone suggest anything about improving query performance on remote/virtual cubes? Analysis shows query performance is sufferring at the database level.
    I am looking for some advise beyond hardware and database parameters. It seems current hardware and database parametrs work fine with basis cubes.
    Another solution is datamart.But can anything be done before/without going to datamart?
    Any help will be appreciated.
    Thanks

    Hi,
    In this case try to find where more time is utilized by using the ST03.
    if there is more time consuming in front end, rearrange the query by using less in the row and using more free chars and filter areas. using the variable also will work.
    use reporting agent and schedule the query in background to fill the cache and rerun the query.
    Reg,
    Vishwa

  • Virtual chars performance issues ?

    Hello BW Experts,
    Does virtual chars have performance issues ?
    It has small effect or an adverse effect ?
    please suggest.
    Thanks,
    BWer

    hi BWer,
    it depends on the 'logic' in virtual chars exit, may have performance issue e.g if you select/populate from
    table, but may not if only filled with other characteristic value or only simple logic.
    the 'how to' doc sample from sap market place have no performance issue.
    hope this helps.

  • Service Manager 2012 SP1 consoles hanging or slow performance issue in Virtual Environment

    Hii,
    We are facing SCSM SP1 console performance issue which is utilizing high CPU and working deadly.
    For information, our SCSM is in Virtual environment via Hyper-V.
    When running the console over an RDP session to a Hyper-V virtual machine, we have to be careful not to maximize the console so that it will remain fast.  If we maximize on the VM, the console is so slow as for it to be unusable.
    Can someone share his experience, please?
    Regards, Syed Fahad Ali

    Hi Sayed,
    This is a bug and hopefully Microsoft team will solve it soon. if you can to vote for this bug here
    https://connect.microsoft.com/WindowsServer/feedback/details/810667/scsm-console-consumes-a-lot-of-cpu-when-opened-maximized-on-work-item-view-like-all-incidents
    Mohamed Fawzi | http://fawzi.wordpress.com

  • Performance issues in bw

    Hi All,
    What is buffering number?How it is useful in performance issue? Tell me the option where it is available? To set this what are navigational steps?
    Thanks inadvance.
    Yogeswar

    Hi Yogi,
    A nice weblog by Vikas Please do check this.on number range buffering,
    /people/vikash.agrawal/blog/2006/04/05/load-lots-of-data-147faster148-with-buffering-number-range
    Check these links.
    FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap
    https://www.sdn.sap.com/irj/sdn/developerareas/bi?rid=/webcontent/uuid/b4674415-0b01-0010-ae81-deb009860b7e [original link is broken]
    following are the links that may help you
    http://help.sap.com/search/highlightContent.jsp
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/biw/s-u/sap%20bw%20business%20planning%20and%20simulation%20-%20how%20to%20guides%20list.htm
    http://help.sap.com/search/highlightContent.jsp
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/cccad390-0201-0010-5093-fd9ec8157802
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/afbad390-0201-0010-daa4-9ef0168d41b6
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/b7bdde90-0201-0010-26b1-dcda5e0b394d
    How to improve performance in reporting side?
    Query Performance Techniques:
    1. Check Query properties—Use RSRT tcode
    2. Check whether cube is compressed
    3. Optimize query definition
    4. Analyze query execution
    5. Check for additional indexes
    6. Archive unwanted data
    7. Check for partitioning options
    8. Check for additional aggregates ( Consider DB ratio and KPI ratio)
    9. Check for parallelization options
    10. Use Nav attributes instead of hierarchies, use free char and filters.
    Possible causes for the performance :
    A) High Database Runtime
    B) High OLAP Runtime
    C) High Frontend Runtime
    Depending upon your analysis
    A)Strategy - High Database Runtime
    Check if an aggregate is suitable (use All data to get values "selected records to transferred records", a high number here would be an indicator for query performance improvement using an aggregate)
    Check if database statistics are update to data for the Cube/Aggregate, use Tcode RSRV output (use database check for statistics and indexes)
    Check if the read mode of the query is unfavourable - Recommended (H)
    B)Strategy - High OLAP Runtime
    Check if a high number of Cells transferred to the OLAP (use "All data" to get value "No. of Cells")
    a) Use RSRT technical Information to check if any extra OLAP-processing is necessary (Stock Query, Exception Aggregation, Calc. before Aggregation, Virtual Char. Key Figures, Attributes in Calculated Key Figs, Time-dependent Currency Translation) together with a high number of records transferred.
    b) Check if a user exit Usage is involved in the OLAP runtime?
    c) Check if large hierarchies are used and the entry hierarchy level is as deep as possible. This limits the levels of the hierarchy that must be processed.
    C)Strategy - High Frontend Runtime
    1) Check if frontend PC are within the recommendation (RAM, CPU Mhz)
    2) Check if the bandwidth for WAN connection is sufficient.
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1e553368-0601-0010-49ab-c429607f3eb3
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/5401ab90-0201-0010-b394-99ffdb15235b
    check this, you can download lot of performance materials
    Business Intelligence Performance Tuning [original link is broken] [original link is broken] [original link is broken]
    and e-learning -> intermediate course and advance course
    https://www.sdn.sap.com/irj/sdn/developerareas/bi?rid=/webcontent/uuid/fe5b0b5e-0501-0010-cd88-c871915ec3bf [original link is broken]
    e.g
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/media/uuid/10b589ad-0701-0010-0299-e5c282b7aaad
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/media/uuid/d9fd84ad-0701-0010-d9a5-ba726caa585d
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/media/uuid/8e6183ad-0701-0010-e083-9ab1c6afe6f2
    performance tools in bw 3.5
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/media/uuid/07a4f070-0701-0010-3b91-a6bf7644c98f
    (here also you can download the presentation by righ click the disk drive icon)
    Check the following links,
    FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap
    Business Intelligence Performance Tuning [original link is broken] [original link is broken] [original link is broken]
    http://help.sap.com/saphelp_nw04/helpdata/en/06/b5f8926ba22b45bc9eaa589f1c835b/content.htm
    Some bw docs/ performance material
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1955ba90-0201-0010-d3aa-8b2a4ef6bbb2
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/3a699d90-0201-0010-bc99-d5c0e3a2c87b
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/4c0ab590-0201-0010-bd9a-8332d8b4f09c
    and don't miss bw performance knowledge centre, there are e-learning
    Business Intelligence Performance Tuning [original link is broken] [original link is broken] [original link is broken]
    Hope this Helps.
    <removed>
    Regards,
    Ravikanth.

Maybe you are looking for

  • MOVED: msi s270 series unable to use external HDD casing

    This topic has been moved to Other MSI products. https://forum-en.msi.com/index.php?topic=92843.0

  • Downloading Query result set

    Hi forum, We have a requirement to download query result set and distribute it as a periodic job. Can open hub be used for this purpose? Thanks in advance. Regards,

  • Dreamweaver crashes/freezes when doing find in site

    When I use "find" it works fine for a single document but, if I choose "find next" after setting the search to the entire site (or a folder), it performs the search and starts to bring up the correct file, then it crashes.  Behavior is the rainbow wh

  • Layout:render tag attributes

    Hi all, I'm trying to make a layout for portal having a HTML table with 4 coloumns each 25% wide, and a second row with 25% / 50% / 25% using colspan="2" in the middle section. In the default layouts I have examples for using four coloumns using rend

  • Can a WCF Service return a Datatable ??

    The snippet shown below is not working ??? Why is it  namespace MLA_CRM_Service     [DataContract]     public class MLAContact : DataTable  //(Cant i do this ?)         [DataMember]         private string MLAMemberType;         [DataMember]         p