Usage Tracking performance impact

Hi all,
I just started to implement the OBIEE usage tracking functionalities.
Anyone implemented it and call share some experience on the performance impact with/without usage tracking enabled??
Thanks
BCHK

If you use direct_insert option impact will be minimized. In general usage tracking captures report run activity and logs into table, so there is overhead on the system to insert to table or write to file. trade off is the information you get with usage tracking..
hope this helps..
Edited by: Kasyap on Mar 23, 2013 10:32 PM

Similar Messages

  • Change Tracker Performance Impact

    MDM: 7.1
    CE: 7.2
    ERP: 6.0 EHP4
    Hi,
    We are currently using CE/BPM based central master data management. A custom application is being developed for collaborative master data authoring.
    As part of the MDM configuration, the client wishes to track changes on MDM records for audit purposes. We are looking at the MDM Change Tracking facility provided by SAP but not sure about the performance impact it will have on the MDM server.
    We have over 300 attributes for the object we wish to track for changes. Not all attributes will change all the time but it is expected that the overall number of changes every month will be over 1000, each change including approx 20-30 fields. The number of users is expected to be approx 15 initially but will increase over time.
    I have seen on SCN forums people talking about potential performance degrade by enabling change tracking. Has anyone actually experienced performance degrade due to enabling change tracking for MDM records? If so, have you tried any means to keep the impact low e.g. by allocating more resources to MDM server?
    Thanks and regards,
    Shehryar
    Edited by: Shehryar Khan on Dec 2, 2010 1:39 PM

    Hi,
    Change histoey does have impact on system performance. this can be controlled via regular archiving of Change History database. Lets say, Change history table data older than 3 months can be stored in another repository.
    Second option could be to export all Change History data(Older than 1-2 days) to a BI system(using regular scheduled job), and have change hisotry report there. It will bring drastci change in MDM System performance. 1-2 days data can be viewed from MDM Change history database itself.
    Regards,
    Prashant
    Edited by: Prashant Malik on Dec 3, 2010 1:58 AM

  • Usage Tracking - Cant perform insert operation inside a READ ONLY trans

    Hello,
    I have been experiencing the following error message when trying to install a row in our s_nq_acct table for USAGE TRACKING:
    [nQSError: 16001] ODBC error state: S1000 code: 1456 message: [Oracle][ODBC][Ora]ORA-01456: may not perform insert/delete/update operation inside a READ ONLY transaction.
    - I am able to manually insert rows via SQL*Plus whilst logged on as the same user as the Connection Pool
    - The Read Only option is not ticked for the ODBC
    Can anyone please help me with this?
    Thanks,
    Rhys

    Hi rnm1978,
    You were right!
    I was being a bit retarded and had to simply switch the call interface on the connection pool to use the OCI and ensure the dta source name was that of the TNS names entry and not the system DSN.

  • How can I optimize my hard disk drive usage and performance in Windows 8 or Windows 7?

    QuestionHow can I optimize my hard disk drive usage and performance in Windows 8 or Windows 7?
    AnswerThere are a few simple steps you can take to ensure your hard disk drive is used optimally.
    Use Toshiba HDD Protection
    Many Toshiba laptops come with a program called Toshiba HDD Protection pre-installed. This program helps to protect your hard disk drive from being damaged due to falls or impacts. By default, it should already be enabled. You might be tempted to lower the detection levels in this application, but doing so could cause your hard disk drive to be damaged. Remember that while the application can reduce the chance of damage, you should still avoid allowing the laptop to fall or suffer rapid impacts.
    For more information on this utility, see the following article:
    TOSHIBA HDD Protection
    Optimize the drive
    Windows 8 and Windows 7 optimize hard disk drives automatically through a process called defragmentation. Unless you've disabled this, you don't need to do anything. If you have disabled this and want to run the process, you can still do so.
    In Windows 8, search for "Defrag" at the Windows Start screen and select "Defragment and optimize your drives."
    In Windows 7, search for "Defrag" in the Start Menu's search field and select "Disk defragmenter."
    You can use this tool to optimize your hard disk drives, allowing Windows to find needed files faster.
    Remove items from startup
    Some applications run automatically when Windows starts. This can add additional functionality, but it also decreases the performance of your computer. Sometimes you might want to disable certain programs from starting automatically.
    In Windows 8, search for "Task Manager" at the Start screen. Select the "Startup" tab. Select an application you'd like to disable from starting automatically and then click the "Disable" button in the lower-right.
    In Windows 7, type "msconfig" in the Start Menu's search field and press ENTER. Uncheck the boxes next to applications you'd like to disable from starting automatically.
    You should be sure of the purpose of an application before disabling it from starting automatically. Some applications might be important. If in doubt, you might consider searching on the Web to discover more information about a program. Remember that if you find that you disabled something vital, you can always re-enable it.
    For more information, please see the following video:

    QuestionHow can I optimize my hard disk drive usage and performance in Windows 8 or Windows 7?
    AnswerThere are a few simple steps you can take to ensure your hard disk drive is used optimally.
    Use Toshiba HDD Protection
    Many Toshiba laptops come with a program called Toshiba HDD Protection pre-installed. This program helps to protect your hard disk drive from being damaged due to falls or impacts. By default, it should already be enabled. You might be tempted to lower the detection levels in this application, but doing so could cause your hard disk drive to be damaged. Remember that while the application can reduce the chance of damage, you should still avoid allowing the laptop to fall or suffer rapid impacts.
    For more information on this utility, see the following article:
    TOSHIBA HDD Protection
    Optimize the drive
    Windows 8 and Windows 7 optimize hard disk drives automatically through a process called defragmentation. Unless you've disabled this, you don't need to do anything. If you have disabled this and want to run the process, you can still do so.
    In Windows 8, search for "Defrag" at the Windows Start screen and select "Defragment and optimize your drives."
    In Windows 7, search for "Defrag" in the Start Menu's search field and select "Disk defragmenter."
    You can use this tool to optimize your hard disk drives, allowing Windows to find needed files faster.
    Remove items from startup
    Some applications run automatically when Windows starts. This can add additional functionality, but it also decreases the performance of your computer. Sometimes you might want to disable certain programs from starting automatically.
    In Windows 8, search for "Task Manager" at the Start screen. Select the "Startup" tab. Select an application you'd like to disable from starting automatically and then click the "Disable" button in the lower-right.
    In Windows 7, type "msconfig" in the Start Menu's search field and press ENTER. Uncheck the boxes next to applications you'd like to disable from starting automatically.
    You should be sure of the purpose of an application before disabling it from starting automatically. Some applications might be important. If in doubt, you might consider searching on the Web to discover more information about a program. Remember that if you find that you disabled something vital, you can always re-enable it.
    For more information, please see the following video:

  • Error in implementing Usage Tracking in OBIEE

    Hi All,
    I wants to implement Usage Tracking in my project. I have done all the steps as per link http://obiee101.blogspot.com/2008/08/obiee-setting-up-usage-tracking.html
    But i am getting following error in NQServer.log file----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
    [59048] Usage Tracking encountered an insert statement execution error. This error has occurred 1 times and resulted in the loss of 1 insert statements since this message was last logged.
    [nQSError: 16001] ODBC error state: S1000 code: 1456 message: [Oracle][ODBC][Ora]ORA-01456: may not perform insert/delete/update operation inside a READ ONLY transaction.
    [nQSError: 16015] SQL statement execution failed.
    I have also check by changing the Call_Interface in Connection Pool of Admin Tool from ODBC 3.5 to OCI 10g. By this chage i get ollowing error:
    [59048] Usage Tracking encountered an insert statement execution error. This error has occurred 1 times and resulted in the loss of 1 insert statements since this message was last logged.
    [nQSError: 17001] Oracle Error code: 3114, message: ORA-03114: not connected to ORACLE
    at OCI call OCIStmtExecute.
    [nQSError: 17011] SQL statement execution failed.
    Please help....
    Thanks in Advance...
    Regards,
    Avi

    problem is solved....
    Actually in NQServer.ini file i was using actual Database name instead of Admin Tool Physical Layer name at PHSICAL TABLE NAME: <database>.<schema>.<tablename>
    and at CONNECTION POOL: <database name>.<connection pool>
    thanks...

  • Usage Tracking - Access problem when Authentication Mode = Windows

    Hi Everyone,
    I´m working on UPK Usage Tracking configuration, in order to provide the finished training material.
    1) In Server01 (on Window Server 2003) the UPK Usage Tracking is installed
    2) In Server02 (also on Windows Server 2003) the Usage Tracking database is installed
    3) By accessing the configuration file (http://Server01/ODSTrack/configuration/setup.aspx) on Server01,I setup the Authentication Mode = Forms
    Note: The rest of the configurations were done.
    4) Once the configuration from step 3 is done, I execute the traning material (on Server01) from another node of the windows network
    and as a result I´m able to perform it.
    5) I access the Statistc data on Server01 by accessing the file (http://Server01/ODSTrack/admin/default.aspx)and I´m also able to see the results.
    6) When I execute the step 3 but with Authentication Mode = Windows, and including the GROUP name (windows group specially created for this goal where my user is included),
    - I still have access to the training material (step 4)
    - I have NO aceess to the statistics data any more (step 5) and the following message is display
    "You do not have permission to access this page. Please contact your Usage Tracking Server Administrator to update your permissions. "
    I don´t know what else I can do, and I wonder if some other configurations need to be done at windows network and/or explorer lever or any other.
    Any help would be appreciated.
    Best Regards//
    Rubén Zamudio

    Hi All,
    This problem was solved by reconfiguring Usage Tracking in the authentication method (was anonymous and the solution was Windows integrated).
    It is important to count on people from your organization working on Networks with some knowledge in IIS.
    Best Regards
    Ruben

  • Performance Impact When Using SNC Communication

    Hello,
    Does anybody know if and how much performance impact there is if we use SNC for communication between the SAP Server and SAPGUI?
    I think there are two areas that may be impacted; Network and server CPU.
    For network load, I did find a part in "Front-End Network Requirements for SAP Business Solutions" document saying "overhead of roughly 350 bytes per user interaction step" but it does not specify the type of encryption.  I wonder if there is any other info on this?
    For CPU impact, how much overhead should I consider for sapgui access?
    I see no field for this in the quicksizer and I can't seem to find any white papers on this subject.
    Thank you in advance.

    >
    Peter Adams wrote:
    > Ken,
    >
    > if you plan to use SAPcryptlib for SNC between SAP servers, then you should use a SAPcryptolib-compatible solution for the SNC communication between SAPGUI and SAP server, and there is only one vendor who can provide this. Let me know, if you need help finding it. My contact information is in my SDN business card.
    Just so Kan is clear - It is not legal to use the SAP cryptolib provided by SAP for SNC between SAP GUI and SAP servers, so if x.509 is the desired mechanism you need to purchase additional software from the company which Peter works for to provide SAP GUI SNC-based SSO. I think instead, Kan might be using the free SAP supplied SNC Kerberos library, which is why I asked him to confirm this in my last post. I doubt he is interested to buy any third party software.
    > As to the performance discussion: first of all, yes, there will be a small performance impact if SNC is used (no matter which type or implementation), but from our experience with many actual SNC implementations, I can state that this is practically not relevant. It is not noticeable by users. There were never any performance discussions with customers. See also SAP Note 1043694.
    I agree with this - the performance impact is not noticed by users, but the system managers who look after the servers where SAP is installed, and the team responsible for the network need to be aware of any differences (if any) when SNC is turned on and when SNC is turned off. I think this is why Kan is asking these questions, not because he is concerned about users noticing any difference when they logon to SAP.
    > Just a first quick comment on certain statements above: Tim's arguments for proving his overall statement are not conclusive from my perspective. Nor do I think his overall statement itself is correct.
    The facts I mentioned are well known facts, e.g. symmetric crypto is far better from performance point of view than asymmetric. I know the examples I have shown which I found when doing a quick google search were not conclusive, but they were shown as initial examples, not necessarily the best examples. This is why I specifically mentioned that if you search in google yourself you will see many more references where comparisons are done between Kerberos (e.g. symmatric) compared with PKI (e.g. asymmetric).
    > First of all, he only selects one aspect of performance - CPU impact of encryption algorithms.
    No, I didn't. Some of the examples I referred to also discuss other differences. I also mentioend other differences such as memory and what protection level is used when configuring SNC.
    > But for a true comparison, you'd have to look at all relevant aspects (latency, network overhead, ...).
    Yes, I agree. No doubts here.
    >Network performance overhead is usuallly worse with Kerberos than with PKI.
    This is not true. When SAP is using SNC, the GSS-API standard is used and so the only network communication involves SAP software sending a standard GSS token from the workstation to the SAP server, and this GSS token is often about the same size, regardless of which mechanism is used, so any network performance differences are not related to the mechanism, but more related to the complexity of the cryptography used on each end (mostly on the server side).
    >Second, you need to look at the specific usage scenario. For example, the first report referenced by Tim is an analysis about different Token Profile mechanism for WS Security, for one specific implementation. This does not allow to draw any conclusion for the SNC use case in general, and for sure not for a specific implemenation. It does not take the overhead for the encryption of the message content into account. Third, Tim associates PKI exclusively with asymmetric encryption. Yes, it is well known that asymmetric algorithms are slower than symmetric ones, but it is also well known that the encryption of the message content (by far the majority of the data) happens with symmetric encryption algorithms in the PKI scenario. With PKI-based SNC, you can even select a symmetric algorithm and use a more performant one that the ones that Kerberos prescribes.
    Kerberos works with many different symmetric algorithms as well, so mentioning that the alg is selectable is not relavent to any comparison.
    > To summarize, I will try and collect facts that will support the opposite point of view. From our practical experience, the performance overhead is not relevant, and criteria like consistency with SAPcryptolib, strength of security, ease of administration, choice of authentication and encryption mechanism, etc. are much more important.
    >
    > Peter

  • Can we improve performence using usage tracking

    hi all
    wat is the use of usage tracking in obiee
    by using this can we improve report performance
    Thanks
    Sreedhar

    Hi,
    We can use it to track the performance but not to improve the performance...
    Check this....http://sranka.wordpress.com/2010/06/22/the-myth-usage-tracking-measures-for-tracking-the-performance-of-report-retrieval-time/
    In order to improve performance...http://obieetalk.com/obiee-performance-tuning
    http://www.rittmanmead.com/2008/11/28/thoughts-on-obiee-performance-optimization-diagnostics/
    Regards,
    Srikanth

  • Error with OBIEE Usage Tracking

    Hi,
    I have set up Usage Tracking in Oracle BI Enterprise Edition with the Direct Insert Option. I have created the S_NQ_ACCT table in an Oracle schema that is referenced as a connection pool in the Oracle BI Repository physical layer. When I run my requests from Oracle BI Answers and check my server logs (NQServer.log), I get the following message:
    [59048] Usage Tracking encountered an insert statement execution error. This error has occurred 9 times and resulted in the loss of 9 insert statements since this message was last logged.
    [nQSError: 16001] ODBC error state: S1000 code: 1456 message: [Oracle][ODBC][Ora]ORA-01456: may not perform insert/delete/update operation inside a READ ONLY transaction.
    [nQSError: 16015] SQL statement execution failed.
    2008-04-25 04:34:18
    [59048] Usage Tracking encountered an insert statement execution error. This error has occurred 9 times and resulted in the loss of 9 insert statements since this message was last logged.
    [nQSError: 16001] ODBC error state: S1000 code: 1456 message: [Oracle][ODBC][Ora]ORA-01456: may not perform insert/delete/update operation inside a READ ONLY transaction.
    [nQSError: 16015] SQL statement execution failed.
    Does anyone know how to fix this problem?
    Regards,
    Peter

    Thanks for the response.
    You're right. The response seems to imply that the account does not have the authorization. However, the account I am using to access the database from OBIEE Repository owns the schema where the S_NQ_ACCT table is stored. So I am puzzled as to why this response is generated.

  • Usage Tracking Customization

    Hello,
    Can we customize the usage tracking tables?
    I want to capture how long a user had a session open. Is this achievable?
    If yes, how?
    Thanks,
    Deep

    You said :
    The query used to get trimmed outI thought that you may use a trigger on the table "OBI"."S_NQ_ACCT".
    Check here for a before insert trigger :
    http://www.techonthenet.com/oracle/triggers/before_insert.php
    Or as example,
    connect you on oracle on the OBI schema and start this code :
        CREATE OR REPLACE TRIGGER S_NQ_ACCT_before_insert
        BEFORE INSERT
            ON  S_NQ_ACCT
            FOR EACH ROW
           BEGIN
            -- Substr
            :new.QUERY_TEXT := substr(:old.QUERY_TEXT,1,1024);
        END;Then for each row, a substr will be performed.
    Success
    Nico

  • How to handle Integrated Configuration performance impact on AAE/Java AS

    Hi there,
    Recently I have moved  a configuration scenario from standard flow involving both ABAP and Java stacks, to Integrated Configuration usage. Undoubtedly, this will increase the load on AAE/Java stack. However, do you have link to some clear (official - even better) guidelines - what configurational changes should be done on Java side in order to handle the performance impact of such transition?
    Best Regards,
    Lalo

    Hi Lalo,
    In fact, using AAE generates no traffic in ABAP stack at all (it is ommited when processing a message), while the traffic in Java stack should be lower than for normal scenario. The performance should be noticeably better, thanks to smaller number of persistence steps and no costly HTTP connections between stacks. For more details, please refer to this document:
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/2016a0b1-1780-2b10-97bd-be3ac62214c7
    Important quotation from this document:
    Since the Integration Engine is bypassed for local message processing in the AAE, the resource consumption both in memory and CPU is lower. This leads to higher message throughput, and faster response times which especially is important for synchronous scenarios.
    Moreover, have a look at this document, especially its beginning, for details about the architecture of AAE processing:
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/70066f78-7794-2c10-2e8c-cb967cef407b
    Hope this helps,
    Greg

  • How does the S_NQ_AACT table gets written when working with Usage Tracking?

    Hi,
    We want to monitor the performance of each subject area so we have implemented the Usage Tracking.
    We have followed the steps mentioned on the link http://www.oracle.com/technology/obe/obe_bi/bi_ee_1013/usage_tracking/usage_tracking.htm
    But now the table S_NQ_ACCT is not getting written, whenever any user logs into OBIEE and runs a request his entry is not logged into that table.
    Anybody has encountered such a type of issue help would be appreciated.
    Regards,
    Apoorv

    The part of NQSConfig,ini goes like this:
    [ USAGE_TRACKING ]
    ENABLE = YES; //DEFAULT: NO
    //==============================================================================
    // Parameters used for writing data to a flat file (i.e. DIRECT_INSERT = NO).
    STORAGE_DIRECTORY = "<full directory path>";
    CHECKPOINT_INTERVAL_MINUTES = 5;
    FILE_ROLLOVER_INTERVAL_MINUTES = 30;
    CODE_PAGE = "ANSI"; // ANSI, UTF8, 1252, etc.
    //==============================================================================
    DIRECT_INSERT = YES;
    //==============================================================================
    // Parameters used for inserting data into a table (i.e. DIRECT_INSERT = YES).
    PHYSICAL_TABLE_NAME = "OBI Usage Tracking"."Catalog"."dbo"."S_NQ_ACCT" ; // "<Database>"."<Catalog>"."<Schema>"."<Table>" ; // Or "<Database>"."<Schema>"."<Table>" ;
    CONNECTION_POOL = "OBI Usage Tracking"."Usage Tracking Writer Connection Pool" ;
    BUFFER_SIZE = 10 MB ;
    BUFFER_TIME_LIMIT_SECONDS = 5 ;
    NUM_INSERT_THREADS = 5 ;
    MAX_INSERTS_PER_TRANSACTION = 1 ;
    //==============================================================================
    The User has the previleges to write to the table, I am not sure where I have gone wrong?
    Regards,
    Apoorv

  • Is it recommended to enable Usage Tracking in Production ??

    Hi,
    Just a quick question - Have anyone turned on Usage Tracking in Production? if so at what level ?
    OBIEE 64 bit 11g - 11.1.1.5.0
    OS - Linux 64 bit
    Cherrish Vaidiyan

    Hi,
    We have enabled the usuage tracking to track down the users usuage. And often rectifying the issues by setting the logging level to 7.
    But remembe, until and unless its required (in case to note the performance of the application for some long running queries).
    Though query logging has immeasurable development value, do not use this for regular production users as the runtime logging cost is extremely high. Every log item is flushed to the disk, which in turn hurts query response. Also, note that the query log files are not created on per user or query basis, there is only one query log per OBIEE server and it would have exclusive lock on the log file, which kills concurrent performance. The log file is NQQuery.log which resides in OracleBI Directory.
    Regards
    MuRam

  • Active Table Logging T000 performance impact

    Hi fellow SAP experts,
    I need some advice on system performance impact when switching on Table Logging for T000 - configuration in production please?
    We have decided to turn on Table Logging for auditing purposes, only allowing developer config in production following a volume of evidence being supplied.
    I need to know how much this activation is going to impact the performance of the companies production environments, how much storage, memory, performance, etc. this function is going to consume and how much of the above consumables I need to cater for now and in the future?
    We have a Dual Track environment, BAU want to switch on Table Logging for fix on fail, I want to swich it on for Project deliveries.
    Please advise, with referencing if possible?
    Thank you kindly
    Paul

    Hi Paul,
    There has been a constant debate whether table logging affects the System performance, especially in Produciton environment. Please see my comments below :
    1) To turn on logging for table T000, you will have to activate the parameter rec/client, with values for one client or all the clients in the System depending on your requirement.
    2) This parameter setting will not only log changes to T000, but also for over 28000 tables.
    3) But these are customizing tables which usually contain a relatively small amount of data which is changed occasionally.
    4) After activating, if you suddenly find performance issues, you can check which tables are causing issues via transaction SCU3.
    5) You can go to transaction SE13 and deactivate logging for a table, if you find too many entries for any particular table in SCU3.
    So, logging tables doesn't necessarily impact performance. Hope this helps. Please refer to SAP Note-608835 related to this.
    Best Regards,
    Savitha

  • Usage Tracking Error on Microsoft SQL Server 2012 RCU

    During an upgrade to 11.1.1.7.131017 recently, I ran into the following error in nqserver.log when trying to implement Usage Tracking on a MS SQL Server 2012 RCU DB:
    [2013-11-18T15:19:38.000-08:00] [OracleBIServerComponent] [ERROR:1] [] [] [ecid: 00iIX8GTiKpFw00Fzzw0w00001I00008PG,0:1:1:19:1:1] [tid: 1330]  [nQSError: 16015] SQL statement execution failed. [[[nQSError: 16001] ODBC error state: 22008 code: 0 message: [Microsoft][SQL Server Native Client 11.0]Datetime field overflow. Fractional second precision exceeds the scale specified in the parameter binding..]]
    The workaround I outlined below is not an ideal solution, just a workaround to prevent the bug from occurring until there's a fix:
    http://www.emerald-cube.com/2013/12/adventures-in-obiee-upgrade-land-quirks-and-workarounds-episode-i/http://www.emerald-cube.com/2013/12/adventures-in-obiee-upgrade-land-quirks-and-workarounds-episode-i/
    Just providing as an FYI in case anyone else searches for this error and needs a solution.
    Jeremy

    I like this one written by Itzik Ben-Gan
    http://www.amazon.com/Microsoft-High-Performance-Functions-Developer-Reference/dp/0735658366
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

Maybe you are looking for

  • Is it possible to get the scope name within a running BPEL instance

    I am currently adding some error handling for a business fault and wondered whether there is any way to get the scope name at run time. I am already catching the fault details using getFaultAsString but the support team have asked whether we can log

  • SELECT permission denied with ownership chaining on

      I have a database ('Datamart') that contains views on the tables another database ('Rawdb') on the same server.  All tables and views are in 'dbo' in each database.  I turned cross-database ownership chaining on.  I created the roles RawdbReadRole

  • Undefined: #174

    We are in the process of performing Post Installation steps from SSM Server Installation Guide. While trying to connect to "http://<nwce_server>:<port>/strategy/administration", we are getting these errors undefined: 'undefined' is null or not an obj

  • How do you link a ssl certificate to your website created with adobe muse?

    I would like to know how you can link a ssl certificate to a website created with Muse especially when you have created an ecormmerce website.

  • How can I get Pattern list Photoshop

    Hi I want access to preset files list (Pattern , Style , ...) in Photoshop with ActionScript in ExtentionBuilder 2.1 or JavaScript. I can just get fonts and actions in API. Please Help me.