Best practice excel addin retrievals of attributes of large flat sparse dim

Is anyone aware of a way to quickly lookup attributes of members of large flat sparse dimensions? One would imagine that since there exists a one-to-one relationship between a member and its attribute in the outline, Essbase shouldn't have to parse through the data to populate attributes when drilling down an attribute dimension. However, in practice, Essbase takes forever to return member attributes (depending on the base dimension and other factors, of course). Something similar to the subset rule within member selection would be ideal, but I don't seem to find anything. Any ideas? Thanks in advance!

Using Attribute Associations to generate a list of base dimension members should not have an overly adverse affect upon the retrieval time.
Introducing an Attribute dimension to a spreadsheet retrieval will slow things down because it introduces dynamic calculation to the query. If there are other dynamically calculated items in the outline, the intersecting dynamic calculations will slow retrieval times even further.

Similar Messages

  • Best practice for using messaging in medium to large cluster

    What is the best practice for using messaging in medium to large cluster In a system where all the clients need to receive all the messages and some of the messages can be really big (a few megabytes and maybe more)
    I will be glad to hear any suggestion or to learn from others experience.
    Shimi

    publish/subscribe, right?
    lots of subscribers, big messages == lots of network traffic.
    it's a wide open question, no?
    %

  • Best practices for WC State, Code, Attribute

    Currently, our process for assigning WC Codes is to assign them at the position level once an ee is joined to the position.  The result is a lot of one-off efforts outside of our regular workflow.  Is it possible to assign these attributes at the job key level?  So that way anytime a position is created, it inherits that information from the job key?

    Dear Ryan,         
    It is not possible in the standard to add tab into PPOME/PPOSE, as per notes below
    327614    TA PPOME customer-specific tab pages
    483345    Transaction PPOME Tab for standard infotypes is missing
    410869    PPOME: Infotypes in the detailed area
    385019    PPOME Detail including SAP standard infotypes
    I have also pasted in the relevant paragraph from note 385019:
    Tab pages for additional SAP standard infotypes are not available in the standard system and CANNOT be generated as described in note 327614 (transaction PPOME customer-specific tab pages).
    All the infotypes predefined for PPOME/PPOSE can be found in table T77OCTABUS; it is only necessary to activate or deactivate them.
    Best regards
    Sarah

  • Best Practices needed -- question regarding global support success stories

    My customer has a series of Go Lives scheduled throughout the year and is now concerned about an October EAI (Europe, Asia, International) go live.  They wish to discuss the benefits of separating a European go Live from an Asia/International go live in terms of support capabilities and best practices.  The European business is definitely larger and more important than the Asia/International business and the split would allow more targeted focus on Europe.  My customer does not have a large number of resources to spare and is starting to think that supporting the combined go live may be too much (i.e., too much risk to the businesses) to handle.
    The question for SAP is regarding success stories and best practices.
    From a global perspective, do we recommend this split?  Do most of our global customers split a go live in Europe from a go live in Asia/International (which is Australia, etc.).  Can I reference any of these customers?  If the EAI go live is not split, what is absolutely necessary for success, etc, etc?  For example, if a core team member plus local support is required in each location, then this may not be possible with the resources they have u2026u2026..
    I would appreciate any insights/best practices/success stories/or u201Cwaru201D stories you might be aware of.
    Thank you in advance and best regards,
    Barbara

    Hi, this is purely based on customer requirement.
    I  have a friend in an Organization which went live in 38 centers at the same time.
    With the latest technologies in networking, distances does not make any difference.
    The Organization where I currently work in, has global business locations. In my current organization the go live was in phases. Here they went live in the region where the business was maximum first because this region was their largest and most important as far as revenue was concerned. Then after stabilizing this region, a group of consultants went to the rest of the regions for the go live in that region.
    Both the companies referred above are successfully into SAP and are leading partners with SAP. Unfortunately I am not authorized to give you the names of the Organizations as a reference for you as you requested.
    But in your case if you have shortage of manpower, you can do it in phases by first going live in the European Market and then in phases you can go live in the other regions.
    Warm Regards

  • Best Practice for creating Excel report from SSIS.

    I have a requirement to create an Excel report on a daily basis which pulls data from SQL. I have attempted to resolve this by creating a stored procedure to save the results in SQL, a template in Excel to hold the graphs & pivot tables and an SSIS package
    to copy the data to the template.
    Problem 1: When the data turns up in Excel it is saved as text rather than numbers.
    Problem 2: When the data turns up in Excel it appends the data rather than overwriting it.
    I resolved problem 1 by having another sheet which converts the text to numbers (=int(sheet1!A1))
    I resolved problem 2 by adding some VB script to my SSIS package which clears the existing cells before copying the data
    The job runs fine, however when I schedule the job to run overnight it complains "System.UnauthorizedAccessException: Retrieving the COM class factory for component with CLSID". A little googling tells me that running the client side commands in
    my vb script (workSheet1.Range("A2:F9999").Clear(), workBook.Save(), workBook.Close() etc) from a server side task is bad practice.
    So, I am left wondering how people usually get around this problem; copy a SQL table into an existing Excel file and overwrite the data, without having the numbers turn up as text. My requirements are that the report must display pivot charts with selectable
    options and be automatically updated overnight.
    Help appreciated,
    Bish.
    Office 2013 on my PC, Office 2010 on the server, Windows Server 2008R2 Enterprise, SQL Server 2008R2.

    I think that the best practice in case like this is to Link an excel file to a view or directly to a table. So you don't have to struggle with changing template, with overnight packages, etc. If the data are too much complex and the desiderate too excessive
    then I tend to create a Cube and that's it...dashboard, graph and everyone is happy. In your case if the request is not too much try to don't use SSIS but directly build a view and point directly on SQL.
    SSIS is really strong for the ETL, to run some stored procedure too heavy, to use a cut time scheduled, etcetera , etcetera, etcetera...I love it. But sometimes we need to find the easier solutions...
    I hope this post helped you

  • Select One Choice attribute' LoV based on two bind variables, best practice

    Hello there,
    I am in the process of learning the ADF 11g, I have following requirement,
    A page must contain a list of school names which is needed to be fetched based on two parameters, the parameters are student information been inserted in the previous page.
    I have defined a read only view "SchoolNamesViewRO", it's query depends on two bind variables :stdDegree and stdCateg.
    added that RO View as a view accessor to the entity to which the name attribute belongs, and then add LoV for the name attribute using the ReadOnly view,
    added the name attribute as Select One Choice to page2,
    and now I need to pass the values of the bind variables of the ReadOnly view,
    the information needed to be passed as the bind variables is inserted in the previous page, I could have the data as bindings attribute values in the page2 definition
    I have implemented the next two appraoches but both resulted in an empty list :
    * added ExecuteWithParams Action to the bindings of the page and then defined an Invoke Action (set refresh condition) in the executable s, set the default values of the parameters to be the attributes values' input value,
    in the trace I code see that the binding fetches correct values as supposed , but the select list appears empty, does the this execution for the query considered to be connected to the list ?
    * added a method to the ReadOnly view Imp java class to set the bind variables, then I define it as a MethodAction in the bindings , and then create an Invoke action for it , also the select is empty,
    if the query been executed with the passed variables, then why the list is empty? is it reading data from another place than the page!
    and what is the best practice to implement that requirement?
    would the solution be : by setting the default value of the bind variables to be some kind of Expression!
    please notice that query execution had the bound variables ( I see in the trace) are set to the correct values.
    would you give some hints or redirect me to a useful link,
    Thanks in advance
    Regards,

    please give me any example using backing bean .for example
    <?xml version='1.0' encoding='UTF-8'?>
    <jsp:root xmlns:jsp="http://java.sun.com/JSP/Page" version="2.1"
    xmlns:f="http://java.sun.com/jsf/core"
    xmlns:h="http://java.sun.com/jsf/html"
    xmlns:af="http://xmlns.oracle.com/adf/faces/rich">
    <jsp:directive.page contentType="text/html;charset=UTF-8"/>
    <f:view>
    <af:document id="d1">
    <af:form id="f1">
    <af:selectOneChoice label="Label 1" id="soc1" binding="#{Af.l1}"
    autoSubmit="true">
    <af:selectItem label="A" value="1" id="si1"/>
    <af:selectItem label="B" value="2" id="si2"/>
    </af:selectOneChoice>
    <af:selectOneChoice label="Label 2" id="soc2" disabled="#{Af.l1=='2'}"
    partialTriggers="soc1">
    <af:selectItem label="C" value="3" id="si3"/>
    <af:selectItem label="D" value="4" id="si4"/>
    </af:selectOneChoice>
    </af:form>
    </af:document>
    </f:view>
    </jsp:root>
    package a;
    import oracle.adf.view.rich.component.rich.input.RichSelectOneChoice;
    public class A {
    private RichSelectOneChoice l1;
    public A() {
    public void setL1(RichSelectOneChoice l1) {
    this.l1 = l1;
    public RichSelectOneChoice getL1() {
    return l1;
    is there any mistake

  • Import data from excel file - best practice in the CQ?

    Hi,
    I have question related to importing data from excel file and creates from those data a table in the CQ page. Is inside CQ some OOTB component which provides this kind of functionalities? Maybe somebody implement this kind of functionality or there is best practice to do this kind of functionalities?
    Thanks in advance for any answer,
    Regards
    kasq

    You can check a working example package [1] (use your Adobe ID to log in)
    After installing it, go to [2] for immediate example.
    Unfortunately it only supports the old OLE-2 Excel format (.xls and not .xlsx)
    [1] - http://dev.day.com/content/packageshare/packages/public/day/cq540/demo/xlstable.html
    [2] - http://localhost:4502/cf#/content/geometrixx/en/company/news/pressreleases/my_personal_bes ts.html

  • Is it possible to add Attribute(column) to entity via Excel addin?

    I can create an entity via the excel addin and update data in it but i cannot add additional columns (attributes)  to it via the excel addin once its created. am i missing something or is it just not possible?

    Winston, is this still an issue?
    Thanks!
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!
    I worked around it using the excel addin, but the question is to see if its possible to do this in teh MDS web UI, without using excel, as its not very good design to have to jump from the browser based designer to excel and back in order to manage an MDS model.

  • FI document Excel upload best practice?

    Hello,
    What's the current best practice for financial document upload from Excel? (uploading data from excel to run FB01/FB50)
    Especially I'm looking for a comprehensive and easy-to-implement solution for error handling. The simulation/testrun in foreground should find out all master data and combination mistakes and give detail feedback for the user. Additionally the production run should be straight forward, also with status/error handling (read, user should not be asked to go to SM35).
    Is there any how-to docs released by SAP?
    Thanks!
    Aki

    Hi Aki
    www.mkpg.ch developed an Excel add-in called xls2Sap. With this add-in you can post from Excel directly into SAP, without an upload file, without leaving Excel. Occuring errors are written into a seperate log-tab in the Excel sheet. Furthermore the user can choose to store he posting document in SAP. It seems that our solution fullfills your requirements. Let me know, if you have any other questions.
    There are some videos on www.mkpg.ch showing some of the possibilities. However the are in Swiss German
    Kind regards,

  • Retrieve attribute description to Excel Addin for MDS

    I propose a new feature:
    retrieve attribute description to Excel Addin for MDS - as
    separate row, which would be above attribute-row.
    This is especially useful for users - translation into other languages.

    Do you think, it would help if we add description as comments of the column name. So user have to hover the mouse to see it?
    Also, we already have a work item to support multiple language in the model/entity/attribute name. So based on the client OS, it would show different language for the column name. But I cannot promise when we will deliver it at this moment. 

  • EJB + "transaction attributes" + "best practice"

    hi
    For EJBs, when you specify multiple transaction attributes for a single
    method, is the latest attribute taken or the first?
    for e.g. In Jbuilder8 in the container transaction section for an EJB,
    if i were to put
    Method Transaction attribute
    * Required
    getSomeThing() Supported
    As you can see all the methods were initially given the required attribute,
    and then the method getSomeThing() was given Supported. will the
    getSomething() method use Required or else Supported. Secondly, will this
    cause an extra overload (Transaction attributes are set and the again
    reset), is it a recommended best practice?
    suren

    getSomeThing() Supported
    As you can see all the methods were initially given
    the required attribute,
    and then the method getSomeThing() was given
    Supported. will the
    getSomething() method use Required or else Supported.It will use supported.
    Secondly, will this
    cause an extra overload (Transaction attributes are
    set and the againIt won't cause extra overload.
    reset), is it a recommended best practice?It really depends on your needs - what you're trying to achieve. There is no best practice for that it depends on the application you're building.
    Regards,
    Dimitar

  • Best Practice for Significant Amounts of Data

    This is basically a best-practice/concept question and it spans both Xcelsius & Excel functions:
    I am working on a dashboard for the US Military to report on some basic financial transactions that happen on bases around the globe.  These transactions fall into four categories, so my aggregation is as follows:
    Year,Month,Country,Base,Category (data is Transaction Count and Total Amount)
    This is a rather high level of aggregation, and it takes about 20 million transactions and aggregates them into about 6000 rows of data for a two year period.
    I would like to allow the users to select a Category and a country and see a chart which summarizes transactions for that country ( X-axis for Month, Y-axis Transaction Count or Amount ).  I would like each series on this chart to represent a Base.
    My problem is that 6000 rows still appears to be too many rows for an Xcelsius dashboard to handle.  I have followed the Concatenated Key approach and used SUMIF to populate a matrix with the data for use in the Chart.  This matrix would have Bases for row headings (only those within the selected country) and the Column Headings would be Month.  The data would be COUNT. (I also need the same matrix with Dollar Amounts as the data). 
    In Excel this matrix works fine and seems to be very fast.  The problem is with Xcelsius.  I have imported the Spreadsheet, but have NOT even created the chart yet and Xcelsius is CHOKING (and crashing).  I changed Max Rows to 7000 to accommodate the data.  I placed a simple combo box and a grid on the Canvas u2013 BUT NO CHART yet u2013 and the dashboard takes forever to generate and is REALLY slow to react to a simple change in the Combo Box.
    So, I guess this brings up a few questions:
    1)     Am I doing something wrong and did I miss something that would prevent this problem?
    2)     If this is standard Xcelsius behavior, what are the Best Practices to solve the problem?
    a.     Do I have to create 50 different Data Ranges in order to improve performance (i.e. Each Country-Category would have a separate range)?
    b.     Would it even work if it had that many data ranges in it?
    c.     Do you aggregate it as a crosstab (Months as Column headings) and insert that crosstabbed data into Excel.
    d.     Other ideas  that Iu2019m missing?
    FYI:  These dashboards will be exported to PDF and distributed.  They will not be connected to a server or data source.
    Any thoughts or guidance would be appreciated.
    Thanks,
    David

    Hi David,
    I would leave your query
    "Am I doing something wrong and did I miss something that would prevent this problem?"
    to the experts/ gurus out here on this forum.
    From my end, you can follow
    TOP 10 EXCEL TIPS FOR SUCCESS
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/204c3259-edb2-2b10-4a84-a754c9e1aea8
    Please follow the Xcelsius Best Practices at
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/a084a11c-6564-2b10-79ac-cc1eb3f017ac
    In order to reduce the size of xlf and swf files follow
    http://myxcelsius.com/2009/03/18/reduce-the-size-of-your-xlf-and-swf-files/
    Hope this helps to certain extent.
    Regards
    Nikhil

  • Best Practice Table Creation for Multiple Customers, Weekly/Monthly Sales Data in Multiple Fields

    We have an homegrown Access database originally designed in 2000 that now has an SQL back-end.  The database has not yet been converted to a higher format such as Access 2007 since at least 2 users are still on Access 2003.  It is fine if suggestions
    will only work with Access 2007 or higher.
    I'm trying to determine if our database is the best place to do this or if we should look at another solution.  We have thousands of products each with a single identifier.  There are customers who provide us regular sales reporting for what was
    sold in a given time period -- weekly, monthly, quarterly, yearly time periods being most important.  This reporting may or may not include all of our product identifiers.  The reporting is typically based on calendar-defined timing although we have
    some customers who have their own calendars which may not align to a calendar month or calendar year so recording the time period can be helpful.
    Each customer's sales report can contain anything from 1,000-20,000 rows of products for each report.  Each customer report is different and they typically have between 4-30 columns of data for each product; headers are consistently named.  The
    product identifiers included may vary by customer and even within each report for a customer; the data in the product identifier row changes each week.  Headers include a wide variety of data such as overall on hand, overall on order, unsellable on hand,
    returns, on hand information for each location or customer grouping, sell-through units information for each location or customer grouping for that given time period, sell-through dollars information for each location or customer grouping for that given time
    period,  sell-through units information for each location or customer grouping for a cumulative time period (same thing for dollars), warehouse on hands, warehouse on orders, the customer's unique categorization of our product in their system, the customer's
    current status code for that product, and so on.
    Currently all of this data is stored in a multitude of Excel spreadsheets (by customer, division and time period).  Due to overall volume of information and number of Excel sheets, cross-referencing can take considerable time.  Is it possible to
    set-up tables for our largest customers so I can create queries and pivot tables to more quickly look at sales-related information by category, by specific product(s), by partner, by specific products or categories across partners, by specific products or
    categories across specific weeks/months/years, etc.  We do have a separate product table so only the product identifier or a junction table may be needed to pull in additional information from the product table with queries.  We do need to maintain
    the sales reporting information indefinitely.
    I welcome any suggestions, best practice or resources (books, web, etc).
    Many thanks!

    Currently all of this data is stored in a multitude of Excel spreadsheets (by customer, division and time period).  Due to overall volume of information and number of Excel sheets, cross-referencing can take considerable time.  Is it possible to
    set-up tables .....
    I assume you want to migrate to SQL Server.
    Your best course of action is to hire a professional database designer for a short period like a month.
    Once you have the database, you need to hire a professional DBA to move your current data from Access & Excel into the new SQL Server database.
    Finally you have to hire an SSRS professional to design reports for your company.
    It is also beneficial if the above professionals train your staff while building the new RDBMS.
    Certain senior SQL Server professionals may be able to do all 3 functions in one person: db design, database administration/ETL & business intelligence development (reports).
    Kalman Toth Database & OLAP Architect
    SELECT Video Tutorials 4 Hours
    New Book / Kindle: Exam 70-461 Bootcamp: Querying Microsoft SQL Server 2012

  • Best practice for searching on surname/lastname/name in Dutch

    I'm looking for a best practice to store names of persons, but also names of companies, in my database.
    I always store them as is (seems logical since you need to be able to display the original input-name) but I also want to store them transformed in some sort of way so I can easily search on them with LIKE! (Soundex, Metaphone, Q-Gram, ...)
    I know SOUNDEX and DIFFERENCE are included in SQLServer, but they don't do the trick.
    If somebody searches for the phrase "BAKKER", you should find names like "Backer", "Bakker", ... but also "De Backer", "Debecker", ... and this is where SOUNDEX fails ...
    Does someone know some websites to visit, or someone already wrote a good function to transform a string that I can use to store the names but also to transform my search data?
    (Example:  (Pseudo lang :-))
    function MakeSearchable (sString)
      sString = sString.Replace(" ", ""); //Remove spaces
      sString = sString.Replace("CK", "K");
      sString = sString.Replace("KK", "K");
      sString = sString.Replace("C", "S");
      sString = sString.Replace("SS", "S");
      return sString;
    Greetz,
    Tim

    Thanks for the response, but unfortunately the provided links are not much help:
    - The first link is about an article I don't have access to (i'm not a registered user)
    - The second link is about Integration Services. This is nice for Integration stuff, but I need to have a functionality within a frontend. 
    - The third link is for use in Excel.
    Maybe I'm looking for the wrong thing when wanting to create an extra column with "cleaned" up data. Maybe there's another solution from within my frontend or business layer, but I simply want a textbox on a form where users can type a search-value like
    "BAKKER". The result of the search should return names like "DEBACKER", "DE BEKKER", "BACKER", "BAKRE", ...
    I used to work in a hospital where they wrote their own SQL-function (on an Interbase database) to do this: They had a column with the original name, and a column with a converted name:
    => DEBACKER => Converted = DEBAKKER
    => DE BEKKER => Converted = DEBEKKER
    => BACKER => Converted = BAKKER
    => BAKRE => Converted = BAKKER
    When you searched for "BAKKER", you did a LIKE operation on the converted column ...
    What I am looking for is a good function to convert my data as above.
    Greetz,
    Tim

  • Home Movie Cataloging - BEST PRACTICES

    I have about 200 hours of old home movies on VHS which I am in the process of adding to my iMac. I am wondering about 'best practices' on how much video can be stored inside of iMovie '08, when how much video becomes too much inside of the program, etc.
    In a perfect world, I'd like to simply import all of my home videos into iMovie, leave them in the 'library' section, and make 2-5 minute long clips in the 'projects' section for sharing with family members, but never deleting anything from the 'library'. Is this a good way to store original data? Would it be smarter to export all of the original video content to .DV files or something like that for space saving, etc?
    Can I use iMovie to store and catalog all of my old home movies in the same way I use iPhoto to store ALL of my photos, and iTunes to store ALL of my music/hollywood-movies, etc?

    We-ell, since no-one else has replied:
    1 hour of DV (digital video in the file system which iMovie uses) needs 13GB of hard disc space.
    You have 200 hours of video. 200 x 13 = 2,600 gigabytes. Two point six terabytes. If you put all that on one-and-a-bit 2TB hard discs, and a hard disc fails - oops! - where's your backup? ..Ah, on another one-and-a-bit 2TB hard discs ..or, preferably, spread over several hard discs, so that if one fails you haven't lost everything!
    iMovie - the program - can handle video stored on external discs. But are you willing to pay the price for those discs? If so; fine! Digitise all your VHS and store it on computer discs (prices come down month by month).
    Yes, you can "mix'n'match" clips between different projects, making all sorts of "mash-ups" or new videos from all the assorted video clips. But you'll need more hard disc space for the editing, too. You could use your iMac's internal hard disc for that ..or use one of the external discs for doing the editing on. That's how professionals edit: all the video "assets" on external discs, and edit onto another disc. That's what I do with my big floorstander PowerMac, or whatever those big cheesegraters were called..
    So the idea's fine, as long as you have all the external storage you'd need, plus the backup in case one of those discs fails, and all the time and patience to digitise 200 hours of VHS.
    Note that importing from VHS will import material as one long, continuous take - there'll be no automatic scene breaks between different shots - so you'll have to spend many hours chopping up the material into different clips after importing it.
    Best way to index that? Dunno; there have been several programs which supposedly do the job for you (..I can't remember their names; I've tried a few: find them by Googling..) but they've been more trouble - and taken up more disc space - than I've been prepared to bother with. I'd jot down the different clips as you create them, either by jotting in TextEdit (simplest) or in a database or spreadsheet program such as Excel or Numbers or similar ..or even in a notebook.
    Jot down the type of footage (e.g; 16th Birthday party), name of clip (e.g; 016 party), duration (e.g; 06:20 mins and seconds) and anything else you might need to identify each clip.
    Best of luck!

Maybe you are looking for

  • Data in single row

    thanks Edited by: 786304 on May 25, 2011 12:22 PM

  • Can't play songs on iPhone

    I can play most of my songs on the iPhone, but there are certain ones that will not. It will show the album art work, go through the motions of playing, but there is no sound and then it will skip to the next song after about 5 seconds. These same so

  • Dual head setup (mach64 - ATI Rage Mobility)

    hello everyone! I want to use a new extra monitor I got recently, with my laptop, I use aweosme. I have installed: xf86-video-mach64 (recompiled with dri support) mach64-dri (from community) mach64drm (from AUR) https://wiki.archlinux.org/index.php/M

  • MacBook Pro power adapters dying, keys frying, mysterious heat problem

    I've got a very frustrating and mysterious major situation going on with my MBPro. I'm living in Peru in the jungle. I've got a basic electric hookup that's caused no problems for the few electrical devices I've got: small dorm-size fridge, a lightbu

  • Brochure pages appearance in Acrobat Reader

    Hello! I would like to create a brochure much like this one: http://imagesrv.gartner.com/summits/docs/emea/enterprise-architecture/ea13_summit_brochure _generic.pdf from Gartner. They have the front page (when viewing the PDF document in Acrobat/Read