Can we load data for all levels in ASO cube

Hi All,
Can we load data for all levels of members in ASO cube in 9.3.1.
Regards

Yes you can load data for all levels in an ASO cube in any version HOWEVER, none of the upper level data in any cube will be there when you look for it. You will get a warning message in the load because ASO cubes don't store data at upper levels. It is the same as loading data into dynamic calc members in BSO cube. It will do the load without compalints, but there will be no data there (At least you get the warning in ASO)

Similar Messages

  • Can we load data for all levels in ASO?

    Hi All,
    Im creating cube in ASO
    can i load data for all levels in ASO
    we can load data for all Levels In BSO but in ASO i need confirmation????
    and one more
    wat is the consider all levels option in ASO is used for ? wat is the purpose?
    Can any one help ,it would be appriciate.
    Thanks

    In an ASO cube you can only load to level zero
    The consider all levels if used for aggregation hints. It allows you to tell the aggregation optimizer to look at all levels when deciding if aggregation needs to be done on the dimension

  • Is there a way to allow setting of "Due Date for All Tasks" to current date in OOB Workflow

    Is there a way to allow setting of "Due Date for All Tasks" to current date in OOB Workflow
    I need to set it today because there are tasks that is needed to be completed on the same day.
    Is there a way without modifying the page/form. like a configuration or something?
    ----------------------- Sharepoint Newbie

    Hi ,
    As Ajeet said,  for implement your requirement, you can change the data validation on the Due data field.
    You can do as the following :
    Open the site with SharePoint Designer 2010
    Click on Workflows on the left launch, then find ‘Approval-SharePoint 2010’, click on it
    On the Forms section, click ‘ReviewApproval_***.xsn’
    Then you can find Due Data for All Tasks, and change the validation
    In addition, if you create an approval workflow with SharePoint Designer, you can use a specific date, it contains hour and minute.
    I hope this helps.
    Thanks,
    Wendy
    Wendy Li
    TechNet Community Support

  • HOW TO MIGRATE DUNNING LEVEL & LAST DUNNED DATE FOR ALL OPEN LINE ITEMS

    Hi,
    We need to upload dunning level & last dunned date for all open items (to be done separately i.e. after all open items migrated) which is different for each and every line item so if we use mass upload (fbl5n) we have to do for each and every open line item which is not feasible as open line items in thousands..I tried LSMW also,problem is that for Batch input recording we need to run one complete cycle in sm30 but there is no maintenance dialog for BSEG & BSID table.. Please share your valuable Inputs how we can go ahead..
    Thanks in advance

    Hi Mark,
    Thanks for your reply.
    It is complete a Zcopy of a standard one, but we are facing one or two records are not updating correctly out of hundreds of customers in the worklist.

  • Purchased MLB via 1 of my 3 appletv's. My oldest 2nd gen 720P shows "error loading content" for all games. What's up?  Also I can't get MLB to work on my iPad2, iPhone 4S or iPod 5th gen. Arrhhgg

    Purchased MLB via iTunes on 1 of my 3 appletv's. My oldest 2nd gen 720p shows "unable to load content" for all games , but my other 2 appletv's work fine, restarted problem appletv to no luck.  Also can't watch games on my ipad2 "unable to restore from iTunes" , iPhone 4s " unable to connect to MLB.com" or iPod 5th gen.

    Purchased MLB via iTunes on 1 of my 3 appletv's. My oldest 2nd gen 720p shows "unable to load content" for all games , but my other 2 appletv's work fine, restarted problem appletv to no luck.  Also can't watch games on my ipad2 "unable to restore from iTunes" , iPhone 4s " unable to connect to MLB.com" or iPod 5th gen.

  • Is it possible to load data at aggregated level directly

    Hi All,
    My question may sound vague but still would like to clarify this.
    Is it possible to load data at some higher level than the leaf level? The reason for asking this question is that we are facing severe performance issues while loading the cube.
    We are trying to reduce the number of members which inturn can reduce the loading time. On this attempt to reduce the number of members client said that there are 3 million leaf members(out of 4.3 million total members) for which they do not bother if they appear on report or not but they want to see the total value correct. i.e. the dimension will be only loaded with 1.3 million(4.3 - 3) members but the total can still be correct(including the 3 million leaf members).
    Is it possible to have two loads one at leaf level only and other at parents and above levels.
    DB - 10.2.0.4
    Also I want to know when do we use allocmap? how does this work?
    Regards,
    Brijesh
    Edited by: BGaur on Feb 13, 2009 3:33 PM

    Hi Carey,
    Thanks for your response.
    I worked on your suggestion and I could able to load data at higher level for a value based hierarchy. But I met with a other problem while doing this.
    I am describing this as a level based but while loading we are converting to value based hierarchy.
    we have following levels in the customer dimension having two hierarchy
    hier1
    lvl1
    lvl2
    lvl3
    lvl4
    lvl5
    prnt
    leaf
    hier2
    level1
    level2
    level3
    level4
    prnt
    leaf
    so prnt and leaf is common levels are common in both the hierarchies but we were facing multipath issues in second hierarchy and we work around it by concatenating the level name.
    In attemp to decrease the number of this dimension member(currrently 4.3 million) business suggested that there are some kind of customer for which they dont want to see the data at leaf and prnt level instead the level4 or lvl5 should have the total consisting of those members. So by this way we need not to load those members and they were making up 2.4 million out of 4.3
    To implement above I did following.
    1. Created six hierarchies.
    one to have members from level4 till level1
    second to have members from lvl5 till lvl1
    third to have all members of hier1
    fourth to have all members of hier2
    fifth will have leaf and prnt of hier1
    sixth will have leaf and prnt of hier2
    In fifth and sixth hierarchy leaf is common but prnt is different due to the concatenation.
    In the cube I am selecting the aggregation hierarhies as first,second,fifth and sixth. hierarchies third and fourth will be used for viewing the data.
    Fact will have data corresponding for leaf level data,level4 and lvl5.
    The challenge I am facing is that if there is some fact value for leaf level loaded through relational fact but no value being loaded through fact for lvl5 or level4 I am not seeing the total value as leaf value is aggregating till prnt level and no value at level4 or lvl5 so same is propagated till lvl1 and level1.
    I tried to be as clear as possible but still if there is any confussion pls update the thread. I understand that the approach is vague but I am not seeing any other way.
    Thanks
    Brijesh

  • Loading data for Cube Characteristics

    Hi Experts,
    I have basic questions on loading of data for InfoCube characterisctis.
    1.Suppose I have 6 dimensions and total 26 characteristics are in dimensions.Do I need to load master data for 26 characteristis.
    2.I have master data InfoProvider say 0Customer ,it does have navigational attributes and display attributes.Do I need to define Navigational attribute as a Master data Infoprovider and load data for them or any other way.
    Please clarify my doubts which I am facing in my project now.I will awarded for all inputs.
    Thanks,
    Rani
    Edited by: Rani on Aug 24, 2008 4:18 AM

    Hi,
    Actually its not really clear what your doubt is... If you question is whether you can load data to infocube without maintaining master data for the characteristics - the answer is yes. You can do that; just make sure that you do not tick the option of checking master data while loading.

  • Why files in Lightroom mobile and files in the creative cloud not the same? idea-  one place(stored date) for all application ??????

    Why files in Lightroom mobile and files in the creative cloud not the same? idea- one place(stored date) for all application ??????

    Lightroom Mobile is not a cloud storage service. You shouldn't treat it as a way of backing up your files. You are merely storing Smart Previews of your files in the cloud space, high-quality JPEGs of your files regardless of their original format on your desktop. The point is that they are there so you can continue editing them in a Lightroom-like environment even while away from your desktop/laptop computer. The files in Lightroom Mobile can only be used in Lightroom Mobile.
    This is very different from what is offered by the Creative Cloud storage, which can be used to synchronize your files between any device (that can support the individual files).
    The vast difference in how each service works and its intended use is why they are separate.

  • Problem in Loading data for clob column using sql ldr

    Hi,
    I am having problem in loading data for tables having clob column.
    Could anyone help me in correcting the below script for ctrl file inorder to load the data which is in mentioned format.
    Any help really appreciated.
    Table Script
    Create table samp
    no number,
    col1 clob,
    col2 clob
    Ctrl File
    options (skip =1)
    load data
    infile 'c:\1.csv'
    Replace into table samp
    fields terminated by ","
    trailing nullcols
    no,
    col1 Char(100000000) ,
    col2 Char(100000000) enclosed by '"' and '"'
    Data File(1.csv)
    1,asdf,"assasadsdsdsd""sfasdfadf""sdsdsa,ssfsf"
    2,sfjass,"dksadk,kd,ss""dfdfjkdjfdk""sasfjaslaljs"
    Error Encountered
    ORA-01461: can bind a LONG value only for insert into a LONG column
    Table sampThanks in advance

    I can't reproduce it on my 10.2.0.4.0. CTL file:
    load data
    INFILE *
    Replace into table samp
    fields terminated by ","
    trailing nullcols
    no,
    col1 Char(100000000) ,
    col2 Char(100000000) enclosed by '"' and '"'
    BEGINDATA
    1,asdf,"assasadsdsdsd""sfasdfadf""sdsdsa,ssfsf"
    2,sfjass,"dksadk,kd,ss""dfdfjkdjfdk""sasfjaslaljs"Loading:
    SQL> Create table samp
      2  (
      3  no number,
      4  col1 clob,
      5  col2 clob
      6  );
    Table created.
    SQL> host sqlldr scott/tiger control=c:\temp\samp.ctl log=c:\temp\samp.log
    SQL> select * from samp
      2  /
            NO
    COL1
    COL2
             1
    asdf
    assasadsdsdsd"sfasdfadf"sdsdsa,ssfsf
             2
    sfjass
    dksadk,kd,ss"dfdfjkdjfdk"sasfjaslaljs
            NO
    COL1
    COL2
    SQL> SY.

  • Difference between Due Date for All Tasks and Duration per task in an approval workflow

    Hello,
    I'm starting to read up and learn about approval workflows in Designer 2010.  I am confused by the fields "Due Date for all Tasks" and the "Duration per task".  I understand that the Due date for all Tasks is the date that all
    of the tasks in the workflow must be completed by.  And I understand that the duration means each task has x number of days to complete.
    But I dont know whey we need them both.  If I have the task duration set to two days and there are 3 tasks involved, then wouldn't the due date be 6 days from the start?  Why would I have a due date for all tasks that is completely different than
    the sum of the duration of days for each task? 
    Please help shed some light!

    Hi Michelle,
    Due date for all tasks, this setting specifies the date by which all tasks are due. It’s used for serial and parallel
    task.
    Duration per tasks, this setting specifies the date by which an individual
     task is due, it’s used for serial workflow participants.
    Note, when you create an approval workflow, select task process participants(approvers oob approval workflow), we can also
    select once at a time(serial) or all at once(parallel).
    For more information, see
    http://office.microsoft.com/en-us/sharepoint-server-help/use-an-approval-workflow-HA101793831.aspx
    Best Regards.
    Kelly Chen
    TechNet Community Support

  • In mdx how to get max date for all employees is it posible shall we use group by in mdx

    in mdx how to get max date for all employees is it posible shall we use group by in mdx
    example
    empno  ename date
    1         hari        12-01-1982
    1         hari        13-06-2000
    by using above data i want to get max data

    Hi Hari3109,
    According to your description, you want to get the max date for the employees, right?
    In your scenario, do you want to get the max date for all the employees or for each employee? In MDX, we have the Max function to achieve your requirement. You can refer to Naveen's link or the link below to see the details.
    http://www.sqldbpros.com/2013/08/get-the-max-date-from-a-cube-using-mdx/
    If this is not what you want, please provide us more information about the structure of you cube, so that we can make further analysis.
    Regards,
    Charlie Liao
    TechNet Community Support

  • How to fetch data for all OUs

    Hi,
    Run below query:
    select * FROM PO_headers
    Result: no data
    EXECUTE MO_GLOBAL.set_POLICY_CONTEXT('S','120');
    Run below query:
    select * FROM PO_headers
    Result:data for org_id=120
    EXECUTE MO_GLOBAL.set_POLICY_CONTEXT('A','');
    select * FROM PO_headers
    Result: no data
    So how can we fetch data from all OUs whothout using PO_HEADERS_ALL tables?
    why do we have this MO_GLOBAL.set_POLICY_CONTEXT('A','') option?

    in 11i PO_HEADERS is view with org_id policy
    in r12 PO_HEADERS is synonym with policies based on multi-org
    so for r12
    SQL> select count(*) from po_headers;
      COUNT(*)
             0
    SQL> begin
      2    mo_global.set_policy_context('S', 101);
      3  end;
      4  /
    PL/SQL procedure successfully completed
    SQL> select count(*) from po_headers;
      COUNT(*)
           110
    SQL> begin
      2    mo_global.set_policy_context('S', 82);
      3  end;
      4  /
    PL/SQL procedure successfully completed
    SQL> select count(*) from po_headers;
      COUNT(*)
           235
    SQL>
    SQL> begin
      2      mo_global.init('SQLAP');
      3  end;
      4  /
    PL/SQL procedure successfully completed
    SQL> select count(*) from po_headers;
      COUNT(*)
           345
    SQL> select unique org_id from po_headers;
        ORG_ID
            82
           101
    SQL> mo_global.set_policy_context('S',<org_id>) - single operating unit
    mo_global.init(<app_short_name>) - multi operating unit

  • How to generate a list of  pick dates for all scheduled deliveries?

    Hi!
      How to generate a list of  pick dates for all scheduled deliveries?
      What all the tables involved?
    Thanks
    Imran.

    Thanks for the suggestion concerning the file path.  And certainly it would have been nice to have done this before beginning.  However this is a project that has been around for quite a while, and the files have been moved into different bins.  And now the project sequence is being revised.
    So the problem is, worded slightly differently, how can I search all of the bins for the files that are used just by this sequence, ignoring the files which are used by other sequences?  Or, how can I get a list of the file paths of the files that are used in the sequence?

  • SPP - Unable to load data for numeric parts

    Hi all,
    We have been trying to upload data from a flat file for a set of numeric parts and have been unsuccessful in doing so. We get an error "No valid item to load into APO-BI". We have been successful in loading data for alphanumeric parts.
    We suspected that this may be due to the output format for product number we specified in SPRO (not checked lexicographic). This has resulted in the system attaching leading zeros to the product number. The setting however cannot be changed once a numeric product has been created in the system. We were not able to delete the product (for which we started a separate thread).
    Would appreciate if someone could share if there is a workaround to this problem ?
    Best Regards
    Mitesh

    The Problem got resolved as we were able to delete the numeric part numbers and then checking the "lexicographic" option in "Specify Output format of Product Numbers".

  • Can i load data directly into DSO without PSA?

    hi friends,
    can i load data into DSO directly without PSA(without infopackage), if possible ple let me know where can i do that in dtp?
    regards
    ss

    Suneel,
    Why don't you try it to see if error DTP is possible with a flat file ds or not?
    Or, extrapolate logically -
    - An error DTP is based on a regular DTP.
    - In any situation where you can have a DTP, why can you not have an error DTP? I would think there would be no reason for this.
    At this point of my reasoning I will give it a try to validate if my assumption (DTP possible==>Error DTP possible) is correct or not.
    regards,

Maybe you are looking for