HIerarchy Variable in Selection Screen not taken into account.

Hi,
Our users choose a cost center hierarchy in the selection screen and run the report. As we have suppressed Unassigned Nodes, they do not appear. However when users then deactivate the hierarchy, unassigned nodes are shown. Their argument is that the unassigned nodes should still not be displayed because a hierarchy was entered on the selection screen. Is there a way of stopping this?
Thanks

Hi Shawn,
I think I get what you did in your query, though it took me some time...
Did you select the cost center in your query and set the display hierarchy using settings, hierarchy tab, select hierarchy, hierarchy variable (or something like that)? That would explain why the query isn't filtered on that hierarchy, only the display of cost elements is affected. You should also have the same hierarchy variable in your filter settings in the query designer. That way, the user filters on a hierarchy and the unassigned nodes will never be in the result.
Hope it helps, tell me if I got it wrong,
Kind regards,
Alex

Similar Messages

  • New values not taken into account in MASS / XD99

    I am trying to use a background job to block Customers for Billing at month end.
    In MASS/XD99  I created a variant to update field KNVV-FAKSD (Sales Area Billing Block).
    But running the variant in background (using MASSBACK), it seems the new values I thought I saved in the variant in MASS/XD99 are not taken into account.
    I opened the variant again in MASS/XD99 and indeed it seems it had not saved the new values, because the screen again shows the current values.
    Tried a couple of times again, but everytime I open the variant I have saved again, the new values revert back to 'current values'.
    Is this the reason MASSBACK is also making no updates? What am I doing wrong?
    Thanks in advance for your comments & advice.

    Here is the second method  -
    Regarding your issue:
    Retraction is based on the records of the SEM buffer. Zero records are
    in general filtered in the buffer, so that they cannot be considered forplanning functions that execute the retraction. I regret but this is
    the normal system behaviour in the standard. As workaround, please use adummy key figure that is set to a value <> 0.
    The only way is a work around. If you fill a dummy key figure with an
    arbitrary value, the frame work recognizes this as not 0, even if the
    proper key figure is 0. In that case the 0 retraction would work. So if
    not existing, create an additional dummy key figure to the cube and workon as described.
    For all the of retractions (CO-OM, IM, PS) BPS planning functions are
    used and they run in the general BPS framework. This framework is
    designed to explicitly not process zero records because of performance
    reasons.
    If zero records would be processed, planning functions would run longer
    but they can't deal with those records anyway. Therefore the developmentdecided on the current design, to delete all zero records first, before
    processing them with a planning function. This decision has been
    discussed several times before and it not likely to be changed.
    Regarding SAP note you can check the work aroud I explained can be foundin the oss note 768822 SEM-BPS: Retraction of WBS element w/ Pushback.
    I hope this explains the reason.
    Please kindly confirm the message if your issue is answered.
    Thanks for your support and cooperation.

  • Index not taken into account

    Dear experts,
    I have the following problem. Even though table VEPO has the necessary index (by unvel), it is not taken into account in my select statement:
            SELECT SINGLE *
              INTO wa_vepo_line
              FROM *vepo
             WHERE  unvel = wa_vepo_line-venum .
    the index is standard and is called : A : Selection by subordinate shipping unit  and it is certainly activated.
    What might be wrong?
    Thanks in advance for your help.
    Roxani

    hi
    the system picks up the primary keys or index at the runtime...it is not possible to find out which one it will pick!
    try to add vepos in ur where condition.
    avoid select * . specify the required fields IF possible..
    regards,
    madhu

  • WebLogic 11g: Weblogic-Application-Version not taken into account

    Hello,
    I'm trying to set up Production Redeployment on my EAR, under WebLogic 10.3.6.
    However, I can't figure out how to make Weblogic-Application-Version work in MANIFEST.MF
    Here is my EAR structure:
    MyApplication
    --- APP-INF
    --- --- lib
    --- --- classes
    --- META-INF
    --- --- application.xml
    --- --- weblogic-application.xml
    --- --- MANIFEST.MF <- file I update
    --- WebApp1
    --- --- META-INF
    --- --- --- MANIFEST.MF
    --- --- WEB-INF
    --- --- --- classes
    --- --- --- lib
    --- --- --- web.xml
    --- --- --- weblogic.xml
    --- WebApp2
    --- --- same as WebApp1
    MANIFEST.MF content is the following:
    Manifest-Version: 1.0
    Weblogic-Application-Version: VERSION_1
    The problem is that it's not taken into account when I deploy the application in the WebLogic console.
    When I use staged mode, the line Weblogic-Application-Version: VERSION_1 is removed from the staged MANIFEST.MF file.
    The only way to manage to version my application is to add the flag -appversion to weblogic.Deployer, but I don't want to use that on a Production environment (and it's not recommended in the documentation!).
    Any idea why Weblogic-Application-Version not taken into account? I tried to add debug mode to node weblogic/debug but with no success.
    Thanks by advance,
    Julien

    I tested it at my end and was able to redeploy it successfully with the application versions.
    Are you saying that by deploying through weblogic console,the archive version is not honoring/getting picked up .is that correct?
    if yes, what does your weblogic console shows for archive version,when you first try to deploy your application with the VERSION_1.
    I mean,when you are on the "Optional Settings" Page.Do you see the Archive Version as VERSION_1?
    Deployments >> Click on the Install >> select your application ( MyAPP)>> Install this deployment as an application >> Select deployment targets >> Optional Settngs
    Example from the Optional settings Page for My application
    What do you want to name this deployment?
    Name: MyAPP     
    Archive Version:     
    VERSION_1
    This can confirm that the application being deployed is versioned one and if it doesn't shows up,then you need to check why its not getting picked up.?
    Moreover,You can enable the "deploy" debug flag to see whats going wrong in your case during the deployment process
    Server > Debug > Weblogic > Deploy
    Do refer the below viewlet on the Production Redeployment
    http://download.oracle.com/otn_hosted_doc/wls/redeployment/wls-side-by-side-non-annotation_viewlet_swf.html

  • Zero values not taken into account during retraction

    Hello,
    we are having an issue with data retraction.
    When we reset a value to zero, the change in the value is not taken into account and the corresponding value in R/3 remains unchanged.
    There are some posts that say that the function modules UPR_COST_PLAN_EXEC and UPR_COST_PLAN_INIT should be modified in order to populate <xth_data> with zero values, but debug mode shows that the data are read before calling these functions, and the zero values are
    already ignored at this stage.
    Could anyone help me, please?
    Thank you in advance!

    Here is the second method  -
    Regarding your issue:
    Retraction is based on the records of the SEM buffer. Zero records are
    in general filtered in the buffer, so that they cannot be considered forplanning functions that execute the retraction. I regret but this is
    the normal system behaviour in the standard. As workaround, please use adummy key figure that is set to a value <> 0.
    The only way is a work around. If you fill a dummy key figure with an
    arbitrary value, the frame work recognizes this as not 0, even if the
    proper key figure is 0. In that case the 0 retraction would work. So if
    not existing, create an additional dummy key figure to the cube and workon as described.
    For all the of retractions (CO-OM, IM, PS) BPS planning functions are
    used and they run in the general BPS framework. This framework is
    designed to explicitly not process zero records because of performance
    reasons.
    If zero records would be processed, planning functions would run longer
    but they can't deal with those records anyway. Therefore the developmentdecided on the current design, to delete all zero records first, before
    processing them with a planning function. This decision has been
    discussed several times before and it not likely to be changed.
    Regarding SAP note you can check the work aroud I explained can be foundin the oss note 768822 SEM-BPS: Retraction of WBS element w/ Pushback.
    I hope this explains the reason.
    Please kindly confirm the message if your issue is answered.
    Thanks for your support and cooperation.

  • Complaint wth ref. to invoice in CRM - Copy control not taken into account!

    Hi all,
    We are facing problems when replicating return orders (business object Complaint)  with reference to invoices in ERP.
    The copy control in ERP is not taken into account. Does anyone know if it is possible to check the copy control when replicating the complaint to ERP with reference to the invoice?
    The return order is being replicated correctly and the link to the invoice is shown (so, the flow is correct).
    But:
    1.- The manual pricing conditions in the invoice in ERP are not being copied in the return order.
    2.- Some data as the 'Incoterms' is being redetermined from the customer master data and is not being copied from the invoice.
    3.- Data related to material is also being redetermined when the order is replicated
    The copy control is not being taken into account and we can replicate return orders from CRM to ERP with reference to invoices for which the copy control is not maintained in ERP...
    Please help...
    Thanks in advance!
    Yolanda

    Try something like this, in place of  ls_input_field_names-fieldname = 'ZZDEP2PVNDPGM'..
    ls_input_field_names-objectname = 'ORDERADM_H'.
      ls_input_field_names-ref_handle = '100'.
      insert ls_input_field_names into table lt_input_field_names.
    Edited by: Kevin Alcock on Aug 10, 2009 4:03 PM

  • CK11N  Follow-up material not taken into account when costing

    We are running the standard cost with quantity structure and we have detected that all the follow-up information is not taken into account to determine the bill of materials that will be used to calculate the standard cost.
    SAP helpdesk has answered me the following:
    "The situation in product cost planning is we do not check availability of any material stocks, and therefore we cannot provide the same functionality as MRP. You will have to create a new BOM that includes
    the follow-up material or apply the modification to take into account the follow-up part."
    Please, if anyone knows any user-exit we could use to develop something to fix it, we would appreciate it very much. It is a normal process and we can not say it does not work.
    Regards, and waiting for any solution.
    Yolanda.

    Dear,
    Are you getting these reservation in MD04?
    Check the Planning file entry for these material in MD21 if not then create it through MD20 and take MRP run with planning mode -3 Deleted and recreated. And Processing key NETCH.
    Also check the The firmed receipt is within the rescheduling horizon. The rescheduling horizon is set in Customizing for the MRP group (Transaction OPPR, Table T438M). A default value can be predefined in the Customizing of plant data (Transaction OPPQ, Table T399D). define the longer duration for it.
    The firmed receipt (the MRP element) participates in the rescheduling check. You can set which MRP elements participate in the rescheduling check in the Customizing of the plant data (Transaction OPPQ).
    Else you can carry out modification A using the MD_CHANGE_MRP_DATA BAdI and the CONSIDER_RESB method without making a modification.
    Regards,
    R.Brahmankar

  • Corruption of User privileges on folder hierarchy (not taken into account ANYMORE)

    Hello.
    We use Portal 3.0.6.3.3 (still EA release) on NT with Oracle 8iR2.
    We have set a Content Area and a very classical folder hierarchy to be managed by a single user (called administrator having all privileges on all folders) and other users having just the "Manage Items" and "View Content" for all folders, starting with the root folder.
    It has worked perfectly for a while but after further work on privileges that, unfortunately, we cannot trace back, we have our folders not taking into account the privileges set (under the Access Tab) anymore.
    Actually all users have now all privileges on the whole folder hierarchy. It means, an "Edit Folder" not only provides the "Add Item" link option but also all other options (Folder Properties, etc) are available.
    We have checked again and again the privileges on the CA, root folder, folders etc, tried and removed and re-created the users or new users etc without success.
    We imagine some privilege management-related table of the Content Area/Folder are "corrupted".
    How could we return to a normal situation? On which tables of Portal schema should we investigate and try to find out what is going on?
    Thanks for any help.
    null

    ... still me.
    OK, forget my question. It appears that some "generic" group priviliges (AUTHENTICATED_USERS) had been changed by mistake thus overriding any other settings of the users and groups we were contemplating. So everything is under control now...
    Thanks for having read anyway ;-)

  • [ACCEPT][PROMPT] Input not taken into account in the script

    Hi,
    I'm preparing sql scripts to create schemas, one of the script is the main one.
    In that one I'm requesting the passwords for two users, and a path to create an oracle directory.
    The password for the first user and the path are taken into account.
    But the second password requested seems not to be retained.
    Here is the script:
    SET ECHO OFF;
    spool epss_schema.log;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' hide;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    create user epss10 identified by &EPSS_PASSWD default tablespace users temporary tablespace temp;
    grant dba to epss10;
    connect epss10/&EPSS_PASSWD;
    @EPSS_SCHEMA.sql
    @EPSS_ACTIVITY_CODES_DATA.sql
    @EPSS_BCP_TABLE_DATA.sql
    @EPSS_CALL_SETUP_TYPE.sql
    @EPSS_COUNTRY_CD_DATA.sql
    @EPSS_ERC_KEYWORDS_DATA.sql
    @EPSS_ERC_KEYWORDS_LS_DATA.sql
    @EPSS_ERC_KEYWORDS_PE_DATA.sql
    @EPSS_ERC_KEYWORDS_SH_DATA.sql
    @EPSS_GEN_ACTIVITY_CODES_DATA.sql
    @EPSS_INSTRUMENT_LIST_DATA.sql
    @EPSS_KEYWORDS_DATA.sql
    @EPSS_NACE_CD_DATA.sql
    @EPSS_REVIEW_PANELS_DATA.sql
    @EPSS_REVIEW_PANELS_GEN_DATA.sql
    @EPSS_STATE_DATA.sql
    create or replace directory PIC_IMP_DIR as '&PIC_DIRECTORY';
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    create user epss_dump10 identified by &EPSS_DUMP_PASSWD default tablespace users temporary tablespace temp;
    grant connect,resource,JAVAUSERPRIV to epss_dump10;
    grant CREATE DATABASE LINK to epss_dump10;
    grant CREATE PUBLIC SYNONYM to epss_dump10;     
    grant CREATE SYNONYM to epss_dump10;
    grant CREATE TABLE to epss_dump10;
    grant CREATE VIEW to epss_dump10;
    grant UNLIMITED TABLESPACE to epss_dump10;
    grant read,write on directory PIC_IMP_DIR to epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    @EPSS_DUMP_SCHEMA.sql
    spool off;
    exit;
    I don't understand why the two first work and not the last one.
    I already tried by moving the third ACCEPT at the top of the script without any success.
    I had thought, as I'm frst connecting as sysdba asking for passwords and then connect as another user meaning another session, the script would have lost the variables values. But no as now I'm rerquesting the second password after the second connection.
    Advice are welcome.
    Thanks and regards.

    Hi,
    I just made some extended tests.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' hide;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    The variables have the right information.
    Now create the users will work without any problem
    If I first connect with the first user (EPSS_DUMP) and then connect directly with the other one, that will work.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' ;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    CREATE user epss_dump10 IDENTIFIED BY &EPSS_DUMP_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE user epss10 IDENTIFIED BY &EPSS_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE OR REPLACE directory PIC_IMP_DIR AS '&PIC_DIRECTORY';
    GRANT connect,resource TO epss_dump10;
    GRANT dba TO epss10;
    GRANT connect,resource,JAVAUSERPRIV TO epss_dump10;
    GRANT CREATE DATABASE LINK TO epss_dump10;
    GRANT CREATE PUBLIC SYNONYM TO epss_dump10;     
    GRANT CREATE SYNONYM TO epss_dump10;
    GRANT CREATE TABLE TO epss_dump10;
    GRANT CREATE VIEW TO epss_dump10;
    GRANT UNLIMITED TABLESPACE TO epss_dump10;
    GRANT READ,WRITE ON directory PIC_IMP_DIR TO epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    PROMPT 'EPSS PASSWORD: &EPSS_PASSWD';
    connect epss10/&EPSS_PASSWD;
    Now If I run additional script after I'm connected with the first user and then I connect with the second user, the connection will faill telling wrong login/password.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' ;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    CREATE user epss_dump10 IDENTIFIED BY &EPSS_DUMP_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE user epss10 IDENTIFIED BY &EPSS_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE OR REPLACE directory PIC_IMP_DIR AS '&PIC_DIRECTORY';
    GRANT connect,resource TO epss_dump10;
    GRANT dba TO epss10;
    GRANT connect,resource,JAVAUSERPRIV TO epss_dump10;
    GRANT CREATE DATABASE LINK TO epss_dump10;
    GRANT CREATE PUBLIC SYNONYM TO epss_dump10;     
    GRANT CREATE SYNONYM TO epss_dump10;
    GRANT CREATE TABLE TO epss_dump10;
    GRANT CREATE VIEW TO epss_dump10;
    GRANT UNLIMITED TABLESPACE TO epss_dump10;
    GRANT READ,WRITE ON directory PIC_IMP_DIR TO epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    @EPSS_DUMP_SCHEMA.sql
    @COMPILE_INVALID_OBJECTS.sql
    PROMPT 'EPSS PASSWORD: &EPSS_PASSWD';
    connect epss10/&EPSS_PASSWD;
    My script EPSS_DUMP_SCHEMA.sql is just a script to create tables with constraints, synonyms, views, packages and so on.
    Nothing seems strange.
    Well the only strange thing is that if you call a script inside the main script, it seems that all the variables have lost their values.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' ;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    CREATE user epss_dump10 IDENTIFIED BY &EPSS_DUMP_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE user epss10 IDENTIFIED BY &EPSS_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE OR REPLACE directory PIC_IMP_DIR AS '&PIC_DIRECTORY';
    GRANT connect,resource TO epss_dump10;
    GRANT dba TO epss10;
    GRANT connect,resource,JAVAUSERPRIV TO epss_dump10;
    GRANT CREATE DATABASE LINK TO epss_dump10;
    GRANT CREATE PUBLIC SYNONYM TO epss_dump10;     
    GRANT CREATE SYNONYM TO epss_dump10;
    GRANT CREATE TABLE TO epss_dump10;
    GRANT CREATE VIEW TO epss_dump10;
    GRANT UNLIMITED TABLESPACE TO epss_dump10;
    GRANT READ,WRITE ON directory PIC_IMP_DIR TO epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    @EPSS_DUMP_SCHEMA.sql
    PROMPT 'EPSS PASSWORD: &EPSS_PASSWD';
    PROMPT 'EPSS_DUMP PASSWORD: &EPSS_DUMP_PASSWD';
    PROMPT 'PIC DIRECTORY: &PIC_DIRECTORY';
    The three prompts return
    &EPSS_PASSWD
    &EPSS_DUMP_PASSWD
    &PIC_DIRECTORY
    Thanks for your help.

  • Profiles are not taken into account with DIS

    Hi,
    I created several profiles in the content management.
    They work correctly.
    Now, I have to use Desktop Integration Suite.
    The installation worked correctly.
    The problem is, when a user saves and closes the document, the "check-in" screen is the standard one.
    So, all the metadata are displayed on the screen (nearly 40 - it is becoming a mess for the users to known which ones to fill).
    I would like to use the profiles to reduce the metadata (as it is done for the check-in and search in the content server).
    How can I do it ?
    Does anybody encounter this problem ?
    Any help is welcome.
    Kind regards
    Pierre

    Hi Jason,
    I understood my mistake (but I installed the patch).
    The profile is correctly taken into account.
    But the trigger of my profile is the document type.
    With the DIS, I don't know how to transfer metadata before to open the "standard check-in" page.
    Once the check-in page is opened, if I change the document type, the profile is not activated.
    Do you know if it is possible :
    - To transfer metadata to the content server that could be used to determine the profile, for example by defining properties on the Word document.
    - To change of profiles by changing the value of the trigger used for the profile.
    For the moment, the only trick I've found was to create a "server" by value of the trigger and to put the document type as a data to be remembered.
    At that moment, the profile is correctly loaded.
    But, this solution is not very user-friendy. Do you see another one ?
    Pierre

  • Sales orders ans reservations not taken into account on MPS / MRP

    Hello,
    I am implementing the MPS / MRP and I have the following problem:
    I have modified the necessary material master fields to take into account some materials on the MPS / MRP execution. These fields are MRP type, Lot Size, Strategy group and Availability Check.
    Then, executing the MPS / MRP I have realised that only the sales orders and reservations created after the material master modifications are taken into account when I execute the MPS / MRP.
    I suppose this is because the sales orders and reservations created before the material master modifications did not create the requirements because the availability check was not set for the material and so on... but I need to take into account these "old" reservations and sales orders .
    Is there any way to "update" these orders and reservations?? Or I need to delete them and create new ones??
    There are several sales orders and reservations created before the material master modifications. I think there have to be an standard way to update them.
    Thank you very much in advanced,
    Jordi.

    Dear,
    Are you getting these reservation in MD04?
    Check the Planning file entry for these material in MD21 if not then create it through MD20 and take MRP run with planning mode -3 Deleted and recreated. And Processing key NETCH.
    Also check the The firmed receipt is within the rescheduling horizon. The rescheduling horizon is set in Customizing for the MRP group (Transaction OPPR, Table T438M). A default value can be predefined in the Customizing of plant data (Transaction OPPQ, Table T399D). define the longer duration for it.
    The firmed receipt (the MRP element) participates in the rescheduling check. You can set which MRP elements participate in the rescheduling check in the Customizing of the plant data (Transaction OPPQ).
    Else you can carry out modification A using the MD_CHANGE_MRP_DATA BAdI and the CONSIDER_RESB method without making a modification.
    Regards,
    R.Brahmankar

  • GR processing time not taken into account for exception message in MD04

    Hello Gurus!
    In MD04 when we toggle between displaying at GR date and AV date sometimes there is a shortage in AV date view but none in GR date view due to GR processing time.
    How can we include the GR processing time to be taken into account so that an exception message is shown for the shortage during AV view?
    Thanks in advance!
       -Alvin

    Hi There
    What you meant by shortage,,,
    AV- Available date,
    GR- Availble date+GR processing time
    if you have not set up the GR processing time in Master, it may come as ZERO days,, so AV and GR date may be same,,
    if this is not your quesry, please provide more details on your query,,,
    Thanks
    Senthil

  • Old Purchase orders and template - Price Information not taken into account

    Hi experts,
    In MDM catalog, I have a product with 'Price Information' and with 'Contract price'. I create a shopping cart with the 'Price Information'.
    Then, I create another shopping cart using the anterior througth the functionality 'Old Purchase orders and template'.
    The price taken into account is the 'Contract Price'
    Is this the standard procedure?
    Thanks in advance,
    R Silva

    as far as i know the price preference is in this order
    1. Contract price
    2. Purchase Info record
    3. list price
    so in that case - this is standard procedure
    you can confirm - by doing the same process for an item that  has ONLY price information (list price) and see if that price is copied from old to new SC
    thanks
    -Adrivit

  • Movement type 281, scrap not taken into account

    Dear Experts,
    We have a BOM with 100 components transfered into a project network with transaction CN33.
    In the BOM, on item level, we have stated component scrap %.
    MRP takes these component scrap % into account.
    However, if we confirm a project with transaction CN25 the proposed/booked goods movement 281 does not take the scrap % into account.
    Example material BOM A (base quantity 100) contains component B (quantity 100 and component scrap 10%). MRP calculates with requirement 100 for material BOM A a requirement of 110 for component B. However, if we explode material BOM A to a project with transaction CN33 and we book the confirmation with transaction CN25 the the proposed/booked goods movement 281for component B is 100 (and not the desired 110).
    We would like to take the component scrap % into account at goods issue of the project.
    Is there a setting in customizng or the material masterfor this (or is the only option to use two BOM types for this)?
    regards, Herman Alkemade

    Hi Alex,
    The problem is the Apply button doesn't become active and if you click the O.K bottom the dialog box closes and the setting doesn't change, my work around is to click on the  Trace Fixed radio button and then Trace Can be Move bottom again, this should activate the Apply.  This is a known bug and it's in our system.
    Tien P.
    National Instruments

  • Setting default member of attribute hierarchy in MDX script. Default member not taken into account while using user defined hierarchy

    Hi all
    I have a date dimension that (type time) with attributes
    - [FiscalYear] (type years)
    - [FiscalMonth] (type months)
    - [FiscalWeek] (type weeks)
    In addition to the attributes used in the natural hierarchy, I have an attribute [PerType] containing one member comming from the relational table 'WTD' which corresponds to 'Current Date'. All other members of this attribute hierarchy are calculated members (defined in the MDX script). Examples:
    --Last year
    CREATE MEMBER CURRENTCUBE.[Date].[PerType].[LY] AS NULL,
    VISIBLE = 1;
    SCOPE([Date].[PerType].[LY]);
    SCOPE(DESCENDANTS([Date].[Fiscal].[All], [Date].[Fiscal].[Year], SELF_AND_AFTER));
    THIS = ([Measures].CurrentMember, [Date].[PerType].[WTD], ParallelPeriod([Date].[Fiscal].[Year], 1));
    END SCOPE;
    END SCOPE;
    --Month to date
    CREATE MEMBER CURRENTCUBE.[Date].[PerType].[MTD] AS NULL,
    VISIBLE = 1;
    SCOPE([Date].[PerType].[MTD]);
    SCOPE(DESCENDANTS([Date].[Fiscal].[All], [Date].[Fiscal].[Week], SELF_AND_AFTER));
    THIS = Aggregate(CrossJoin({[Date].[PerType].[WTD]}, MTD([Date].[Fiscal].CurrentMember)));
    END SCOPE;
    END SCOPE;
    --Year to date
    CREATE MEMBER CURRENTCUBE.[Date].[PerType].[YTD] AS NULL,
    VISIBLE = 1;
    SCOPE([Date].[PerType].[YTD]);
    SCOPE(DESCENDANTS([Date].[Fiscal].[All], [Date].[Fiscal].[Period], SELF_AND_AFTER));
    THIS = Aggregate(CrossJoin({[Date].[PerType].[WTD]}, YTD([Date].[Fiscal].CurrentMember)));
    END SCOPE;
    END SCOPE;
    The defalut member of FiscalWeek attribute hierarchy is set dynamically in the MDX script:
    ALTER CUBE CURRENTCUBE UPDATE DIMENSION [Date].[FiscalWeek], DEFAULT_MEMBER =
    Filter(
    [Date].[FiscalWeek].Members,
    [Date].[FiscalWeek].Properties( "FiscalWeekStartDate", TYPED) <= DateAdd("d", -2, CDate(CStr(Month(Now())) + "/" + CStr(Day(Now())) + "/" + CStr(Year(Now()))))
    AND
    [Date].[FiscalWeek].Properties( "FiscalWeekEndDate", TYPED) >= DateAdd("d", -2, CDate(CStr(Month(Now())) + "/" + CStr(Day(Now())) + "/" + CStr(Year(Now()))))
    )(0).PrevMember;
    If I run the following query:
    with member
    measures.x as [Date].[Fiscal].DefaultMember.Name
    measures.y as [Date].[FiscalWeek].DefaultMember.Name
    select
    measures.x,
    measures.y
    } on axis(0)
    from [GLWeekly]
    it gives me back correctly the default member set over the MDX script.
    I order the statements in the MDX Script so that the default period (week) is set at the beginning of the script (just after the calculate).
    I do not understand why creating the following calculated member I am obliged to specify [Date].[Fiscal].CurrentMember in the tuple to have correct results:
    MEMBER [Account].[CoA].[Standard Engagement Revenue (MTD)] AS ([Account].[CoA].[Standard Engagement Revenue], [Date].[PerType].[MTD], [Date].[Fiscal].CurrentMember)
    I would expect that:
    ([Account].[CoA].[Standard Engagement Revenue], [Date].[PerType].[MTD])
    is sufficient.
    If the default week is specified in the slicer using a member of the natural hierachy (=> [Date].[Fiscal].x) it works.
    Why can't SSAS use the default member if it is must defined in the MDX script?
    Can someone explains me this. Thanks a lot in advance.

    Hi Ina,
    have you thought about adding a dynamic statement inside the MDX script? You could define the default member like this:
    ... DEFAULT_MEMBER = iif( Day( Now() ) = 3, <expression for previous month>, <expression for current month> );
    This way you don't need to change it everytime by running a script.
    By the way, what do you mean it doesn't update the default member? When you execute this MDX what does it says?
    with member measures.x as [Dimension].[HierarchyName].DefaultMember.Name
    select { measures.x } on 0 from Cubename
    If this returns the correct name, then the problem is somewhere else. I believe it should return you the correct name. Look here, test this on Adventure Works, statement by statement and see what happens.
    ALTER
    CUBE [Adventure Works]
    UPDATE
    DIMENSION [Product].[Product Categories],
    DEFAULT_MEMBER = [Product].[Product Categories].[Category].&[1]
    with
    member measures.x
    as [Product].[Product Categories].DefaultMember.Name
    select measures.x on 0
    from [Adventure Works]
    ALTER
    CUBE [Adventure Works]
    UPDATE
    DIMENSION [Product].[Product Categories],
    DEFAULT_MEMBER = [Product].[Product Categories].[All Products]
    with
    member measures.x
    as [Product].[Product Categories].DefaultMember.Name
    select measures.x on 0
    from [Adventure Works]
    I think you can see which members are default (on related hierarchies) using
    MDX Studio. This should help you detect which attributes have not moved accordingly and hence cause problems in your report. The usual suspects are those attributes used in your last month reports. If that's too much for you, just copy paste the definition
    of the measure x and use .CurrentMember instead .DefaultMember. And so for all related hierarchies of your dimension. You can run it as one query, just put enough measures (x1, x2, ...), one for each hierarchy, ok?
    Here's a test for Day():
    with
    member measures.y
    as
    iif( Day(Now()) = 28, 'Yes', 'No' )
    select
    measures.y on 0
    from [Adventure Works]
    Today this returns Yes, tomorrow it will be No.
    Ups, I just checked one more thing. When you run the script, it sets the default member only for that session. If you execute the first two of the four statements that I've sent you, it will set the default member on Bikes and show you that.
    But, if you open another query windows and execute that select statement (only), you'll see All member instead. So, it has set it to Bikes only for the currect session. Consequence? You reports are not aware of it. So, better use dynamic statement in
    your MDX script.
    Regards,
    Tomislav Piasevoli
    Business Intelligence Specialist
    www.softpro.hr

Maybe you are looking for

  • Multiple projects in one web role using Azure cloud?

    How to run One webrole with multiple projects using azure? In my solution One WCF service and ane website i want run both on one solution, How to run both in one domain? anilbabu

  • Activating A JComboBox used as a TableCellRender

    In my table there are some columns where a cell can contain multiple rows, so I have created by own CellRenderer which subclasses JPanel and contains a JComboBox showing the first value. (I dont subclass JComboBox directly because there are other thi

  • What problems should be expected after purchasing a macbook in america?

    My friend is purchasing a Mac Book in the united states and then moving to england. Apart from the obvious need for a plug adapter, what other problems will he face? How might this affect the warranty or cover plan? Any advice would be appreciated! i

  • Control Mandatory fields on custom screen in Module pool program

    Hello all, I have developed a custom screen on which I have a table control with 5 fields. I wanted to make the first three fields mandatory. So instead of selecting the entry as "required" from the screen attributes for these input fields, I have lo

  • Problem Related to MIR7 and MIRO?

    Hi All, I want to hide the amount field in MIRO and MIR7, so that no bone could change the amount while passing it. please tell me from where can I hide the same?