BBP: Items not taken into BBP after VAT change

Hello,
I am new in FI-CA and I am facing one issue.
1. We changed VAT from 01.01.2011 by creating new VAT code
2. We are using BBP procedure 2 - with 11 due dates
3. In the amount of the BBP were taken only the items from the old VAT code.
The posting items with new VAT code from the billing document were not taken into budget billing plan.
Any advice where to look to find out problem?
Thanks in advance.
Vlado

hello Amlan,
We have tested these steps
1. The dunning procedure releases as an activity the documents that reach a certain dunning level.
2. The items are submitted to the collection agency using FP03D
3. The users manually recall some of these documents
5. Execute FPCI and like this the recall is included in the file.
If we do the previous everything works fine,but I believe the standard SAP is missing something for cancel release cases, because an inconsisty is being created between SAP and the collection agency. If you are reseting something a notice for the collection agencies would have to be created.
On top of that we cannot exclude recalled items (09) in the selection of FP03D (or FP03DM), because we also have systematic recalls that are done according to the longevity of the documents and according to that longevity placed in different collection agencies.
Do you think this can be reported to SAP as an error? The system in any case should prevent you to do the cancel release if the record is not still in any update file (table dfkkcoli_log) to prevent the inconsistency. Just like the system does not allow you to do the cancel release if the item has not been recall first.
Regards,
Sergio

Similar Messages

  • Zero values not taken into account during retraction

    Hello,
    we are having an issue with data retraction.
    When we reset a value to zero, the change in the value is not taken into account and the corresponding value in R/3 remains unchanged.
    There are some posts that say that the function modules UPR_COST_PLAN_EXEC and UPR_COST_PLAN_INIT should be modified in order to populate <xth_data> with zero values, but debug mode shows that the data are read before calling these functions, and the zero values are
    already ignored at this stage.
    Could anyone help me, please?
    Thank you in advance!

    Here is the second method  -
    Regarding your issue:
    Retraction is based on the records of the SEM buffer. Zero records are
    in general filtered in the buffer, so that they cannot be considered forplanning functions that execute the retraction. I regret but this is
    the normal system behaviour in the standard. As workaround, please use adummy key figure that is set to a value <> 0.
    The only way is a work around. If you fill a dummy key figure with an
    arbitrary value, the frame work recognizes this as not 0, even if the
    proper key figure is 0. In that case the 0 retraction would work. So if
    not existing, create an additional dummy key figure to the cube and workon as described.
    For all the of retractions (CO-OM, IM, PS) BPS planning functions are
    used and they run in the general BPS framework. This framework is
    designed to explicitly not process zero records because of performance
    reasons.
    If zero records would be processed, planning functions would run longer
    but they can't deal with those records anyway. Therefore the developmentdecided on the current design, to delete all zero records first, before
    processing them with a planning function. This decision has been
    discussed several times before and it not likely to be changed.
    Regarding SAP note you can check the work aroud I explained can be foundin the oss note 768822 SEM-BPS: Retraction of WBS element w/ Pushback.
    I hope this explains the reason.
    Please kindly confirm the message if your issue is answered.
    Thanks for your support and cooperation.

  • New values not taken into account in MASS / XD99

    I am trying to use a background job to block Customers for Billing at month end.
    In MASS/XD99  I created a variant to update field KNVV-FAKSD (Sales Area Billing Block).
    But running the variant in background (using MASSBACK), it seems the new values I thought I saved in the variant in MASS/XD99 are not taken into account.
    I opened the variant again in MASS/XD99 and indeed it seems it had not saved the new values, because the screen again shows the current values.
    Tried a couple of times again, but everytime I open the variant I have saved again, the new values revert back to 'current values'.
    Is this the reason MASSBACK is also making no updates? What am I doing wrong?
    Thanks in advance for your comments & advice.

    Here is the second method  -
    Regarding your issue:
    Retraction is based on the records of the SEM buffer. Zero records are
    in general filtered in the buffer, so that they cannot be considered forplanning functions that execute the retraction. I regret but this is
    the normal system behaviour in the standard. As workaround, please use adummy key figure that is set to a value <> 0.
    The only way is a work around. If you fill a dummy key figure with an
    arbitrary value, the frame work recognizes this as not 0, even if the
    proper key figure is 0. In that case the 0 retraction would work. So if
    not existing, create an additional dummy key figure to the cube and workon as described.
    For all the of retractions (CO-OM, IM, PS) BPS planning functions are
    used and they run in the general BPS framework. This framework is
    designed to explicitly not process zero records because of performance
    reasons.
    If zero records would be processed, planning functions would run longer
    but they can't deal with those records anyway. Therefore the developmentdecided on the current design, to delete all zero records first, before
    processing them with a planning function. This decision has been
    discussed several times before and it not likely to be changed.
    Regarding SAP note you can check the work aroud I explained can be foundin the oss note 768822 SEM-BPS: Retraction of WBS element w/ Pushback.
    I hope this explains the reason.
    Please kindly confirm the message if your issue is answered.
    Thanks for your support and cooperation.

  • Index not taken into account

    Dear experts,
    I have the following problem. Even though table VEPO has the necessary index (by unvel), it is not taken into account in my select statement:
            SELECT SINGLE *
              INTO wa_vepo_line
              FROM *vepo
             WHERE  unvel = wa_vepo_line-venum .
    the index is standard and is called : A : Selection by subordinate shipping unit  and it is certainly activated.
    What might be wrong?
    Thanks in advance for your help.
    Roxani

    hi
    the system picks up the primary keys or index at the runtime...it is not possible to find out which one it will pick!
    try to add vepos in ur where condition.
    avoid select * . specify the required fields IF possible..
    regards,
    madhu

  • Complaint wth ref. to invoice in CRM - Copy control not taken into account!

    Hi all,
    We are facing problems when replicating return orders (business object Complaint)  with reference to invoices in ERP.
    The copy control in ERP is not taken into account. Does anyone know if it is possible to check the copy control when replicating the complaint to ERP with reference to the invoice?
    The return order is being replicated correctly and the link to the invoice is shown (so, the flow is correct).
    But:
    1.- The manual pricing conditions in the invoice in ERP are not being copied in the return order.
    2.- Some data as the 'Incoterms' is being redetermined from the customer master data and is not being copied from the invoice.
    3.- Data related to material is also being redetermined when the order is replicated
    The copy control is not being taken into account and we can replicate return orders from CRM to ERP with reference to invoices for which the copy control is not maintained in ERP...
    Please help...
    Thanks in advance!
    Yolanda

    Try something like this, in place of  ls_input_field_names-fieldname = 'ZZDEP2PVNDPGM'..
    ls_input_field_names-objectname = 'ORDERADM_H'.
      ls_input_field_names-ref_handle = '100'.
      insert ls_input_field_names into table lt_input_field_names.
    Edited by: Kevin Alcock on Aug 10, 2009 4:03 PM

  • WebLogic 11g: Weblogic-Application-Version not taken into account

    Hello,
    I'm trying to set up Production Redeployment on my EAR, under WebLogic 10.3.6.
    However, I can't figure out how to make Weblogic-Application-Version work in MANIFEST.MF
    Here is my EAR structure:
    MyApplication
    --- APP-INF
    --- --- lib
    --- --- classes
    --- META-INF
    --- --- application.xml
    --- --- weblogic-application.xml
    --- --- MANIFEST.MF <- file I update
    --- WebApp1
    --- --- META-INF
    --- --- --- MANIFEST.MF
    --- --- WEB-INF
    --- --- --- classes
    --- --- --- lib
    --- --- --- web.xml
    --- --- --- weblogic.xml
    --- WebApp2
    --- --- same as WebApp1
    MANIFEST.MF content is the following:
    Manifest-Version: 1.0
    Weblogic-Application-Version: VERSION_1
    The problem is that it's not taken into account when I deploy the application in the WebLogic console.
    When I use staged mode, the line Weblogic-Application-Version: VERSION_1 is removed from the staged MANIFEST.MF file.
    The only way to manage to version my application is to add the flag -appversion to weblogic.Deployer, but I don't want to use that on a Production environment (and it's not recommended in the documentation!).
    Any idea why Weblogic-Application-Version not taken into account? I tried to add debug mode to node weblogic/debug but with no success.
    Thanks by advance,
    Julien

    I tested it at my end and was able to redeploy it successfully with the application versions.
    Are you saying that by deploying through weblogic console,the archive version is not honoring/getting picked up .is that correct?
    if yes, what does your weblogic console shows for archive version,when you first try to deploy your application with the VERSION_1.
    I mean,when you are on the "Optional Settings" Page.Do you see the Archive Version as VERSION_1?
    Deployments >> Click on the Install >> select your application ( MyAPP)>> Install this deployment as an application >> Select deployment targets >> Optional Settngs
    Example from the Optional settings Page for My application
    What do you want to name this deployment?
    Name: MyAPP     
    Archive Version:     
    VERSION_1
    This can confirm that the application being deployed is versioned one and if it doesn't shows up,then you need to check why its not getting picked up.?
    Moreover,You can enable the "deploy" debug flag to see whats going wrong in your case during the deployment process
    Server > Debug > Weblogic > Deploy
    Do refer the below viewlet on the Production Redeployment
    http://download.oracle.com/otn_hosted_doc/wls/redeployment/wls-side-by-side-non-annotation_viewlet_swf.html

  • CK11N  Follow-up material not taken into account when costing

    We are running the standard cost with quantity structure and we have detected that all the follow-up information is not taken into account to determine the bill of materials that will be used to calculate the standard cost.
    SAP helpdesk has answered me the following:
    "The situation in product cost planning is we do not check availability of any material stocks, and therefore we cannot provide the same functionality as MRP. You will have to create a new BOM that includes
    the follow-up material or apply the modification to take into account the follow-up part."
    Please, if anyone knows any user-exit we could use to develop something to fix it, we would appreciate it very much. It is a normal process and we can not say it does not work.
    Regards, and waiting for any solution.
    Yolanda.

    Dear,
    Are you getting these reservation in MD04?
    Check the Planning file entry for these material in MD21 if not then create it through MD20 and take MRP run with planning mode -3 Deleted and recreated. And Processing key NETCH.
    Also check the The firmed receipt is within the rescheduling horizon. The rescheduling horizon is set in Customizing for the MRP group (Transaction OPPR, Table T438M). A default value can be predefined in the Customizing of plant data (Transaction OPPQ, Table T399D). define the longer duration for it.
    The firmed receipt (the MRP element) participates in the rescheduling check. You can set which MRP elements participate in the rescheduling check in the Customizing of the plant data (Transaction OPPQ).
    Else you can carry out modification A using the MD_CHANGE_MRP_DATA BAdI and the CONSIDER_RESB method without making a modification.
    Regards,
    R.Brahmankar

  • Call list: calling hours not taken into consideration

    Hi Gurus,
    I have a question. We are using CRM 7.0 EHP1, and especially marketing functionality.
    We are generating call lists from a marketing campaign. What we see is that the customer calling hours are not taken into consideration when creating the call list.
    However, if we create the call list the "old-fashioned" way (using CRMD_CALL_LIST), we have the option "scheduling... based on calling hours".
    How can I achieve that calling hours are also considered when creating the call list from the campaign?
    Is there a customizing setting, or do I need to activate a BAdi?
    thanks.
    regards,
    Wim Olieman

    Also, keep in mind that MRP calculates availability of inspection stock in two ways.
    Before an inspection lot is created, (i.e. a purchase order or production order is out there and planned already, MRP uses the expected GR date plus the Avg. GR days on the main QM view for the material to determine the availability date.
    Once the inspection lot is created however, MRP uses the end date of the inspection lot which is the lot creation date plus the avg insp. duration in the inspection type setup in the material master.  This can result in some minor differences if you aren't careful in setting your various fields in the material master.
    If a material is going to reside in QI for an extended time, the lab should use QA02 to modify the end date of the inspection lot.
    FF

  • Apple tv not accepting itunes password after i changed it on the computer

    Apple tv not accepting itunes password after i changed it on my mac. Mac opening itunes account with that password

    I don't have Apple TV, but are you able to log out of your account on it and then log back in with the new password so as to 'refresh' the account on it ?

  • Taxes not taken into consideration on consignment items

    We are having an issue when we are transacting consignment material from consignment to unrestricted inventory.  When the movement 411K is used no taxes are transacted causing an issue with our moving average price (MAP).  When a purchase order (PO) is received into unrestricted from a regular PO using a 101 movement type taxes are taken into consideration and applied to the MAP.  If we do the same with a consignment PO is used the material is transacted to consignment stock and no taxes are considered then when the material is moved to unrestricted, as pointed out above, using a 411K no taxes are taken into consideration causing the MAP to be understated.
    Has anyone else seen this issue and if so what was done to alleviate the problem.

    Hi Amit,
    Thank you very much for your response.  Yes, I understand this portion of the issue that IM does not take tax into consideration.  Our problem is that when the materials are received into inventory via a consignment PO taxes are never considered and posted to our moving average price.  Taxes are only considered at invoicing or settlement time and never posted properly to the material.  Then a manual posting must take place to clear the imbalance.  Does this make sense?
    Thank you for your time.
    Best regards,
    Steve

  • Sales orders ans reservations not taken into account on MPS / MRP

    Hello,
    I am implementing the MPS / MRP and I have the following problem:
    I have modified the necessary material master fields to take into account some materials on the MPS / MRP execution. These fields are MRP type, Lot Size, Strategy group and Availability Check.
    Then, executing the MPS / MRP I have realised that only the sales orders and reservations created after the material master modifications are taken into account when I execute the MPS / MRP.
    I suppose this is because the sales orders and reservations created before the material master modifications did not create the requirements because the availability check was not set for the material and so on... but I need to take into account these "old" reservations and sales orders .
    Is there any way to "update" these orders and reservations?? Or I need to delete them and create new ones??
    There are several sales orders and reservations created before the material master modifications. I think there have to be an standard way to update them.
    Thank you very much in advanced,
    Jordi.

    Dear,
    Are you getting these reservation in MD04?
    Check the Planning file entry for these material in MD21 if not then create it through MD20 and take MRP run with planning mode -3 Deleted and recreated. And Processing key NETCH.
    Also check the The firmed receipt is within the rescheduling horizon. The rescheduling horizon is set in Customizing for the MRP group (Transaction OPPR, Table T438M). A default value can be predefined in the Customizing of plant data (Transaction OPPQ, Table T399D). define the longer duration for it.
    The firmed receipt (the MRP element) participates in the rescheduling check. You can set which MRP elements participate in the rescheduling check in the Customizing of the plant data (Transaction OPPQ).
    Else you can carry out modification A using the MD_CHANGE_MRP_DATA BAdI and the CONSIDER_RESB method without making a modification.
    Regards,
    R.Brahmankar

  • [ACCEPT][PROMPT] Input not taken into account in the script

    Hi,
    I'm preparing sql scripts to create schemas, one of the script is the main one.
    In that one I'm requesting the passwords for two users, and a path to create an oracle directory.
    The password for the first user and the path are taken into account.
    But the second password requested seems not to be retained.
    Here is the script:
    SET ECHO OFF;
    spool epss_schema.log;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' hide;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    create user epss10 identified by &EPSS_PASSWD default tablespace users temporary tablespace temp;
    grant dba to epss10;
    connect epss10/&EPSS_PASSWD;
    @EPSS_SCHEMA.sql
    @EPSS_ACTIVITY_CODES_DATA.sql
    @EPSS_BCP_TABLE_DATA.sql
    @EPSS_CALL_SETUP_TYPE.sql
    @EPSS_COUNTRY_CD_DATA.sql
    @EPSS_ERC_KEYWORDS_DATA.sql
    @EPSS_ERC_KEYWORDS_LS_DATA.sql
    @EPSS_ERC_KEYWORDS_PE_DATA.sql
    @EPSS_ERC_KEYWORDS_SH_DATA.sql
    @EPSS_GEN_ACTIVITY_CODES_DATA.sql
    @EPSS_INSTRUMENT_LIST_DATA.sql
    @EPSS_KEYWORDS_DATA.sql
    @EPSS_NACE_CD_DATA.sql
    @EPSS_REVIEW_PANELS_DATA.sql
    @EPSS_REVIEW_PANELS_GEN_DATA.sql
    @EPSS_STATE_DATA.sql
    create or replace directory PIC_IMP_DIR as '&PIC_DIRECTORY';
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    create user epss_dump10 identified by &EPSS_DUMP_PASSWD default tablespace users temporary tablespace temp;
    grant connect,resource,JAVAUSERPRIV to epss_dump10;
    grant CREATE DATABASE LINK to epss_dump10;
    grant CREATE PUBLIC SYNONYM to epss_dump10;     
    grant CREATE SYNONYM to epss_dump10;
    grant CREATE TABLE to epss_dump10;
    grant CREATE VIEW to epss_dump10;
    grant UNLIMITED TABLESPACE to epss_dump10;
    grant read,write on directory PIC_IMP_DIR to epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    @EPSS_DUMP_SCHEMA.sql
    spool off;
    exit;
    I don't understand why the two first work and not the last one.
    I already tried by moving the third ACCEPT at the top of the script without any success.
    I had thought, as I'm frst connecting as sysdba asking for passwords and then connect as another user meaning another session, the script would have lost the variables values. But no as now I'm rerquesting the second password after the second connection.
    Advice are welcome.
    Thanks and regards.

    Hi,
    I just made some extended tests.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' hide;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    The variables have the right information.
    Now create the users will work without any problem
    If I first connect with the first user (EPSS_DUMP) and then connect directly with the other one, that will work.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' ;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    CREATE user epss_dump10 IDENTIFIED BY &EPSS_DUMP_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE user epss10 IDENTIFIED BY &EPSS_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE OR REPLACE directory PIC_IMP_DIR AS '&PIC_DIRECTORY';
    GRANT connect,resource TO epss_dump10;
    GRANT dba TO epss10;
    GRANT connect,resource,JAVAUSERPRIV TO epss_dump10;
    GRANT CREATE DATABASE LINK TO epss_dump10;
    GRANT CREATE PUBLIC SYNONYM TO epss_dump10;     
    GRANT CREATE SYNONYM TO epss_dump10;
    GRANT CREATE TABLE TO epss_dump10;
    GRANT CREATE VIEW TO epss_dump10;
    GRANT UNLIMITED TABLESPACE TO epss_dump10;
    GRANT READ,WRITE ON directory PIC_IMP_DIR TO epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    PROMPT 'EPSS PASSWORD: &EPSS_PASSWD';
    connect epss10/&EPSS_PASSWD;
    Now If I run additional script after I'm connected with the first user and then I connect with the second user, the connection will faill telling wrong login/password.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' ;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    CREATE user epss_dump10 IDENTIFIED BY &EPSS_DUMP_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE user epss10 IDENTIFIED BY &EPSS_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE OR REPLACE directory PIC_IMP_DIR AS '&PIC_DIRECTORY';
    GRANT connect,resource TO epss_dump10;
    GRANT dba TO epss10;
    GRANT connect,resource,JAVAUSERPRIV TO epss_dump10;
    GRANT CREATE DATABASE LINK TO epss_dump10;
    GRANT CREATE PUBLIC SYNONYM TO epss_dump10;     
    GRANT CREATE SYNONYM TO epss_dump10;
    GRANT CREATE TABLE TO epss_dump10;
    GRANT CREATE VIEW TO epss_dump10;
    GRANT UNLIMITED TABLESPACE TO epss_dump10;
    GRANT READ,WRITE ON directory PIC_IMP_DIR TO epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    @EPSS_DUMP_SCHEMA.sql
    @COMPILE_INVALID_OBJECTS.sql
    PROMPT 'EPSS PASSWORD: &EPSS_PASSWD';
    connect epss10/&EPSS_PASSWD;
    My script EPSS_DUMP_SCHEMA.sql is just a script to create tables with constraints, synonyms, views, packages and so on.
    Nothing seems strange.
    Well the only strange thing is that if you call a script inside the main script, it seems that all the variables have lost their values.
    Code :
    SET ECHO OFF;
    spool epss_dump_schema.log;
    ACCEPT EPSS_DUMP_PASSWD char PROMPT 'Enter EPSS_DUMP Schema password: ' hide;
    ACCEPT EPSS_PASSWD char PROMPT 'Enter EPSS Schema password: ' ;
    ACCEPT PIC_DIRECTORY char PROMPT 'Enter The Directory where the PIC dump files will be stored:' ;
    prompt &EPSS_DUMP_PASSWD;
    prompt &EPSS_PASSWD;
    prompt &PIC_DIRECTORY;
    CREATE user epss_dump10 IDENTIFIED BY &EPSS_DUMP_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE user epss10 IDENTIFIED BY &EPSS_PASSWD DEFAULT tablespace users TEMPORARY tablespace temp;
    CREATE OR REPLACE directory PIC_IMP_DIR AS '&PIC_DIRECTORY';
    GRANT connect,resource TO epss_dump10;
    GRANT dba TO epss10;
    GRANT connect,resource,JAVAUSERPRIV TO epss_dump10;
    GRANT CREATE DATABASE LINK TO epss_dump10;
    GRANT CREATE PUBLIC SYNONYM TO epss_dump10;     
    GRANT CREATE SYNONYM TO epss_dump10;
    GRANT CREATE TABLE TO epss_dump10;
    GRANT CREATE VIEW TO epss_dump10;
    GRANT UNLIMITED TABLESPACE TO epss_dump10;
    GRANT READ,WRITE ON directory PIC_IMP_DIR TO epss_dump10;
    connect epss_dump10/&EPSS_DUMP_PASSWD
    @EPSS_DUMP_SCHEMA.sql
    PROMPT 'EPSS PASSWORD: &EPSS_PASSWD';
    PROMPT 'EPSS_DUMP PASSWORD: &EPSS_DUMP_PASSWD';
    PROMPT 'PIC DIRECTORY: &PIC_DIRECTORY';
    The three prompts return
    &EPSS_PASSWD
    &EPSS_DUMP_PASSWD
    &PIC_DIRECTORY
    Thanks for your help.

  • Corruption of User privileges on folder hierarchy (not taken into account ANYMORE)

    Hello.
    We use Portal 3.0.6.3.3 (still EA release) on NT with Oracle 8iR2.
    We have set a Content Area and a very classical folder hierarchy to be managed by a single user (called administrator having all privileges on all folders) and other users having just the "Manage Items" and "View Content" for all folders, starting with the root folder.
    It has worked perfectly for a while but after further work on privileges that, unfortunately, we cannot trace back, we have our folders not taking into account the privileges set (under the Access Tab) anymore.
    Actually all users have now all privileges on the whole folder hierarchy. It means, an "Edit Folder" not only provides the "Add Item" link option but also all other options (Folder Properties, etc) are available.
    We have checked again and again the privileges on the CA, root folder, folders etc, tried and removed and re-created the users or new users etc without success.
    We imagine some privilege management-related table of the Content Area/Folder are "corrupted".
    How could we return to a normal situation? On which tables of Portal schema should we investigate and try to find out what is going on?
    Thanks for any help.
    null

    ... still me.
    OK, forget my question. It appears that some "generic" group priviliges (AUTHENTICATED_USERS) had been changed by mistake thus overriding any other settings of the users and groups we were contemplating. So everything is under control now...
    Thanks for having read anyway ;-)

  • Old Purchase orders and template - Price Information not taken into account

    Hi experts,
    In MDM catalog, I have a product with 'Price Information' and with 'Contract price'. I create a shopping cart with the 'Price Information'.
    Then, I create another shopping cart using the anterior througth the functionality 'Old Purchase orders and template'.
    The price taken into account is the 'Contract Price'
    Is this the standard procedure?
    Thanks in advance,
    R Silva

    as far as i know the price preference is in this order
    1. Contract price
    2. Purchase Info record
    3. list price
    so in that case - this is standard procedure
    you can confirm - by doing the same process for an item that  has ONLY price information (list price) and see if that price is copied from old to new SC
    thanks
    -Adrivit

  • GR processing time not taken into account for exception message in MD04

    Hello Gurus!
    In MD04 when we toggle between displaying at GR date and AV date sometimes there is a shortage in AV date view but none in GR date view due to GR processing time.
    How can we include the GR processing time to be taken into account so that an exception message is shown for the shortage during AV view?
    Thanks in advance!
       -Alvin

    Hi There
    What you meant by shortage,,,
    AV- Available date,
    GR- Availble date+GR processing time
    if you have not set up the GR processing time in Master, it may come as ZERO days,, so AV and GR date may be same,,
    if this is not your quesry, please provide more details on your query,,,
    Thanks
    Senthil

Maybe you are looking for

  • Trying to track my stolen Macbook pro...any ideas will help?

    Our home was recently broken into they made off with 2 Mac. One was a Macbook Pro purchased in 2009, and a New IMac desk top monitor (dummies left the mouse and key board). We hadent downloaded any type of "ifind" apps to either computer, that we kno

  • Message -- MIME -- Combine Messages

    I encountered the text of the subject line in an email that I sent to myself from a copy machine. The copy machine was suppose to scan the pages, convert them to PDF, and send them to my email account. Apparently the PDF attachment was too large for

  • Has anyone tried to use the example XMLResourceAdapter?

    I get various java errors when trying to use the sample XML file resource adapter code supplied in the /idm2525/REF folder. I believe my java environment is Ok. I compiled the XMLResourceAdapter.java as per instructions. I get the 3 class files. I ad

  • Can we pass temporary result set to the  procedure?

    Hi, The result set is stored in a temporary storage. Can we pass that resultset to the procedure? Thanks and regards Gowtham Sen.

  • Question-Star Rating in Elements 11

    Please Help--Question about Star ratings...There are no stars under any of my photos.  It has always been that way.  Can't figure out how to activate stars under my downloaded photos.  Is something turned off or did I get a broken version of photosho