Issue with Delta Request

Hi Friends,
I have an issue with loading.
1. From source 2lis_13_VDITM, data loads to 2 targets.
ZIC_SEG, 0SD_C03.and from 0sd_C03 it again loads to ZSD_C03W and ZSD_C03M through DTP.
I have done a repair full load on 08.08.2011 to PSA and loaded this request manually to cube ZIC_seg.
I forgoted to delete this request in PSA as i dont want to load this to other targets.
Before i notice ,already delta requests got loaded and has pulled my repair full request also to other targets.
As i have not done any selective deltions on other cubes there may be double entries.
I am planning to do the below steps inorder to rectify the issue.
1. Do a selective deletion of the delta request in all 3 targets which got loaded on 8th along with repair full.
2. Delete the repair full request from PSA.
So now delta request which got loaded after repair full request, is left in PSA.
My question is if my PC runs today will this delta request in PSA also pulls again to cube through DTP?
Kindly share any other ideas please urgent...
Regards,
Banu

Hi Banu,
If the data in the CUBE's is not compressed, then follow the below steps
1)Delete the latest request in CUBE ZIC_SEG
2)Delete the latest request from cubes  ZSD_C03W and ZSD_C03M and then from 0SD_C03.
3)load the delta request manually using DTP from your cubes(in DTP you can load by giving request number). to cubes ZIC_SEG and 0SD_C03.
3)now run the delta DTP to load the delta request from 0SD_C03 to ZSD_C03W and ZSD_C03M.
Next when your PC runs it will load only that particular day delta request.
It will work
Regards,
Venkatesh

Similar Messages

  • Issue with Delta in Function Module

    Hi Team,
    I have an issue with delta in Genric extraction using function module.Full load is working fine and i have taken post_date as delta field.plz chk the code if any delta related statements are missing.
    FUNCTION ZRSAX_BIW_MANGEMENT_RAT .
    ""Local interface:
    *"  IMPORTING
    *"     VALUE(I_REQUNR) TYPE  SRSC_S_IF_SIMPLE-REQUNR
    *"     VALUE(I_DSOURCE) TYPE  SRSC_S_IF_SIMPLE-DSOURCE OPTIONAL
    *"     VALUE(I_MAXSIZE) TYPE  SRSC_S_IF_SIMPLE-MAXSIZE OPTIONAL
    *"     VALUE(I_INITFLAG) TYPE  SRSC_S_IF_SIMPLE-INITFLAG OPTIONAL
    *"     VALUE(I_READ_ONLY) TYPE  SRSC_S_IF_SIMPLE-READONLY OPTIONAL
    *"  TABLES
    *"      I_T_SELECT TYPE  SRSC_S_IF_SIMPLE-T_SELECT OPTIONAL
    *"      I_T_FIELDS TYPE  SRSC_S_IF_SIMPLE-T_FIELDS OPTIONAL
    *"      E_T_DATA STRUCTURE  ZQMBW_FUJ_MANAGEMENT OPTIONAL
    *"  EXCEPTIONS
    *"      NO_MORE_DATA
    *"      ERROR_PASSED_TO_MESS_HANDLER
    Example: DataSource for table MANAGEMENT RATING
      TABLES: ZQMBW_MANAGEMENT.
    Auxiliary Selection criteria structure
      DATA: L_S_SELECT TYPE SRSC_S_SELECT.
    Maximum number of lines for DB table
      STATICS: S_S_IF TYPE SRSC_S_IF_SIMPLE,
    counter
              S_COUNTER_DATAPAKID LIKE SY-TABIX,
    cursor
              S_CURSOR TYPE CURSOR.
      RANGES: POST_DATE FOR ZMMTVEND_RATING-POST_DATE,
              VENDOR FOR ZMMTVEND_RATING-VENDOR.
    Initialization mode (first call by SAPI) or data transfer mode
    (following calls) ?
      IF I_INITFLAG = SBIWA_C_FLAG_ON.
    Initialization: check input parameters
                    buffer input parameters
                    prepare data selection
    Check DataSource validity
        CASE I_DSOURCE.
          WHEN 'ZQMMANAGEMENT_DS'.
          WHEN OTHERS.
            IF 1 = 2. MESSAGE E009(R3). ENDIF.
    this is a typical log call. Please write every error message like this
            LOG_WRITE 'E'                  "message type
                      'R3'                 "message class
                      '009'                "message number
                      I_DSOURCE   "message variable 1
                      ' '.                 "message variable 2
            RAISE ERROR_PASSED_TO_MESS_HANDLER.
        ENDCASE.
        APPEND LINES OF I_T_SELECT TO S_S_IF-T_SELECT.
    Fill parameter buffer for data extraction calls
        S_S_IF-REQUNR    = I_REQUNR.
        S_S_IF-DSOURCE = I_DSOURCE.
        S_S_IF-MAXSIZE   = I_MAXSIZE.
    Fill field list table for an optimized select statement
    (in case that there is no 1:1 relation between InfoSource fields
    and database table fields this may be far from beeing trivial)
        APPEND LINES OF I_T_FIELDS TO S_S_IF-T_FIELDS.
      ELSE.                 "Initialization mode or data extraction ?
        IF S_COUNTER_DATAPAKID = 0.
    Fill range tables BW will only pass down simple selection criteria
    of the type SIGN = 'I' and OPTION = 'EQ' or OPTION = 'BT'.
          LOOP AT S_S_IF-T_SELECT INTO L_S_SELECT WHERE FIELDNM = 'VENDOR'.
            MOVE-CORRESPONDING L_S_SELECT TO VENDOR.
            CALL FUNCTION 'CONVERSION_EXIT_ALPHA_INPUT'
              EXPORTING
                INPUT  = VENDOR-LOW
              IMPORTING
                OUTPUT = VENDOR-LOW.
            CALL FUNCTION 'CONVERSION_EXIT_ALPHA_INPUT'
              EXPORTING
                INPUT  = VENDOR-HIGH
              IMPORTING
                OUTPUT = VENDOR-HIGH.
            APPEND VENDOR.
          ENDLOOP.
          LOOP AT S_S_IF-T_SELECT INTO L_S_SELECT WHERE FIELDNM = 'POST_DATE'.
            MOVE-CORRESPONDING L_S_SELECT TO POST_DATE.
            CONCATENATE L_S_SELECT-LOW6(4)  L_S_SELECT-LOW3(2) L_S_SELECT-LOW+0(2) INTO POST_DATE-LOW.
            CONCATENATE L_S_SELECT-HIGH6(4)  L_S_SELECT-HIGH3(2) L_S_SELECT-HIGH+0(2) INTO POST_DATE-HIGH.
            APPEND POST_DATE.
          ENDLOOP.
    **Get Management rating details
          OPEN CURSOR WITH HOLD S_CURSOR FOR
          SELECT VENDOR POST_DATE OVERALL_MNGT_RAT OVERALL_DEV_RAT FROM ZMMTVEND_RATING WHERE VENDOR IN VENDOR AND POST_DATE IN POST_DATE .
        ENDIF.
    Fetch records into interface table.
        FETCH NEXT CURSOR S_CURSOR
                   APPENDING CORRESPONDING FIELDS
                   OF TABLE E_T_DATA
                   PACKAGE SIZE S_S_IF-MAXSIZE.
        S_COUNTER_DATAPAKID = S_COUNTER_DATAPAKID + 1.
        IF SY-SUBRC <> 0.
          CLOSE CURSOR S_CURSOR.
          RAISE NO_MORE_DATA.
        ENDIF.
      ENDIF.              "Initialization mode or data extraction ?
    ENDFUNCTION.

    Hi
    Check URLs:
    How to populate the ranges using FM for the SELECTs
    Re: Generic Delta Function Module

  • Issue with Delta Load in BI 7.0... Need resolution

    Hi
    I am having difficulty in Delta load which uses a Generic Extractor.  The generic extractor is based on a view of two Tables.  I use the system date to perform the delta load.  If the system date increaes by a day, the load is expected to pick up the extra records.  One of the tables used in the view for master data does not have the system date in it.
    the data does not even come up to PSA.  It keeps saying there are no records....  Is it because I loaded the data for yesterday and manually adding today's data...? 
    Not sure what is the cuase of delta failing....
    Appreciate any suggestions to take care of the issue.
    Thanks.... SMaa

    Hi
    The Generic DataSource supports following delta types:
    1. Calender day
    2. Numeric Pointer
    3. Time stamp
    Calday u2013 it is based on a  calday,  we can run delta only once per day that to at the end of the clock to minimize the missing of delta records.
    Numeric pointer u2013 This type of delta is suitable only when we are extracting data from a table which supports only creation of new records / change of existing records.
    It supports
    Additive Delta: With delta type additive delta, the record to be loaded only returns the respective changes to key figures for key figures that can be aggregated. The extracted data is added in BI (Targets: DSO and InfoCube)
    New status for changed records: With delta type new status for changed records, each of the records to be
    loaded returns the new status for all key figures and characteristics. The values in BI are overwritten (Targets: DSO and Master Data)
    Time stamp u2013 Using timestamp we can run delta multiple times per day but we need to use the safety lower limit and safety upper limit with minimum of 5 minutes.
    As you specified, the DS is based on VIEW (of two tables, with one containing date and other does not).
    Kindly check the above lines and verify, if the view (primary key) could be used to determine the Delta for your DS.
    Also let us the if any standard SAP tables used in creating VIEW and if so, what is the size of the DS load every day?
    Thanks.
    Nazeer

  • Issue with delta load from R/3 to BW

    Hi frnds,
    There is a standarded D.S  2LIS_05_ITEM in R/3. Evereday We extraceted data from R/3 to BW based on this standard Data Source with delta load.There are two fields ZX and ZY in BW ,some times the data haven't extracted for this 2 fields in BW with the delta update.But some times this 2 fields properly extracted in to BW with the same delta update.
    Why is it happen like this ? Please give some inputs on this.
    Regards,
    Satya.

    Hello,
    if its a standard field then its getting populated in correct way.
    if its custom field then you need to analyze the records for which its getting populated and for one which its not.
    Quite possible that some cutomization in CMOD results in this behaviour.
    Also,check underlying tables to see the correct values.
    Regards
    Ajeet

  • Issue with delta queue initialization

    Hello
    I am having a problem with an initialization of the delta queue for a certain extractor. The initialization was successfull, and I can see the extractor in the delta queue maintenance (RSA7) in ECC, but when users are posting transactions in ECC, they are not coming into RSA7.
    The extractor is 0CO_OM_WBS_6, so its not standard LO cockpit where we have V3 (background) jobs which have to be triggered.
    One more thing, the record which was posted was backdated, ie the date is sometime in the past. But I think the delta queue still picks the records. Can i know how the delta queue funtions based on the time stamp?
    Please advise.

    Here is the OOS note fro FI Datasources.
    OSS 485958:
    Symptom
    The FI line item delta DataSources can identify new and changed data only to the day because the source tables only contain the CPU date and not the time as time characteristic for the change.
    This results in a safety interval of at least one day.
    This can cause problems if the batch processing of the period-end closing is still running on the day following the period-end closing and all data of the closed period is to be extracted into BW on this day.
    Example:
    The fiscal year of the company ends on December 31, 2002.
    On January 01, 2003, all data of the closed fiscal year is to be extracted into BW.In this case, the FI line item delta DataSources only extract the data with a safety interval of at least one day.
    This is the data that was posted with a CPU date up to December 31, 2002.
    However, the batch programs for closing the 2002 fiscal year are still running on January 01, 2003.
    The postings created by these programs with a January 01 date for the 2002 fiscal year can therefore only be extracted with delta extraction on January 02, 2003. However, extraction should already be possible on January 01, 2003.
    Other terms
    FI line item delta DataSources
    0FI_AP_4; 0FI_AR_4; 0FI_GL_4
    Reason and Prerequisites
    This problem is caused by the program design.
    Solution
    The functions are enhanced as of PI 2002.1, standard version.
    Due to a customized setting, it is now possible for the FI line item extractors to read the data up to the current date during the delta runs.
    During the next delta run, the data is then read again from the day when the previous extraction was called, up to the current day of the call for the new extraction.This means that the delta extraction of the data always occurs with time intervals overlapping by one day.
    The delta calculation by the ODS objects ensures that the repeatedly extracted data does not cause duplicate values in the InfoCube.
    Activate the new functions as follows:
    1. PI 2001.1 or PI 2001.2:
    Implement the new functions with the correction instructions in this note.As of PI 2002.1, these changes are already implemented in the standard version.
    1. Start a test run for an FI line item DataSource via transaction RSA3.
    2. Two new parameters are now available in the BWOM_SETTINGS table.
    3. Parameter BWFIOVERLA (Default _)
    Set this parameter to "X" using transaction SE16, to activate the new function of overlapping time intervals.
    Caution
    The new function also uses a safety interval of one day to determine the To value of the selection, if the time limit set by BWFITIMBOR is not reached.This means that only data up to the day before the extraction was called is selected, if the extraction starts before 02:00:00 (default for BWFITIMBOR).
    Correction: Parameter BWFITIMBOR (Format HH:MM:SS, default 020000)
    The default value for this parameter is 02:00:00.Do not change this value.
    If this time limit is not reached, only data up to the previous day (with BWFIOVERLA = 'X') or the day before this day (with BWFIOVERLA = ' ') is selected.
    This logic prevents records with a CPU date of the previous day being contained in the posting when the extraction is executed.
    Check any OSS notes for your data source.

  • Issue with deltas

    hello, i am having problems with rsa7 queue. i have to delete the init everyday and recreate every day. all day long it works fine but then it stops working the next day. in my monitor all i see is a yellow request. cna some one tell me how to fix this issue? thanks.

    Why are you deleting INIT everyday and recreating  ? What is the error msg you have seen in Monitor ?
    Is the present load is Init ? or Delta ? what did you see in monitor screen ? Is it still running ?
    can check the jobs in SM37 in BW and R/3 , also check in SM58 / WE02 if any Idocs are stucked

  • Issue with Service Request email Notification

    Hi All,
    Whenever we change the status of SR an email should be triggered. But it is not happening. Is there any setups we have to look for. When ever an event is raised there is no notification sent.
    Please let me know your comments at the earliest as it is an priority issue.
    Thanks,
    Mahesh

    Hi Mohammed,
    Thanks for the reply. I will explain the issue in detail. The issue is with the iSupport.
    User creates an SR in iSupport. Now the status will be in Open. Again user updates the SR by adding some notes by which the status of the SR will be changed to Updated by Customer. At this point we are getting an error notification. This error is as follows
    Event Error Name: WFENG_EVENT_NOTFOUND
    Event Error Message: 3142: Process 'SERVEREQ/100794754-1997529-EVT' has no activities waiting to receive event 'oracle.apps.cs.sr.ServiceRequest.statuschanged'.
    Event Error Stack:
    Wf_Engine.Event(SERVEREQ, 100794754-1997529-EVT, CSEVTPROC, oracle.apps.cs.sr.ServiceRequest.statuschanged)
    Wf_Rule.Default_Rule(oracle.apps.cs.sr.ServiceRequest.statuschanged, CEEC6B02CCB86CD4E0340003BA4D7D08)
    But If we update the same SR from service request form in customer support everything is working fine. \
    Please let me know your feedback
    Thanks,
    Mahesh

  • Seeing issue with XAI request

    Can anyone help me with a XAI request, having 1 payor a/cs, paying for 2 or more payee a/cs using 1 tender. This should create 1 PaymentEventID, 2 PaymentIDs and 1 TenderIDs.
    Thanks in advance.
    Edited by: user8978681 on Feb 3, 2012 1:18 AM

    Hi
    Try to do this online first, making sure you got all the Ids in place. THen mimic this in XAI, using the XAI Dynamic submission. Use SHOW XML to "learn" from the request that is built dynamically. THen try your request in XAI Submission (both under ADMIN)
    Good luck
    Arjen

  • Issues with transport request

    Hi,
    I have this situation:
    There is a table. And multiple changes have been made in the development system (under multiple transport requests). Now what I want to do is, create just one transport request(to be transported to production) which will club all these changes.
    How can I do this? Are there any specific issues that I need to take care of?
    Thanks..

    Hi one simple way u can do it
    In SE01 first u release all the TO that u have created for that table
    Now u goto changemode for the table in SE11
    then click on save, it will ask another req number
    u create new one and save
    and transport this TO number to production.
    that will work
    reward points if helpful
    regs
    Manas Ranjan Panda

  • Issue with Delta upload in Extractor

    Hi all,
    We are using the Extractor 2LIS_11_VAITM to upload Sales Overview cube through Delta Process. It so happens that when ever a new record is entered (To be specific, whenever a new sales order is created) data is extracted and uploaded properly in to the cube.
    But when an existing record is changed(existing sales order), the record is not being extracted to R/3 extractor (2LIS_11_VAITM) which means any CHANGED RECORDS are not being extracted as per the schedule.
    Any hints to solve the issue would be helpful.
    Regards,
    Ram.

    Hello,
    Load is using Direct Delta or Queued Delta or Unserialized V3 Update ?
    Is Data from InfoSource 2LIS_11_VAITM directly feeding the InfoCube or is into a ODS ? If it is directly feeding into InfoCube, do you have document number InfoObject in the Cube ? How are you checking whether existing data is been updated or not or new record arrived into your Cube ? Take any document # which you think got changed in R3 and check in the PSA too.
    problem of not updating existing documents was always there or just happened today ?
    GSM.

  • Issues with Delta Loads.

    Hi All,
    I hope you don’t mind to provide a suggestion about the delta load of Purchase Orders.
    There was a request from Procurement team and they complaint that the data is not looking right.
    For example: for a given purchase order the exemption flag is displayed as ADE.
    Before change:
    Purchase Doc Number
    Puchase Exemption flag
    90002106
    ADE
    They want this to be displayed as:
    After Change :
    Purchase Doc Number
    Puchase Exemption flag
    90002106
    So ABAPER has included the logic in ECC Functional Module to populate the Extract Structure for Data Source to display above result. But the overnight run didn’t pick up the changes to above record in BW. The loads are delta loads.
    So my Question is:
    ·         This load for the Infocube Purchase Orders is delta load.  Does change to Function Module will pick up the changes to old records as well? I mean if we are doing delta load does the old records gets changed and updated in CUBE or should I delete the data from all the PSA and Infocube and run the load into CUBE to pick up the changes. Am bit worried about deleting the data from Infocube as it may have history data (because of delta load).  Total records in cube are 800000 records.
    ·         Another question does the process chain trigger the function module to load the data into Extract Structure of data source in ECC (am bit confused about this process in ECC)?
    Please suggest me how can I capture those changed records?
    Any help would be much appreciated. Please let me know if you need more info.
    Thanks in advance
    Sandeep

    Hi Sandeep,
    I am little confused from your post. You are using  2lis_02_SCL or the  generic FM datasource?.
    Also it looks like new  custom field added in your client. I never saw any purchase exemption flag even though i am working in purchasing flow for  last 2  years.
    W.r.t purchasing , deletion flag will be there  which  will come from EKPO and have  the values S or L (  Deleted or Blocked).
    You can block or delete a PO aslong as GR was not happened. Other wise you can do it. So use the EKPO-LOEKZ for  your use.
    Regards,
    Rajesh

  • Issue with Delta Queue

    Dear All,
    We have scheduled Delta Q on developement and it was working fine.
    But on Quality side all the jobs for Application 11 are failing while application 13 are running successfully.
    Please help on the same immediately.
    Regards,
    Sohil Shah.

    Hi Sohil,
    It seems that you have changed the structure of the datasource. When you change the structure of the datasource you have to make sure that there is no data in the delta queue or SMQ1 queue of the target system before you transport it.  Delete the data from the delta queue and SMQ1 queue and reintialise the datasource.
    Let me know if you have any doubt.
    Regards,
    Viren

  • Issue with deleting a group using Request APIs in OIM 11g R1

    Hi,
    I am facing an issue with Request Based provisioning in OIM 11g R1.
    I am currently testing a scenario where i have imported a data set for 'Modify Provisioned Resource' and am able to add a group/entitlement to an already provisioned resource by using the following code :
            RequestBeneficiaryEntityAttribute childEntityAttribute= new RequestBeneficiaryEntityAttribute();
            childEntityAttribute.setName("AD User Group Details");
            childEntityAttribute.setType(TYPE.String);
            List<RequestBeneficiaryEntityAttribute> childEntityAttributeList=new ArrayList<RequestBeneficiaryEntityAttribute>();
            RequestBeneficiaryEntityAttribute attr = new RequestBeneficiaryEntityAttribute("Group Name", <group>,                                                                       RequestBeneficiaryEntityAttribute.TYPE.String);
            childEntityAttributeList.add(attr);
            childEntityAttribute.setChildAttributes(childEntityAttributeList);
            childEntityAttribute.setAction(RequestBeneficiaryEntityAttribute.ACTION.Add);
            beneficiaryEntityAttributeList = new ArrayList<RequestBeneficiaryEntityAttribute>();   
            beneficiaryEntityAttributeList.add(childEntityAttribute);
            beneficiarytEntity.setEntityData(beneficiaryEntityAttributeList);
    This works fine for adding a group but if i try to remove a group by changing the action to Delete in the same code, the request fails. The only change made is in the following line:
    childEntityAttribute.setAction(RequestBeneficiaryEntityAttribute.ACTION.Delete);
    Could you please suggest where can this possibly be wrong.
    Thanks for your time and help

    Hi BB,
    I am trying to follow up your response.
    You are suggestng to use prepopulate adapter for to populate respource object name, that means We have to just use an sql query from obj tabke to get the resource object name. right ?? it could be like below, what should I have entity-type value here ??
    <AttributeReference name="Field1" attr-ref="act_key"
    available-in-bulk="false" type="Long" length="20" widget="ENTITY" required="true"
    entity-type="????"/>
    <PrePopulationAdapter name="prepopulateResurceObject"
    classname="my.sample.package.prepopulateResurceObject" />
    </AttributeReference>
    <AttributeReference name="Field2" attr-ref="Field2" type="String" length="256" widget="lookup-query"
    available-in-bulk="true" required="true">
    <lookupQuery lookup-query="select lkv_encoded as Value,lkv_decoded as Description from lkv lkv,lku lku
    where lkv.lku_key=lku.lku_key and lku_type_string_key='Lookup.xxx.BO.Field2'
    and instr(lkv_encoded,concat('$Form data.Field1', '~'))>0" display-field="Description" save-field="Value" />
    </AttributeReference>
    Then I need think about the 'Lookup.xxx.BO.Field2' format.
    Could you please let me know if my understanding is correct?? What is the entity-type value of the first attribute reference value?
    Thanks for your all help.

  • Issue with loading of Delta data

    Hi all,
    I have constructed a generic datasource using the tables ESLH,ESLL,ESKL tables with delta field as AEDAT(Changed on) from ESKL table.Created a Infocube and DSO basing on the datasource and loading is completed.Delta loads are running daily.
    Now my question is how to delete the data from the infocube in the BW system if the fields ESLL-DEL(deletion indicator) and ESLL-STOKZ(reversal document) is set to active in the ESLL table in the R/3 system when the delta loading of data is completed.
    can anyone have an idea of how to resolve this issue.
    Thanks in Advance,
    Vinay Kumar

    There are a couple of ways to do this, but I would question your logic as to why you need to delete them first:
    For records that are marked as "deleted" you should just filter these records in the Query designer. Doing it this way is non-destructive, and later if your business needs to know some KPI based around number of deleted records - then you will be able to answer this. If you permanently remove the data, then you cannot answer this KPI.
    For records marked "reversal" you should never remove these because they affect your key figures. A reversal can also be a partial reversal and so your KPIs will be wrong if you don't include them, and in any case in reporting you never see the reversal because almost all KFs are aggregated. The only time you will see a reversal record is when you include the characteristic that identifies it as a reversal, and in most reports that is unnecessary.
    Right - so now to your solution(s) - although I repeat I advise not doing any of these.
    As these are custom data sources, make sure that in your source system you create abap code to fill in thr RECORDMODE for each record.
    This way when you load your data to the DSO, the activation step will take care of processing deletions and reversals.
    Next compress your cube and use selective deletion on the cube to remove the deleted and reversal records.
    Alternatively:
    Compress your cube and use selective deletion on the cube to remove the deleted and reversal records.
    Then in your transformation to the cube, create a start routine which removes the records from the source package. ie the records are never allowed to go to the cube.
    I hope that helps.

  • Request management service issue with SSL Sites

    Hi guys,
    I've configured up the request management service, but after start the service on our wfe servers (even before to configure specific web applications),  our ssl sites begin to fail and the event viewer start to alert several errors, by the other hand,
    sites over 80 port with no SSL works as expected.
    I've found some similiar scenarios without a happy ending. Below you can find some of the main errors in event viewer and ULS Logs. In
    this post  solved a very similar issue using ssl host headers with a unique ip but we are using fqdn certificates so it is not an option for us.
    I've already checked the
    good posts from Spencer Harbar and He says that is good idea to use host named site collection, but this is not an option for me because we need to use Self Service Creation and mixed authetnication, besides, None of the Technet literature that i've reviewed
    says  something about don't support path based site collection.
    So far, the only way to avoid this errors is not using the service(discouraging finding), These are some errors we got on Event viewer and logs:
    Machine 'MACHINENAME (AppPool(_LM_W3SVC_515444293_ROOT))' failed ping validation and has been unavailable since '3/28/2014 3:55:48 PM'.  (Just starting the service Event viewer begis to have a bunch of this errors)
    03/28/2014 13:34:01.96 w3wp.exe (0x1184)
    0x154C SharePoint Foundation
    Request Management ai2q3
    High Reached maximum number of failed machines based on ping results for this routing group
    53c2819c-8216-20f3-68c0-c0a3e55c92d5
    03/28/2014 13:34:01.96 w3wp.exe (0x1184)
    0x154C SharePoint Foundation
    Request Management ai2q4
    Medium Unavailable machines based on ping results: MACHINENAME
    53c2819c-8216-20f3-68c0-c0a3e55c92d5
    03/28/2014 13:34:01.97 w3wp.exe (0x1184)
    0x1C2C SharePoint Foundation
    Request Management adc7u
    Medium Mapping URI from 'https://HOSTNAME:443/Style Library/somos-sura-css/inicio/inicio.css' to 'https://MACHINENAME/Style%20Library/somos-sura-css/inicio/inicio.css'
    53c2819c-8217-20f3-68c0-cb2f392c388b
    Seems like it tries to find a site with the machine name, not the host name registered in the Alter access mappings (like portal.acme.com) .
    I would appreciate some feedback about this. Another posts with similar issues:
    http://amolmeshe.blogspot.com/2013/05/sharepoint-2013-request-management.html 
    http://www.akspug.org/Blog/Post/4/Request-Management-and-Error-8316 
    Regards,

    Hi,
    According to your post, my understanding is that you had an issue about the Request management service with SSL sites.
    It’s a known issue that the request management service could not work with the SSL site.
    We had already reported the issue to the product team, as a workaround, if you want to use the request management service, you can change the https site back to http.
    What’s more, the SharePoint 2013 SP1 has been relased, you can install it to check whether it works.
    http://support.microsoft.com/kb/2817429/en-us
    Thanks & Regards,
    Jason
    Jason Guo
    TechNet Community Support

Maybe you are looking for

  • Why don't some of my .ttf fonts appear in the font menu?

    Hello, I have a number of .ttf fonts installed on my system (located in Win7/Windows/Fonts folder) which do not appear in Flash or other Adobe applications such as InDesign. I did try putting them in the Fonts folder of these programs instead of the

  • Can I use Adobe CS 5.5 Production Premium legally in Europe (Poland)?

    I have bought the above product second-hand in Poland and I am waiting for the transfer of license to complete. Adobe Poland support gave me an information that the Universal English version is not considered legal outside US. There is however no suc

  • Return statement and Try Catch problem

    Hi!! I've got the next code: public ResultSet DBSelectTeam(String query) { try { Statement s = con.createStatement(); ResultSet rs = s.executeQuery(query); return rs; } catch (Exception err) { JOptionPane.showMessageDialog(null, "ERROR: " + err); But

  • Flags and Follow-up

    Unless I am going loopy and suffering a flashback to my Entourage days, there used to be a feature in Mail that enabled me, when flagging a message for follow-up, to also set a reminder to follow-up it up by a certain date and time (automatically set

  • 80g Classic: Cover Flow Not In Alphabetical Order

    Actually, most of it is in alphabetical order. But I have an album by "Mandalay" that shows up last (yup, after the "Z" artists) when in cover flow. Also an album by "Ohgr" shows up twice, both with the complete set of songs (1st instance is correct