DMVPN: Too many spoke to spoke tunnels created!

Hi all,
We have a customer wiht a DMVPN network phase 2. There are around 40 spokes. What I can see on all the spokes is between 10 and 15 IPsec spoke to spoke tunnels which is really strange as most of the spoke locations should only have between 1 to 4 spoke to spoke tunnels maximum.
My question is what could cause the spoke routers to create an IPsec spoke to spoke tunnel? Some weired applications?
IOS version: 12.4(15)T11 Platform: Cisco 1812.
Best regards,
Laurent

Hi,
But there may be an apllication that trigger this Spoke to spoke tunnels?
I cannot in the inspect sessions the IP of the destination spokes. So it looks it is a quick trigger, then inspect session time ud og IPsec SAs stayded for the defaut life time.
Regards,
Laurent

Similar Messages

  • Hdiutil -- trying to create a sparseimage ends with "too many processes"

    Hello
    I've been trying to create a sparseimage file using both the Disk Util app and the hdiutil command-line tool to use with the Time Machine from a network folder. However, both fail with a "too many processes" error.
    % hdiutil create -library SPUD -size 1g -fs HFS+J -type SPARSEBUNDLE -volname lala  lala
    hdiutil: create failed - Too many processes
    I've tried to search for it on Google without success. Is anybody else getting such errors?
    Kind regards,
    Mauricio

    Printing different images on one page

  • DMVPN - Spoke to spoke direct tunnel

    For some reason, my spoke to spoke tunnel doesn't stay up when my hub goes down. Not sure what I am missing. Please help!!
    Hub:
    interface Tunnel0
     ip address 172.16.1.1 255.255.255.0
     no ip redirects
     no ip next-hop-self eigrp 10
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp network-id 1
     no ip split-horizon
     ip summary-address eigrp 10 10.1.0.0 255.255.0.0
     tunnel source 1.1.1.1
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    spoke 1: 
    interface Tunnel0
     ip address 172.16.1.7 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.7.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    Spoke 2:
    interface Tunnel0
     ip address 172.16.1.6 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.6.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end

    For some reason, my spoke to spoke tunnel doesn't stay up when my hub goes down. Not sure what I am missing. Please help!!
    Hub:
    interface Tunnel0
     ip address 172.16.1.1 255.255.255.0
     no ip redirects
     no ip next-hop-self eigrp 10
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp network-id 1
     no ip split-horizon
     ip summary-address eigrp 10 10.1.0.0 255.255.0.0
     tunnel source 1.1.1.1
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    spoke 1: 
    interface Tunnel0
     ip address 172.16.1.7 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.7.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    Spoke 2:
    interface Tunnel0
     ip address 172.16.1.6 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.6.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end

  • DMVPN - Spoke to Spoke tunnel

    Hi,
    Once a Spoke to Spoke tunnel is established, what happens if the Hub goes down? Does the Spoke to Spoke tunnel remain active? 

    For some reason, my spoke to spoke tunnel doesn't stay up when my hub goes down. Not sure what I am missing. Please help!!
    Hub:
    interface Tunnel0
     ip address 172.16.1.1 255.255.255.0
     no ip redirects
     no ip next-hop-self eigrp 10
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp network-id 1
     no ip split-horizon
     ip summary-address eigrp 10 10.1.0.0 255.255.0.0
     tunnel source 1.1.1.1
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    spoke 1: 
    interface Tunnel0
     ip address 172.16.1.7 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.7.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    Spoke 2:
    interface Tunnel0
     ip address 172.16.1.6 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.6.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end

  • DMVPN spoke to spoke not established

    In the topology attached I am advertising tunnel ip and loopbacks for the specific hub and spokes in their eigrp , there is connectivity between all but still Eigrp routes are not coming and its flapping.ISAKMP is on , tunnel is also up.I am also attaching hub and spoke configs and topology  for your reference.
    This is the error message:
    7:09.791: %DUAL-5-NBRCHANGE: IP-EIGRP(0) 145: Neighbor 172.16.145.5 (Tunnel0) is down: retry limit exceeded
    *Mar  1 00:57:11.351: %CRYPTO-4-RECVD_PKT_INV_SPI: decaps: rec'd IPSEC packet has invalid spi for destaddr=10.1.12.1, prot=50, spi=0x77DF7896(2011134102), srcaddr        in use settings ={Transport, }
            conn id: 29, flow_id: SW:29, crypto map: Tunnel0-head-0
            sa timing: remaining key lifetime (k/sec): (4405256/3495)
            IV size: 8 bytes
            replay detection support: Y
            Status: ACTIVE
    *Mar  1 00:57:21.895: YPTO-6-PRINTABORT: deletion caused early termination of show output for identity
    R1#
    *Mar  1 00:09:49.443: %DUAL-5-NBRCHANGE: IP-EIGRP(0) 145: Neighbor 172.16.145.4 (Tunnel0) is down: retry limit exceeded
    *Mar  1 00:09:50.051: %DUAL-5-NBRCHANGE: IP-EIGRP(0) 145: Neighbor 172.16.145.4 (Tunnel0) is up: new adjacency
    *Mar  1 00:11:00.311: %DUAL-5-NBRCHANGE: IP-EIGRP(0) 145: Neighbor 172.16.145.5 (Tunnel0) is down: retry limit exceeded
    *Mar  1 00:11:00.775: %DUAL-5-NBRCHANGE: IP-EIGRP(0) 145: Neighbor 172.16.145.5 (Tunnel0) is up: new adjacency
    *Mar  1 00:11:09.575: %DUAL-5-NBRCHANGE: IP-EIGRP(0) 145: Neighbor 172.16.145.4 (Tunnel0) is down: retry limit exceeded
    *Mar  1 00:11:11.551: %DUAL-5-NBRCHANGE: IP-EIGRP(0) 145: Neighbor 172.16.145.4 (Tunnel0) is up: new adjacency

    For some reason, my spoke to spoke tunnel doesn't stay up when my hub goes down. Not sure what I am missing. Please help!!
    Hub:
    interface Tunnel0
     ip address 172.16.1.1 255.255.255.0
     no ip redirects
     no ip next-hop-self eigrp 10
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp network-id 1
     no ip split-horizon
     ip summary-address eigrp 10 10.1.0.0 255.255.0.0
     tunnel source 1.1.1.1
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    spoke 1: 
    interface Tunnel0
     ip address 172.16.1.7 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.7.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end
    Spoke 2:
    interface Tunnel0
     ip address 172.16.1.6 255.255.255.0
     no ip redirects
     no ip split-horizon eigrp 10
     ip nhrp authentication _GW_
     ip nhrp map multicast dynamic
     ip nhrp map 172.16.1.1 1.1.1.1
     ip nhrp map multicast 1.1.1.1
     ip nhrp network-id 1
     ip nhrp nhs 172.16.1.1
     ip summary-address eigrp 10 10.6.0.0 255.255.0.0
     tunnel source GigabitEthernet0/0
     tunnel mode gre multipoint
     tunnel protection ipsec profile dmvpn
    end

  • Unable to create report. Query produced too many results

    Hi All,
    Does someone knows how to avoid the message "Unable to create report. Query produced too many results" in Grid Report Type in PerformancePoint 2010. When the mdx query returns large amount of data, this message appears. Is there a way to get all
    the large amount in the grid anyway?
    I have set the data Source query time-out under Central Administration - Manager Service applications - PerformancePoint Service Application - PerformancePoint Service Application Settings at 3600 seconds.
    Here Event Viewer log error at the server:
    1. An exception occurred while running a report.  The following details may help you to diagnose the problem:
    Error Message: Unable to create report. Query produced too many results.
            <br>
            <br>
            Contact the administrator for more details.
    Dashboard Name:
    Dashboard Item name:
    Report Location: {3592a959-7c50-0d1d-9185-361d2bd5428b}
    Request Duration: 6,220.93 ms
    User: INTRANET\spsdshadmin
    Parameters:
    Exception Message: Unable to create report. Query produced too many results.
    Inner Exception Message:
    Stack Trace:    at Microsoft.PerformancePoint.Scorecards.Server.PmServer.ExecuteAnalyticReportWithParameters(RepositoryLocation analyticReportViewLocation, BIDataContainer biDataContainer)
       at Microsoft.PerformancePoint.Analytics.ServerRendering.OLAPBase.OlapViewBaseControl.ExtractReportViewData()
       at Microsoft.PerformancePoint.Analytics.ServerRendering.OLAPBase.OlapViewBaseControl.CreateRenderedView(StringBuilder sd)
       at Microsoft.PerformancePoint.Scorecards.ServerRendering.NavigableControl.RenderControl(HtmlTextWriter writer)
    PerformancePoint Services error code 20604.
    2. Unable to create report. Query produced too many results.
    Microsoft.PerformancePoint.Scorecards.BpmException: Unable to create report. Query produced too many results.
       at Microsoft.PerformancePoint.Scorecards.Server.Analytics.AnalyticQueryManager.ExecuteReport(AnalyticReportState reportState, DataSource dataSource)
       at Microsoft.PerformancePoint.Scorecards.Server.PmServer.ExecuteAnalyticReportBase(RepositoryLocation analyticReportViewLocation, BIDataContainer biDataContainer, String formattingDimensionName)
       at Microsoft.PerformancePoint.Scorecards.Server.PmServer.ExecuteAnalyticReportWithParameters(RepositoryLocation analyticReportViewLocation, BIDataContainer biDataContainer)
    PerformancePoint Services error code 20605.
    Thanks in advance for your help.

    Hello,
    I would like you to try the following to adjust your readerquotas.
    Change the values of the parameters listed below to a larger value. We recommend that you double the value and then run the query to check whether the issue is resolved. To do this, follow these steps:
    On the SharePoint 2010 server, open the Web.config file. The file is located in the following folder:
    \Program Files\Microsoft Office Servers\14.0\Web Services\PpsMonitoringServer\
    Locate and change the the below values from 8192 to 16384.
    Open the Client.config file. The file is located in the following folder:
    \Program Files\Microsoft Office Servers\14.0\WebClients\PpsMonitoringServer\
    Locate and change the below values from 8192 to 16384.
    After you have made the changes, restart Internet Information Services (IIS) on the SharePoint 2010 server.
    <readerQuotas
    maxStringContentLength="2147483647"
    maxNameTableCharCount="2147483647"
    maxBytesPerRead="2147483647"
    maxArrayLength="2147483647"
                  maxDepth="2147483647"
    />
    Thanks
    Heidi Tr - MSFT
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • We have too many apple devices, how can I creat another apple ID and divide the devices among 2 Apple Ids?

    Hello all,
    We started off with the original iPhone a few years ago and now between my wife and I (plus Kids) we have too many apple devices under the same apple-ID (2 X iPhone 6+, 2X macbook Pro, 2X iPod-touch and 1X iPad)
    We are constantly running into issues because everyone in the house is using the same apple id, downloading Apps using the same ID, sharing the same cloud and receiving each other's iMessages.
    I know that I can create another Apple-ID, what I am unclear on is once I do have a new apple ID, how do I assign that to half the devices (1X iPhone 6+, 1 X MBP and 1 Ipad)
    Please respond if you are able to help.
    Thank you

    I'm not sure what they told your at the Apple store, but each Apple ID can only create a single iCloud account.  If you want separate accounts for all these devices, they will have to be created using separate IDs.  Note: you can do this and still share the same ID for other services such as your iTunes store account.  The iCloud ID does not need to be the same as the ID used for other services.
    To create an account, go to Settings>iCloud and sign in with a different Apple ID.  If you want to migrate a copy of the current iCloud data to the different accounts, start by saving any photo stream photos that you want to keep to your camera roll (unless already there) by opening your my photo stream album, tapping Select, tapping the photos, tap the share icon (box with upward facing arrow), then tapping Save to Camera Roll.  If you are syncing notes with iCloud that you want to keep, you'll need to open each of your notes and email them to yourself so you can later copy and paste the text into new notes created in your new account.  Then go to Settings>iCloud, tap Delete Account (which only deletes it from this device, not from iCloud; the devices keeping the current account will not be effected by this), provide the password to turn off Find My iDevice and choose Keep on My iDevice when prompted.  Then sign back in with a different Apple ID to create your new account and choose Merge to upload your data.

  • Too Many Line Items While Creating a PO

    Dear All,
    Here we have a scenario:
    We are creating a PO which is having around 10000(ten thousand) line items, and while saving system is showing error message "Too Many PO Line Items" Message No: FI899. and not able to save it.
    Kindly Let me know is there any limit for line item creation in a PO.
    Regards,
    Ravi B

    Limitations are limitations, you need a workaround or pay SAP for an extra developement. Probably worth, if you have such POs everyday.
    ( I cannot even imagine for what reason a PO should have more than 10000 items. Manual maintenance is impossible, the person would sit for a week and more to finish one PO. the CSR at your vendor is as well not able to put the items from your PO into their Sales order within a reasonable time frame - order entry should be ready before delivery date is arrived.
    This sounds like you want buy a company and want create just one purchase order to buy any single material they have on stock. But for that purpose the PO is certainly not designed.
    Can you give more background that makes my mind open ?   )

  • Create procedure is generating too many archive logs

    Hi
    The following procedure was run on one of our databases and it hung since there were too many archive logs being generated.
    What would be the answer? The db must remain in archivelog mode.
    I understand the nologging concept, but as I know this applies to creating tables, views, indexes and tablespaces. This script is creating procedure.
    CREATE OR REPLACE PROCEDURE APPS.Dfc_Payroll_Dw_Prc(Errbuf OUT VARCHAR2, Retcode OUT NUMBER
    ,P_GRE NUMBER
    ,P_SDATE VARCHAR2
    ,P_EDATE VARCHAR2
    ,P_ssn VARCHAR2
    ) IS
    CURSOR MainCsr IS
    SELECT DISTINCT
    PPF.NATIONAL_IDENTIFIER SSN
    ,ppf.full_name FULL_NAME
    ,ppa.effective_date Pay_date
    ,ppa.DATE_EARNED period_end
    ,pet.ELEMENT_NAME
    ,SUM(TO_NUMBER(prv.result_value)) VALOR
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAf.ASSIGNMENT_ID ASSG_ID
    ,paf.ORGANIZATION_ID
    FROM
    pay_element_classifications pec
    , pay_element_types_f pet
    , pay_input_values_f piv
    , pay_run_result_values prv
    , pay_run_results prr
    , pay_assignment_actions paa
    , pay_payroll_actions ppa
    , APPS.pay_all_payrolls_f pap
    ,Per_Assignments_f paf
    ,per_people_f ppf
    WHERE
    ppa.effective_date BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND ppa.payroll_id = pap.payroll_id
    AND paa.tax_unit_id = NVL(p_GRE, paa.tax_unit_id)
    AND ppa.payroll_action_id = paa.payroll_action_id
    AND paa.action_status = 'C'
    AND ppa.action_type IN ('Q', 'R', 'V', 'B', 'I')
    AND ppa.action_status = 'C'
    --AND PEC.CLASSIFICATION_NAME IN ('Earnings','Alien/Expat Earnings','Supplemental Earnings','Imputed Earnings','Non-payroll Payments')
    AND paa.assignment_action_id = prr.assignment_action_id
    AND prr.run_result_id = prv.run_result_id
    AND prv.input_value_id = piv.input_value_id
    AND piv.name = 'Pay Value'
    AND piv.element_type_id = pet.element_type_id
    AND pet.element_type_id = prr.element_type_id
    AND pet.classification_id = pec.classification_id
    AND pec.non_payments_flag = 'N'
    AND prv.result_value &lt;&gt; '0'
    --AND( PET.ELEMENT_INFORMATION_CATEGORY LIKE '%EARNINGS'
    -- OR PET.element_type_id IN (1425, 1428, 1438, 1441, 1444, 1443) )
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PET.EFFECTIVE_START_DATE AND PET.EFFECTIVE_END_DATE
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PIV.EFFECTIVE_START_DATE AND PIV.EFFECTIVE_END_DATE --dcc
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN Pap.EFFECTIVE_START_DATE AND Pap.EFFECTIVE_END_DATE --dcc
    AND paf.ASSIGNMENT_ID = paa.ASSIGNMENT_ID
    AND ppf.NATIONAL_IDENTIFIER = NVL(p_ssn, ppf.NATIONAL_IDENTIFIER)
    ------------------------------------------------------------------TO get emp.
    AND ppf.person_id = paf.person_id
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN ppf.EFFECTIVE_START_DATE AND ppf.EFFECTIVE_END_DATE
    ------------------------------------------------------------------TO get emp. ASSIGNMENT
    --AND paf.assignment_status_type_id NOT IN (7,3)
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN paf.effective_start_date AND paf.effective_end_date
    GROUP BY PPF.NATIONAL_IDENTIFIER
    ,ppf.full_name
    ,ppa.effective_date
    ,ppa.DATE_EARNED
    ,pet.ELEMENT_NAME
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAF.ASSIGNMENT_ID
    ,paf.ORGANIZATION_ID
    BEGIN
    DELETE cust.DFC_PAYROLL_DW
    WHERE PAY_DATE BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND tax_unit_id = NVL(p_GRE, tax_unit_id)
    AND ssn = NVL(p_ssn, ssn)
    COMMIT;
    FOR V_REC IN MainCsr LOOP
    INSERT INTO cust.DFC_PAYROLL_DW(SSN, FULL_NAME, PAY_DATE, PERIOD_END, ELEMENT_NAME, ELEMENT_INFORMATION_CATEGORY, CLASSIFICATION_ID, ELEMENT_INFORMATION1, VALOR, TAX_UNIT_ID, ASSG_ID,ELEMENT_TYPE_ID,ORGANIZATION_ID)
    VALUES(V_REC.SSN,V_REC.FULL_NAME,v_rec.PAY_DATE,V_REC.PERIOD_END,V_REC.ELEMENT_NAME,V_REC.ELEMENT_INFORMATION_CATEGORY, V_REC.CLASSIFICATION_ID, V_REC.ELEMENT_INFORMATION1, V_REC.VALOR,V_REC.TAX_UNIT_ID,V_REC.ASSG_ID, v_rec.ELEMENT_TYPE_ID, v_rec.ORGANIZATION_ID);
    COMMIT;
    END LOOP;
    END ;
    So, how could I assist our developer with this, so that she can run it again without it generating a ton of logs ? ?
    Thanks
    Oracle 9.2.0.5
    AIX 5.2

    The amount of redo generated is a direct function of how much data is changing. If you insert 'x' number of rows, you are going to generate 'y' mbytes of redo. If your procedure is destined to insert 1000 rows, then it is destined to create a certain amount of redo. Period.
    I would question the <i>performance</i> of the procedure shown ... using a cursor loop with a commit after every row is going to be a slug on performance but that doesn't change the fact 'x' inserts will always generate 'y' redo.

  • Unable to create virtual folder. Too many folders

    Hi,
    We are getting this error while trying to create new folder in UCM (even trying directly from UCM throws same error)
    oracle.stellent.ridc.protocol.ServiceException: Unable to create virtual folder. Too many folders
    Has anyone experienced this? If yes, please let us know how you resolved it.
    Thanks!

    You don't "store" anything in contribution folders. They are just webdav access points. I recommend zero folders personally.
    If you want to upload files using webdav, then I suggest using one staging folder, where files can enter the system. When the document is workflowed and/or verified, it is moved out of the folder. It's more likely however that you'll need a few access points for various types of content as each folder enforces specific metadata.
    Now I imagine you are after a way to store all the HR bits and pieces of one person together as a unit. Some people use Content Folios. Personally, I'd just use a metadata field that contains the ContentID of the person's master record. Any new content for that person just gets the master ContentID stamped on it. Just search for that ContentID to get all their stuff.
    Regarding best practices, there are lots of quality blogs floating around. Stellent/Oracle don't talk about best practice because each system is used by their clients in a unique way... well, that's what they say.

  • F110 -Too many payment mediums created for this payment group

    Hello Gurus,
    Am getting below error after executing payment run F110.
    BFIBL02160               Too many payment mediums created for this payment group
    BFIBL02616               End of log for payment run 20140409 / DPA04 , PE1 / 500
    We are not able to see the Payment advice note in the spool request.
    Can you please share your ideas to resolve the issue
    Thanks in advacne

    HI,
    Please let us know for account like vendor , customer  you running the payment advice.
    check if the customer ,vendor document  is posted or not.
    which customer, vendor number you had maintained in the f110 .
    check that before you run it.
    else.. create new document  and start run the payment.
    for single document fi you try  more than once it will give the above error.
    and also the check the program and variant as well.
    let me know any corrections.
    thanks.

  • DMEE error (Error Log -  Too many payment mediums created for this payment group)

    Hi Experts,
    Please see the below error while generating the DMEE file.
    Error Log -  Too many payment mediums created for this payment group
    Error is - Too many payment mediums created for this payment group
    Please explain how to resolve this error.Send detail documentation how to rectify this error
    Regards
    Sreedhar

    Hi Nimish Agarwal,
    With reference to the error faced i.e. BFIBL02160 "too many payment
    mediums created for this payment group", please review the information
    provided in the long text of this error in detail.
    Please ensure that the orders are not carried out several times.
    Alternatively you can apply the optional usable message 166 (same
    class: BFIBL02), so that you can prevent creating duplicates in future.
    Please customize this message as an error so as to prevent the system
    from creating duplicate files. You can customize this message in OBA5.
    The message BFIBL02 160 always occurs, if you create more than one
    file for the same payment run, in order to warn you not to send the
    same file twice to the bank.
    The SAPFPAYM program does not create 2 files at once. But if you
    run this program more than once for the same payment run, more than
    one file is created. From the second run on you receive the error
    message BFIBL02 160. You can easily test it, by creating a new
    payment proposal and creating a new file with this proposal and
    SAPFPAYM. The first time you run this program you will not receive
    the message.
    Hope this clarifies.
    Kind Regards,
    Fernando Evangelista

  • Special Ledger Roll up: too many records created in receiver table

    Hello,
    When I execute roll up from 1 special ledger to another special ledger (leaving out a number of fields) 4 records (postings) are selected but in the receiver table there are 92 records created (4*23) so 88 records too many. For each selected record in the sender table 7 records contain only 0 amounts and have an incorrect RPMAX value (starting with 16 till 336). The 8th record has an amount and has the correct RPMAX value: 352). In Sender and Receiver Ledger fiscal year variant Daily Balance is assigned (= each day is a period). Any hint how to prevent the empty records?

    Installing the patch, re- importing the SAP Provisioning framework (I selected 'update') and recreating the jobs didn't yield any result.
    When examining pass 'ReadABAPRoles' of Job 'AS ABAP - Initial Load' -> tab 'source', there are no scripts used .
    After applying the patch we decided anyway to verify the scripts (sap_getRoles, sap_getUserRepositories) in our Identity Center after those of 'Note 1398312 - SAP NW IdM Provisioning Framework for SAP Systems' , and they are different
    File size of SAP Provisioning Framework_Folder.mcc of SP3 Patch 0 and Patch 1 are also exactly the same.
    Opening file SAP Provisioning Framework_Folder.mcc with Wordpad : searched for 'sap_getRoles'  :
    <GLOBALSCRIPT>
    <SCRIPTREVISIONNUMBER/>
    <SCRIPTLASTCHANGE>2009-05-07 08:00:23.54</SCRIPTLASTCHANGE>
    <SCRIPTLANGUAGE>JScript</SCRIPTLANGUAGE>
    <SCRIPTID>30</SCRIPTID>
    <SCRIPTDEFINITION> ... string was too long to copy
    paste ... </SCRIPTDEFINITION>
    <SCRIPTLOCKDATE/>
    <SCRIPTHASH>0940f540423630687449f52159cdb5d9</SCRIPTHASH>
    <SCRIPTDESCRIPTION/>
    <SCRIPTNAME>sap_getRoles</SCRIPTNAME>
    <SCRIPTLOCKSTATE>0</SCRIPTLOCKSTATE>
    -> Script last change 2009-05-07 08:00:23.54 -> that's no update !
    So I assume the updates mentioned in Note 1398312 aren't included in SP3 Patch 1. Manually replaced the current scripts with those of the note and re- tested : no luck. Same issue.
    Thanks again for the help,
    Kevin

  • Portal creates too many database sessions. fix doesn't work!!

    Oracle Database 9i, Application Server 9iAS,
    Operating System SUSE Linux 7.2
    CPU - Athlon 1400
    Ram - 1GB
    There is a modification for the http server that aims to eliminate a problem on unix that causes the database to create too many sessions. The script can be found at
    http://portalstudio.oracle.com/servlet/page?_pageid=1787&_dad=ops&_schema=OPSTUDIO
    However the script fails to work. The http server is on port 80 with the redirect on port 7778. However going to http://myhost/pls/ results in a server error - connection refused. Is it possible that there may be an error in the script.
    Thanks in anticipation

    ok so my sound is ok now I can check that off of the list... I just looked around and found some external USB speakers that work just right. My built in computer speakers are just not the best quality.
    As for the microphone, I still haven't been able to find out what is causing it to not work. I notice that it isn't actually broken since if I make a really loud noise right next to the input it will register a little but only a little bit.
    $ arecord -L
    null
    Discard all samples (playback) or generate zero samples (capture)
    pulse
    PulseAudio Sound Server
    default
    Default ALSA Output (currently PulseAudio Sound Server)
    sysdefault:CARD=Intel
    HDA Intel, ALC269 Analog
    Default Audio Device
    front:CARD=Intel,DEV=0
    HDA Intel, ALC269 Analog
    Front speakers
    surround40:CARD=Intel,DEV=0
    HDA Intel, ALC269 Analog
    4.0 Surround output to Front and Rear speakers
    surround41:CARD=Intel,DEV=0
    HDA Intel, ALC269 Analog
    4.1 Surround output to Front, Rear and Subwoofer speakers
    surround50:CARD=Intel,DEV=0
    HDA Intel, ALC269 Analog
    5.0 Surround output to Front, Center and Rear speakers
    surround51:CARD=Intel,DEV=0
    HDA Intel, ALC269 Analog
    5.1 Surround output to Front, Center, Rear and Subwoofer speakers
    surround71:CARD=Intel,DEV=0
    HDA Intel, ALC269 Analog
    7.1 Surround output to Front, Center, Side, Rear and Woofer speakers
    $
    here is some additional information about my sound. Also, when I open alsamixer, I find that when I go to "select sound card"(F6) I see default and HDA Intel. If the default is pulseaudio, then is it possible that pulseaudio is causing the problem?

  • Well, I created a new Apple ID and I put money on it, I wanted to sign into ICloud and it wouldn't let me sign in due to the usage of too many Apple IDs created. The point for logging into the ICloud was to save my photos before I reset my phone.

    Well, I created a new Apple ID and I put money on it, I wanted to sign into ICloud and it wouldn't let me sign in due to making too many Apple IDs created. The point for logging into the ICloud was to save my photos before I reset my phone. Is there any way I could fix this? I also want to transfer my money because I think it's a waste to just forget about the $13.75 on my account. Thanks.

    I recently created a new apple ID
    Bad idea.
    Content bought with an Apple ID is forever associated with that Apple ID. Apple will not transfer content from one Apple ID to another and Apple will not merge Apple IDs. Unless you are prepared to forfeit all the previously purchased content and buy it all again with the new Apple ID, you will need to maintain the old ID to update and redownload your content.

Maybe you are looking for