Cookie - Bad Request - Size of a request header field exceeds server limit -

We are on cq5.5. We see this error intermittently. What is the best way to fix this? Cookie size seems to be adding to the issue.
Bad Request
Your browser sent a request that this server could not understand.
Size of a request header field exceeds server limit.
Cookie: cq-mrss=path%3D%252Fcontent%252Fdam%26p.limit%3D-1%26mainasset%3Dtrue%26type%3Ddam%3AAsse t; __unam=acfbce4-13b8ffd6084-6070cfe6-4; __utma=16528299.1850197993.1355330446.1361568697.1362109625.3; __utmz=16528299.1355330446.1.1.utmcsr=(direct)|utmccn=(direct)|utmcmd=(none); REM_ME=1004; SessionPersistence-author-lx_qa_author2=CLIENTCONTEXT%3A%3DvisitorId%3Danonymous%2Cvisito rId_xss%3Danonymous%7CPROFILEDATA%3A%3DauthorizableId%3Danonymous%2CformattedName%3DAnonym ous%20Surfer%2Cpath%3D%2Fhome%2Fusers%2Fa%2Fanonymous%2Cavatar%3D%2Fetc%2Fdesigns%2Fdefaul t%2Fimages%2Fcollab%2Favatar.png%2Cage%3D%2Cage_xss%3D%7CTAGCLOUD%3A%3Dtopic%3Aworkflow%3D 14%2Cindustry%3Aprocess_management%3D2%2Ctopic%3Aprocess_mining%3D3%2Ctopic%3Aprocess_docu mentation%3D1%2Ctopic%3Aintelligent_capture%3D5%2Cindustry%3Acapture%3D5%2Ctopic%3Adocumen t_imaging%3D2%2Ctopic%3Adistributed_intelligent_capture%3D2%2Ctopic%3Adocument_output_mana gement%3D4%2Cindustry%3Acontent_management%3D14%2Cindustry%3Asoftware_solutions_hardware%3 D4%2Cindustry%3Adevice_management%3D2%2Ctopic%3Ahelp_desk_services%3D2%2Cindustry%3Aintera ct%3D15%2Ctopic%3Asecure_content_monitor%3D2%2Ctopic%3Aelectronic_forms%3D2%2Ctopic%3Ainte lligent_forms%3D2%2Ctopic%3Adocument_accounting%3D2%2Ctopic%3Aerp_output_management%3D2%2C topic%3Aprint_release%3D2%2Cindustry%3Aoutput_management%3D4%2Ctopic%3Aerp_printing%3D4%2C topic%3Aenterprise_search%3D4%2Ctopic%3Amicrosoft_sharepoint%3D6%2Ctopic%3Adocument_filter s%3D4%2Cindustry%3Asearch%3D4%2Ctopic%3Ahuman_services_case_management%3D2%2Cindustry%3Aca se_management%3D2%2Cindustry%3Aimprove_business_processes%3D6%2Ctopic%3Abusiness_process_m odeling%3D1%2Ctopic%3Alawson%3D1%2Ctopic%3Aapplication_integration%3D8%2Cindustry%3Asoluti on%3D4%2Ctopic%3Amicrosoft_dynamics_crm%3D2%2Cindustry%3Ahealthcare%3D13%2Cindustry%3Areta il%3D8%2Cindustry%3Abanking%3D3%2Cindustry%3Aincrease_efficiency%3D7%2Cindustry%3Agovernme nt%3D8%2Ctopic%3Amicrosoft_outlook%3D2%2Ctopic%3Aesri%3D2%2Ctopic%3Ajd_edwards%3D2%2Ctopic %3Asap%3D1%2Cindustry%3Adrive_business_growth%3D1%2Cindustry%3Abusiness_challenges%3D6%2Ci ndustry%3Aconnect_distributed_workforce%3D1%2Ctype%3Alanding_page%3D2%2Ctopic%3Aconsulting _services%3D2%2Ctopic%3Aretail_pharmacy%3D2%2Cindustry%3Aindustry_solutions%3D5%2Ctopic%3A health_information_management%3D3%2Ctopic%3Apatient_scheduling%3D3%2Ctopic%3Aclinical_depa rtment_solutions%3D3%2Ctopic%3Aclinical_hit_integration%3D3%2Ctopic%3Apatient_admissions_r egistration%3D3%2Ctopic%3Ahealthcare_forms_management%3D3%2Ctopic%3Apatient_access%3D3%2Ct opic%3Aenterprise_print_management_software%3D2%2Ctopic%3Aprint_queue_management%3D2%2Ctop ic%3Aadvanced_print_management%3D2%2Ctopic%3Aemployee_onboarding%3D3%2Ctopic%3Ahuman_resou rces%3D1%2Cindustry%3Ahuman_resources%3D3%2Ctopic%3Aemployee_recruitment%3D1%2Cindustry%3A manufacturing%3D2%2Ctopic%3Aplatform_integration%3D1%2Ctopic%3Awealth_management%3D2%2Cind ustry%3Afinancial_services%3D2%2Ctopic%3Aaccount_opening%3D2%2Ctopic%3Acompliance%3D1%2Cin dustry%3Acompliance%3D1%2Ctopic%3Abusiness_operations_solutions_for_banking%3D2%2Ctopic%3A retail_delivery%3D1%2Ctopic%3Aloan_processing%3D1%2Ctopic%3Aon_demand_negotiable_documents %3D1%2Ctopic%3Anew_account_openings%3D1%2Ctopic%3Aon_demand_forms_customer_communications% 3D1%2Cindustry%3Ainsurance%3D1%2Ctopic%3Amicr_printing%3D1%2Ctopic%3Abank_branch_capture%3 D1%2Ctopic%3Aagency_capture%3D1%7C; ys-cq-damadmin-tree=o%3Awidth%3Dn%253A240%5EselectedPath%3Ds%253A/content/dam; ys-cq-damadmin-grid-assets=o%3Acolumns%3Da%253Ao%25253Aid%25253Ds%2525253Anumberer%25255E width%25253Dn%2525253A23%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253At humbnail%25255Ewidth%25253Dn%2525253A45%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25 253Ds%2525253Atitle%25255Ewidth%25253Dn%2525253A78%25255Ehidden%25253Db%2525253A1%25255Eso rtable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Aname%25255Ewidth%25253Dn%2525253A3 37%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Apublished%25255Ewidth%2 5253Dn%2525253A37%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Amodified %25255Ewidth%25253Dn%2525253A78%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%25 25253Ascene7Status%25255Ewidth%25253Dn%2525253A78%25255Ehidden%25253Db%2525253A1%25255Esor table%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Astatus%25255Ewidth%25253Dn%2525253A 71%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Dn%2525253A8%25255Ewidth%25253Dn%2 525253A78%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Aworkflow%25255Ew idth%25253Dn%2525253A78%25255Ehidden%25253Db%2525253A1%25255Esortable%25253Db%2525253A1%25 5Eo%25253Aid%25253Ds%2525253Awidth%25255Ewidth%25253Dn%2525253A37%25255Esortable%25253Db%2 525253A1%255Eo%25253Aid%25253Ds%2525253Aheight%25255Ewidth%25253Dn%2525253A37%25255Esortab le%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Asize%25255Ewidth%25253Dn%2525253A37%25 255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Areferences%25255Ewidth%25253 Dn%2525253A199%25255Esortable%25253Db%2525253A1%5Esort%3Do%253Afield%253Ds%25253Alabel%255 Edirection%253Ds%25253AASC; amlbcookie=04; ObLK=0x82abacf3a5e3b1e2|0x1cf34305ac210c7e9b2b07e3725392e2; iPlanetDirectoryPro=AQIC5wM2LY4Sfcw0UQ2MST5NlqDAsUi2dscer0wO7VMy9pE.*AAJTSQACMDYAAlMxAAIw NA..*; renderid=rend01; login-token=c9c0d027-c5f9-4e5a-9a90-09d1cf21cfd2%3a0279e369-1689-433c-80ef-d8411040efe5_6 15c2fd1eba8fd42%3acrx.default; ys-cq-siteadmin-grid-pages=o%3Acolumns%3Da%253Ao%25253Aid%25253Ds%2525253Anumberer%25255E width%25253Dn%2525253A23%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253At humbnail%25255Ewidth%25253Dn%2525253A50%25255Ehidden%25253Db%2525253A1%25255Esortable%2525 3Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Atitle%25255Ewidth%25253Dn%2525253A386%25255Es ortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Aname%25255Ewidth%25253Dn%2525253A 148%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Apublished%25255Ewidth% 25253Dn%2525253A25%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Amodifie d%25255Ewidth%25253Dn%2525253A86%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2 525253Ascene7Status%25255Ewidth%25253Dn%2525253A86%25255Ehidden%25253Db%2525253A1%25255Eso rtable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Astatus%25255Ewidth%25253Dn%2525253 A76%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Aimpressions%25255Ewidt h%25253Dn%2525253A86%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Atempl ate%25255Ewidth%25253Dn%2525253A86%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds %2525253Aworkflow%25255Ewidth%25253Dn%2525253A86%25255Ehidden%25253Db%2525253A1%25255Esort able%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2525253Alocked%25255Ewidth%25253Dn%2525253A8 6%25255Ehidden%25253Db%2525253A1%25255Esortable%25253Db%2525253A1%255Eo%25253Aid%25253Ds%2 525253AliveCopyStatus%25255Ewidth%25253Dn%2525253A86%25255Ehidden%25253Db%2525253A1%25255E sortable%25253Db%2525253A1%5Esort%3Do%253Afield%253Ds%25253Atitle%255Edirection%253Ds%2525 3AASC; ys-cq-siteadmin-tree=o%3Awidth%3Dn%253A306%5EselectedPath%3Ds%253A/content/homesite/en-US /insights/video_unum-group-accelerates-workflows-with-solutions-; ys-cq-cf-clipboard=o%3Acollapsed%3Db%253A1; ys-cq-cf-tabpanel=o%3AactiveTab%3Ds%253AcfTab-Images-QueryBox; JSESSIONID=ad311ac3-7c24-4e62-ae8a-0ebacd8e8188; SessionPersistence-author-lx_qa_author1=CLIENTCONTEXT%3A%3DvisitorId%3Danonymous%2Cvisito rId_xss%3Danonymous%7CPROFILEDATA%3A%3DauthorizableId%3Danonymous%2CformattedName%3DAnonym ous%20Surfer%2Cpath%3D%2Fhome%2Fusers%2Fa%2Fanonymous%2Cavatar%3D%2Fetc%2Fdesigns%2Fdefaul t%2Fimages%2Fcollab%2Favatar.png%2Cage%3D%2Cage_xss%3D%7CGEOLOCATION%3A%3D%7CTAGCLOUD%3A%3 Dindustry%3Aconnect_distributed_workforce%3D1%2Cindustry%3Abusiness_challenges%3D1%2Cindus try%3Acontent_management%3D1%2Cindustry%3Ahealthcare%3D1%2Ctopic%3Afinance%3D1%2Ctopic%3Ap rocurement_processing%3D1%2Cindustry%3Afinancial_services%3D2%2Cindustry%3Ainsurance%3D2%2 Cindustry%3Aindustry_solutions%3D2%2Ctopic%3Aagency_capture%3D2%7C; s_cc=true; s_sq=lxmtest%3D%2526pid%253Dinsights%25253Avideo_unum-group-accelerates-workflows-with-so luti

Hi EbodaWill,
File daycare for fp 2324 where in you can configure & allow you to increase the request header size and avoid the bad request error OR for a package that improves client side persistence & does not use cookies.
Thanks,
Sham

Similar Messages

  • Cannot retrieve my e-mail "browser sent request server could not understand. Size of request header field exceeds server limit"

    Upgraded to Firefox 5.0.1 yesterday. No, after logging on to firefox, which takes me to my comcast page and when I try to get
    my e-mail I get this message "your browser sent a request this server could not understand. Size of request header field exceeds server limit" Then it says something about "cookies" I also tried to connect to other sites and get similar messages. Just to let you know I am not a guru, and 80 years old, but I did not have this problem with the previous version. Question, why are the headers repeated? Could that be the problem???

    This issue can be caused by corrupted cookies.
    Clear the cache and the cookies from sites (e.g. comcast) that cause problems.
    "Clear the Cache":
    * Firefox > Preferences > Advanced > Network > Offline Storage (Cache): "Clear Now"
    "Remove Cookies" from sites causing problems:
    * Firefox > Preferences > Privacy > Cookies: "Show Cookies"

  • I can no longer access Bejeweled Blitz through my facebook account.  I get the message that says, "your browser sent a request that this server could not understand. Size of a request header field exceeds server limit".  Help please.

    I can no longer access Bejeweled Blitz through facebook.  I get the message, "your browser sent a request that this server could not understand. Size of a request header field exceeds server limit". I can access Bejeweled through FB using my husband's log in so to me that suggests the problem is with my log in. Help please.

    Contact FB or use another browser. 

  • When I try to access stories from the main page of yahoo news...this is what I get...Sorry, Bad Request. Your browser sent a request that this server could not understand. Size of a request header field exceeds server limit.

    Very infrequently it will allow the page to open but 90% of the time I get the error message. It also happens if I click on a link to a yahoo news story from another web site. I had tried Firefox 4, but it was giving me all kinds of errors so I reverted back to 3.6

    This issue can be caused by corrupted cookies.
    Clear the cache and the cookies from sites that cause problems.
    "Clear the Cache":
    * Tools > Options > Advanced > Network > Offline Storage (Cache): "Clear Now"
    "Remove Cookies" from sites causing problems:
    * Tools > Options > Privacy > Cookies: "Show Cookies"

  • When I try to go to certain pages on PayPal, I get an error message which says 400 Bad Request: Size of request exceeds server limit. Didn't do it before the lasted Foxfire update. Can I revert to an older version?

    1. Using Windows XP home
    2.PayPal is the only website I have problems with Foxfire; works fine on IE8
    3. I can get to some of the desired pages through other links on PayPal so it's not the pages themselves
    4. The error message contains this:
    Cookie: KHcl0EuY7AKSMgfvHl7J5E7hPtK=STMZXVPEjOuzek-HGHcBJRjmRXIQgXpML8uy9fV13oPeEYcAUnBOhKFzvNJqcb-rsc2S6nlYOklSdX6P; cookie_check=yes; LANG=en_US%3bUS; INSIDE
    When I saw Cookie, I did a disk clean including temp files and cookies. No difference. I also did a System Restore, to no avail.
    5. The 400 message actually reads: Size of a request header field exceeds server limit.
    6. Talk to PayPal customer service and they have not heard about any Foxfire conflicts.
    7. Windows Firewall and Kaspersky Firewall turned ON but this is the same configuration before the Foxfire software update. PS. I'll give up Foxfire before I give up Kaspersky

    This issue can be caused by corrupted cookies.
    Clear the cache and the cookies from sites that cause problems.
    * "Clear the Cache": Tools > Options > Advanced > Network > Offline Storage (Cache): "Clear Now"
    * "Remove the Cookies" from sites causing problems: Tools > Options > Privacy > Cookies: "Show Cookies"

  • How to avoid WRAP of EXCEL Column HEADER Fields

    We have 3 tablix in the rdl file with 11 inch * 8.5 inch (to print in A4)
    Tablix 1 - HTML viewer
    Tablix 2 - EXCEL
    Tablix 3 - PDF 
    The size of the rdl is fixed , if size altered then printing becomes an issue.
    So we have to keep the rdl size fix
    but at the same time the WRAP of column header of the EXCEL need to be avoided ,
    because if the size of the column header is not reduced then we cannot be able to accomodate in the defined page size
    Eaxmple of SOme Header Field Name 
    1.OccupationCode 
    2.ChargeJobPhase which are quiet large fields, so it need to be wrapped(to reduce the length of the column to accomodate ) to fit in the page.
    So as an Alternative option  I use the COncept of Subreport and called it by adding row outside group otherwise the
    Report Server treats subreport as new report for each instance of the parent report.
    The challenge we faced is when export to EXCEL some of the columns is getting merged , I believe the size of the child report (sub report) is having the length
    greater than the parent report (Called Report )  and the size has to be large otherwise the COLUMN HEADER need to reduced which makes it Wrapped 
    So  is there any alternative without changing the parent report size and without WRAPPING the HEADER text of CHild report the MERGING of cell can be avoided.
    Since MERGING is happening which is not allowing us to PUT  FILTER in EXCEL but we need to have FILTER applicable since the END user use it to sort, find the records.

    Hi SubhadipRoy,
    In Reporting Services, whether to wrap text or not is depended on the CanGrow property. If we set the CanGrow property of tablix header row to true, then the header row of tablix will be expanded. If we set the CanGrow property of tablix header row
    to false, the header row will be shrunk as the designed size.
    So if we want to avoid wrap text, we can try to set the CanGrow property of tablix header row to false in Reporting Services level. Alternatively, we can click the Wrap Text button to enable the expended extra-long text property in Excel level.
    If there are any other questions, please feel free to ask.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • I am getting an error Bad Request Request too long HTTP Error 400. The size of the request headers is too long.

    When I sign into Methodintegration.com my data base I can get in ok. When I sign out I get the following error: I am getting an error Bad Request Request too long HTTP Error 400. The size of the request headers is too long. If I close Fire Fox and go back in it will work one time and then I get the same error message.

    That is usually a problem with corrupted cookies.
    Clear the cache and the cookies from sites that cause problems.
    * "Clear the Cache": Tools > Options > Advanced > Network > Offline Storage (Cache): "Clear Now"
    * "Remove the Cookies" from sites causing problems: Tools > Options > Privacy > Cookies: "Show Cookies"

  • If I access the website of Cebu Pacific I get as reply upon entering a search request "HTTP Error 400. The size of the request headers is too long.", but if i use private browsing i do not get this reply. How do i resolve that?

    If I access the website of Cebu Pacific Air I get as reply upon entering a search request "HTTP Error 400. The size of the request headers is too long.", but if i use private browsing i do not get this reply. This refers to Firefox 5.0 on ubuntu 11.04
    I have tried to resolve this through deleting history and the cookies, but i always get the same reply. Although if i use firefox (ubuntu) private browsing or Firefox on Windows 7 i do not get this error message.
    What causes this error and how do i resolve this?

    Copied from the link mentioned below.
    ''That is usually a problem with corrupted cookies. Clear the cache and the cookies from sites that cause problems.
    "Clear the Cache": Tools > Options > Advanced > Network > Offline Storage (Cache): "Clear Now"
    "Remove the Cookies" from sites causing problems: Tools > Options > Privacy > Cookies: "Show Cookies" ''
    https://support.mozilla.com/en-US/questions/785186?s=http+error+400&as=s
    Check and tell if its working.

  • IP SLA Request size (ARR data portion) UDP Jitter

    Looking through the documentation for ip sla udp jitter, Cisco says these are the default values -
    "By default, ten packet frames (N), each with a payload size of 10 bytes (S), are generated every 10 ms (T), and the operation is repeated every 60 seconds (F)."
    however, looking at sh ip sla configuration on my router I get different values -
    ip sla 1001
    udp-jitter 2.2.2.2 20000 source-ip 1.1.1.1
    ip sla schedule 1001 life 3000 start-time now
    sh ip sla configuration
    Entry number: 1001
    Owner:
    Tag:
    Operation timeout (milliseconds): 5000
    Type of operation to perform: udp-jitter
    Target address/Source address: 1.1.1.1/2.2.2.2
    Target port/Source port: 20000/0
    Type Of Service parameter: 0x0
    Request size (ARR data portion): 32
    Packet Interval (milliseconds)/Number of packets: 20/10
    This is confusing to me. According to this, the operation is sending 10 frames, at 32 bytes each with 20 ms between frames. Is this correct? Or are these default parameters for UDP Jitter? If so, this definitely does not match up with the data presented on cisco.com
    thanks,
    Alex

    Hi Alex ,
    for UDP-jitter ,  Request Data size is :32( By Default) ,however you can change it .
    NMS-6500(config-ip-sla-jitter)#request-data-size ?
      <16-1500>  Number of bytes in payload
    for ICMP ECHO ,
    Request Data size is :28  ( By Default)
    Thanks-
    Afroz
    [Do rate the useful post]

  • Change Font size in Spool request

    Hi,
       Report output list is generated in spool.
       used GET_PRINT_PARAMETERS and write statements.
      The requirment is to increase the font size in spool.
      Please let me know how can i approch regarding this.
    Thanks & Regards,
    Padmaja.

    Hi ,
    Font size in spool request can't be changed .  It can be changed while printing using format settings .
    I Hope u will  got my point.
    Regards,
    Anish

  • Get a specific Http header field from an http request in a component

    Hello,
    In a custom component that I've created ( which extends ServiceHandler), I'd like to access a specific http header field from the http requests that this component receives.
    I assume this could be accessible through m_binder or m_service, but I didn't find any solution so far.
    Any Idea?
    Thanks in advance,
    Leo

    As it may help others, I answer what I figured out.
    Every details of the http request is put into the m_binder.
    So a specific field can be accessed with m_binder.getEnvironmentValue("My_data")
    You can also list all the available data as it is a standard properties set.

  • Java.lang.OutOfMemoryError: Requested array size exceeds VM limit

    Hi!
    I've a this problem and I do not know how to reselve it:
    I' ve an oracle 11gr2 database in which I installed the Italian network
    when I try to execute a Shortest Path algorithm or a shortestPathAStar algorithm in a java program I got this error.
    [ConfigManager::loadConfig, INFO] Load config from specified inputstream.
    [oracle.spatial.network.NetworkMetadataImpl, DEBUG] History metadata not found for ROUTING.ITALIA_SPAZIO
    [LODNetworkAdaptorSDO::readMaximumLinkLevel, DEBUG] Query String: SELECT MAX(LINK_LEVEL) FROM ROUTING.ITALIA_SPAZIO_LINK$ WHERE LINK_LEVEL > -1
    *****Begin: Shortest Path with Multiple Link Levels
    *****Shortest Path Using Dijkstra
    [oracle.spatial.network.lod.LabelSettingAlgorithm, DEBUG] User data categories:
    [LODNetworkAdaptorSDO::isNetworkPartitioned, DEBUG] Query String: SELECT p.PARTITION_ID FROM ROUTING.ITA_SPAZIO_P_TABLE p WHERE p.LINK_LEVEL = ? AND ROWNUM = 1 [1]
    [QueryUtility::prepareIDListStatement, DEBUG] Query String: SELECT NODE_ID, PARTITION_ID FROM ROUTING.ITA_SPAZIO_P_TABLE p WHERE p.NODE_ID IN ( SELECT column_value FROM table(:varray) ) AND LINK_LEVEL = ?
    [oracle.spatial.network.lod.util.QueryUtility, FINEST] ID Array: [2195814]
    [LODNetworkAdaptorSDO::readNodePartitionIds, DEBUG] Query linkLevel = 1
    [NetworkIOImpl::readLogicalPartition, DEBUG] Read partition from blob table: partition 1181, level 1
    [LODNetworkAdaptorSDO::readPartitionBlobEntry, DEBUG] Query String: SELECT BLOB, NUM_INODES, NUM_ENODES, NUM_ILINKS, NUM_ELINKS, NUM_INLINKS, NUM_OUTLINKS, USER_DATA_INCLUDED FROM ROUTING.ITA_SPAZIO_P_BLOBS_TABLE WHERE PARTITION_ID = ? AND LINK_LEVEL = ? [1181,1]
    [oracle.spatial.network.lod.LabelSettingAlgorithm, WARN] Requested array size exceeds VM limit
    [NetworkIOImpl::readLogicalPartition, DEBUG] Read partition from blob table: partition 1181, level 1
    [LODNetworkAdaptorSDO::readPartitionBlobEntry, DEBUG] Query String: SELECT BLOB, NUM_INODES, NUM_ENODES, NUM_ILINKS, NUM_ELINKS, NUM_INLINKS, NUM_OUTLINKS, USER_DATA_INCLUDED FROM ROUTING.ITA_SPAZIO_P_BLOBS_TABLE WHERE PARTITION_ID = ? AND LINK_LEVEL = ? [1181,1]
    Exception in thread "main" java.lang.OutOfMemoryError: Requested array size exceeds VM limit
    I use the sdoapi.jar, sdomn.jar and sdoutl.jar stored in the jlib directory of the oracle installation path.
    When I performe this query : SELECT BLOB, NUM_INODES, NUM_ENODES, NUM_ILINKS, NUM_ELINKS, NUM_INLINKS, NUM_OUTLINKS, USER_DATA_INCLUDED FROM ROUTING.ITA_SPAZIO_P_BLOBS_TABLE WHERE PARTITION_ID = ? AND LINK_LEVEL = ? [1181,1]
    I got the following result
    BLOB NUM_INODES NUM_ENODES NUM_ILINKS NUM_ELINKS NUM_INLINKS NUM_OUTLINKS USER_DATA_INCLUDED
    (BLOB) 3408 116 3733 136 130 128 N
    then the java code I use is :
    package it.sistematica.oracle.spatial;
    import it.sistematica.oracle.network.data.Constant;
    import java.io.InputStream;
    import java.sql.Connection;
    import oracle.spatial.network.lod.DynamicLinkLevelSelector;
    import oracle.spatial.network.lod.GeodeticCostFunction;
    import oracle.spatial.network.lod.HeuristicCostFunction;
    import oracle.spatial.network.lod.LODNetworkManager;
    import oracle.spatial.network.lod.LinkLevelSelector;
    import oracle.spatial.network.lod.LogicalSubPath;
    import oracle.spatial.network.lod.NetworkAnalyst;
    import oracle.spatial.network.lod.NetworkIO;
    import oracle.spatial.network.lod.PointOnNet;
    import oracle.spatial.network.lod.config.LODConfig;
    import oracle.spatial.network.lod.util.PrintUtility;
    import oracle.spatial.util.Logger;
    public class SpWithMultiLinkLevel
         private static NetworkAnalyst analyst;
         private static NetworkIO networkIO;
         private static void setLogLevel(String logLevel)
         if("FATAL".equalsIgnoreCase(logLevel))
         Logger.setGlobalLevel(Logger.LEVEL_FATAL);
         else if("ERROR".equalsIgnoreCase(logLevel))
         Logger.setGlobalLevel(Logger.LEVEL_ERROR);
         else if("WARN".equalsIgnoreCase(logLevel))
         Logger.setGlobalLevel(Logger.LEVEL_WARN);
         else if("INFO".equalsIgnoreCase(logLevel))
         Logger.setGlobalLevel(Logger.LEVEL_INFO);
         else if("DEBUG".equalsIgnoreCase(logLevel))
         Logger.setGlobalLevel(Logger.LEVEL_DEBUG);
         else if("FINEST".equalsIgnoreCase(logLevel))
         Logger.setGlobalLevel(Logger.LEVEL_FINEST);
         else //default: set to ERROR
         Logger.setGlobalLevel(Logger.LEVEL_ERROR);
         public static void main(String[] args) throws Exception
              String configXmlFile =                "LODConfigs.xml";
              String logLevel =           "FINEST";
              String dbUrl =                Constant.PARAM_DB_URL;
              String dbUser =                Constant.PARAM_DB_USER;
              String dbPassword =                Constant.PARAM_DB_PASS;
              String networkName =                Constant.PARAM_NETWORK_NAME;
              long startNodeId = 2195814;
              long endNodeId = 3415235;
         int linkLevel = 1;
         double costThreshold = 1550;
         int numHighLevelNeighbors = 8;
         double costMultiplier = 1.5;
         Connection conn = null;
         //get input parameters
         for(int i=0; i<args.length; i++)
         if(args.equalsIgnoreCase("-dbUrl"))
         dbUrl = args[i+1];
         else if(args[i].equalsIgnoreCase("-dbUser"))
         dbUser = args[i+1];
         else if(args[i].equalsIgnoreCase("-dbPassword"))
         dbPassword = args[i+1];
         else if(args[i].equalsIgnoreCase("-networkName") && args[i+1]!=null)
         networkName = args[i+1].toUpperCase();
         else if(args[i].equalsIgnoreCase("-linkLevel"))
         linkLevel = Integer.parseInt(args[i+1]);
         else if(args[i].equalsIgnoreCase("-configXmlFile"))
         configXmlFile = args[i+1];
         else if(args[i].equalsIgnoreCase("-logLevel"))
         logLevel = args[i+1];
         // opening connection
         System.out.println("Connecting to ......... " + Constant.PARAM_DB_URL);
         conn = LODNetworkManager.getConnection(dbUrl, dbUser, dbPassword);
         System.out.println("Network analysis for "+networkName);
         setLogLevel(logLevel);
         //load user specified LOD configuration (optional),
         //otherwise default configuration will be used
         InputStream config = (new Network()).readConfig(configXmlFile);
         LODNetworkManager.getConfigManager().loadConfig(config);
         LODConfig c = LODNetworkManager.getConfigManager().getConfig(networkName);
         //get network input/output object
         networkIO = LODNetworkManager.getCachedNetworkIO(
         conn, networkName, networkName, null);
         //get network analyst
         analyst = LODNetworkManager.getNetworkAnalyst(networkIO);
         double[] costThresholds = {costThreshold};
         LogicalSubPath subPath = null;
         try
              System.out.println("*****Begin: Shortest Path with Multiple Link Levels");
              System.out.println("*****Shortest Path Using Dijkstra");
              String algorithm = "DIJKSTRA";
              linkLevel = 1;
              costThreshold = 5000;
              subPath = analyst.shortestPathDijkstra(new PointOnNet(startNodeId), new PointOnNet(endNodeId),linkLevel, null);
              PrintUtility.print(System.out, subPath, true, 10000, 0);
              System.out.println("*****End: Shortest path using Dijkstra");
              catch (Exception e)
              e.printStackTrace();
              try
              System.out.println("*****Shortest Path using Astar");
              HeuristicCostFunction costFunction = new GeodeticCostFunction(0,-1, 0, -2);
              LinkLevelSelector lls = new DynamicLinkLevelSelector(analyst, linkLevel, costFunction, costThresholds, numHighLevelNeighbors, costMultiplier, null);
              subPath = analyst.shortestPathAStar(
              new PointOnNet(startNodeId), new PointOnNet(endNodeId), null, costFunction, lls);
              PrintUtility.print(System.out, subPath, true, 10000, 0);
              System.out.println("*****End: Shortest Path Using Astar");
              System.out.println("*****End: Shortest Path with Multiple Link Levels");
              catch (Exception e)
              e.printStackTrace();
         if(conn!=null)
         try{conn.close();} catch(Exception ignore){}
    At first I create a two link level network with this command
    exec sdo_net.spatial_partition('ITALIA_SPAZIO', 'ITA_SPAZIO_P_TABLE', 5000, 'LOAD_DIR', 'sdlod_part.log', 'w', 1);
    exec sdo_net.spatial_partition('ITALIA_SPAZIO', 'ITA_SPAZIO_P_TABLE', 60000, 'LOAD_DIR', 'sdlod_part.log', 'w', 2);
    exec sdo_net.generate_partition_blobs('ITALIA_SPAZIO', 1, 'ITA_SPAZIO_P_BLOBS_TABLE', true, true, 'LOAD_DIR', 'sdlod_part_blob.log', 'w', false, true);
    exec sdo_net.generate_partition_blobs('ITALIA_SPAZIO', 2, 'ITA_SPAZIO_P_BLOBS_TABLE', true, true, 'LOAD_DIR', 'sdlod_part_blob.log', 'w', false, true);
    Then I try with a single level network but I got the same error.
    Please can samebody help me?

    I find the solution to this problem.
    In the LODConfig.xml file I have:
    <readPartitionFromBlob>true</readPartitionFromBlob>
                   <partitionBlobTranslator>oracle.spatial.network.lod.PartitionBlobTranslator11g</partitionBlobTranslator>
    but when I change it to
    <readPartitionFromBlob>true</readPartitionFromBlob>
                   <partitionBlobTranslator>oracle.spatial.network.lod.PartitionBlobTranslator11gR2</partitionBlobTranslator>
    The application starts without the obove mentioned error.

  • Error Code - client cache is smaller than the size of the requested content

    Even though we have increased the size of the ccmcache via Control Panel > Configuration Manager, we still get the Error Code 0x87D01202 (-2016407038) "the content
    download cannot be performed because the total size of the client cache is smaller than the size of the requested content"  The CCMEXEC Service and computer have both been restarted, after increasing the ccmcache size.  Which local log
    file under C:\Windows\CCM\Logs should we check for more information ?
    Thanks

    so when you re deploying the client go into your settings and set the variable below:
    smscachesize=10240
    note:
    SMSCACHESIZE
    Specifies the size of the client cache folder in megabyte (MB) or as a percentage when used with the PERCENTDISKSPACE or PERCENTFREEDISKSPACE property. If this property is not set, the folder defaults to a maximum size of 5120 MB. The lowest value that you
    can specify is 1 MB.
    Note
    If a new package that must be downloaded would cause the folder to exceed the maximum size, and if the folder cannot be purged to make sufficient space available, the package download fails, and the program or application will not run.
    This setting is ignored when you upgrade an existing client and when the client downloads software updates.
    Example: CCMSetup.exe SMSCACHESIZE=100
    Note
    If you reinstall a client, you cannot use the SMSCACHESIZE or SMSCACHEFLAGS installation properties to set the cache size to be smaller than it was previously. If you try to do this, your value is ignored and the cache size is automatically set to the last
    size it was previously.
    For example, if you install the client with the default cache size of 5120 MB, and then reinstall the client with a cache size of 100 MB, the cache folder size on the reinstalled client is set to 5120 MB.
    Twitter: @dguilloryjr LinkedIn: http://www.linkedin.com/in/dannyjr Facebook: http://www.facebook.com/#!/dguilloryjr

  • Routeserver-java.lang.OutOfMemoryError:Requested array size exceeds VM limi

    Well,
    When I started to try running routeserver, I was using "false" in <param-name>long_ids</param-name> (web.xml). When I tried to use "true", an OutOfMemoryError error was occuring. Now I know that "false" is wrong. So, I walked a bit more...
    The error now is:
    09/02/03 15:25:29.547 web: Error initializing servlet
    java.lang.OutOfMemoryError: Requested array size exceeds VM limit
         at oracle.spatial.router.engine.NonBoundaryEdge.readNonBoundaryEdge(NonBoundaryEdge.java:74)
         at oracle.spatial.router.engine.Partition.readPartition(Partition.java:103)
         at oracle.spatial.router.engine.PartitionCache.loadPartitionFromDatabase(PartitionCache.java:286)
         at oracle.spatial.router.engine.PartitionCache.obtainPartitionReference(PartitionCache.java:244)
         at oracle.spatial.router.engine.Network.<init>(Network.java:77)
         at oracle.spatial.router.server.RouteServerImplementation.<init>(RouteServerImplementation.java:136)
         at oracle.spatial.router.server.RouteServerServlet.init(RouteServerServlet.java:299)
         at com.evermind.server.http.HttpApplication.loadServlet(HttpApplication.java:2379)
         at com.evermind.server.http.HttpApplication.findServlet(HttpApplication.java:4830)
         at com.evermind.server.http.HttpApplication.findServlet(HttpApplication.java:4754)
         at com.evermind.server.http.HttpApplication.initPreloadServlets(HttpApplication.java:4942)
         at com.evermind.server.http.HttpApplication.initDynamic(HttpApplication.java:1144)
         at com.evermind.server.http.HttpApplication.<init>(HttpApplication.java:741)
         at com.evermind.server.ApplicationStateRunning.getHttpApplication(ApplicationStateRunning.java:431)
         at com.evermind.server.Application.getHttpApplication(Application.java:586)
         at com.evermind.server.http.HttpSite$HttpApplicationRunTimeReference.createHttpApplicationFromReference(HttpSite.java:1987)
         at com.evermind.server.http.HttpSite$HttpApplicationRunTimeReference.<init>(HttpSite.java:1906)
         at com.evermind.server.http.HttpSite.initApplications(HttpSite.java:643)
         at com.evermind.server.http.HttpSite.setConfig(HttpSite.java:290)
         at com.evermind.server.http.HttpServer.setSites(HttpServer.java:270)
         at com.evermind.server.http.HttpServer.setConfig(HttpServer.java:177)
         at com.evermind.server.ApplicationServer.initializeHttp(ApplicationServer.java:2493)
         at com.evermind.server.ApplicationServer.setConfig(ApplicationServer.java:1042)
         at com.evermind.server.ApplicationServerLauncher.run(ApplicationServerLauncher.java:131)
         at java.lang.Thread.run(Thread.java:595)
    09/02/03 15:25:29.547 web: Error preloading servlet
    javax.servlet.ServletException: Error initializing servlet
         at com.evermind.server.http.HttpApplication.findServlet(HttpApplication.java:4857)
         at com.evermind.server.http.HttpApplication.findServlet(HttpApplication.java:4754)
         at com.evermind.server.http.HttpApplication.initPreloadServlets(HttpApplication.java:4942)
         at com.evermind.server.http.HttpApplication.initDynamic(HttpApplication.java:1144)
         at com.evermind.server.http.HttpApplication.<init>(HttpApplication.java:741)
         at com.evermind.server.ApplicationStateRunning.getHttpApplication(ApplicationStateRunning.java:431)
         at com.evermind.server.Application.getHttpApplication(Application.java:586)
         at com.evermind.server.http.HttpSite$HttpApplicationRunTimeReference.createHttpApplicationFromReference(HttpSite.java:1987)
         at com.evermind.server.http.HttpSite$HttpApplicationRunTimeReference.<init>(HttpSite.java:1906)
         at com.evermind.server.http.HttpSite.initApplications(HttpSite.java:643)
         at com.evermind.server.http.HttpSite.setConfig(HttpSite.java:290)
         at com.evermind.server.http.HttpServer.setSites(HttpServer.java:270)
         at com.evermind.server.http.HttpServer.setConfig(HttpServer.java:177)
         at com.evermind.server.ApplicationServer.initializeHttp(ApplicationServer.java:2493)
         at com.evermind.server.ApplicationServer.setConfig(ApplicationServer.java:1042)
         at com.evermind.server.ApplicationServerLauncher.run(ApplicationServerLauncher.java:131)
         at java.lang.Thread.run(Thread.java:595)
    09/02/03 15:25:29.547 web: 10.1.3.4.0 Started
    I start OC4J with:
    C:\Java\jdk1.5.0_16\bin>java -server -Xms1024m -Xmx1024m -XX:NewSize=512m -XX:Max NewSize=512m -Dsun.rmi.dgc.server.gcInterval=3600000 -Dsun.rmi.dgc.client.gcInterval=3600000 -verbose:gc -jar c:\oc4j\j2eehome\oc4j.jar -config c:\oc4j\j2ee\home\config\server.xml
    My computer has 2 Gb of RAM, AMD Turion 64 Mobile 2.20 GHz
    Any ideas?
    Thanks a lot again!
    Regards,
    Daniel

    Well,
    I am using Router from 11g. The web.xml from 10g does not have logs_id parameter and servlet mapping.... May I add logs_id parameter in web.xml?
    Here we have contract with oracle. But I could NOT find any thing about patchs for routeserver... I am downloading "configuration manager" to update my metalink. Could you tell me tip about where are routeserver patchs?
    Tks a lot,
    Daniel
    Edited by: user10788592 on 03/02/2009 12:09

  • When trying to render i get the error "your hardware cannot render at the requested size and depth"

    when trying to render i get the error "your hardware cannot render at the requested size and depth" and does not render,, it is the first time, and happens only when i added a filter from magic bullet and won't render.

    Do you have 2 graphics cards in your system?
    You can check out your hardware profile using the "System Profiler" app which lives in the /Applications/Utilities folder ... you can also launch it from the Apple menu by choosing "About This Mac..." and then clicking the "More Info..." button. When the app launches you can choose Graphics/Displays from the Contents/Hardware list at left.

Maybe you are looking for

  • How to Retrieve Data From Web Service with Optional Arguments

    This should be a simple one. I have a web service that returns a list of books. There are 3 optional arguments that can be passed to this (ex. book type, author, id) to filter this list. Do I still need to use the <mx:request> tags if I am not going

  • How do i use Jukebox drivers to install firmwa

    I have searched these forums and tried numerous things to get my Zen 20gb to update its firmware. I searched for appropriate updates and found the file JB3MV2_PCWDRV_US__30_03.EXE. It appears that this is the correct update, so i'm wondering what i d

  • How do you fix the Health App from crashing?

    If you put too many steps in the Health App, the Health App crashes and won't let you get back in it. How do you reset it, or fix it?

  • DVI Signal Irregularities

    When I connect my MacBook Pro to my HDTV (Sharp Aquos 37") via DVI cable, the signal looks great. However, every 30 seconds or so, the signal gets a strange irregularity. It's hard to explain, but the displayed image seems to "split" -- with the bott

  • WRT54G V8 firmware upgrade, easylink issue

    I was having the typical dropped  connection issues  that these routers seem to be famous for (WRT54G).   The easylink application was installed and working.  I upgraded firmware to the newly released 8.00.4.  After the firmware upgrade the easylink