Command sequence too long

I am putting together a web template with a button that should automatically filter structures on two different dataproviders.
Here's my command sequence URL:
<SAP_BW_URL DATA_PROVIDER='DP'
FILTER_IOBJNM='9DLWWASJARRDZ6WOY7WWA39G7'
FILTER_VALUE_1='71IMI689IBYS87IHVTZ0TCYDA'
FILTER_VALUE_2='F5C7W9WFPUU3S8JSGJRGIO5F5'
FILTER_VALUE_3='80B8DZ6DOAPN41LVBDRVFH2EE'
FILTER_VALUE_4='9NWL81CA7S8535FW9ZMYIQBMX'
FILTER_VALUE_5='7YB1YNE13ZYI6F29HNTOJBPKZ'
FILTER_VALUE_6='F0I4NP9MRXGNY7TN2B29R57C4'
CMD_1='DATA_PROVIDER=DP_2&FILTER_IOBJNM=0YV5RI7DNTJKFZARAAGAIBDPU&
FILTER_VALUE_1=4SOUXJTCNPHLKJ53AYM0B1LSK&
FILTER_VALUE_2=EXYHBWXTKPFNV5EDOPEY4DY4M&
FILTER_VALUE_3=D92NJZD7QBK48T4G1Y4M3JS67&
FILTER_VALUE_4=8GJ7M67YXFFUE86IONBF2099B&
FILTER_VALUE_5=81S1ATFOAI6638S7NTQ23TY4U&
FILTER_VALUE_6=15KVM3ABYR6LKK22TLJUG1XV2'>
The first DP filter works, but DP_2 only sets the first four filters, since the command only recognizes the first 250 characters. I tried breaking it up into multiple commands: for example, CMD_1 sets the first three, CMD_2 sets the last three, but it only shows the results from the last command.
How can I get this to work? Is there any way around the 250-character limit on a command? If not, is there some kind of attribute to the filter command that tells BW to keep the existing filters in place while adding additional filters?
Thanks,
Jason

There are only nine total columns in both of these structures, so I might be able to exclude the columns I don't want to see within 250 characters...the only problem is that FILTER_SIGN doesn't seem to work. Here's my new URL:
<SAP_BW_URL DATA_PROVIDER='DP'
FILTER_IOBJNM='9DLWWASJARRDZ6WOY7WWA39G7'
FILTER_VALUE_1='71IMI689IBYS87IHVTZ0TCYDA'
FILTER_VALUE_2='F5C7W9WFPUU3S8JSGJRGIO5F5'
FILTER_VALUE_3='80B8DZ6DOAPN41LVBDRVFH2EE'
FILTER_VALUE_4='9NWL81CA7S8535FW9ZMYIQBMX'
FILTER_VALUE_5='7YB1YNE13ZYI6F29HNTOJBPKZ'
FILTER_VALUE_6='F0I4NP9MRXGNY7TN2B29R57C4'
CMD_1='DATA_PROVIDER=DP_2&FILTER_IOBJNM=44GMJ8XO8WAH6F39TQUQ33C97&
FILTER_SIGN=E&FILTER_VALUE_1=8P7FXUFEQHUJL9KLAT6VFUSS6&
FILTER_VALUE_2=5AKSQD4FJA2JRWKH2DT5EB8B9&
FILTER_VALUE_3=18KK6ZB9LV0D24TZBYGL70GNM'>
I've tried FILTER_SIGN=E and FILTER_SIGN_1=E, but neither seem to work.
Message was edited by: Jason Kraft

Similar Messages

  • 11gR2:crsctl, srvctl commands takes too long to respond

    Hi,
    I have successfully configured *11gR2 two node RAC on ASM on Win 2008 64bit.*
    Everything work very well like connecting to database, querying database. Node restart also takes acceptable time to go down & restart the clusterware & database.
    But when I execute crsctl status resource -t or srvctl status database -d db_name commands from any node takes 10-15min to give output.
    They give output & everything completes successfully but takes too long to respond.
    The questions are:
    - If everything works fine then why crsctl, srvctl takes too long to respond?
    - what could be blocking these command to gather clusterware and database status on all nodes?
    - what additional info would be helpful that I can provide?

    dag wrote:
    I dont have this issue either. are you auto starting mpd?
    that time is how long it is up in other words you opened it then closed it at that time
    I'm not sure if you are referring to me or not, but in my screenshot I am timing the lag in ncmpcpp by pressing 'q' in the terminal during the delay, so it quits ncmpcpp immediately after the lag. The lag is longer for ncmpc because the interface loads before the program connects to mpd, so I have to stop it manually immediatly after it connects.
    WonderWoofy wrote:I don't have this problem... if you are starting it as a systemd user service, maybe there is relevant information in the journal.
    The journal did not reveal anything relevant sadly. I have now tried launching mpd without systemd, and the lag remains. I have also noticed that a small mpd programming project I am writing also experiences the same lag when it tries to connect to mpd.

  • The input line is too long. The syntax of the command is incorrect.

    I am a prentice on J2EE by using the Tutorial 1.4.
    I keep getting this error message: "The input line is too long. The syntax of the command is incorrect." from the windows's terminal no matter what arguments I issued for the command asant right after I successfully executed command "asant build" under the directory of both <Install>\j2eetutorial14\examples\web\bookstore and bookstore1 when I practice the topic of The Example Servlets on Chapter 11: Java Servlet Technology.
    The command and the arguments are:
    asant create-db_common
    asant create-bookstore-war
    asant package-bookstore
    asant deploy-war
    All of them have the same outcome:
    "The input line is too long.
    The syntax of the command is incorrect."
    I have tried to do the research on the google and ased some Java developer friends and school teachers; unfortunately, none of them can solve this problem.
    The OS I am using is win2k professional. The commands were issued on DOS terminal. I have successfully built application server, and pointbase, deploy tool. I have utilize this tools and environment to run all the examples very well on Chapter 8: Building Web Services with JAX-RPC.
    Even worse, after failing on running the Servlets Examples, I got the same error message when I go back to the Chapter 8's Examples and run the command "asant build" on the directory <install>\j2eetutorial14\examples\jaxrpc\staticstub.
    The Server and PointBase were activated when I issued those commands.
    Can someone or the Tutorial's authors kindly help? Thank you very much!

    Hi eric.jendrock,
    Appreciate your help; unfortunately, in my asant.bat script I cannot find the line you indicated even though I found the solution addressed in the FAQ as you instructed, too. The only ines I found are:
    :runAnt
    "%_JAVACMD%" %ANT_OPTS% -Dcom.sun.aas.installRoot="%AS_INSTALL%" -Djava.library.path="%AS_INSTALL%\bin;%AS_ICU_LIB%" -classpath "%AS_INSTALL%/lib/appserv-se.jar";"%AS_INSTALL%/lib/admin-cli.jar";"%AS_INSTALL%/lib/appserv-admin.jar";"%AS_INSTALL%/lib/j2ee.jar";"%AS_INSTALL%/lib/appserv-ext.jar";"%AS_INSTALL%/lib/appserv-rt.jar";"%AS_INSTALL%/lib/commons-launcher.jar";"%AS_INSTALL%/lib/install/applications/jmsra/imqjmsra.jar";"%LOCALCLASSPATH%" "-Dant.home=%ANT_HOME%" org.apache.tools.ant.Main %ANT_ARGS% %ANT_CMD_LINE_ARGS%
    goto end
    :runAntWithJikes
    "%_JAVACMD%" %ANT_OPTS% -Dcom.sun.aas.installRoot="%AS_INSTALL%" -Djava.library.path="%AS_INSTALL%\bin;%AS_ICU_LIB%" -classpath "%AS_INSTALL%/lib/appserv-se.jar";"%AS_INSTALL%/lib/admin-cli.jar";"%AS_INSTALL%/lib/appserv-admin.jar";"%AS_INSTALL%/lib/j2ee.jar";"%AS_INSTALL%/lib/appserv-ext.jar";"%AS_INSTALL%/lib/appserv-rt.jar";"%AS_INSTALL%/lib/commons-launcher.jar";"%AS_INSTALL%/lib/install/applications/jmsra/imqjmsra.jar";"%LOCALCLASSPATH%" "-Dant.home=%ANT_HOME%" "-Djikes.class.path=%JIKESPATH%" org.apache.tools.ant.Main %ANT_ARGS% %ANT_CMD_LINE_ARGS%
    goto end
    I however did tried to remove those lines with -Dcom and -Djava but kept the variables of arg0s, "%_JAVACMD%" %ANT_OPTS%. Here is the result I got:
    Usage: java [-options] class [args...]
    (to execute a class)
    or java [-options] -jar jarfile [args...]
    (to execute a jar file)
    where options include:
    -client to select the "client" VM
    -server to select the "server" VM
    Looks funnier. Do you haver any clue or better solutions? Thanks!

  • Compile error "input line is too long" when using useLegacyAOT

    I'm using FlashBuilder 4.7 and the release of AIR 4 on Windows 7 to package for iOS...As soon as I include 3 or more ANEs I can no longer compile to iOS (and also use the new "useLegacyAOT no" command).  When I try I get an error "The input line is too long.   Compilation failed while executing : compile-abc"
    I found a couple other similar issues.  One post suggested instead of pointing to individual ANEs (in Package Contents) to point to a folder instead--but this doesn't fix the
    Another post for a slightly different problem said it was fixed in the release of AIR 4.
    Any ideas?  I wonder if I move my AIR SDK to a simple path (like "c:\air\adt.jar" instead of "C:\Program Files\Adobe\Adobe Flash Builder 4.7 (64 Bit)\eclipse\plugins\com.adobe.flash.compiler_4.7.0.349722\AIRSDK\lib\adt.jar") if it'd help. 
    Thanks in advance!

    Awesome!  I got used to the new useLegacyAOT thing very fast.  In the meantime, I have to decide whether to remove all my ANEs and test or sit through an interminable build sequence.
    What's the current plan for a new release?  Is there any way I could get a beta copy just to use during development?
    Thanks!

  • ORA-00494: enqueue [CF] held for too long (more than 900 seconds)

    Hi ,
    I am getting this error , the database is in no archive mode and there is frequent log switch.
    However following few metalink notes I have set the hidden parameters like
    killcontrolfile_enqueue_blocker=false
    killenqueue_blocker=1
    I wanted to reduce the frequece log switch how to reduce it.
    ORA-00494: enqueue [CF] held for too long (more than 900 seconds) by 'inst 1

    One way is to make the size of the redo logs bigger and another way is to do some less work :)
    Changing Redo Log File SizesEach Redo Log File member in a Group must be identical in size. If you need to make your Redo Log Files larger, use the following steps.
    1. Use the V$LOG view to identify the current active Redo Log Group.
    SQL> SELECT group#, status FROM v$log;
    GROUP# STATUS
    1 INACTIVE
    2 INACTIVE
    3 CURRENT
    4 INACTIVE
    2. Drop one or more of the inactive Redo Log Groups keeping at least two current On-line Redo Log Groups.
    3. Use operating system commands to delete the files that stored the dropped Redo Log Files.
    4. Recreate the groups with larger file sizes. Continue this sequence until all groups have been resized.
    Edited by: TSharma on Jan 17, 2013 11:52 AM

  • Sometimes my computer takes too long to connect to new website. I am running a pretty powerful work program at same time, what is the best solution? Upgrading speed from cable network, is it a hard drive issue? do I need to "clean out" the computer?

    Many times my computer takes too long to connect to new website. I have wireless internet (time capsule) and I am running a pretty powerful real time financial work program at same time, what is the best solution? Upgrading speed from cable network? is it a hard drive issue? do I only need to "clean out" the computer? Or all of the above...not to computer saavy.  It is a Macbook Pro  osx 10.6.8 (late 2010).

    Almost certainly none of the above!  Try each of the following in this order:
    Select 'Reset Safari' from the Safari menu.
    Close down Safari;  move <home>/Library/Caches/com.apple.Safari/Cache.db to the trash; restart Safari.
    Change the DNS servers in your network settings to use the OpenDNS servers: 208.67.222.222 and 208.67.220.220
    Turn off DNS pre-fetching by entering the following command in Terminal and restarting Safari:
              defaults write com.apple.safari WebKitDNSPrefetchingEnabled -boolean false

  • Cannot delete file: filename too long

    Sometime ago, I made a copy of a hierarchy of folders under my "Users" folder to another drive as a quick backup. I no longer need this copy of folders and files. So when I tried to delete these files I got a message that some filenames are too long. When
    I look at these files they mostly seem to be part of Internet Explorer temp directory, and also some files used by Mozilla Thunderbird, so they were likely created by these utilities on purpose. I cannot rename them to shorter names, or even move them to lower
    folders. I can't even copy their filenames to show you what these filenames are! When I use the DOS prompt's "dir /x" command these files do not even have any 8.3 short filenames either, which I could've used to delete them with. There are over 400 of these
    files!
    I know what the problem is, the Windows OS has a limit of somewhere around 250 characters for filenames and complete pathnames. I've seen some utilities advertised saying that they can delete these files, but they all require money. Anything that's free?
    Yousuf Khan

    Hi,
    It is just for testing. It is good since the files can be deleted in Linux.
    However, if you don't have Linux, it needs to find the way to delete the files under Windows.
    Since the files are located in Internet Explorer temp directory, you could use Disk Cleanup to clear the Temporary Internet Files.
    Also, check if the following KB is helpful. It has several methods. Go to check if 8.3 name creation is disabled.
    http://support.microsoft.com/default.aspx?scid=kb;en-us;320081
    Best Regards,
    L.S
    Regarding it being in the Internet Explorer temp directory, it actually wasn't. Remember I said in the original posting that this was in a backup directory of these directories, so Disk Cleanup has no way of knowing that this belongs to Internet Explorer
    or anything else. Also there was a Thunderbird directory having the same issues as well, which Disk Cleanup wouldn't have touched no matter what.
    Okay, I checked out the state of the filesystem registry. The "NtfsDisable8dot3NameCreation" entry in the "HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\FileSystem" is set to non-zero, meaning the feature is in effect. Interestingly it was set to 2
    rather than 1, but I doubt that makes a difference.
    Even if I disable this feature right now, all previously created files still won't be assigned the 8.3 filenames, just only those files that are newly created will have the 8.3 filenames available.
    Yousuf Khan

  • Browser times out when trying to view my website - says the server is taking too long. And no, I don't have a firewall.

    I can't view my website at www.artisancandies.com, even though it's working and everyone else seems to see it. No, I don't have a firewall, and it's not because of my internet provider - I have AT&T at work, and Comcast at home. My husband can see the site on his laptop. I tried dumping my cache in both Firefox and Safari, but it didn't work. I looked at it through proxify.com, and can see it that way, so I know it works. This is so frustrating, because I used to only see it when I typed in artisancandies.com - it would never work for me if I typed in www.artisancandies.com. Now it doesn't work at all. This is the message I get in Firefox:
    "The connection has timed out. The server at www.artisancandies.com is taking too long to respond."
    Please help!!!
    Kristen Scott

    Linc, here's what I've got from what you asked me to do. I hope you don't mind, but it was simple enough to leave everything in, so you could see the progression:
    Kristen-Scotts-Computer:~ kristenscott$ kextstat -kl | awk ' !/apple/ { print $6 $7 } '
    Kristen-Scotts-Computer:~ kristenscott$ sudo launchctl list | sed 1d | awk ' !/0x|apple|com\.vix|edu\.|org\./ { print $3 } '
    WARNING: Improper use of the sudo command could lead to data loss
    or the deletion of important system files. Please double-check your
    typing when using sudo. Type "man sudo" for more information.
    To proceed, enter your password, or type Ctrl-C to abort.
    Password:
    com.microsoft.office.licensing.helper
    com.google.keystone.daemon
    com.adobe.versioncueCS3
    Kristen-Scotts-Computer:~ kristenscott$ launchctl list | sed 1d | awk ' !/0x|apple|edu\.|org\./ { print $3 } '
    com.google.keystone.root.agent
    com.adobe.ARM.202f4087f2bbde52e3ac2df389f53a4f123223c9cc56a8fd83a6f7ae
    Kristen-Scotts-Computer:~ kristenscott$ ls -1A {,/}Library/{Ad,Compon,Ex,Fram,In,La,Mail/Bu,P*P,Priv,Qu,Scripti,Sta}* 2> /dev/null
    /Library/Components:
    /Library/Extensions:
    /Library/Frameworks:
    Adobe AIR.framework
    NyxAudioAnalysis.framework
    PluginManager.framework
    iLifeFaceRecognition.framework
    iLifeKit.framework
    iLifePageLayout.framework
    iLifeSQLAccess.framework
    iLifeSlideshow.framework
    /Library/Input Methods:
    /Library/Internet Plug-Ins:
    AdobePDFViewer.plugin
    Disabled Plug-Ins
    Flash Player.plugin
    Flip4Mac WMV Plugin.plugin
    Flip4Mac WMV Plugin.webplugin
    Google Earth Web Plug-in.plugin
    JavaPlugin2_NPAPI.plugin
    JavaPluginCocoa.bundle
    Musicnotes.plugin
    NP-PPC-Dir-Shockwave
    Quartz Composer.webplugin
    QuickTime Plugin.plugin
    Scorch.plugin
    SharePointBrowserPlugin.plugin
    SharePointWebKitPlugin.webplugin
    flashplayer.xpt
    googletalkbrowserplugin.plugin
    iPhotoPhotocast.plugin
    npgtpo3dautoplugin.plugin
    nsIQTScriptablePlugin.xpt
    /Library/LaunchAgents:
    com.google.keystone.agent.plist
    /Library/LaunchDaemons:
    com.adobe.versioncueCS3.plist
    com.apple.third_party_32b_kext_logger.plist
    com.google.keystone.daemon.plist
    com.microsoft.office.licensing.helper.plist
    /Library/PreferencePanes:
    Flash Player.prefPane
    Flip4Mac WMV.prefPane
    VersionCue.prefPane
    VersionCueCS3.prefPane
    /Library/PrivilegedHelperTools:
    com.microsoft.office.licensing.helper
    /Library/QuickLook:
    GBQLGenerator.qlgenerator
    iWork.qlgenerator
    /Library/QuickTime:
    AppleIntermediateCodec.component
    AppleMPEG2Codec.component
    Flip4Mac WMV Export.component
    Flip4Mac WMV Import.component
    Google Camera Adapter 0.component
    Google Camera Adapter 1.component
    /Library/ScriptingAdditions:
    Adobe Unit Types
    Adobe Unit Types.osax
    /Library/StartupItems:
    AdobeVersionCue
    HP Trap Monitor
    Library/Address Book Plug-Ins:
    SkypeABDialer.bundle
    SkypeABSMS.bundle
    Library/Internet Plug-Ins:
    Move_Media_Player.plugin
    fbplugin_1_0_1.plugin
    Library/LaunchAgents:
    com.adobe.ARM.202f4087f2bbde52e3ac2df389f53a4f123223c9cc56a8fd83a6f7ae.plist
    com.apple.FolderActions.enabled.plist
    com.apple.FolderActions.folders.plist
    Library/PreferencePanes:
    A Better Finder Preferences.prefPane
    Kristen-Scotts-Computer:~ kristenscott$

  • OCDM  - Sample Report Installation issue - Runs too longer with No progress

    Hello,
    As a part of OCDM environment setup for Practice we were successful installing following :
    * Supported Enterprise Linux
    * Oracle DB 11g R2 (OLAP & MINING option)
    * OWB was already installed as a part of Oracle DB 11gR2 - just unlocked the OWB accounts.
    * Installed OBIEE 11g R2 installation - Installation successful and able to open url's. (analytics and BI Publisher etc.) - Issue with login account.
    * Installation of Communication Data Model (First Installation type from OCDM Installer)
    When we try to install Sample Reports at same stage the screen remain open and runs too longer with No progress.
    The Installation process for OCDM Sample Reports is given in the attached doc with screenshot and the log file is also included in the attached document.
    Please let us know the following two issue:
    1) How to login to Analytics after OBIEE 11gR2 installed on Linux - ( in our case we use user as Administrator and tried password same as username, even checked with the user weblogic and other but no success. Please provide us the information which give details(document) about the Administartion of OBIEE 11gR2 on Linux).
    2) The issue with installation of OCDM Sampple Reports though successful installation of OCDM Data Model.(First Installation Type from OCDM Installer)
    Thanks & Best Regards,
    Amol Thite
    I think, I cant Attach the file here and character limit is 30K for the post here so putting below the few lines of Log File content:
    I can email the screen shots docs and complete log file on request if needed.
    It stuck in Configuration Assistance screen -
    After more than 1 hour the same screen – No progress – I tried 3 times but same issue. Log file also not growing and seems its waiting for some dependent thing that need to happen before it proceed.
    =====start few lines of Log File=======================================
    The file oraparam.ini could not be found at /u01/app/oracle/product/11.1.0/db_1/oui/bin/oraparam.ini
    Using paramFile: /u01/app/oracle/product/11.1.0/db_1/oui/oraparam.ini
    Checking swap space: must be greater than 500 MB. Actual 1027 MB Passed
    Checking monitor: must be configured to display at least 256 colors. Actual 65536 Passed
    The number of files bootstrapped for the jre is 689.
    The number of files bootstrapped for the oui is 77.
    Using the umask value '022' available from oraparam.ini
    =====end few lines of Log File==================================
    INFO: Copying Aggr XML for: Oracle Communications Data Model
    INFO: The Top level Aggreage File = /u01/app/oracle/product/11.1.0/db_1/inventory/ContentsXML/ConfigXML/oracle.ocdm.11_2_3_0_0.xml
    INFO: deleted all the required instance files
    INFO: OUI_CAPlugIn is not found in XML
    INFO: cf session will be created for OH: /u01/app/oracle/product/11.1.0/db_1/ TLAggr: oracle.ocdm instancePath: inventory/ContentsXML/ConfigXML/
    INFO: cf session for OH: /u01/app/oracle/product/11.1.0/db_1/ TL Aggr: [oracle.ocdm] instancePath: inventory/ContentsXML/ConfigXML/
    INFO: aggr ref length : 2
    INFO: cf session hashcode: 22855989
    INFO: cf session saved with key: OraDb11g_home1 oracle.ocdm
    INFO: cf session is ok
    INFO: created and saved cf session for oh: OraDb11g_home1
    INFO: passing params to cf
    INFO: Handling the storing of variables for aggr name oracle.ocdm
    INFO: This variable sl_ASMSelectableDiskGroups is not added to the global context map
    INFO: This variable s_scratchPath is not added to the global context map
    INFO: exitonly tools to be excuted passed: 0
    INFO: Starting to execute configuration assistants
    INFO: Command = oracle.ocdm.OCDMCfgPlugIn /u01/app/oracle/product/11.1.0/db_1/ocdm/ocdm_install.sh ${s_dbSysPasswd}
    OCDMCfgPlugIn: Starting OCDM configuration...
    OCDMCfgPlugIn: OCDM configuration initialized, waiting for script response...
    OCDMCfgPlugIn: Receiving SYSTEM password...
    OCDMCfgPlugIn: Passing SYSTEM credentials to configuration script...
    OCDMCfgPlugIn: SYSTEM credentials passed to configuration script.
    OCDMCfgPlugIn: SYSTEM password received.
    OCDMCfgPlugIn: Initializing config parameteres...
    OCDMCfgPlugIn: Creating log folder...
    OCDMCfgPlugIn: ocdm log folder exist
    OCDMCfgPlugIn: Exporting config env...
    OCDMCfgPlugIn: Performing OWB check...
    OCDMCfgPlugIn: Installing OWB...
    OCDMCfgPlugIn: WARNING: OWB OWNER already exists.
    OCDMCfgPlugIn: Archive: /u01/app/oracle/product/11.1.0/db_1/ocdm/pdm/relational/sample_schema/ocdm_sample.dmp.zip
    OCDMCfgPlugIn: inflating: /u01/app/oracle/product/11.1.0/db_1/ocdm/install_tmp/ocdm_sample.dmp
    OCDMCfgPlugIn: 2011_02_28_11_24_30 10 % complete
    OCDMCfgPlugIn: Importing OCDM sample schema
    ==================end of log File ========================================

    Hi Please check my answers
    1) How to login to Analytics after OBIEE 11gR2 installed on Linux - ( in our case we use user as Administrator and tried password same as username, even checked with the user weblogic and other but no success. Please provide us the information which give details(document) about the Administartion of OBIEE 11gR2 on Linux).
    The current OCDM version does not support BIEE 11g, so you need to install BIEE 10g. In next release(will be available soon), OCDM will support BIEE 11g.
    2) The issue with installation of OCDM Sampple Reports though successful installation of OCDM Data Model.(First Installation Type from OCDM Installer)
    The importing will take a long time, you can login database with ocdm_sample_sys/ocdm_sample_sys, and use "select count(*) from tabs;" to monitor the import process.

  • Database open (recovery) taking too long

    Hi,
    Ive been using your awesome BerkeleyDB Java Edition for a couple of years, and have been very happy with it.
    I am currently facing an issue with trying to open the database after a disk-full issue (which resulted in the database being unable to write, and hence not closed properly).
    While recovery seems to be operating, it has been taking an inordinate amount of time - 16 hours so far. My database has data of around 200GB, which inflated to over 450GB during deletion of entries, hence gobbling up all free space on disk.
    My questions are:
    * Should i continue to wait for recovery?
    * Is there any chance that recovery is looping?
    * Is there an easier way (DBDump?) to extract data from the database without having to perform recovery?
    Some other information that may help:
    * The recovery has decreased the size of the last significant file, and created 3 new files since it started running.
    * I have been monitoring the open files (using lsof), and they change every now and then to other files, though a good amount of its time is spent near the end of the database.
    Thus, i feel like recovery is running normally, just taking too long. Please let me know your opinion.
    A few other things i should mention regarding my issue:
    * The database was, till yesterday, running on bdb java 3.3.75. After running several hours of recovery, i upgraded to 4.1.10 (since i read about a possible recovery looping bug in one of the versions)
    * Once 4.1.10 started recovery, it spat out errors regarding the last 2 files. Only on deleting those 2 files (the last being 0 bytes, the 2nd-last being about 5k) did the recovery start. Note that the older 3.3.75's recovery never complained about those files. I can post the errors here if relevant.
    * Some of the jdb files (about 500 files out of the 47,000 files that make up the database) are 100 MB files, since i had experimented with larger sized files for a few days, then reverted the setting.
    Would any of these above affect a successful recovery?
    My setup is:
    OS:Linux CentOS 5.2, 64-bit, kernel 2.6.18-92.el5
    JVM: Sun Java 1.6.0_20, 64-bit
    Memory: 16 GB RAM, of which 8 GB is allocated to the java process (-Xmx8000M -Xms8000M)
    BDB cache set to use 6GB RAM (envconfig.setCacheSize(6000000000))
    Only the BDB basic API is being used (Environment, database, cursors). We do not use DPL, or HA features.
    Awaiting your kind response,
    Sushant A

    Hi Sushant,
    * Should i continue to wait for recovery?* Is there any chance that recovery is looping?>
    I'm not aware of a bug that would cause recovery to loop, however, you may want to take thread dumps to see if it is progressing. It isn't easy to tell, however, since each phase of recovery is in fact a loop. What you can tell easily from the thread dumps is whether recovery is blocked (completely stopped) for some reason. I don't know of a bug that would cause this, but it's something I would check for.
    Assuming it is not blocked, I suggest that you leave recovery running, and additionally (in parallel) try to obtain some information about your log. While recovery is running you can run the DbPrintLog utility, which does not itself run recovery. I suggest running the following command, which will tell us in general what your log looks like and in particular how far apart the checkpoints are:
    java -jar je-x.y.z.jar DbPrintLog -h <envHome> -S > <output>Please post the output.
    If checkpoints are not running in your application for some reason, or they are running very infrequently, this can cause VERY long recoveries. Unfortunately, you may have such a problem in your app and not be aware of it, until you crash and have to recover. To guard against this sort of thing in the future, you should keep an eye on the checkpoint frequency. EnvironmentStats.getNCheckpoints and getEndOfLog can together be used to tell how much log is written between checkpoints. We will also be able to see this from the DbPrintLog -S output.
    * Is there an easier way (DBDump?) to extract data from the database without having to perform recovery?DbDump normally runs recovery. DbDump with the -r or -R option does not run recovery, but has other drawbacks. With -r, a large amount of memory may be necessary to dump an accurate representation of your data set. If this fails because you run out of memory, -R can be used, but this will dump multiple versions of each record and it will be up to you to interpret the output.
    If regular recovery does not succeed, then DbDump -r is the next thing to try.
    Would any of these above affect a successful recovery?No, I don't believe so.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • "ORA-01704: string literal too long" in Oracle 10g

    I would like to explain my problem with an example, let's create the following table:
    CREATE TABLE longtest
    (text LONG);
    I have to insert lots of data into this table and when I run this command:
    INSERT INTO LONGTEST VALUES ('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    ..................more than 4000 characters..................
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa');
    I get the following error:
    ORA-01704: string literal too long
    To workaround this, I used bind variables, like this:
    BEGIN
    EXECUTE IMMEDIATE 'INSERT INTO LONGTEST VALUES (:a) '
    using
    'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    ..............(repeat more than 32768 characters)..............
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa';
    COMMIT;
    END;
    This worked fine with Oracle 9.2.0.4 but when I do the same thing
    in Oracle 10.2.0.2 there is a 32768 characters limit.
    I am planning to upgrade the database but I need a workaround for
    this, anyone knows why Oracle behaves differently in 10g??

    So, I wonder if anyone can explain why the following code works fine in Oracle 9.2.0.4 but not in Oracle 10.2.0.2:
    CREATE TABLE longtest
    (text LONG);
    BEGIN
    EXECUTE IMMEDIATE 'INSERT INTO LONGTEST VALUES (:a) '
    using
    'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    ..............(repeat more than 32768 characters)..............
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
    aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa';
    COMMIT;
    END;
    Regards,
    Cassiano

  • Large XML Publisher reports fail with client-error-request-value-too-long

    We are running BI Publisher 5.6.3 with Oracle Applications 11.5.10.2 (ATG.H RUP 5), and are having frequent failures with client-error-request-value-too-long messages on large PDF reports that cause the entire report to not print.
    The exact error message is as follows:
    lp: unable to print file: client-error-request-value-too-long
    Pasta: Error: Print failed. Command=lp -c -dPDX_B6_1LJ5200NEAST /logs/temp/pasta3262_0.tmp
    Pasta: Error: Check printCommand/ntPrintCommand in pasta.cfg
    Pasta: Error: Preprocess or Print command failed!!!
    APP-FND-00500: AFPPRN received a return code of failure from routine FDUPRN. Program exited with status 1
    Cause: AFPPRN received a return code of failure from the OSD routine FDUPRN. Program exited with status 1.
    Action: Review your concurrent request log file for more detailed information.
    I have Google'd client-error-request-value-too-long, and found it is a common CUPS issue on Linux. The popular solutions (having /var/spool/cups and /var/spool/cups/tmp) are already in place. Also, the temp files are nowhere near 2 GB (44MB).
    Has anyone had this issue with BI Publisher on Linux?

    Thanks for the link. It looks like the sysadmins have throttled cups to low to allow large bitmapped print jobs:
    grep MaxRequestSize cupsd.conf
    # MaxRequestSize: controls the maximum size of HTTP requests and print files.
    MaxRequestSize 10M
    I am trying to get a more reasonable size limit.

  • Large BI Publisher reports fail with client-error-request-value-too-long

    We are running BI Publisher 5.6.3 with Oracle Applications 11.5.10.2, and are having frequent failures with client-error-request-value-too-long messages on large PDF reports that cause the entire report to not print.
    The exact error message is as follows:
    lp: unable to print file: client-error-request-value-too-long
    Pasta: Error: Print failed. Command=lp -c -dPDX_B6_1LJ5200NEAST /logs/temp/pasta3262_0.tmp
    Pasta: Error: Check printCommand/ntPrintCommand in pasta.cfg
    Pasta: Error: Preprocess or Print command failed!!!
    APP-FND-00500: AFPPRN received a return code of failure from routine FDUPRN. Program exited with status 1
    Cause: AFPPRN received a return code of failure from the OSD routine FDUPRN. Program exited with status 1.
    Action: Review your concurrent request log file for more detailed information.
    I have Google'd client-error-request-value-too-long, and found it is a common CUPS issue on Linux. The popular solutions (having /var/spool/cups and /var/spool/cups/tmp) are already in place. Also, the temp files are nowhere near 2 GB (44MB).
    Has anyone had this issue with BI Publisher on Linux?

    The linux sysadmins set the cups max_request_size at 10 MB, which was causing this error. Once the restriction was lifted, the reports ran without error.

  • SharePoint 2010, Visual Studio 2010, Packaging a solution - The specified path, file name, or both are too long. The fully qualified file name must be less than 260 characters, and the directory name must be less than 248 characters.

    Hi,
    I have a solution that used to contain one SharePoint 2010 project. The project is named along the following lines:
    <Company>.<Product>.SharePoint - let's call it Project1 for future reference. It contains a number of features which have been named according
    to their purpose, some are reasonably long and the paths fairly deep. As far as I am concerned we are using sensible namespaces and these reflect our company policy of "doing things properly".
    I first encountered the following error message when packaging the aforementioned SharePoint project into a wsp:
    "The specified path, file name, or both are too long. The fully qualified file name must be less than 260 characters, and the directory name must be less than 248 characters."
    I went through a great deal of pain in trying to rename the project, shorten feature names and namespaces etc... until I got it working. I then went about gradually
    renaming everything until eventually I had what I started with, and it all worked. So I was none the wiser...not ideal, but I needed to get on and had tight delivery timelines.
    Recently we wanted to add another SharePoint project so that we could move some of our core functinality out into a separate SharePoint solution - e.g. custom workflow
    error logging. So we created another project in Visual Studio called:
    <Company>.<Product>.SharePoint.<Subsystem> - let's call it Project2 for future reference
    And this is when the error has come back and bitten me! The scenario is now as follows:
    1. project1 packages and deploys successfully with long feature names and deep paths.
    2. project2 does not package and has no features in it at all. The project2 name is 13 characters longer than project1
    I am convinced this is a bug with Visual Studio and/or the Package MSBuild target. Why? Let me explain my findings so far:
    1. By doing the following I can get project2 to package
    In Visual Studio 2010 show all files of project2, delete the obj, bin, pkg, pkgobj folders.
    Clean the solution
    Shut down Visual Studio 2010
    Open Visual Studio 2010
    Rebuild the solution
    Package the project2
    et voila the package is generated!
    This demonstrates that the package error message is in fact inaccurate and that it can create the package, it just needs a little help, since Visual Studio seems to
    no longer be hanging onto something.
    Clearly this is fine for a small time project, but try doing this in an environment where we use Continuous Integration, Unit Testing and automatic deployment of SharePoint
    solutions on a Build Server using automated builds.
    2. I have created another project3 which has a ludicrously long name, this packages fine and also has no features contained within it.
    3. I have looked at the length of the path under the pkg folder for project1 and it is large in comparison to the one that is generated for project2, that is when it
    does successfully package using the method outlined in 1. above. This is strange since project1 packages and project2 does not.
    4. If I attempt to add project2 to my command line build using MSBuild then it fails to package and when I then open up Visual Studio and attempt to package project2
    from the Visual Studio UI then it fails with the path too long error message, until I go through the steps outlined in 1. above to get it to package.
    5. DebugView shows nothing useful during the build and packaging of the project.
    6. The error seems to occur in
    CreateSharePointProjectService target called at line 365 of
    Microsoft.VisualStudio.SharePoint.targetsCurrently I am at a loss to work out why this is happening? My next task is to delete
    project2 completely and recreate it and introduce it into my Visual Studio solution.
    Microsoft, can you confirm whether this is a known issue and whether others have encountered this issue? Is it resolved in a hotfix?
    Anybody else, can you confirm whether you have come up with a solution to this issue? When I mean a solution I mean one that does not mean that I have to rename my namespaces,
    project etc... and is actually workable in a meaningful Visual Studio solution.

    Hi
    Yes, I thought I had fixed this my moving my solution from the usual documents  to
    c:\v2010\projectsOverflow\DetailedProjectTimeline
    This builds ok, but when I come to package I get the lovely error:
    Error 2 The specified path, file name, or both are too long. The fully qualified file name must be less than 260 characters, and the directory name must be less than 248 characters. C:\VS2010\ProjectsOverflow\DetailedProjectTimeline\VisualDetailedProjectTimelineWebPart\Features\Feature1\Feature1.feature VisualDetailedProjectTimeline
    Now, the error seems to be related to 
    Can anyone suggest what might be causing this. Probably some path in an XML file somewhere. Here is my prime suspect!
    <metaData>
    <type name="VisualDetailedProjectTimelineWebPart.VisualProjectTimelineWebPart.VisualProjectTimeline, $SharePoint.Project.AssemblyFullName$" />
    <importErrorMessage>$Resources:core,ImportErrorMessage;</importErrorMessage>
    </metaData>
    <data>
    <properties>
    <property name="Title" type="string">VisualProjectTimelineWebPart</property>
    <property name="Description" type="string">My Visual WebPart</property>
    </properties>
    </data>
    </webPart>
    </webParts>
    .... Unless I can solve this I will have to remove the project and recreate but with simple paths. Tho I will be none the wiser if I come across this again.
    Daniel

  • File name too long cannot copy (cont'd)

    This is a continuation of the post started September 01, 2009, with the last post on
    October 17, 2011 named File name too long cannot copy
    Since this is an ongoing unsolved issue, and the thread was locked, it continues here.
    It is ever so easy to create a long file/path using explorer.exe as will be shown below.
    What needs to be solved is, what is an easy way (no, not by listing out all the files in a directory to a text file and counting characters), perhaps using a shell extension, to find out which files in a directory and it's subdirectories will be (or were)
    too long to copy, (and then perhaps even copying those over either intact or renamed)?
    Maflagulator said:
    I'm running the 7100 build...enjoying it except for one big thing:
    While attempting to copy 402gb from my main storage volume onto a spare 500gb drive (for the purpose of changing to a new RAID array) I've come across something that I would expect a Windows 98 OS to give me.
    It tells me that a file has TOO LONG of a file name, then provides with two unhelpful options: SKIP or CANCEL
    I never had XP give me an issue like this at all, so what gives? And while some specific files did have long file names (such as for songs, etc.) it had 7 issues with folders stating that their name was too long, but in fact they were not since they were
    titled '07-06-07' for the date that I dumped the audio files in them. However, they may have contained FILES with long file names though.
    Anyone else get this same situation? Perhaps the RTM version does not do this? Can anyone verify this regarding their install of the RC or the RTM?
    It made it through 400gb out of the 402gb transfer.
    I'm just happy to see that it doesn't spazz out about an issue like this until it has done all the other transfers that it can do because it saves the issues it has with files until the very end. In XP it would spazz about it the moment it came across it
    causing the transfer process to halt.
    Since long path/file names can so easily be created on Win7, it might be useful to see a typical way this happens, which might then give clues how to work with them.
    In Windows Vista, we learnt from:
    File names and file name extensions: frequently asked questions that:
    Windows usually limits file names to 260 characters. But the file name must actually be shorter than that, since the complete path (such as C:\Program Files\filename.txt) is included in this character count.
    In Windows 7, we are told here:
    File names and file name extensions: frequently asked questions that:
    It depends on the length of the complete path to the file (such as C:\Program Files\filename.txt). Windows limits a single path to 260 characters. This is why you might occasionally get an error when copying a file with a very long file name to a location
    that has a longer path than the file's original location.
    From the Windows Dev Center - Desktop, we read about Maximum Path Length Limitation here:
    Naming Files, Paths, and Namespaces
    This helps us understand why a folder can be a maximum of 244 characters, from the defined 260 length of MAX_PATH as follows:
    260 minus C:\ (3) minus <NUL> (1) = 256
    256 minus 8.3 file name (12) = 244
    We also learn there that: The Windows API has many functions that also have Unicode versions to permit an extended-length path for a maximum total path length of 32,767 characters.
    And we read the claim that: The shell and the file system have different requirements. It is possible to create a path with the Windows API that the shell user interface is not be able to interpret properly.
    There is also a comment below this document that reads: In a previous iteration of this document, it is mentioned that The Unicode versions of several functions permit a maximum path length of approximately 32,000 characters composed of components up to
    255 characters in length. This information is now gone.
    So we are in a position where the file system and Windows API can create long path/flies that the shell cannot handle.
    But then we need to be able to handle it, so a little exploration might lead to a better understanding of how to do this.
    For most tasks being performed on long folder/files, Windows 7 and other Windows programs balk when the Path+Filename length > 260
    Let's create a long path/file.
    Create a folder called A at the root of a Drive.
    Create a sub-folder of A called: B
    Create a sub-folder of B called: C
    Make a FILE in sub-folder C called (no spaces or break, one long continuous string): 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T
    123456789U123456789V123456789W123456789X123456.txt
    Rename sub-folder C to the string (no spaces or break, one long continuous string) (The actual directory created will be slightly shorter than this full length): 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J
    123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X123456789Y123456789Z
    Rename sub-folder B to the same full string above. (The actual directory created will be slightly shorter than this full length but 2 characters longer than the step above.)
    Rename folder A to that same full original string. (Again the actual directory created will be slightly shorter than this full length but 2 characters longer than the step above.)
    You now have the lovely file placed at (the breaks are just so it fits into the screen):
    C:\123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X1234\ 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J
    123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X12\ 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T
    123456789U123456789V123456789W123456789X\ 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X123456.txt
    You have a folder length of over 700 and a file length of over 250 for a total of over 950
    However you will notice that each folder, when created, could only be a maximum of 247 charachters including the path (example C:\ , & C:\A , & C:\A\B
    This only applies backwards, that is up the path. It did not matter what was further down the path.
    Now, you can't easily access or rename the file, but you can rename the folders easily.
    For best results, start renaming from the top of the Tree, working down the subfolders, because renaming from down up will limit you, and in fact won't work if the folder lengths are too long.
    So how might knowing this help us?
    Well, to copy this long_file from the C:\ drive to the D:\ drive, and keeping the path structure, this should work:
    Note the name of the top folder. Rename it to something very short, say: A (Make sure C:\A does not exist)
    Note the name of the 2nd folder. Rename it to something very short, say: B (Make sure C:\A\B does not exist)
    Note the name of the 3rd folder. Rename it to something very short, say: C (Make sure C:\A\B\C does not exist)
    Make sure D:\A does not exist - then copy the A folder on disk C: to disk D: (which gives you D:\A\B\C\long_file
    Rename D:\A\B\C to D:\A\B\Original_3rd_Folder_name
    Rename D:\A\B to D:\A\B\Original_2nd_Folder_name
    Rename D:\A to D:\Original_top_Folder_name
    Rename C:\A\B\C back to their original names, in this same reverse order starting with C, then B, then A
    Note: If using Explorer, at some points you might have to press the F5 refresh key.
    This is of course how you might copy such long path/files without using the other more "easy" techniques for the "normal" everyday user like:
    sharing a sub-folder
    using the commandline to assign a drive letter by means of SUBST
    using AddConnectionunder VB to assign a drive letter to a path
    using the "\\?\" prefix to a path string to tell the Windows APIs to disable all string parsing and to send the string that follows it straight to the file system
    and so on.
    See how simple Windows can be for Dummies!
    But then, how can we know that files to be copied exceed this MAX_PATH? Or also after a copy has taken place, know exactly which files that have NOT been copied because of exceeding the MAX_PATH limit, and then a procedure to copy these either by renaming
    them, or by copying them intact as they are?
    There have been suggestions to use
    LongPathTool, but this does not have a facility to check a series of folders and tell you which files are going to be caught by the error when copying. So once a copy has taken place using Windows 7, one does not know which files did not get copied, and
    where exactly they are located.
    Neither does the free
    Old Path Scanner do that. It can only check for overly long directory paths, but misses out when the directory path is within limits, but adding in the file name puts it out of bounds.
    So, as shown above, it is ever so easy to create a long file/path using explorer.exe
    So, what then is an easy way (no, not by listing out all the files in a directory to a text file and counting characters), perhaps using a shell extension, to find out which files in a directory and it's subdirectories will be (or were) too long to copy,
    (and then perhaps even copying those over either intact or renamed)?

    This is a not a "solution" ....but a "work around": a low tech fix....for error message "your file name is too long to be copied, deleted, renamed, moved" :
    1.   problem is this: the "file name" has a limit on number of characters.....the sum of characters really includes the entire path name; you gotta shorten it first (i.e, if the total number of characters in the file name + Path name are over the
    limit, the error appears).  The deeper your file folder sub levels are, the more this problem will come up, especially when you copy to a subfolder of a subfolder/subfolder of another path ...adds to character limit)
    2.  How do you know which combined file names + path names are too long if  you are in the  middle of a copy operation and this error pops up?  Some files copied but the "long files error message" says "skip" or "cancel" ... but not which
    files are the "too long" ones.  If you hit "skip" or "cancel" the "too long" files are left behind...but are mixed in with non-offender "good" "short name" files.   Sorting thru 1000s of "good" files to find a few "bad" ones manually is impractical.
    3.   Here's how you sort out the "bad" from the "good":
    4.    Let's say you want to copy a folder ..."Football" ...that has five layers of subfolders; each subfolder contains numerous files:
      C:/1 Football / 2 teams/ 3 players/ 4 stats/ 5 injuriessidelineplayerstoolong 
           There are five levels root "1 football" with subfolders  2, 3, 4 and lastly "5 injuries"
    5.    Use "cut" and "paste"  (for example to backup all five levels to a new backup folder):
           select "1 football" ....cut....select target folder....paste 
           ("cut" command means as the files are copied to the target destination, the file is deleted from the source location)
          Hint: avoid "cut" and "paste" to a target folder that is itself a sub/sub/sub folder ...that compounds the "characters over the limit" problem ...because the characters in the sub/sub/sub folder are included in the "file name
    character limit"...instead "paste" to a C:/ root directory.
           Suppose in the middle of this operation...error pops up: "5 files have file names that are too long"  Skip or cancel?
           select "skip"  ...and let operation finish
    6.    Now go back and look at the source location: since the software allows only the "good" "short name" files to be copied (and because you "skipped" the "bad" "Long name" files so they are not copied or deleted) ...all that remains
    in the source location are the "bad" "long name files" (because "good" ones were deleted from the source location after the "cut" operation...the bad ones stick out like a sore thumb.
    7.   You will find ....all that remains in source folders are: the "bad" "too long" files; in this example the "bad" file is in level 5:
          C:/ 1 football / 2 teams /3 players /4 stats /5 injuriessidelineplayerstoolong
    8.   select folder 5 injuriessidelineplayerstoolong (that's right...select folder, not file) gotta rename the folder first.
    9.  hit F2 rename folder..folder name highlighted...delete some of the letters in the folder name:
           like this:   5 injuriessidelineplayers  ....you should delete 'toolong'....from the folder name
    10.  then go into folder 5....and do the same operation ...with the too long file name:
            hit F2 rename file....file name hightlighted...delete some of the letters
               like this:  injuriessidelineplayers.....you should delete 'toolong' from the file name
    11.  Now..."cut and paste"  the renamed file to the target backup folder.  
    The Error message will pop up again if you missed any "bad" files....for example, it will indicate "5 files too long" ....then repeat process 5 times until you fix all of them
    12.     Finally, copy the target destination files back to the source location (when you are certain all source location file folder locations are empty) 
    Of course, this "makeshift" solution would not be necessary if MSFT would fix the problem...

Maybe you are looking for

  • Crashes while trying to reload Snow Leopard from the disc.

    I am trying to reload Snow Leopard onto a MacBook Pro. It crashes every time I attempt it. When starting up without the disc, I get a folder with a question mark. What is wrong? When attempting to start up using the Snow Leopard disc, it crashes. Any

  • How to automatically initiate a dial-up connection from Lookout and send an attachment

    Folks: I am trying to send (as an attachment) a spreadsheet file via email. How can I programatically (from Lookout) initiate the default dial-up, send the email (and attachment) and after success (btw, how do I measure this?), ... then close the dia

  • T61 Nvidia internal display vs. dock question

    I've read the threads on the Nvidia issue and I'm grasping at straws that my problem may relate to a different component. The internal screen on my T61 blanked out a few weeks ago but both analog and digital outputs are fine on the dock. The backligh

  • Master Scheduling/MRP

    Is Master Scheduling/MRP included in e-business suite 2003 professional user? Where can I find list of licenses included in e-business suite 2003 professional user?

  • Screen flickers with multi coloured lines... then dies

    I really hope someone out there can help... My iphone is only a couple of months old (bought it first day of release in Australia) and now it is starting to break down. When i press the home key, the screen flickers and has these strange multi colour