Internal File Access

I was wondering if it a java application being executed from a JAR is allowed to acces files within that same JAR? I mean read-and-write, besides maybe the actual CLASS files being executed.

You can definitely read files that are inside a jar (but since they are inside a jar, they aren't really "files" so you need to treat them as a URL).
As far as I know, there is no way to write to an existing jar. Using the command line, you can update a jar's manifest file but for all I know, this might extract and copy all the files from the existing jar into a new jar, adding the new manifest in the process, then delete the existing jar and rename the new jar.

Similar Messages

  • Log file (access.log) of the internal ITS

    Hello,
    anybody know how to access the logfiles of the internal ITS. In particularly im looking for the log file access.log which you had for the external ITS accessable over the ITS admin page http://<servername>/scripts/wgate/admin/!
    The log file loged all users and the transaction they accessed over the time in the format
    2006/10/21 18:39:16.093, 0 #197349: IP ???.???.???.???, -its_ping
    Thanks in advance,
    Kai Mattern

    hi
    good
    go through these links, i hope these ll help you to solve your problem.
    http://www.hp.com/hpbooks/prentice/chapters/0130280844.pdf
    http://help.sap.com/saphelp_46c/helpdata/en/5d/ca5237943a1e64e10000009b38f8cf/content.htm
    thanks
    mrutyun^

  • Trouble installing Photoshop CC 64 bit windows 8 - internal file signature incorrect

    I had a problem and fixed it, and thought I'd post in case others have the problem. I have a new Win 8 machine.  I subscribed to CC (LR+PS), but couldn't install PS 64 bit. Well, when I installed, I got a message it installed ok but had a problem. Adobe showed a log file I could access, but I couldn't access it, and I didn't want to look at a log file anyway. I was hoping for an answer in English.  I tried to open PS 64 and got this message:  "internal file signature is incorrect".  I googled the message, and the majority opinion was to try and reinstall.  Well, I had already tried that and failed.  Then I got an idea I wish I had had before.  When you get a new PC, first thing to do is update all the drivers, etc, and then go to Windows update.  There were many Windows update.  My PC is HP, so I went into it's support center, and clicked the button to update drivers, etc.  There were a lot of updates, including the bios.  After restarting the computer, the install of PS 64 worked, but then I got the other error I posted today (and fixed).  Hope these posts help some poor souls.

    Hi ArtfulToad,
    Thank you for sharing the information.
    Regards,
    Romit Sinha

  • Very very slow file access iSCSI NSS on SLES11/XEN/OES11

    Hi,
    Like many Novell customers while carrying out a hardware refresh we are moving off traditional Netware 6.5 to OES11 and at the same time virtualising our environment.
    We have new Dell Poweredge 620 serves attached by 10gig iSCSI to Equalogic SAN
    Installed SLES will all patches and updates and XEN and then created OES11 SP2 virtual machines, these connect to NSS volume by iSCSI
    Migrated files from traditional netware server to new hardware and stated testing and ran into very very slow files access times
    A 3.5mb pdf file takes close to 10 minutes to open from local PC with Novell Client installed, same if no client and open via cifs. Opening same file off traditional NW6.5 server takes 3-4 seconds.
    We have had a case open with Novell for almost 2 months but they are unable to resolve.
    To test other options we installed VMWare ESXi on the internal usb flash drive and booted off that, created same OES11 VM and connected to NSS on SAN and same pdf open in seconds.
    The current stack of SLES11/XEN/OES11 is not able to be put into production
    Any ideas where the bottleneck might be? We think is in XEN.
    Thanks

    Originally Posted by idgandrewg
    Waiting for support to tell me what the implications are of this finding and best way to fix
    Hi,
    As also mentioned in the SUSE forums, there is the option of using the Equallogic Hit Kit. One of the tools, next to the great autoconfigure options it has, is the eqltune tool.
    Some of the stuff that I've found important:
    -gro is known read performance killer. Switch if off on the iSCSI interfaces.
    - if possible (meaning you have decent hardware), disable flowcontrol as this generally offers stability but at the cost of performance. If your hardware is decent, this form of traffic control should not be needed.
    -To have multipath work correctly over iSCSI and starting SLES 11 SP1. Make sure kernel routing and arp handling are set correctly (not directly relevant if you only have 1 10 GB link):
    net.ipv4.conf.[iSCSI interfaceX name].arp_ignore = 1
    net.ipv4.conf.[iSCSI interfaceX name].arp_announce = 2
    net.ipv4.conf.[iSCSI interfaceX name].rp_filter = 2
    Test if traffic is actively routed over both iSCSI interfaces:
    ping -I [iSCSI interfaceX name] [iSCSI Group IP EQL]
    -Make sure network buffers etc are adequately set as recommended by Dell (set in /etc/sysctl.conf):
    #NetEyes Increase network buffer sizes for iSCSI
    net.core.rmem_max = 16777216
    net.core.wmem_max = 16777216
    net.ipv4.tcp_rmem = 8192 87380 16777216
    net.ipv4.tcp_wmem = 4096 65536 16777216
    net.core.wmem_default = 262144
    net.core.rmem_default = 262144
    -Settings for the /etc/iscsi/iscsid.conf I'm using:
    node.startup = automatic # <--- review and set according to environment
    node.session.timeo.replacement_timeout = 60
    node.conn[0].timeo.login_timeout = 15
    node.conn[0].timeo.logout_timeout = 15
    node.conn[0].timeo.noop_out_interval = 5
    node.conn[0].timeo.noop_out_timeout = 5
    node.session.err_timeo.abort_timeout = 15
    node.session.err_timeo.lu_reset_timeout = 20 #Default is 30
    node.session.err_timeo.tgt_reset_timeout=20 #Default is 30
    node.session.initial_login_retry_max = 12 # Default is 4
    node.session.cmds_max = 1024 #< --- Default is 128
    node.session.queue_depth = 128 #< --- Default is 32
    node.session.iscsi.InitialR2T = No
    node.session.iscsi.ImmediateData = Yes
    node.session.iscsi.FirstBurstLength = 262144
    node.session.iscsi.MaxBurstLength = 16776192
    node.conn[0].iscsi.MaxRecvDataSegmentLength = 131072 #A lower value improves latency at the cost of higher IO throughput
    discovery.sendtargets.iscsi.MaxRecvDataSegmentLeng th = 32768
    node.session.iscsi.FastAbort = No # < --- default is Yes
    -Have Jumbo frames configured on the iSCSI interfaces & iSCSI switch.
    If you are using multipathd instead of the dm-switch provided with the Equallogic Hit kit, make sure the /etc/multipath.conf holds the optimal settings for the Equallogic arrays.
    Ever since Xen with SLES 11 SP1 we have been seeing strong performing virtual servers. We still use 1GB connections (two 1GB connections for each server, serving upto 180~190Mb/s).
    There could be a difference with the 10GB setup, where multipath is not really needed or used (depending on the scale of your setup). One important thing is that the iSCSI switches are doing their thing correctly. But seeing you've already found better results tuning network parameters on the Xen host, seems to indicate that's ok.
    Cheers,
    Willem

  • Setting up g4 as internal file server

    I work for a small company who happens to have a nearly full file server. I have been assigned the task to set up a Dual 867 PowerPC G4 as our main internal file server. The server would serve only internally, providing a place to store our files where anyone in the building can access the file server.
    One slight problem: I haven't really got a clue how to do this.
    We want the server to contain two (2) 750 GB drives... one mirroring the other automatically.
    I contacted Seagate to find out what the biggest hard drive they have for this particular machine and they said 120 GB. They said the bios on a G4 will not allow a 750GB drive. They told me to go to a company called www.attotech.com and get a controller card/host adapter that will provide 48-bit addressing... they said this would bypass the system bios and allow the machine to house larger drives (possibly up to 750 GB).
    Anyway... if anyone could give me some direction on this project, it would help out immensely. I just need some instruction on how to set this machine up... what i need to do... what hardware i need, etc. From the ground up, basically.
    Please help!!!
    Thanks a million!
    Machine specs:
    -OSx 10.4.8
    -Dual 1.2 GHz PowerPC G4
    -1 MB L3 cache per processor
    -768 MB DDR SDRAM

    Maooley,
    I am surprised you did not get a response from this post.
    I am not sure of the 120GB limit on that G4. I know an earlier version was limited to 80GB, so your info might be accurate. In any event going with a SATA card will definitely improve performance. While you are shopping I would also suggest replacing the internal HD. Drives fail. If this is the original drive then replace it before deploying your new server.
    Once you get all the hardware in place you can mirror the internal drives with Disk Utility (check out the RAID tab).
    You might also consider OS X Server. It's the same core OS but adds a set of services and the tools to managed the server. The 10 client version is $499. The 10 client limits concurrent AFP connections, but all other services are unlimited (SMB, FTP, web, etc)
    Often new admins set up a server and give one login to everyone, sometimes with no password. I strongly recommend each user have their own login name and password. This way you can keep track of who is connected, who is creating file, and you can give dedicated server space to each user.
    Have you considered backup? I sure hope so. Now that you have the company goods in one place you have the opportunity loose it all in one quick trash session. You also have the opportunity to pack it up and take it offsite site so you can recover from operator error, sabotage, flood, fire etc.

  • Manually add hidden Shares to Microsoft UAG File Access - You do not have permission

    Have made little progress adding hidden shares for Home Folder access, followed steps listed here
    http://itcalls.blogspot.ca/2012/10/manual-add-of-shares-to-microsoft-uag.html
    Manually entered the server and share name into "ShareAccessCfg.xml", activated the configuration once the shares showed up in file access console.
    However, our users still cannot access their Home Folder.
    Error displayed internally:
    Message displayed externally:

    We finally figured out what might be the cause, seems our hidden file shares are hosted on Windows Server 2012 R2 which uses SMB 3.02 dialect, looking into the server manager a little deeper we noticed that the shares had SMB encryption enabled which is
    not supported on server 2008 R2 / UAG...

  • Not Working: BlackBerry Link Remote File Access Remove Old BlackBerry from BB ID

    I have: removed my blackberry q10 from blackberry link, wiped my q10, uninstalled and reinstalled blackberry link, have link version 1.1.1.32, blackberry q10 os version 10.1.0.2309 - which whenever I plug in the q10 after removing the device from link, says there is an os update, so I update it, and then it does nothing apart from rebooting the q10, or gives an error saying 'updates not available. Please check internet connection and retry.'.
    Also, I have Logout/login on blackberry link, I know my q10 is activated as I have purchased items from app world, and on the computer the app world site shows that I have a q10 when I sign in. I am using the same BBID and have changed the password on my q10 to see if that made a difference, then signed out and back into link on computer with updated password. I have administrator rights on my computer.
    I can access the q10 from my computer over the same wifi network, and it syncs files fine with wireless sync turned on. It is remote file access not giving the option to connect to my q10. 'BBID *email* isn't associated with a device running Blackberry 10 OS or an error occurred when trying to retrieve your information from the Blackberry ID servers.'
    The problem is that although I have the same blackberry id for link and my q10, no devices show up on the remote file access screen, and I cannot access my files on my computer from my q10, yet I can sync them on my computer over the wifi network. How can I enable Remote File Access?
    A problem could be that I have a Curve 9300 on my BB ID account, so if I could disable this (which I have no idea how to do after a lot of research), this might rectify why link cannot set up remote file access.
    What I haven't done is wiped my q10 and uninstalled all user data from link at the same time, as then I would have to manually reload everything on my q10.
    My q10 shows my computer in the settings\blackberry link\paired devices ,  but I 'cannot view files on *computer*'. Also, link does not show in the file manager. I set up my q10 as a permanent link not a one-off connection.

    No one cares because it's not a z10 (I know because I am having the same problem and no one knows how to fix it and no threads for q10 fix)
    Slàinte.
    Not all things are good.
    Not all things are evil.
    But with an open mind...
    You can see the difference.
    Knowledge is power.

  • Read-only file access from network volume

    I get an read-only file access from network volume problem while sharing a drive from Snow Leopard to a Tiger install. Most of files were opening well, but *.fp7 (FileMaker) and *.xls (Excel) files won't open dealing with a read-only error.
    As descibed in the last post of http://discussions.apple.com/thread.jspa?threadID=1406977 the client have the same share name of the server. Renomming it resolved the error!
    Thanks!

    right then, as it looks like I'm talking to myself....
    I have just wiped clean the Macbook Pro.
    I installed Leopard from scratch, then installed Office 2008.
    Logged back onto the network share, and the read-only error came up again, ONLY in Excel.
    bugger.
    Did the same thing with my Macbook and all is fine.
    Copy the file to the local hard drive, opens ok.
    I then copied the file to another Mac on the network.
    mmmm, opens fine.
    what's the difference....
    mmmm, the machine it opens fine from is running 10.4
    the machine which hosts all the data is running 10.3.9
    could this be the problem.
    Just done a software update check on the 10.3 machine and there are some security updates that need doing.
    Going to run that now and see what happens, otherwise I think the iMac running 10.3.9 is going to need to come up to 10.4 and fingers crossed this will solve it.

  • Sql agent job getting file access denied error

    I'm not sure if this question belongs in this forum. Please move it if you want to.
    Here is my question. I have an ssis package that is running into an error at the file system task trying to move a file. The package is deployed to the catalog and I am running the package using the stored procedure
    [SSISDB].[catalog].[start_execution] @execution_id
    When I execute this stored proc in Management Studio while logged in under a sysadmin, everything works fine. But when I call the same TQL in SQL Agent job, I get a file access denied error. This has something to do with the id that is getting used
    to run the package and I am not sure how to track that down. Any help would be appreciated.
    I've check the windows permission on both the id that is running the SQL Agent and SQL SSIS Service. Both seem to have the right windows permission.

    Please see:
    http://support.microsoft.com/kb/918760

  • MaxDB KNLDIAG file access denied

    Hi,
    I'm a DBA recently assigned to support MaxDB database (ver 7.6) used by
    SAP Content Server in our windows 2003/32 bit environments.
    Initially, our windows support group granted the DBA group access to
    the files on E:\sapdb\data\wrk\CD1 where the KNLDIAG and other files
    are allocated. We were able to view the KNLDIAG file when errors
    occurred.
    However, when the database instance is placed offine and back online,
    we loose access to the KNLDIAG file, but not the other files on the
    same directory/subdirectory.
    Can you explain what is happening?
    Is MaxDB doing something behind the scenes?
    What can be done so that we don't loose access to this file?
    I would appreciate any help you can provide.
    Regards,
    Bill

    Hi Bill!
    > I'm a DBA recently assigned to support MaxDB database (ver 7.6) used by
    > SAP Content Server in our windows 2003/32 bit environments.
    > Initially, our windows support group granted the DBA group access to
    > the files on E:\sapdb\data\wrk\CD1 where the KNLDIAG and other files
    > are allocated. We were able to view the KNLDIAG file when errors
    > occurred.
    > However, when the database instance is placed offine and back online,
    > we loose access to the KNLDIAG file, but not the other files on the
    > same directory/subdirectory.
    >
    > Can you explain what is happening?
    Yes, no problem!
    > Is MaxDB doing something behind the scenes?
    Of course not - it's well documented
    Unlike Oracle MaxDB does not use the same file all the time to write out the error messages.
    Instead, with every restart the last file 'KLNDIAG' (and/or 'KNLMSG'  in more recent versions) is renamed to 'KNLDIAG.OLD'. If there is already an 'KNLDIAG.OLD' this file is gone afterwards.
    Then a new file is 'KNLDIAG' is created by the MaxDB Kernel with the default permission of the folder.
    Usually these are dictated by the Windows account that runs the MaxDB-Service.
    If you did not change this, this would be the "local systemaccount".
    On my test system this results in default permissions that allow members of "Adminstrators", "Backup Operators", "System" and of course the "Owner/Creator" to access the file.
    > What can be done so that we don't loose access to this file?
    Simple: if you want to access it via OS tools, make sure that the user is a member of the "Administrators" or the "Backup Operators" group.
    If you don't want to do that, just use the file access via the DBMGUI or DB Studio (or DBMCLI) to get access to the file.
    regards,
    Lars
    p.s.
    Maybe you didn't notice that by now - the KNLDIAG file is not simply written at the end of the file, but consists of a startup header part and a body which is cyclically overwritten.
    For any newbies I cannot overstate the importance of reading the documentation [http://maxdb.sap.com] and/or get a training for it (SAP course ADM 515 really pays of here...)

  • Server 2008r2 and windows 7 pro slow to no networkshare file searching and file access (green progress bar)

    we are running a windows server 2008r2 sp1 domain controller and a 3com gigabit switch to which 7 windows 7 pro computers are connected. We are only using the windows server to distribute a few printers and as a network share (z drive).
    we are storing all files (word documents only) centrally on the server (z share), and files are accessed and written to the z share.
    Problem is that file access from the z share on the windows 7 client computers is at best buggy. we are able to browse the z-share, but as soon as we want to search the z-share for files (f.i. use file explorer with "content: memorandum") the green
    progress bar shows and no search is performed. something like alphabetically sorting a directory on map/file name on the z-share also shows a green bar with no sorting. manually going to the files works.
    One thing i noticed, sometimes not all files that one user created and stored on the server is visible and can be found by another user. I prefer to disable all offline file caching setting, because that feature is not used. no files should be made accessible
    offline. i don't know where to change that on the server side. 
    I am reluctant to experiment on the server end (2008r2 sp1), so i tried a few suggested solutions on the client end (windows 7 client)
    I tried disabling offline file caching
    i tried properties (z share) ==> advanced and map optimizing and several options there
    i tried disabling the smbv2 protocol and disabling the smbv1 protocol (using
    sc.exe config lanmanworkstation depend= bowser/mrxsmb10/nsi command etc. posted by microsoft)
    i tried disabling windows defender
    i tried accessing the z share via ip and server name (the share is mapped via its servername, e.g. dennis-dc1\share). still the same
    file searching on the clients c drive ('own harddrive') works much better, file searching on the server also works.
    any suggestions or solutions that i can test on the client side, as mentioned i am reluctant to change something on the server side. if a solutions works, i prefer to change it on the server side.

    Hi,
    The issue could be due to windows indexing service taking long time in sorting and searching in mapped drives. Please try the steps belwo to disable windows indexing service or search service on the windows 7 client to resolve the issue.
    You can click on Start and select Control Panel, click on
    Programs and Features, go into the Turn Windows Features on or off section Scroll down the list and uncheck the box next to Indexing Service or Windows Search.
    Best Regards,
    Mandy
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact [email protected]

  • Problem with file access in other computer in jsp

    I have problem with file accessing in other computer in jsp.
    The follow code
    File folder=new File("Z:"+File.separator+"sharefolder");//Z is a net share driver
    File[] files=folder.listFiles();
    System.out.println("test");
    System.out.println("length="+files.length);
    will throw exception at the second print.
    but it works well in main funtion.
    Is anybody know what is the problem.
    JSP works on windows2003 server,tomcat 5.0.28 JDK1.4 net share folder on windows2000 server

    no error code for this.But when I start tomcat I get the follow error.
    java.lang.IllegalArgumentException: Document base Z:\ does not exist or is not a readable directory
         at org.apache.naming.resources.FileDirContext.setDocBase(FileDirContext.java:138)
         at org.apache.catalina.core.StandardContext.resourcesStart(StandardContext.java:3910)
         at org.apache.catalina.core.StandardContext.start(StandardContext.java:4138)
         at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:823)
         at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:807)
         at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:595)
         at org.apache.catalina.core.StandardHostDeployer.addChild(StandardHostDeployer.java:903)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at org.apache.commons.beanutils.MethodUtils.invokeMethod(MethodUtils.java:216)
         at org.apache.commons.digester.SetNextRule.end(SetNextRule.java:256)
         at org.apache.commons.digester.Rule.end(Rule.java:276)
         at org.apache.commons.digester.Digester.endElement(Digester.java:1058)
         at org.apache.catalina.util.CatalinaDigester.endElement(CatalinaDigester.java:76)
         at org.apache.xerces.parsers.AbstractSAXParser.endElement(Unknown Source)
         at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanEndElement(Unknown Source)
         at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown Source)
         at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown Source)
         at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
         at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
         at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
         at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source)
         at org.apache.commons.digester.Digester.parse(Digester.java:1567)
         at org.apache.catalina.core.StandardHostDeployer.install(StandardHostDeployer.java:488)
         at org.apache.catalina.core.StandardHost.install(StandardHost.java:863)
         at org.apache.catalina.startup.HostConfig.deployDescriptors(HostConfig.java:483)
         at org.apache.catalina.startup.HostConfig.deployApps(HostConfig.java:427)
         at org.apache.catalina.startup.HostConfig.start(HostConfig.java:983)
         at org.apache.catalina.startup.HostConfig.lifecycleEvent(HostConfig.java:349)
         at org.apache.catalina.util.LifecycleSupport.fireLifecycleEvent(LifecycleSupport.java:119)
         at org.apache.catalina.core.ContainerBase.start(ContainerBase.java:1091)
         at org.apache.catalina.core.StandardHost.start(StandardHost.java:789)
         at org.apache.catalina.core.ContainerBase.start(ContainerBase.java:1083)
         at org.apache.catalina.core.StandardEngine.start(StandardEngine.java:478)
         at org.apache.catalina.core.StandardService.start(StandardService.java:480)
         at org.apache.catalina.core.StandardServer.start(StandardServer.java:2313)
         at org.apache.catalina.startup.Catalina.start(Catalina.java:556)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:287)
         at org.apache.catalina.startup.Bootstrap.main(Bootstrap.java:425)
    Jun 5, 2006 6:55:41 PM org.apache.catalina.core.StandardContext start
    SEVERE: Error in resourceStart()
    Jun 5, 2006 6:55:41 PM org.apache.catalina.core.StandardContext start
    SEVERE: Error getConfigured
    Jun 5, 2006 6:55:41 PM org.apache.catalina.core.StandardContext start
    SEVERE: Context startup failed due to previous errors
    Jun 5, 2006 6:55:41 PM org.apache.catalina.core.StandardContext start
    SEVERE: Exception during cleanup after start failed
    LifecycleException: Container StandardContext[msgstore] has not been started
         at org.apache.catalina.core.StandardContext.stop(StandardContext.java:4466)
         at org.apache.catalina.core.StandardContext.start(StandardContext.java:4371)
         at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:823)
         at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:807)
         at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:595)
         at org.apache.catalina.core.StandardHostDeployer.addChild(StandardHostDeployer.java:903)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at org.apache.commons.beanutils.MethodUtils.invokeMethod(MethodUtils.java:216)
         at org.apache.commons.digester.SetNextRule.end(SetNextRule.java:256)
         at org.apache.commons.digester.Rule.end(Rule.java:276)
         at org.apache.commons.digester.Digester.endElement(Digester.java:1058)
         at org.apache.catalina.util.CatalinaDigester.endElement(CatalinaDigester.java:76)
         at org.apache.xerces.parsers.AbstractSAXParser.endElement(Unknown Source)
         at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanEndElement(Unknown Source)
         at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown Source)
         at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown Source)
         at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
         at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
         at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
         at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source)
         at org.apache.commons.digester.Digester.parse(Digester.java:1567)
         at org.apache.catalina.core.StandardHostDeployer.install(StandardHostDeployer.java:488)
         at org.apache.catalina.core.StandardHost.install(StandardHost.java:863)
         at org.apache.catalina.startup.HostConfig.deployDescriptors(HostConfig.java:483)
         at org.apache.catalina.startup.HostConfig.deployApps(HostConfig.java:427)
         at org.apache.catalina.startup.HostConfig.start(HostConfig.java:983)
         at org.apache.catalina.startup.HostConfig.lifecycleEvent(HostConfig.java:349)
         at org.apache.catalina.util.LifecycleSupport.fireLifecycleEvent(LifecycleSupport.java:119)
         at org.apache.catalina.core.ContainerBase.start(ContainerBase.java:1091)
         at org.apache.catalina.core.StandardHost.start(StandardHost.java:789)
         at org.apache.catalina.core.ContainerBase.start(ContainerBase.java:1083)
         at org.apache.catalina.core.StandardEngine.start(StandardEngine.java:478)
         at org.apache.catalina.core.StandardService.start(StandardService.java:480)
         at org.apache.catalina.core.StandardServer.start(StandardServer.java:2313)
         at org.apache.catalina.startup.Catalina.start(Catalina.java:556)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:585)
         at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:287)

  • File access to /res folder before decompression

    Hi, anyone please assist on how i would access a compressed text file in the /res folder before I can decompress it. I also need to know if its possible to create a file since the decompression algorithm creates a new decompressed file. The following code runs on a computer but I need to use it in the J2ME platform by calling it from the midlet class. Any assistance will be highly appreciated:
    import java.io.*;
    import java.util.*;
    public class Decompress{
    final static int MAX_CODES = 4096;
    final static int BYTE_SIZE = 8;
    final static int EXCESS = 4;
    final static int ALPHA= 256;
    final static int MASK = 15;
    static int [] s;
    static int size;
    static Element [] h;
    static int leftOver;
    static boolean bitsLeftOver;
    static BufferedInputStream in;
    static BufferedOutputStream out;
    private static void setFiles(String[] args) throws IOException{
    String inputFile, outputFile;
    if(args.length >= 1){
    inputFile = args[0];
    if(!inputFile.endsWith(".lzw")){
    System.out.println("The filename must end with \"lzw\" extension");
    System.exit(1);
    in = new BufferedInputStream(new FileInputStream(inputFile));
    outputFile = inputFile.substring(0, inputFile.length()-4);
    out = new BufferedOutputStream(new FileOutputStream(outputFile));
    else{
    System.out.print("usage:java Decompress <filename>");
    System.exit(1);
    private static void output(int code)throws IOException{
    size = -1;
    while(code>=ALPHA){
    s[++size]=h.suffix;
                code = h.prefix;
    s[++size]=code;
    for(int i=size; i>=0; i--)
    out.write(s);
    private static int getCode() throws IOException{
    int c = in.read();
    if(c == -1)return -1;
    int code;
    if(bitsLeftOver)
    code = (leftOver<<BYTE_SIZE)+c;
    else{
    int d = in.read();
    code = (c<<EXCESS)+(d>>EXCESS);
    leftOver = d&MASK;
    bitsLeftOver = !bitsLeftOver;
    return code;
    private static void decompress() throws IOException{
    int codeUsed = ALPHA;
    s = new int[MAX_CODES];
    h = new Element[MAX_CODES];
    int pcode = getCode(), ccode;
    if(pcode>=0){
    s[0] = pcode;
    out.write(s[0]);
    size = 0;
    do{
    ccode = getCode();
    if(ccode<0)break;
    if(ccode<codeUsed){
    output(ccode);
    if(codeUsed<MAX_CODES)
    h[codeUsed++] = new Element(pcode, s[size]);
    else{
    h[codeUsed++] = new Element(pcode, s[size]);
    output(ccode);
    pcode = ccode;
    }while(true);
    out.close();
    in.close();
    public static void main(String [] args) throws IOException{
    setFiles(args);
    decompress();

    if(messageNotWellFormatted("File access to /res folder before decompression"){
      retry();
    }else{
      answerTo(post.getProfile());
    }retry with the formatting tips !!!!!

  • URL file-access is disabled in the server configuration

    My hosting company has made a server config change and my xslt pages have stopped working with the following error.
    It says url file access is disabled but my phpinfo() still shows allow_url_fopen = on
    What else might be the problem?
    Warning: require_once() [
    function.require-once]: URL file-access is disabled in the server configuration in
    /homepages/0/xxx/htdocs/xxxx/inc/horoscope_xml.php on line
    3
    Warning: require_once(
    http://www.xxxx.co.uk/includes/MM_XSLTransform/MM_XSLTransform.class.php) [
    function.require-once]: failed to open stream: no suitable wrapper could be found in
    /homepages/0/xxxx/htdocs/xxxxx/inc/horoscope_xml.php on line
    3
    Fatal error: require_once() [
    function.require]: Failed opening required 'http://www.xxxxx.co.uk/includes/MM_XSLTransform/MM_XSLTransform.class.php' (include_path='.:/usr/lib/php5') in
    /homepages/0/xxxx/htdocs/xxxxx/inc/horoscope_xml.php on line
    3

    bikeman01 wrote:
    I don't understand your use of the word 'class' in your reply - I am using php not asp.net.
    The name of the file that you are trying to access is MM_XSLTransform.class.php. It is a PHP class created by Dreamweaver for the XSL Transformation server behavior.
    The servers php is 5.2.11, as it has been for sometime, so I know that it previously worked with allow_url_fopen = on and allow_url_include = off
    Judging from the error message, you are trying to include the file using a URL, rather than a file path:
    Fatal error: require_once() [
    function.require]: Failed opening required 'http://www.xxxxx.co.uk/includes/MM_XSLTransform/MM_XSLTransform.class.php' (include_path='.:/usr/lib/php5') in
    /homepages/0/xxxx/htdocs/xxxxx/inc/horoscope_xml.php on line
    3
    Change the URL to a file path:
    require_once('/homepages/0/xxxx/htdocs/includes/MM_XSLTransform/MM_XSLTransform.class.php');
    Is it possible for phpinfo() to show allow_url_fopen = on yet be actually off on the server?
    No.
    [Edited to correct name of server behavior]

  • Problem with default file access permission

    Hi,
    I am accessing a common area '/NFS_DATA' by both my java and oracle codes by the users 'javaUsr' and 'oraUsr' respectively.
    As per the requirement, the oracle code (oraUsr) needs to create some file in the specified location and then the java code (javaUsr) needs to update those files (created by oraUsr) with some new data.
    At present scenario the 'oraUsr' creates files with default access permission 644, which does not permit 'javaUsr' to update them.
    Constraints : I am not supposed to set umask at the .profile of 'oraUsr'.
         Execution of any shell script from oracle procedure is not permitted.
    Is it possible to specify file-system specific default file access permissions??
    Any idea to overcome this issue??

    Hi,
    I am accessing a common area '/NFS_DATA' by both my
    java and oracle codes by the users 'javaUsr'
    and 'oraUsr' respectively.
    As per the requirement, the oracle code
    (oraUsr) needs to create some file in the
    specified location and then the java code
    (javaUsr) needs to update those files (created
    by oraUsr) with some new data.
    At present scenario the 'oraUsr' creates files
    with default access permission 644, which does
    not permit 'javaUsr' to update them.
    Constraints : I am not supposed to set
    umask at the .profile of 'oraUsr'.
    Execution of any shell script from oracle
    oracle procedure is not permitted.
    Is it possible to specify file-system specific
    default file access permissions??
    Any idea to overcome this issue??You might like to try using File ACLs
    man setfacl(1)
    as oraUsr
    setfacl -s user:oraUsr:rw-,user:javaUsr:rwx,group::r--,other:---,mask:rwx file
    This way the oraUsr can not execute file
    but javausr can
    getfacl will show the ACL
    user::rw-
    user:javaUsr:rwx #effective:rwx
    group::rw- #effective:rw-
    mask:rwx
    other:---
    hope this helps a bit

Maybe you are looking for