Error writing input to command  11GR2

Hi
My daily backup job is stuck in a running state on the step; Backup with the error mentioned in the title. No matter what I do in terms of trying to stop, suspend or delete the run of this job nothing seems to clear the step so I can resechedule the job. Does anyone have any ideas what I can try?
Thanks

Hi,
Log a SR, it's a bug, but you need to install the emdiag kit and run some SQL to remove the stuck job,
Cheers
M

Similar Messages

  • IO error writing block to file  (Apex 4.01, 11gR2, RH 64 bit & Using Gatew

    I have Apex 4.01 running on a Linux Red Hat 64 bit server and the database is 11gR2. I'm using the gateway and receiving numerous io errors on files that do not exist in the database. Has any one else experienced this? Should I scrape the gateway and load https ?
    File 201 does not exists in the database (error below). I have verified this the files do not exists using ( dba_data_files or dba_temp_files) . Is this a Gateway issue?
    I have run dbverify and there are not corruptions within in the database. We have run a disk check on the SAN and there are not issues either.
    Example of Error:
    ===========
    </msg>
    <msg time='2010-11-09T16:18:25.446-06:00' org_id='oracle' comp_id='rdbms'
    client_id='ADMIN:1753485708135952' type='UNKNOWN' level='16'
    host_id='houlnxora01d.nblenergy.com' host_addr='10.10.200.26' module='APEX:APPLICATION 4500'
    pid='26431'>
    <txt>Errors in file /u01/app/oracle/diag/rdbms/bdapx01d/bdapx01d/trace/bdapx01d_s000_26431.trc:
    ORA-01114: IO error writing block to file 201 (block # 36609)
    ORA-27072: File I/O error
    Linux-x86_64 Error: 9: Bad file descriptor
    Additional information: 4
    Additional information: 36609
    Additional information: -1
    </txt>
    </msg>

    Hello,
    I think this is a database-related problem which also occurs when using APEX, but I don't think it's caused by APEX.
    The error basically says that something is wrong with your filesystem.
    I suggest you post this error description to the forum where the experienced dbas could help you more: {forum:id=61}
    From what I know, these errors may occur when your filesystem is full or the file was deleted/moved or set offline for some other reason.
    You could get a hint for troubleshooting by locating the file on your filesystem:
    select file_name
    from dba_data_files
    where file_id=201;If it actually doesn't exist, something has messed up your data dictionary. You could then try to find out which tablespace and object/s have a hold on that file:
    select
       owner,
       tablespace_name,
       segment_type,
       segment_name
    from
       dba_extents
    where
       file_id = 201
    and
       block_id = 36609;I hope it's just the "disk full"-case.
    -Udo

  • ST22 Short Dump DATASET_WRITE_ERROR Error text: "Input/output error" Error

    Dear All,
    Please kindly help. I have the following error. Is it possibility due to disk full?
    DATASET_WRITE_ERROR
    An error occurred when writing to the file
    "/capps/AP0/edi/CCM/565500_001105200_00126_20080715_042803.tmp".
    Error text: "Input/output error"
    Error code: 5
    Many thanks in advance,
    Yan

    Hi,
    Check the return code of the OPEN DATASET. May be autorizations for writing might not be there.
    If the file is written partially it would be due to memory problem.
    Also check the OPEN DATASET Parameters if they are correct.
    Hope this helps.

  • Lost all photos on my iPhoto and get the ERROR WRITING TO DISK when trying to reload from Time Machine. iPhoto cannot import your photos because there was a problem writing to the volume containing your iPHOTO LIBRARY  . Looking for solution.

    Lost all photos on my iPhoto and get the ERROR WRITING TO DISK when trying to reload from Time Machine. iPhoto cannot import your photos because there was a problem writing to the volume containing your iPHOTO LIBRARY  . Looking for solution.

    Problems such as yours are sometimes caused by files that should belong to you but are locked or have wrong permissions. This procedure will check for such files. It makes no changes and therefore will not, in itself, solve your problem.
    First, empty the Trash, if possible.
    Triple-click anywhere in the line below on this page to select it, then copy the selected text to the Clipboard by pressing the key combination command-C:
    find ~ $TMPDIR.. \( -flags +sappnd,schg,uappnd,uchg -o ! -user $UID -o ! -perm -600 \) 2>&- | wc -l | pbcopy
    Launch the built-in Terminal application in any of the following ways:
    ☞ Enter the first few letters of its name into a Spotlight search. Select it in the results (it should be at the top.)
    ☞ In the Finder, select Go ▹ Utilities from the menu bar, or press the key combination shift-command-U. The application is in the folder that opens.
    ☞ Open LaunchPad. Click Utilities, then Terminal in the icon grid.
    Paste into the Terminal window (command-V). The command may take a noticeable amount of time to run. Wait for a new line ending in a dollar sign ($) to appear.
    The output of the command will be a number. It's automatically copied to the Clipboard. Please paste it into a reply.
    The Terminal window doesn't show the output. Please don't copy anything from there.

  • Error: Error writing stream: Could not find prefix for namespace:

    Hi all,
    I am getting this error while doing topdownAssemble using WSA. The error seems as if the name space definitions do not have proper prefixs, but I have prefixes for all the name spaces.
    Any idea what else might be causing the problem ?
    Thanks,
    --venkat
    + java -jar /opt/oc4j_extended_101310/webservices/lib/wsa.jar -topDownAssemble -wsdl parlayx_third_party_call_service.wsdl -unwrapParameters false -className oracle.sdp.parlayx.ThirdPartyCallImpl -input classes -output gen -ear dist/tpcall.ear -mappingFileName src/ThirdPartyCallService-java-wsdl-mapping.xml -packageName oracle.sdp.parlayx -fetchWsdlImports true -debug
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_third_party_call_service.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_third_party_call_interface.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_common_faults.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_third_party_call_service.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_third_party_call_interface.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_common_faults.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_third_party_call_interface.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_common_faults.wsdl
    FINE: OWS-11015 Trying to Access Resource at URL: file:/home/vayyagar/IdeaProjects/ParlayX/parlayx_common_faults.wsdl
    Error: Error writing stream: Could not find prefix for namespace: http://www.csapi.org/wsdl/parlayx/common/v2_0/faults
    Error writing stream: Could not find prefix for namespace: http://www.csapi.org/wsdl/parlayx/common/v2_0/faults: java.io.IOException: Could not find prefix for namespace: http://www.csapi.org/wsdl/parlayx/common/v2_0/faults
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeAttValFromQName(WSDLWriterImpl.java:660)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeFault(WSDLWriterImpl.java:577)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeOperation(WSDLWriterImpl.java:563)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writePortType(WSDLWriterImpl.java:422)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeDefinition(WSDLWriterImpl.java:186)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeWSDL(WSDLWriterImpl.java:93)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeWSDL(WSDLWriterImpl.java:83)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.writeOutWSDL(WSDLImportLocalizer.java:637)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.writeOutTo(WSDLImportLocalizer.java:568)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.localize(WSDLImportLocalizer.java:278)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.localize(WSDLImportLocalizer.java:162)
    at oracle.j2ee.ws.tools.wsa.Assembler.topDownAssemble(Assembler.java:828)
    at oracle.j2ee.ws.tools.wsa.cli.Processor.execute(Processor.java:320)
    at oracle.j2ee.ws.tools.wsa.cli.Processor.execute(Processor.java:169)
    at oracle.j2ee.ws.tools.wsa.Main.mainNoSystemExit(Main.java:83)
    at oracle.j2ee.ws.tools.wsa.Main.main(Main.java:48)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeWSDL(WSDLWriterImpl.java:95)
    at oracle.j2ee.ws.wsdl.xml.WSDLWriterImpl.writeWSDL(WSDLWriterImpl.java:83)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.writeOutWSDL(WSDLImportLocalizer.java:637)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.writeOutTo(WSDLImportLocalizer.java:568)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.localize(WSDLImportLocalizer.java:278)
    at oracle.j2ee.ws.common.wsdl.WSDLImportLocalizer.localize(WSDLImportLocalizer.java:162)
    at oracle.j2ee.ws.tools.wsa.Assembler.topDownAssemble(Assembler.java:828)
    at oracle.j2ee.ws.tools.wsa.cli.Processor.execute(Processor.java:320)
    at oracle.j2ee.ws.tools.wsa.cli.Processor.execute(Processor.java:169)
    at oracle.j2ee.ws.tools.wsa.Main.mainNoSystemExit(Main.java:83)
    at oracle.j2ee.ws.tools.wsa.Main.main(Main.java:48)

    HI,
    Seems known issue. I would suggest to raise an Service Request with the test case.
    GD,
    Ashish

  • FODC0002 [{bea-err}FODC0002a]: Error parsing input XML: Error at line:2 col

    I have an ODSI Physical Service that is based on a Java Function. The Java Function builds a SQL statement and uses JDBC to query for a ResultSet. One of the columns that is queried is a Clob. Sometimes, the data in this column causes an XMLBeans validation exception in ODSI: {err}XQ0027: Validation failed: error: decimal: Invalid decimal value: unexpected char '114'
    The issue is not consistently replicable with particular database record, the database records that present this issue at one point in time will be resolved after a restart of ODSI and replaced by another list of records that present the same error.
    As can be seen from the stack trace, it looks like the issue is happening after the database query has returned and while the process is assembling the SOAP response.
    Error at line:2 col:481 Line:2 '=' expected, got char[99]
    at weblogic.xml.babel.scanner.ScannerState.expect(ScannerState.java:241)
    at weblogic.xml.babel.scanner.OpenTag.read(OpenTag.java:60)
    at weblogic.xml.babel.scanner.Scanner.startState(Scanner.java:251)
    at weblogic.xml.babel.scanner.Scanner.scan(Scanner.java:178)
    at weblogic.xml.babel.baseparser.BaseParser.accept(BaseParser.java:533)
    at weblogic.xml.babel.baseparser.BaseParser.accept(BaseParser.java:510)
    at weblogic.xml.babel.baseparser.EndElement.parse(EndElement.java:34)
    at weblogic.xml.babel.baseparser.BaseParser.parseElement(BaseParser.java:457)
    at weblogic.xml.babel.baseparser.BaseParser.parseSome(BaseParser.java:326)
    at weblogic.xml.stax.XMLStreamReaderBase.advance(XMLStreamReaderBase.java:195)
    at weblogic.xml.stax.XMLStreamReaderBase.next(XMLStreamReaderBase.java:237)
    at weblogic.xml.stax.XMLEventReaderBase.parseSome(XMLEventReaderBase.java:189)
    at weblogic.xml.stax.XMLEventReaderBase.nextEvent(XMLEventReaderBase.java:122)
    at weblogic.xml.query.parsers.StAXEventAdaptor.queueNextTokens(StAXEventAdaptor.java:136)
    at weblogic.xml.query.parsers.StAXEventAdaptor.queueNextTokens(StAXEventAdaptor.java:124)
    at weblogic.xml.query.parsers.BufferedParser.fetchNext(BufferedParser.java:79)
    at weblogic.xml.query.iterators.GenericIterator.next(GenericIterator.java:104)
    at weblogic.xml.query.runtime.navigation.ChildPath.fetchNext(ChildPath.java:308)
    at weblogic.xml.query.iterators.GenericIterator.hasNext(GenericIterator.java:133)
    at weblogic.xml.query.schema.BestEffortValidatingIterator$OpenedIterator.hasNext(BestEffortValidatingIterator.java:224)
    at weblogic.xml.query.schema.ValidatingIterator.fetchNext(ValidatingIterator.java:82)
    at weblogic.xml.query.iterators.GenericIterator.next(GenericIterator.java:104)
    at weblogic.xml.query.xdbc.iterators.ItemIterator.fetchNext(ItemIterator.java:86)
    at weblogic.xml.query.iterators.LegacyGenericIterator.next(LegacyGenericIterator.java:109)
    at weblogic.xml.query.schema.BestEffortValidatingIterator.fetchNext(BestEffortValidatingIterator.java:85)
    at weblogic.xml.query.iterators.GenericIterator.next(GenericIterator.java:104)
    at weblogic.xml.query.xdbc.iterators.ItemIterator.fetchNext(ItemIterator.java:86)
    at weblogic.xml.query.iterators.LegacyGenericIterator.next(LegacyGenericIterator.java:109)
    at weblogic.xml.query.runtime.typing.SeqTypeMatching.fetchNext(SeqTypeMatching.java:137)
    at weblogic.xml.query.iterators.GenericIterator.next(GenericIterator.java:104)
    at com.bea.dsp.wrappers.jf.JavaFunctionIterator.fetchNext(JavaFunctionIterator.java:273)
    at weblogic.xml.query.iterators.GenericIterator.next(GenericIterator.java:104)
    at weblogic.xml.query.runtime.querycide.QueryAssassin.fetchNext(QueryAssassin.java:54)
    at weblogic.xml.query.iterators.GenericIterator.peekNext(GenericIterator.java:163)
    at weblogic.xml.query.runtime.qname.InsertNamespaces.fetchNext(InsertNamespaces.java:247)
    at weblogic.xml.query.iterators.GenericIterator.next(GenericIterator.java:104)
    at weblogic.xml.query.runtime.core.ExecutionWrapper.fetchNext(ExecutionWrapper.java:88)
    at weblogic.xml.query.iterators.GenericIterator.next(GenericIterator.java:104)
    at weblogic.xml.query.xdbc.iterators.ItemIterator.fetchNext(ItemIterator.java:86)
    at weblogic.xml.query.iterators.LegacyGenericIterator.hasNext(LegacyGenericIterator.java:130)
    at weblogic.xml.query.xdbc.util.Serializer.serializeItems(Serializer.java:251)
    at com.bea.ld.server.ResultPusher$DSP25CompatibilityPusher.next(ResultPusher.java:236)
    at com.bea.ld.server.ResultPusher.pushResults(ResultPusher.java:112)
    at com.bea.ld.server.XQueryInvocation.execute(XQueryInvocation.java:770)
    at com.bea.ld.EJBRequestHandler.invokeQueryInternal(EJBRequestHandler.java:624)
    at com.bea.ld.EJBRequestHandler.invokeOperationInternal(EJBRequestHandler.java:478)
    at com.bea.ld.EJBRequestHandler.invokeOperation(EJBRequestHandler.java:323)
    at com.bea.ld.ServerWrapperBean.invoke(ServerWrapperBean.java:153)
    at com.bea.ld.ServerWrapperBean.invokeOperation(ServerWrapperBean.java:80)
    at com.bea.ld.ServerWrapper_s9smk0_ELOImpl.invokeOperation(ServerWrapper_s9smk0_ELOImpl.java:63)
    at com.bea.dsp.ws.RoutingHandler$PriviledgedRunner.run(RoutingHandler.java:96)
    at com.bea.dsp.ws.RoutingHandler.handleResponse(RoutingHandler.java:217)
    at weblogic.wsee.handler.HandlerIterator.handleResponse(HandlerIterator.java:287)
    at weblogic.wsee.handler.HandlerIterator.handleResponse(HandlerIterator.java:271)
    at weblogic.wsee.ws.dispatch.server.ServerDispatcher.dispatch(ServerDispatcher.java:176)
    at weblogic.wsee.ws.WsSkel.invoke(WsSkel.java:80)
    at weblogic.wsee.server.servlet.SoapProcessor.handlePost(SoapProcessor.java:66)
    at weblogic.wsee.server.servlet.SoapProcessor.process(SoapProcessor.java:44)
    at weblogic.wsee.server.servlet.BaseWSServlet$AuthorizedInvoke.run(BaseWSServlet.java:285)
    at weblogic.wsee.server.servlet.BaseWSServlet.service(BaseWSServlet.java:169)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
    at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:292)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:175)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3498)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(Unknown Source)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2180)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2086)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1406)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    <Apr 29, 2011 12:47:01 PM EDT> <Notice> <ODSI> <BEA-000000> <LabOrderDataServices> <Error occurred performing ODSI operation: {ld:LabOrder/logical/LabOrderReport}getLabOrderDetails:1
    weblogic.xml.query.exceptions.XQueryDynamicException: ld:LabOrder/logical/LabOrderReport.ds, line 34, column 6: {err}FODC0002 [{bea-err}FODC0002a]: Error parsing input XML: Error at line:2 col:481 Line:2 '=' expected, got char[99]
    at weblogic.xml.query.iterators.AbstractIterator.reportUserError(AbstractIterator.java:95)
    at weblogic.xml.query.iterators.AbstractIterator.reportUserError(AbstractIterator.java:147)
    at weblogic.xml.query.parsers.Parser.reportParseError(Parser.java:157)
    at weblogic.xml.query.parsers.StAXEventAdaptor.queueNextTokens(StAXEventAdaptor.java:225)
    at weblogic.xml.query.parsers.StAXEventAdaptor.queueNextTokens(StAXEventAdaptor.java:124)
    Truncated. see log file for complete stacktrace
    javax.xml.stream.XMLStreamException: Error at line:2 col:481 Line:2 '=' expected, got char[99]
    at weblogic.xml.stax.XMLStreamReaderBase.advance(XMLStreamReaderBase.java:206)
    at weblogic.xml.stax.XMLStreamReaderBase.next(XMLStreamReaderBase.java:237)
    at weblogic.xml.stax.XMLEventReaderBase.parseSome(XMLEventReaderBase.java:189)
    at weblogic.xml.stax.XMLEventReaderBase.nextEvent(XMLEventReaderBase.java:122)
    at weblogic.xml.query.parsers.StAXEventAdaptor.queueNextTokens(StAXEventAdaptor.java:136)
    Truncated. see log file for complete stacktrace
    Error at line:2 col:481 Line:2 '=' expected, got char[99]
    at weblogic.xml.babel.scanner.ScannerState.expect(ScannerState.java:241)
    at weblogic.xml.babel.scanner.OpenTag.read(OpenTag.java:60)
    at weblogic.xml.babel.scanner.Scanner.startState(Scanner.java:251)
    at weblogic.xml.babel.scanner.Scanner.scan(Scanner.java:178)
    at weblogic.xml.babel.baseparser.BaseParser.accept(BaseParser.java:533)
    Truncated. see log file for complete stacktrace
    >
    Can somebody shed some light on this issue?
    Thanks
    Edited by: user738507 on May 6, 2011 7:21 AM

    Here is the java function:
         * Iterate through the search results and build out the XmlBean response
         * @param helper A helper class used to simplify common JDBC commands
         * @param doc The XmlBean document to populate
         * @param isCollectionsIncluded True if Collection info should be included in results, False otherwise
         * @param isFullDetailsIncluded True if Result data should be included in results, False otherwise
         * @throws Exception
         private static void addOrders(XmlBeansJDBCHelper helper, LabOrderReportDocument doc,
                   boolean isCollectionsIncluded, boolean isFullDetailsIncluded) throws Exception {
              int rows = 0;
              ResultSet rs = helper.getResultSet();
              LabOrders labOrders = doc.getLabOrderReport().addNewLabOrders();
              LabOrder record = null;
              HashMap<Long, Collection> parentCollectionMap = null;
              // initialize variable used to track when child elements of the XML should be created
              long previousRowOrderId = 0;
              long previousRowParentOrderCollectionId = 0;
              long previousRowOrderCollectionId = 0;
              long previousRowResultId = 0;
              boolean isRootCollectionNode = false;
              LabOrder.Collections lastParentOuterCollectionsAdded = null;
              com.idexx.services.lde.laborder.Collection.Collections lastParentInnerCollectionsAdded = null;
              com.idexx.services.lde.laborder.Collection lastCollectionAdded = null;
              Result lastResultAdded = null;
              // Loop through the results and build XmlBean nodes for each row
              // Since the SQL is joining Orders to Collections (one-to-many) to Results (one-to-many),
              // and returning a flat structure, there will be duplicate Order data on each row when
              // multiple collections exist on the Order, and duplicate Collection data when multiple
              // Results exist. We can use this fact to determine when to create a new Collection, or
              // Result node.
              while (helper.getResultSet().next())
                   rows++;
                   long currentRowParentOrderCollectionId = 0;
                   long currentRowOrderCollectionId = 0;
                   long currentRowResultId = 0;
                   long currentRowResultRemarkId = 0;
                   //int rowno = helper.getResultSet().getRow();
                   // Get the Order ID
                   logDebug("Getting the OrderId.....");
                   BigInteger dbOrderId = JDBCHelper.getBigInteger(rs, DataConstants.ORDER_ID);
                   logDebug("DONE getting the OrderId.");
                   long currentRowOrderId = dbOrderId.longValue();
                   // Determine the Order ID, Order Collection ID, and Result ID currently being processed.
                   // These will be used to determine whether to start a new LabOrder Bean, Collections Bean, or Results Bean
                   if (isCollectionsIncluded || isFullDetailsIncluded) {
                        // Get the ParentOrderCollectionID
                        logDebug("Getting the Parent Collection Order ID.....");
                        BigInteger dbParentOrderCollectionId = JDBCHelper.getBigInteger(rs, DataConstants.PARENT_ORDER_COLLECTION_ID);
                        if ( dbParentOrderCollectionId != null )
                             currentRowParentOrderCollectionId = dbParentOrderCollectionId.longValue();
                        else
                             currentRowParentOrderCollectionId = 0;
                        // Get the OrderCollectionID
                        logDebug("Getting the Order Collection ID.....");
                        BigInteger dbOrderCollectionId = JDBCHelper.getBigInteger(rs, DataConstants.ORDER_COLLECTION_ID);
                        if ( dbOrderCollectionId != null )
                             currentRowOrderCollectionId = dbOrderCollectionId.longValue();
                        else
                             currentRowOrderCollectionId = 0;
                        if ( isFullDetailsIncluded ) {
                             // Get the ResultID
                             logDebug("Getting the Result Id.....");
                             BigInteger dbResultId = JDBCHelper.getBigInteger(rs, DataConstants.RESULT_ID);
                             if ( dbResultId != null )
                                  currentRowResultId = dbResultId.longValue();
                             else
                                  currentRowResultId = 0;
                             // Get the ResultRemarkID
                             BigInteger dbResultRemarkId = JDBCHelper.getBigInteger(rs, DataConstants.RESULT_REMARK_ID);
                             if ( dbResultRemarkId != null )
                                  currentRowResultRemarkId = dbResultRemarkId.longValue();
                             else
                                  currentRowResultRemarkId = 0;
                   isRootCollectionNode = (currentRowParentOrderCollectionId == 0);
                   logDebug("currentRowOrderId: " + currentRowOrderId);
                   logDebug("previousRowOrderId: " + previousRowOrderId);
                   logDebug("currentRowResultId: " + currentRowResultId);
                   logDebug("previousRowResultId: " + previousRowResultId);
                   logDebug("currentRowResultRemarkId: " + currentRowResultRemarkId);
                   logDebug("previousRowResultRemarkId: N/A");
                   logDebug("currentRowParentOrderCollectionId: " + currentRowParentOrderCollectionId);
                   logDebug("previousRowParentOrderCollectionId: " + previousRowParentOrderCollectionId);
                   logDebug("currentRowOrderCollectionId: " + currentRowOrderCollectionId);
                   logDebug("previousRowOrderCollectionId: " + previousRowOrderCollectionId);
                   if ( currentRowOrderId != previousRowOrderId ) {
                        parentCollectionMap = new HashMap<Long, Collection>();
                        lastParentOuterCollectionsAdded = null;
                        lastParentInnerCollectionsAdded = null;
                        lastCollectionAdded = null;
                        lastResultAdded = null;
                        // This is a new Order, generate a new Lab Order bean
                        record = addOrder(labOrders, helper, dbOrderId, isFullDetailsIncluded);
                        logDebug("Order Added!");
                        // If there is Parent Collection data and it should be included, build a Collections element,
                        // and populate the first one
                        if ( !isRootCollectionNode && (isCollectionsIncluded || isFullDetailsIncluded) ) {
                             lastParentOuterCollectionsAdded = record.addNewCollections();
                             lastCollectionAdded = addCollection(record, helper, lastParentOuterCollectionsAdded, true);
                             logDebug("Collection Added! Is it null? " + (lastCollectionAdded == null));
                        // If there is Collection data and it should be included, build a Collections element,
                        // and populate the first one
                        if ( currentRowOrderCollectionId > 0 && (isCollectionsIncluded || isFullDetailsIncluded) ) {
                             if ( isRootCollectionNode ) {
                                  lastParentOuterCollectionsAdded = record.addNewCollections();
                                  lastCollectionAdded = addCollection(record, helper, lastParentOuterCollectionsAdded, false);
                                  parentCollectionMap.put(new Long(currentRowOrderCollectionId), lastCollectionAdded);
                                  logDebug("parent collection added to map: " + currentRowOrderCollectionId);
                             else {
                                  lastParentInnerCollectionsAdded = lastCollectionAdded.addNewCollections();
                                  lastCollectionAdded = addCollection(record, helper, lastParentInnerCollectionsAdded, false);
                             logDebug("Collection Added! Is it null? " + (lastCollectionAdded == null));
                             // If there is Result data and it should be included, build a Results element,
                             // and populate the first one
                             if ( currentRowResultId > 0 && isFullDetailsIncluded ) {
                                  logDebug("Adding result....");
                                  lastResultAdded = addResult(record, helper, lastCollectionAdded);
                                  logDebug("Result Added!");
                                  // If there is Result Remark data and it should be included, build a ResultRemarks element,
                                  // and populate the first one
                                  if ( currentRowResultRemarkId > 0 && isFullDetailsIncluded ) {
                                       addResultRemark(record, helper, lastResultAdded);
                        logDebug("DONE getting first Collection and Result.");
                   else if ( currentRowParentOrderCollectionId != previousRowParentOrderCollectionId
                             && (isCollectionsIncluded || isFullDetailsIncluded) ) {
                        // This is a new, top level, Order Collection to be included
                        lastParentOuterCollectionsAdded = null;
                        lastParentInnerCollectionsAdded = null;
                        lastCollectionAdded = null;
                        lastResultAdded = null;
                        logDebug("Getting next Order Collection...");
                        // If there is Parent Collection data and it should be included, build a Collections element,
                        // and populate the first one
                        if ( !isRootCollectionNode ) {
                             lastCollectionAdded = (com.idexx.services.lde.laborder.Collection)parentCollectionMap.get(new Long(currentRowParentOrderCollectionId));
                             logDebug("A Collection Added! Is it null? " + (lastCollectionAdded == null));
                        // If there is Collection data and it should be included, build a Collections element,
                        // and populate the first one
                        if ( currentRowOrderCollectionId > 0 ) {
                             if ( isRootCollectionNode ) {
                                  //LabOrder.Collections collections = record.addNewCollections();
                                  lastParentOuterCollectionsAdded = record.getCollections();
                                  lastCollectionAdded = addCollection(record, helper, lastParentOuterCollectionsAdded, false);
                                  parentCollectionMap.put(new Long(currentRowOrderCollectionId), lastCollectionAdded);
                             else {
                                  lastParentInnerCollectionsAdded = lastCollectionAdded.addNewCollections();
                                  lastCollectionAdded = addCollection(record, helper, lastParentInnerCollectionsAdded, false);
                             logDebug("B Collection Added! Is it null? " + (lastCollectionAdded == null));
                             // If there is Result data and it should be included, build a Results element,
                             // and populate the first one
                             if ( currentRowResultId > 0 && isFullDetailsIncluded ) {
                                  lastResultAdded = addResult(record, helper, lastCollectionAdded);
                                  // If there is Result Remark data and it should be included, build a ResultRemarks element,
                                  // and populate the first one
                                  if ( currentRowResultRemarkId > 0 && isFullDetailsIncluded ) {
                                       addResultRemark(record, helper, lastResultAdded);
                   else if ( currentRowOrderCollectionId != previousRowOrderCollectionId
                             && (isCollectionsIncluded || isFullDetailsIncluded) ) {
                        // This is a new Order Collection to be included inside of a parent collection
                        logDebug("Getting next CHILD Order Collection...");
                        logDebug("isRootCollectionNode: " + isRootCollectionNode);
                        logDebug("Order ID: " + helper.getBigInteger(DataConstants.ORDER_ID));
                        logDebug("Order Collection ID: " + helper.getBigInteger(DataConstants.ORDER_COLLECTION_ID));
                        logDebug("Collection ID: " + helper.getBigInteger(DataConstants.COLLECTION_ID));
                        if ( isRootCollectionNode ) {
                             lastCollectionAdded = addCollection(record, helper, lastParentOuterCollectionsAdded, false);
                             parentCollectionMap.put(new Long(currentRowOrderCollectionId), lastCollectionAdded);
                        else {
                             com.idexx.services.lde.laborder.Collection parentCollection = (com.idexx.services.lde.laborder.Collection)parentCollectionMap.get(new Long(currentRowParentOrderCollectionId));
                             if(parentCollection == null) {
                                  log(LOG_LEVEL.WARN, "Parent Collection with id: " + currentRowParentOrderCollectionId + " is null for collection id: " + currentRowOrderCollectionId + " but isRootCollectionNode is " + isRootCollectionNode);
                             } else {
                                  lastParentInnerCollectionsAdded = parentCollection.getCollections();
                                  logDebug("Is lastParentInnerCollectionsAdded null? " + (lastParentInnerCollectionsAdded == null));
                                  lastCollectionAdded = addCollection(record, helper, lastParentInnerCollectionsAdded, false);
                        // If there is Result data and it should be included, build a Results element,
                        // and populate the first one
                        if ( currentRowResultId > 0 && isFullDetailsIncluded ) {
                             lastResultAdded = addResult(record, helper, lastCollectionAdded);
                             // If there is Result Remark data and it should be included, build a ResultRemarks element,
                             // and populate the first one
                             if ( currentRowResultRemarkId > 0 && isFullDetailsIncluded ) {
                                  addResultRemark(record, helper, lastResultAdded);
                   else if ( currentRowResultId != previousRowResultId
                             && isFullDetailsIncluded ) {
                        // There is a new Result to be included
                        logDebug("Getting next Result...");
                        // This is a new result to be included
                        lastResultAdded = addResult(record, helper, lastCollectionAdded);
                        // If there is Result Remark data and it should be included, build a ResultRemarks element,
                        // and populate the first one
                        if ( currentRowResultRemarkId > 0 && isFullDetailsIncluded ) {
                             addResultRemark(record, helper, lastResultAdded);
                   else if ( isFullDetailsIncluded ) {
                        // There is a new Result Remark to include
                        logDebug("Getting next Result Remark...");
                        // This is a new result remark to be included
                        addResultRemark(record, helper, lastResultAdded);
                   logDebug("Done building response.");
                   previousRowResultId = currentRowResultId;
                   previousRowParentOrderCollectionId = currentRowParentOrderCollectionId;
                   previousRowOrderCollectionId = currentRowOrderCollectionId;
                   previousRowOrderId = currentRowOrderId;
              logDebug("Found " + rows + " rows of data.");
         }

  • Why is Premiere 6 giving a "Error writing file (disk full?)" message?

    Hello helpful experts,
    I'm using Premiere 6 on a Windows 7 machine.
    I have a file imported in to the timeline and it is set to play in reverse via the speed setting being set to -100 and it plays back in the preview window correctly, in reverse, when I click the play button.
    But when I do a  "File>Export Timeline>Movie" command I get an error message saying "An error occurred while making the movie. Error writing file (disk full?)"... but my hard drive is not nearly full, with over 30 gigs left open on it... and my RAM usage gadget shows over half of my RAM capacity is unused.
    Now previously, weeks ago on a different project, I was getting this same error message for a while when I would hit enter to preview my movie on the timeline (but that worked itself out somehow and eventually stopped giving me that error message).  But when I went to save and export, then I would not get the error message... now, it is just the opposite... I'm now getting the error message when trying to do an export and save the movie.
    And also now when I hit enter to preview, as opposed to just clicking the play button on the preview window, Premiere is going through the process of saving the movie since it is saving a temporary preview file... hopefully, since I'm getting this error message, I can find the temporary preview file and save it as my exported AVI file.
    But I still shouldn't be getting this error message when trying to save, or when trying to preview, as I did for a while a couple of weeks ago.
    Does anyone know about the erroneous error message that says, "An error occurred while making the movie. Error writing file (disk full?)"
    Thanks,
    numetro

    Depending on your hard drive size, P6 may be confused... since drives were not as large when it was written
    For Win7 64bit Pro or Ultimate (or Enterprise?) this is ONE example of using Virtual XP http://forums.adobe.com/thread/702693?tstart=0
    This is saved from an OLD message that may help
    File size limits by disk format type
    Fat = 2 Gig maximum file size
    Fat32 = 4 Gig
    NTFS = ?Terrabytes?
    Message saved from Premiere forum
    John Acocks - 07:52am Apr 17, 2003 Pacific
    I am unable to complete captures because I get a message
    that says "capture stopped because disk was full." What
    does this mean?
    Daniel Conklin - 08:02am Apr 17, 2003 Pacific (#1 of 2)   
    If you are using a Pinnacle card, a possible solution to
    this is to right click on the recycle bin, and set the
    percentage on your media drive to less than 2%.
    There a several other possible causes as well, but you
    need to supply some info about your system. How are you
    capturing and what format is your video? How big is
    your hard drive?
    Jeff Schell - 09:57am Apr 17, 2003 Pacific (#2 of 2)   
    this is one of my favorite error messages. Basically,
    you've chosen an incorrect setting in your DV settings.
    (Odds are you have DV settings). If so, read on:
    1. The DV format is very rigid. Meaning that if you
       make changes to the format, it cannot export as
       requested, because your movie file must match every
       single spec exactly.
    2. When Premiere goes to export a project, the DV codec
       says "Whoa. I can't do anything, because the settings
       don't match."
    3. Then Premiere comes back to you and has to report an
       error. (Don't shoot the messenger). Unfortunately,
       Premiere doesn't know why it wasn't able to export a
       movie. So it says to you, "Um... I can't make the movie.
       Um... maybe your hard drives are full?"
    It's a complete guess. And has nothing to do with your harddrives.
    Check your project settings and/or your export settings. Make
    sure you have chosen one of the "presets" that will specify
    every correct setting for you. Make sure it is one of the DV
    presets, if you are using DV video, or make sure it is one of
    the presets supplied by your capture card manufacturer (in this
    case I'm going to guess Pinnacle or Matrox.)
    To choose a preset, go to Project > Settings Viewer, and then
    click on the project settings column, and then click 'load' to
    load a correct preset. Do the same with export settings.

  • Error writing output with iview for xml form builder

    Hi,
    I created a Xml Form Builder's project in which I developed an "Edit" and "ListEdit" sheet.
    I also created an iview for theese in which the code link is:"com.sap.km.cm.xmlform",and the field for Style Sheet for List and for single item are set up correctly,but when I tried the preview the following error message happened:
    com.inqmy.lib.xsl.xslt.XSLOutputException: Error writing output. -> org.w3c.dom.DOMException: Root Element is already present, cannot be appended as a child.
    could someone help me?
    thank's a lot!
    Nick.

    Hi,
    Now I'm confused,what do you mean with create new data? Are users editing existing documents
    (as if they go to a document example.xml > edit) or they are creating new documents (as if they go
    on folder > new > forms)?
    The problem is, in both cases a user would need read/write permissions.
    The normal flow content (data) is created in KM is as follows:
    1. user is assigned to a role
    2. role contains KM Navigation iView
    3. KM Navigation iView executes com.sap.km.cm.navigation component
    4. user chooses New > Form UI command (edit_xml_forms)
    5. edit_xml_forms UI command executes its code (com.sapportals.wcm.rendering.uicommand.cm.UIXMLFormsCreateCommand) and open xml edit for the user
    6. user fill the form and click Save, form is created into folder
    For what I understood so far, your requirement basically asks you to go directly to step 5, it is
    possible to pass a URL that goes directly to step 5, the UI command button, but if you do that
    you won't have a context, so chances that it will work are slim, since a context is required to
    fill the parameters asked by the app (like folder, user, permissions, etc). Even though, in some cases you can
    still pass the parameters via post in the URL but you must know which service/parameters the
    app asks for it, also a URL is static...
    That was the create scenario, I think it's more cons than pros, users would still be
    able to bypass the URL iView created for that, I'd suggest evaluating again if it's really
    a problem having users access cm to manage data
    kind regards,
    Rafael

  • Error writing the handshake string to the newly established connection. (02

    Hi
    I am having problems connecting from my Master to the Local Distributor server both on Solaris 10
    I am using ssh with ssl and select encryption, no authentication on the Master
    I was able to ssh from master to LD with prompt
    I check the path it was correct
    When I installed the LD I specified ssh
    I turn on the Debug and got the following results
    Debug from SPS Viewing Host Information
    Trace route to 10.233.12.38:70001 failed. (022167)
    Error writing the handshake string to the newly established connection. (022181)
    Connection handshake failed, invalid handshake string. Ensure that the path to the N1 Service Provisioning System application is correct, the application is configured to accept ssh connections and that you can ssh to the machine without any prompts.
    [debug1: channel 0: obuf empty, debug1: channel 0: close_write, debug1: channel 0: output drain -> closed, debug1: channel 0: almost dead, debug1: channel 0: gc: notify user, debug1: channel 0: gc: user detached, debug1: channel 0: send close, debug1: channel 0: is dead, debug1: channel 0: garbage collecting, debug1: channel_free: channel 0: client-session, nchannels 1, debug3: channel_free: status: The following connections are open:, #0 client-session (t4 r0 i3/0 o3/0 fd -1/-1), , debug3: channel_close_fds: channel 0: r -1 w -1 e 7, debug1: fd 0 clearing O_NONBLOCK, debug1: fd 1 clearing O_NONBLOCK, debug1: fd 2 clearing O_NONBLOCK, debug1: Transferred: stdin 0, stdout 0, stderr 0 bytes in 0.1 seconds, debug1: Bytes per second: stdin 0.0, stdout 0.0, stderr 0.0, debug1: Exit status 1](022138)
    Debug from cr_server.out
    2007-07-12 15:49:57,342 ERROR [Connection [127.0.0.1305]-[10.233.12 .38:70001]:Reader] com.raplix.rolloutexpress.net.transport.Connection$Rea
    dThread (Connection.java:515) - Exception when reading from connection in
    put:Connection [127.0.0.1:98305]-[10.233.0.38:70001]:Closing connecti
    on:
    Unable to make connection handshake using data, "0". (022195)
    Ok what am I mssing

    Hi,
    I am not quite sure where exactly you get an error. Are you trying to prepare an agent that connects to your newly installed LD?
    You did mention the following:
    I was able to ssh from master to LD with promptDo you mean you could login from the MS to the LD without password and got the OS prompt? What you have to make sure is that your ssh authentication works for the n1sps user (or whatever you have chosen during installation) from MS to LD and from LD to RA. That involves copying the public key from the MS n1sps user to the .ssh/authorized_keys files on the LD and RA. And after that, you should check the connections manually:
    1. From the MS: ssh <LD> ls -al
    2. From the MS: ssh <LD> ssh <RA> ls -al
    These both commands should work without any prompts and after that, you shouldn't get any error messages in SPS anymore.
    HTH,
    Michael

  • What are the cases that cause the I/O error writing to output file

    may i ask what are the cases that cause the I/O error writing to output file ??
    cause i have the write path and the input file has data
    any ideas???

    and the permission is to write first time then append to itBy "permission" is meant the operating system controlled rights that let
    particular users do specified things with specified files or directories.
    the error message is a the title of the message
    I/O error writing to output fileThe java runtime actually creates a much more useful error message than
    this but, unfortunately, your program is swallowing this error message and
    replacing it with the message you see.
    Somewhere in your code you have something like:try {
        // all sorts of stuff goes here
        // including the code that is intended to
        // write to the file
    } catch(IOException ioe) {
        System.out.println("I/O error writing to output file");
            // add this line
        //ioe.printStackTrace();
    }It might say Exception rather than IOException.
    To get a much more useful error message add the commented line.

  • Oc4jDcmServlet  : java.io.IOException: Error writing to server

    I've been using the following ant task to do a remote deployment to an OC4J container :
    <tasks>
    <java jar="../../../../../MAVEN-WORK/portal-lib/oc4j_remote_deploy.jar" fork="yes" maxmemory ="512m">
    <jvmarg value="-Djava.protocol.handler.pkgs=HTTPClient"/>
    <arg value="http://wblv-ep-sp05:1824/Oc4jDcmServletAPI/"/>
    <arg value="ias_admin"/>
    <arg value="admin_dev3"/>
    <arg value="redeploy"/>
    <arg value="/app/oracle/product/oc4jext10.1.2-d3"/>
    <arg value="../portal-ear/target/portal.ear"/>
    <arg value="portal"/>
    <arg value="OC4J_portal"/>
    </java>
    </tasks>
    This has worked before , but I'm suddenly getting error when trying to run it .The error message and stack trace are shown below :
    [INFO] Executing tasks
    [java] Initializing log
    [java] Servlet interface for OC4J DCM commands
    [java] Command timeout defined at 600 seconds
    [java] Executing DCM command...
    [java] Executing command redeploy /app/oracle/product/oc4jext10.1.2-d3 ../e
    portal-ear/target/ePortalep2x-QA-V-02_00_5-20070220.ear eportal OC4J_Eportal
    [java] Command = REDEPLOY
    [java] Reading application's ear file
    [java] Ear file was successfully read
    [java] Opening connection to Oc4jDcmServlet
    [java] Setting userName to ias_admin
    [java] Sending command to DCM servlet
    [java] **** Could not check HTTP response code
    [java] ** Thread[main,5,main] ** Sun Feb 25 20:47:34 CAT 2007 ** ** EXCEP
    TION: java.io.IOException: Error writing to server
    [java] java.io.IOException: Error writing to server
    [java] at sun.net.www.protocol.http.HttpURLConnection.writeRequests(Htt
    pURLConnection.java:416)
    [java] at sun.net.www.protocol.http.HttpURLConnection.writeRequests(Htt
    pURLConnection.java:428)
    [java] at sun.net.www.protocol.http.HttpURLConnection.getInputStream(Ht
    tpURLConnection.java:934)
    [java] at java.net.HttpURLConnection.getResponseCode(HttpURLConnection.
    java:367)
    [java] at oracle.j2ee.tools.remote_deploy.Oc4jDcmClient.isHttpResponseO
    k(Unknown Source)
    [java] at oracle.j2ee.tools.remote_deploy.Oc4jDcmClient.sendCommand(Unk
    nown Source)
    [java] at oracle.j2ee.tools.remote_deploy.Oc4jDcmCommand.execute(Unknow
    n Source)
    [java] at oracle.j2ee.tools.remote_deploy.Oc4jDcmCommand.redeployApplic
    ation(Unknown Source)
    [java] at oracle.j2ee.tools.remote_deploy.Oc4jDcmMain.main(Oc4jDcmMain.
    java:63)
    [java] #### HTTP response is NOT ok
    [java] Closing connection to Oc4jDcmServlet
    [java] #### DCM command did not complete successfully (-1)
    [java] #### HTTP return code was -1
    [java] Java Result: -1
    [INFO] Executed tasks
    It seems to stall at the "Sending command to DCM servlet" stage , then fails.
    Does the DCM servlet write out logs that I can look at?Where is the Oc4jDcmServlet.ear located in an OC4J container directory hierarchy?How do I redeploy the Oc4jDcmServlet.ear ?
    This was working a few hours ago as seen here in my previous logs (and I do not understand what could have gone wrong) :
    ** Thread[main,5,main] ** Thu Feb 22 14:50:39 CAT 2007 ** Initializing log
    ** Thread[main,5,main] ** Thu Feb 22 14:50:39 CAT 2007 ** Servlet interface for OC4J DCM commands
    ** Thread[main,5,main] ** Thu Feb 22 14:50:39 CAT 2007 ** Command timeout defined at 600 seconds
    ** Thread[main,5,main] ** Thu Feb 22 14:50:39 CAT 2007 ** Executing command redeploy /app/oracle/product/oc4jext10.1.2-d3 ../portal-ear/target/portal.ear portal OC4J_portal
    ** Thread[main,5,main] ** Thu Feb 22 14:50:39 CAT 2007 ** Command = REDEPLOY
    ** Thread[main,5,main] ** Thu Feb 22 14:50:39 CAT 2007 ** Reading application's ear file
    ** Thread[main,5,main] ** Thu Feb 22 14:50:45 CAT 2007 ** Ear file was successfully read
    ** Thread[main,5,main] ** Thu Feb 22 14:50:45 CAT 2007 ** Opening connection to Oc4jDcmServlet
    ** Thread[main,5,main] ** Thu Feb 22 14:50:45 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:50:45 CAT 2007 ** Sending command to DCM servlet
    ** Thread[main,5,main] ** Thu Feb 22 14:50:50 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:50:50 CAT 2007 ** Command was successfully sent to Oc4jDcmServlet
    ** Thread[main,5,main] ** Thu Feb 22 14:50:50 CAT 2007 ** Receiving session id from servlet to check command status
    ** Thread[main,5,main] ** Thu Feb 22 14:50:50 CAT 2007 ** Session id = 0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:50:50 CAT 2007 ** Please, wait for command to finish...
    ** Thread[main,5,main] ** Thu Feb 22 14:51:00 CAT 2007 ** Checking command status...
    ** Thread[main,5,main] ** Thu Feb 22 14:51:00 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:51:00 CAT 2007 ** Setting Cookie to JSESSIONID=0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:51:00 CAT 2007 ** Checking command status
    ** Thread[main,5,main] ** Thu Feb 22 14:51:00 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:51:00 CAT 2007 ** Command has not finished yet
    ** Thread[main,5,main] ** Thu Feb 22 14:51:10 CAT 2007 ** Checking command status...
    ** Thread[main,5,main] ** Thu Feb 22 14:51:10 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:51:10 CAT 2007 ** Setting Cookie to JSESSIONID=0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:51:10 CAT 2007 ** Checking command status
    ** Thread[main,5,main] ** Thu Feb 22 14:51:10 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:51:10 CAT 2007 ** Command has not finished yet
    ** Thread[main,5,main] ** Thu Feb 22 14:51:20 CAT 2007 ** Checking command status...
    ** Thread[main,5,main] ** Thu Feb 22 14:51:20 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:51:20 CAT 2007 ** Setting Cookie to JSESSIONID=0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:51:20 CAT 2007 ** Checking command status
    ** Thread[main,5,main] ** Thu Feb 22 14:51:20 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:51:20 CAT 2007 ** Command has not finished yet
    ** Thread[main,5,main] ** Thu Feb 22 14:51:30 CAT 2007 ** Checking command status...
    ** Thread[main,5,main] ** Thu Feb 22 14:51:30 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:51:30 CAT 2007 ** Setting Cookie to JSESSIONID=0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:51:30 CAT 2007 ** Checking command status
    ** Thread[main,5,main] ** Thu Feb 22 14:51:30 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:51:30 CAT 2007 ** Command has not finished yet
    ** Thread[main,5,main] ** Thu Feb 22 14:51:40 CAT 2007 ** Checking command status...
    ** Thread[main,5,main] ** Thu Feb 22 14:51:40 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:51:40 CAT 2007 ** Setting Cookie to JSESSIONID=0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:51:40 CAT 2007 ** Checking command status
    ** Thread[main,5,main] ** Thu Feb 22 14:51:40 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:51:40 CAT 2007 ** Command has not finished yet
    ** Thread[main,5,main] ** Thu Feb 22 14:51:50 CAT 2007 ** Checking command status...
    ** Thread[main,5,main] ** Thu Feb 22 14:51:50 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:51:50 CAT 2007 ** Setting Cookie to JSESSIONID=0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:51:50 CAT 2007 ** Checking command status
    ** Thread[main,5,main] ** Thu Feb 22 14:51:50 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:51:50 CAT 2007 ** Command has not finished yet
    ** Thread[main,5,main] ** Thu Feb 22 14:52:00 CAT 2007 ** Checking command status...
    ** Thread[main,5,main] ** Thu Feb 22 14:52:00 CAT 2007 ** Setting userName to ias_admin
    ** Thread[main,5,main] ** Thu Feb 22 14:52:00 CAT 2007 ** Setting Cookie to JSESSIONID=0afb65197201792b582df0d486285eef6aa19c1ab64
    ** Thread[main,5,main] ** Thu Feb 22 14:52:00 CAT 2007 ** Checking command status
    ** Thread[main,5,main] ** Thu Feb 22 14:52:00 CAT 2007 ** HTTP response code = 200, HTTP response msg = OK
    ** Thread[main,5,main] ** Thu Feb 22 14:52:00 CAT 2007 ** Command has finished

    I looked at the log.xml file under the dcm/logs/daemon_logs directory and saw this :
    <MESSAGE>
    <HEADER>
    <TSTZ_ORIGINATING>2007-02-26T08:49:13.203+02:00</TSTZ_ORIGINATING>
    <COMPONENT_ID>DCM</COMPONENT_ID>
    <MSG_TYPE TYPE="ERROR"></MSG_TYPE>
    <MSG_LEVEL>1</MSG_LEVEL>
    <HOST_ID>wblv-ep-sp05</HOST_ID>
    <HOST_NWADDR>10.251.101.25</HOST_NWADDR>
    <MODULE_ID>oracle/defaultLogger/ExceptionLogger</MODULE_ID>
    <THREAD_ID>Deamon Worker for TaskMaster of iAS instance at: /app/oracle/product/oc4jext10.1.2-d3 JVM Id = 1b09468.110df14be01.-8000</THREAD_ID>
    <USER_ID>oracle</USER_ID>
    </HEADER>
    <CORRELATION_DATA>
    <EXEC_CONTEXT_ID><UNIQUE_ID>10.251.101.25:47620:1172472553186:2846</UNIQUE_ID><SEQ>0</SEQ></EXEC_CONTEXT_ID>
    </CORRELATION_DATA>
    <PAYLOAD>
    <MSG_TEXT>principals specified in application is not a valid path: /app/oracle/product/oc4jext10.1.2-d3/j2ee/home/config/../application-deployments/ADFBCManager/principals.xml</MSG_TEXT>
    </PAYLOAD>
    </MESSAGE>
    The same MSG_TEXT appears in the log every time i had a failed deploy.

  • Posfix error writing message: File too large

    I was looking for a reason for my Mac hanging occasionally. In Console I found the following recurring error message.
    10/2/11 8:15:19 PM     postfix/master[656]     daemon started -- version 2.5.5, configuration /etc/postfix
    10/2/11 8:16:00 PM     postfix/pickup[657]     B6095D33A25: uid=501 from=<Al>
    10/2/11 8:16:00 PM     postfix/cleanup[664]     B6095D33A25: message-id=<[email protected]>
    10/2/11 8:16:00 PM     postfix/qmgr[658]     B6095D33A25: from=<[email protected]>, size=668, nrcpt=1 (queue active)
    10/2/11 8:16:00 PM     postfix/local[666]     B6095D33A25: to=<[email protected]>, orig_to=<Al>, relay=local, delay=0.08, delays=0.04/0.02/0/0.02, dsn=5.2.2, status=bounced (cannot update mailbox /var/mail/al for user al. error writing message: File too large)
    10/2/11 8:16:00 PM     postfix/cleanup[664]     C5ACFD33A28: message-id=<[email protected]>
    10/2/11 8:16:00 PM     postfix/bounce[667]     B6095D33A25: sender non-delivery notification: C5ACFD33A28
    10/2/11 8:16:00 PM     postfix/qmgr[658]     C5ACFD33A28: from=<>, size=2366, nrcpt=1 (queue active)
    10/2/11 8:16:00 PM     postfix/qmgr[658]     B6095D33A25: removed
    10/2/11 8:16:00 PM     postfix/local[666]     C5ACFD33A28: to=<[email protected]>, relay=local, delay=0.01, delays=0/0/0/0, dsn=5.2.2, status=bounced (cannot update mailbox /var/mail/al for user al. error writing message: File too large)
    10/2/11 8:16:00 PM     postfix/qmgr[658]     C5ACFD33A28: removed
    10/2/11 8:16:19 PM     postfix/master[656]     master exit time has arrived
    I haven't done anything with Postfix, but I'm guessing one of the monitoring utilities is set up to send me an email message if it finds an error.
    Looking for ideas on how to fix this.

    Well, I think I figured this out.
    I found this posting on a message board http://www.linuxquestions.org/questions/linux-networking-3/file-too-large-in-pos tfix-495988/#post2480213
    I ran this command in terminal: sudo postconf -e "virtual_mailbox_limit=0" then: sudo postfix reload
    That still didn't fix the error messages.
    Then I ran this command in terminal: sudo postconf -e "mailbox_size_limit=0" then: sudo postfix reload
    That stopped the "file too large" error message.
    Now the console message I get is:
    postfix/master[1630]          daemon started -- version 2.5.14, configuration /etc/postfix
    postfix/pickup[1631]          68EA012A66BD: uid=501 from=<Al>
    postfix/cleanup[1643]          68EA012A66BD: message-id=<[email protected]>
    postfix/qmgr[1632]          68EA012A66BD: from=<[email protected]>, size=707, nrcpt=1 (queue active)
    postfix/pickup[1631]          BA07112A66BE: uid=501 from=<Al>
    postfix/cleanup[1643]          BA07112A66BE: message-id=<[email protected]>
    postfix/local[1645]          68EA012A66BD: to=<[email protected]>, orig_to=<Al>, relay=local, delay=0.82, delays=0.5/0.09/0/0.23, dsn=2.0.0, status=sent (delivered to mailbox)
    postfix/master[1630]          master exit time has arrived
    I'm guessing this is normal function of the postfix system. However, I'm still not sure what postfix is doing. Maybe its the under-the-hood stuff for Apple Mail?

  • Insufficient disk space / Error writing to Disk: when using photo library manager

    First I couldn't open my iphoto library (again!) so I am rebuilding my library (again- 4th time). NOW I'm told I donot  have sufficient disk space to import photos.
    Using OSX 10-9-5  Memory 4GB
    And just as I'm writing this I get Error Writing to disk memo
    iPhoto cannot import your photos because there was a problem writing to the volume containing your iPhoto library.

    Freeing Up Space on The Hard Drive
      1. See Lion/Mountain Lion/Mavericks' Storage Display.
      2. You can remove data from your Home folder except for the /Home/Library/ folder.
      3. Visit The XLab FAQs and read the FAQ on freeing up space on your hard drive.
      4. Also see Freeing space on your Mac OS X startup disk.
      5. See Where did my Disk Space go?.
      6. See The Storage Display.
    You must Empty the Trash in order to recover the space they occupied on the hard drive.
    You should consider replacing the drive with a larger one. Check out OWC for drives, tutorials, and toolkits.
    Try using OmniDiskSweeper 1.8 or GrandPerspective to search your drive for large files and where they are located.
    If you select the disk icon for your hard drive on the Desktop, then Press COMMAND-I to open the Get Info window. In the upper panel it will show your drive's space statistics - Capacity, Used, and Available. It is the latter you are interested in to see how much space is left on the drive. If you have less than 1 GB or so, then you must stop trying to write to the drive or you may corrupt it completely.

  • Error writing to field point

    I am trying to connect Field point  FP-1001 by Ni PCI-485/2 multifuction adaptor . when i start the program by Ni scope i found " ERROR WRITING TO FIELD POINT". 
    i have checekd the pci card and its setting is quite Ok. the picture attached for referance. please help .
    Attachments:
    image001.jpg ‏9 KB
    image002.JPG ‏33 KB
    image003.JPG ‏34 KB

    Dear Shane,
           Thanks for yor response. I have tried again yesterday and rectify the probelm. The confilect was between PC port  485 confirgration setting and VISA configration setting. First i treid to change setting but failed to change then i again install all the program again and tried to change the setting.this time i got successs. i per you question the driver and PCI card was installed properly.
    Thanks again.
    Regards,
    wajid
    Shane-C wrote:
    Ok, if it worked on that system before, you should be able to get it working again.  Does your PCI card show up properly in NI MAX?  Can you upload a screen shot of the device there?  Do you have an option in MAX to send a command to the FP-1001 and see if the communication was established properly?  How does your PCI card show up in the Windows Device Manager, any exclamation points indicating that the device driver is not installed or recognized?

  • IO error writing cache during ejb deployment

    I am getting this error a lot when I deploy an ear to standalone oc4j. This happens when I run from the command line or from jdeveloper. Retrying it eventually works, sometimes to speed things up I shutdown oc4j and remove the file. Is there some way to prevent this from happening.
    com.evermind.compiler.CompilationException: Error instantiating compiler: IO error writing cache: D:\oc4j\j2ee\home\application-deployments\WebTSApp\WebTSejb.jar\deployment.cache
         at com.evermind.server.ejb.compilation.Compilation.compileClasses(Compilation.java:500)

    This was a known issue before the preview release.
    Workaround for the problem would be to change the Application name and EAR file name.
    If you have the EJB deployment profile, Right mouse on it and choose Settings.
    Change the name of the EAR file and the Application name and redeploy.
    raghu
    JDev Team

Maybe you are looking for

  • Long Text problem in a BADI

    Hi Everyone, I am implementing BADI "WORKORDER_UPDATE". The requirement is there should be no changes made on the IW32 transaction. I have successfully implemented the BADI and it is now giving an error when we try to change anything on the order in

  • ITunes 11 migration to Windows (yeah, backwards) :)

    I need to migrate my iTunes 11 Library, including all playlists and cue points to a Windows 7 laptop also running iTunes 11. I cannot find a single solution that works. Can anyone tell me specifically what needs to be done for iTunes 11? The solution

  • Itunes Radio on Iphone 4s Stops Playing

    Everytime I go into Itunes radio on my 4s it randomly stop playing. I would be in the middle of a song and it would stop, but when I hit the play button to continue the song there is no sound, sometimes I have to continuously hit the play button for

  • The phone is in Silent mode

    I purchased the phone around 15days back, Since yesterday Automatically the phone has got into silent mode, No music is being palyed no ring tones and no Sound on Notification

  • Highscore list using Flash and PHP

    Hi, I want to make a highscorelist using AS3 and PHP(MySQL). I cant find a good tutorial on internet how to make something like that but I got this so far: import fl.controls.TextInput; import fl.controls.TextArea; var variables:URLVariables = new UR