Translator Error No .dat file

I know this question has been asked a thousand times on here but I have a new twist to the problem.  I don't have a .dat file in my configuration folder. I have looked through every folder in the dreamweaver folder and I don't have a .dat file.
I have tried renaming my configuration folder to "configuration old" but then dreamweaver won't open at all.  Is reinstalling my only solution?

The translator error as I searched seems to be tied to the common Javascript errors.  Follow the below KB article and you should be able to solve your issue:
http://kb2.adobe.com/cps/405/kb405604.html

Similar Messages

  • While importing a request error message' Check-sum error in data file'

    Hi Friends
    I have a problem.
    We are trying to inport a request after putting the files in cofile and data file folders( 4.6C System).While doing so an error message is seen in the log " Check-sum error in data file after XXXX bytes".
    Can some one help me with this?
    Thanks
    Regards
    Ankur

    Hi Ankur,
    It is sure your file is corrupted or not present.
    Check in cofiles and data directory under /usr/sap/trans your transport request number.
    Best Wishes.
    Kumar

  • Clear Data Manager Package Error "The data file is empty."

    Hi,
    When I run the Clear data package in Data Manager, I receive the error "The data file is empty." I selected a very specific set of dimension values (none are calculated) and am on BPC 7.5 SP3. I subsequently turned on debugging to troubleshoot, but do not see any obvious issues leading the the error message. The log file with debugging turned on is below. Any help would be greatly appreciated!
    Thanks.
    Tom
    TOTAL STEPS  3
    1. Export_Zero:        completed  in 1 sec.
    2. Load Cube:          Failed  in 0 sec.
    3. Clear:              completed  in 0 sec.
    [Selection]
    ENABLETASK= Yes
    CHECKLCK= Yes
    (Member Selection)
    Category: ACTUAL
    Time: 2010.C_SEP
    Affiliate: az_swhd
    Account: Donor_DART_ID_1
    Functional: Benchmark_F
    Report: Cons
    Restriction: AnyRestricted
    [Messages]
    The data file is empty. Please check the data file and try again.
    [EvModifyScript Detail]
    12-28-2010  17:30:05 - Debug turned ON
    INFO(%TEMPFILE%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(EXPORT_ZERO, APPSET, ESMetrics)
    TASK(EXPORT_ZERO, APP, CONSOLIDATED)
    TASK(EXPORT_ZERO, USER, NESSGROUP\tbardwil)
    TASK(EXPORT_ZERO, FILE, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(EXPORT_ZERO, SQL,
    select [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) as ZeroTable  group by [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)
    TASK(EXPORT_ZERO, DATATRANSFERMODE, 2)
    TASK(LOAD CUBE, APPSET, ESMetrics)
    TASK(LOAD CUBE, APP, CONSOLIDATED)
    TASK(LOAD CUBE, USER, NESSGROUP\tbardwil)
    TASK(LOAD CUBE, FILE, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(LOAD CUBE, DATATRANSFERMODE, 4)
    TASK(LOAD CUBE, DMMCOPY, 0)
    TASK(LOAD CUBE, PKGTYPE, 0)
    TASK(LOAD CUBE, CHECKLCK, 1)
    TASK(CLEAR COMMENTS, APPSET, ESMetrics)
    TASK(CLEAR COMMENTS, APP, CONSOLIDATED)
    TASK(CLEAR COMMENTS, USER, NESSGROUP\tbardwil)
    TASK(CLEAR COMMENTS, DATATRANSFERMODE, 0)
    TASK(CLEAR COMMENTS, SELECTIONORFILE, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)
    TASK(CLEAR COMMENTS, ENABLETASK, 1)
    TASK(CLEAR COMMENTS, CHECKLCK, 1)
    INFO(%ENABLETASK%, 1)
    INFO(%CHECKLCK%, 1)
    INFO(%SELECTION%, [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED'))
    INFO(%TOSELECTION%, [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED'))
    INFO(%APPSET%, ESMetrics)
    INFO(%APP%, CONSOLIDATED)
    INFO(%CONVERSION_INSTRUCTIONS%, )
    INFO(%FACTCONVERSION_INSTRUCTIONS%, )
    INFO(%SELECTIONFILE%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\FROM_51_.TMP)
    INFO(%TOSELECTIONFILE%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\TO_51_.TMP)
    INFO(%DEFAULT_MEASURE%, PERIODIC)
    INFO(%MEASURES%, Periodic,QTD,YTD)
    INFO(%OLAPSERVER%, ETSCSAP047940.EASTER-SEALS.ORG)
    INFO(%SQLSERVER%, ETSCSAP047940.EASTER-SEALS.ORG)
    INFO(%APPPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\)
    INFO(%DATAPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\DataFiles\)
    INFO(%DATAROOTPATH%, C:\BPC\Data\WebFolders\)
    INFO(%SELECTIONPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\SelectionFiles\)
    INFO(%CONVERSIONPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\ConversionFiles\)
    INFO(%TEMPPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\)
    INFO(%LOGICPATH%, C:\BPC\Data\WebFolders\ESMetrics\Adminapp\CONSOLIDATED\)
    INFO(%TRANSFORMATIONPATH%, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\TransformationFiles\)
    INFO(%DIMS%, [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[Time])
    INFO(%FACTDIMS%, [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID])
    INFO(%CATEGORY_DIM%, [Category])
    INFO(%TIME_DIM%, [Time])
    INFO(%ENTITY_DIM%, [Affiliate])
    INFO(%ACCOUNT_DIM%, [Account])
    INFO(%CURRENCY_DIM%, )
    INFO(%APP_LIST%, Consolidated,ES_INC,GrantMgmt,LegalApp,LRate,Ownership,Rate)
    INFO(%ACCOUNT_SET%, DONOR_DART_ID_1)
    INFO(%AFFILIATE_SET%, AZ_SWHD)
    INFO(%CATEGORY_SET%, ACTUAL)
    INFO(%FUNCTIONAL_SET%, BENCHMARK_F)
    INFO(%REPORT_SET%, CONS)
    INFO(%RESTRICTION_SET%, ANYRESTRICTED)
    INFO(%TIME_SET%, 2010.C_SEP)
    INFO(%ACCOUNT_TO_SET%, DONOR_DART_ID_1)
    INFO(%AFFILIATE_TO_SET%, AZ_SWHD)
    INFO(%CATEGORY_TO_SET%, ACTUAL)
    INFO(%FUNCTIONAL_TO_SET%, BENCHMARK_F)
    INFO(%REPORT_TO_SET%, CONS)
    INFO(%RESTRICTION_TO_SET%, ANYRESTRICTED)
    INFO(%TIME_TO_SET%, 2010.C_SEP)
    INFO(DATAMGRGLOBALBPU, )
    INFO(DATAMGRGLOBALCLIENTMACHINEID, ETSCWLT048794)
    INFO(DATAMGRGLOBALERROR, )
    INFO(DATAMGRGLOBALPACKAGEINFOR, )
    INFO(DATAMGRGLOBALPACKAGENAME, C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\DataManager\PackageFiles\System Files/Clear.dtsx)
    INFO(DATAMGRGLOBALSEQ, 51)
    INFO(DATAMGRGLOBALSITEID, )
    INFO(MODIFYSCRIPT, DEBUG(ON)<BR>PROMPT(SELECTINPUT,[CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED'),,"SELECT THE MEMBERS TO CLEAR",[Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[Time])<BR>PROMPT(RADIOBUTTON,1,"DO YOU WANT TO CLEAR COMMENTS ASSOCIATED WITH DATA REGIONS IN BPC?",1,{"YES","NO"},{"1","0"})<BR>PROMPT(RADIOBUTTON,1,"SELECT WHETHER TO CHECK WORK STATUS SETTINGS WHEN DELETING COMMENTS.",1,{"YES, DELETE COMMENTS WITH WORK STATUS SETTINGS","NO, DO NO DELETE COMMENTS WITH WORK STATUS SETTINGS"},{"1","0"})<BR>INFO(C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempfdla_51_.tmp)<BR>TASK(EXPORT_ZERO,APPSET,ESMetrics)<BR>TASK(EXPORT_ZERO,APP,CONSOLIDATED)<BR>TASK(EXPORT_ZERO,USER,NESSGROUP\tbardwil)<BR>TASK(EXPORT_ZERO,FILE,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)<BR>TASK(EXPORT_ZERO,SQL,
    select [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) as ZeroTable  group by [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)
    )<BR>TASK(EXPORT_ZERO,DATATRANSFERMODE,2)<BR>TASK(LOAD CUBE,APPSET,ESMetrics)<BR>TASK(LOAD CUBE,APP,CONSOLIDATED)<BR>TASK(LOAD CUBE,USER,NESSGROUP\tbardwil)<BR>TASK(LOAD CUBE,FILE,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)<BR>TASK(LOAD CUBE,DATATRANSFERMODE,4)<BR>TASK(LOAD CUBE,DMMCOPY,0)<BR>TASK(LOAD CUBE,PKGTYPE,0)<BR>TASK(LOAD CUBE,CHECKLCK,1)<BR>TASK(CLEAR COMMENTS,APPSET,ESMetrics)<BR>TASK(CLEAR COMMENTS,APP,CONSOLIDATED)<BR>TASK(CLEAR COMMENTS,USER,NESSGROUP\tbardwil)<BR>TASK(CLEAR COMMENTS,DATATRANSFERMODE,0)<BR>TASK(CLEAR COMMENTS,SELECTIONORFILE,C:\BPC\Data\WebFolders\ESMetrics\CONSOLIDATED\PrivatePublications\tbardwil\TempFiles\Tempwbh9_51_.tmp)<BR>TASK(CLEAR COMMENTS,ENABLETASK,1)<BR>TASK(CLEAR COMMENTS,CHECKLCK,1)<BR>BEGININFO(
    select [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 as SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) as ZeroTable  group by [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)
    )<BR><BR><BR><BR><BR><BR><BR>SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM ( SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM TBLFACTCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED') UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM TBLFACTWBCONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')  UNION ALL SELECT [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID],0 AS SIGNEDDATA FROM TBLFAC2CONSOLIDATED WHERE [CATEGORY] in (N'ACTUAL') and [TIMEID] in (N'20100900') and [AFFILIATE] in (N'AZ_SWHD') and [ACCOUNT] in (N'DONOR_DART_ID_1') and [FUNCTIONAL] in (N'BENCHMARK_F') and [REPORT] in (N'CONS') and [RESTRICTION] in (N'ANYRESTRICTED')) AS ZEROTABLE  GROUP BY [Account],[Affiliate],[Category],[Functional],[Report],[Restriction],[TIMEID] OPTION(MAXDOP 1)<BR><BR><BR><BR>ENDINFO<BR><BR><BR>)
    Edited by: Tom Bardwil on Dec 28, 2010 5:20 PM

    You can greatly improve your chance of receiving a helpful answer to your question if you state the version (MS or NW) and the release (5.1, 7.0, 7.5) of BPC which you are using.
    Also notice the sticky [note|Please do not post BPC, SSM or FI/CO questions here!; at the top of this forum whereby we announced new dedicated forums for BPC which are the proper place to post your questions regarding BPC in the future to be able to reach the right audience for your question.
    Thanks and best regards,
    [Jeffrey Holdeman|http://wiki.sdn.sap.com/wiki/display/profile/Jeffrey+Holdeman]
    SAP Labs, LLC
    BusinessObjects Division
    Americas Applications Regional Implementation Group (RIG)

  • Error editing data file, unable to load Region Definition

    Hello
    I'm getting a parseException when I try to edit a data file from UCM.
    The steps followed are:
         1- Create an element definition at Site Studio
         2 - Create a Region Definition refering to the Element Definition
         3 - Create a Content in UCM. Double checked that the region definition and the web site are set. The content of the primary file is:
              <?xml version="1.0" encoding="UTF-8" standalone="no"?>
              <wcm:root xmlns:wcm="http://www.stellent.com/wcm-data/ns/8.0.0" version="8.0.0.0">
              </wcm:root>
         4 - Open the content info and click "Edit data file". Then two pop-up window are showed:
                   A. The contributor pop-up displays a blank page with title containing "'idcService' parameter not found"
                   B. new Page displays the next message:
                        Unable to perform the action due to the following reasons:
                        [+] The Region Definition could not be loaded. The Region Definition 'RD_REGION_ID' could not be loaded. [...] The markup in the document preceding the root element must be well-formed. org.xml.sax.SAXParseException: The markup in the document preceding the root element must be well-formed.
    I've checked the region definition and everything seems correct to me:
         Web browser displays the xml correctly when I open it from "Web Location".
         I've checked the syntax with Notepad++ XML Tools plugin and the result is "No error detected".
         I've also checked the file with hex editor and everything seems to be right.
    Same tests for the primary file
    Both contents, the region definition and the dataFile, are accesible and editable from Site Studio.
    UCM server LOG:
         -translated-
         Could not load 'RD_REGION_ID' of Region Definition. Error parsing Region Definition at line 1, column 2.
         -end of translated log-
         Error: The markup in the document preceding the root element must be well-formed. org.xml.sax.SAXParseException: The markup in the document preceding the root element must be well-formed.. [ Detalles ]
         !csSSWebsiteObjectLoadError,Region Definition,RD_REGION_ID!csSSWebsiteObjectSaxParseException,Region Definition,1,2,The markup in the document preceding the root element must be well-formed.!syJavaExceptionWrapper,org.xml.sax.SAXParseException: The markup in the document preceding the root element must be well-formed.
         org.xml.sax.SAXParseException: The markup in the document preceding the root element must be well-formed.
              at com.sun.org.apache.xerces.internal.util.ErrorHandlerWrapper.createSAXParseException(ErrorHandlerWrapper.java:195)
              at com.sun.org.apache.xerces.internal.util.ErrorHandlerWrapper.fatalError(ErrorHandlerWrapper.java:174)
              at com.sun.org.apache.xerces.internal.impl.XMLErrorReporter.reportError(XMLErrorReporter.java:388)
              at com.sun.org.apache.xerces.internal.impl.XMLScanner.reportFatalError(XMLScanner.java:1414)
              at com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl$PrologDriver.next(XMLDocumentScannerImpl.java:925)
              at com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl.next(XMLDocumentScannerImpl.java:648)
              at com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.next(XMLNSDocumentScannerImpl.java:140)
              at com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:510)
              at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:807)
              at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:737)
              at com.sun.org.apache.xerces.internal.parsers.XMLParser.parse(XMLParser.java:107)
              at com.sun.org.apache.xerces.internal.parsers.DOMParser.parse(DOMParser.java:225)
              at com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderImpl.parse(DocumentBuilderImpl.java:283)
              at weblogic.xml.jaxp.RegistryDocumentBuilder.parse(RegistryDocumentBuilder.java:163)
    Server:
         Version:11gR1-11.1.1.4.0-idcprod1-101229T001824 (Build:7.3.0.180)
         Database Type:Oracle
         Database Version:11.2.0.2.0 ---Oracle Database 11g Enterprise Edition Release --- - Production With the Partitioning,
    Site Studio Info:
         Component Build Version: 11.0.0.648
         Disable Contribution: False
         Web Site Object Type Field: xWebsiteObjectType
         Web Sites Field: xWebsites
         Web Site Section Field: xWebsiteSection
         Do not Show In Lists For Web Sites Field: xDontShowInListsForWebsites
         Region Definition Field: xRegionDefinition
         Contributor Console Enabled: True
         Designer Version: 11.1.1.4.0

    Hi,
    I've tested to obtain the content of the UCM from a webcenter portal server, I create a new Navigation Item of type "Content Item", when I choose any document from the UCM I'm getting the next trace:
    java.lang.NullPointerException
              at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:675)
              at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:700)
              at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:700)
              at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:715)
              at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyString(RegionUtils.java:647)
              at oracle.adf.controller.internal.binding.RegionUtils.getTaskFlowId(RegionUtils.java:577)
    oracle.adf.controller.ControllerException: ADFC-14014: Task flow binding active attribute evaluated to null, binding='pickerWrapper1'.
         at oracle.adf.controller.internal.binding.DCTaskFlowBinding.getConditionalActivationValue(DCTaskFlowBinding.java:974)
         at oracle.adf.controller.internal.binding.TaskFlowRegionController.refreshRegion(TaskFlowRegionController.java:84)
         at oracle.adf.model.binding.DCBindingContainer.internalRefreshControl(DCBindingContainer.java:3211)
         at oracle.adf.model.binding.DCBindingContainer.refresh(DCBindingContainer.java:2883)
    java.lang.NullPointerException
         at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:675)
         at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:700)
         at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:700)
         at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyStringRecurse(RegionUtils.java:715)
         at oracle.adf.controller.internal.binding.RegionUtils.getRegionHierarchyString(RegionUtils.java:647)
         at oracle.adf.controller.internal.binding.RegionUtils.getTaskFlowId(RegionUtils.java:577)
    I don't know if the reason for this error is the same or there are other causes involved
    Regards,
    Rafa

  • What causes Outlook error 0x8004010f "data file cannot be accessed"?

    any indexing running on the machine? is .pst in the exclusion list?

    Hi all, got an interesting one here. A reoccurring problem for one of my clients on all 8 of their desktops when Outlook 2010 comes up with this error. Normally about one machine a month will do it. I know the solution and can repair it pretty quick these days but the manager is now asking why it keeps happening and I haven't got a clue.We've got 8 fairly recent desktops, Windows 7, all built from brand new good quality hardware. Outlook 2010 on each machine has one IMAP account setup. We've got nothing fancy here, the only group policies used are for mapped drives and printers, everything else is very much standard, no 'addins' or integrations with Outlook.The error states the data file cannot be accessed, the file is there and accessible, ScanPST doesn't find any errors, it has not moved, but for some reason Outlook has created an...
    This topic first appeared in the Spiceworks Community

  • How many errors in data file before data load is aborted?

    Is there a limit to how many incorrect records you can have when loading a free form file before Essbase aborts the load process?
    I am trying to export a complete database to another database where the cost center level detail is removed, i.e. all parents exist (and data has been exported using All Levels).
    Thanks

    I have done it this way in the past:
    1-Create a dummy scenario.
    2-Write a calc script to copy data to that scenario, but only fix on the blocks you want to extract.
    3-Delete all other scenarios, thus leaving ONLY the data you copied to your dummy scenario.
    4-Rename the dummy scenario to the scenario name you need.
    5-Export the full database to a file.
    6-Load the export into your destination database.
    This might work but not knowing your exact source and target hierarchies, this may not work.....

  • Clear Data package - "The data file is empty." Error

    Hi,
    When I run the Clear package in Data Manager, I get the error "The data file is empty". We are running BPC 7.5 SP3. There are no calculated members selected. Any suggestions?
    Thanks.
    Tom

    Hi Tom,
    If i'm not mistaken when we run the clear data package, BPC inserts -ve values for the combination of dimensions selected.
    For ex if you have selected time as 2010.DEC it would reverse out all these values by inserting negative values which would then total up to zero.
    So i still think it could be a problem with the Transformation file.
    P.S. Maybe you can try using the Clear from Fact table package.
    Santosh

  • Error in flat file mapping

    I am using flat file as source ,and when i am trying to validate the mapping i am getting the following error:
    No data file specified
    I tried to configure the mapping,I did not find any option asking source file name.
    Can you please help me...
    Kiran

    Hi Kiran
    As far as I remember...when you configure the map, the map configuration property inspector has a tree in the lhs if you navigate to the flat file operator, and right click on that node you should be able to add a data file. The properties for the data file are shown on the rhs when selected in tree.
    Cheers
    David

  • Data files are offline / tablespace offline

    I am confused on the concept of data file offline.
    What happened when you take data file offline?
    Is data file offline because of disk damage? we see the error that data file is haning a problem, we have manually to take data file offline? how?
    Why is tablespace offline but data file is still online?
    Please help!
    Thanks
    S.

    ski123 wrote:
    I am confused on the concept of data file offline.
    What happened when you take data file offline?The offline datafile means simply that the file is not available anymore for the access. All the objects for the offline file become inaccessible for the users to access.
    Is data file offline because of disk damage? we see the error that data file is haning a problem, we have manually to take data file offline? how?Yes it can happen that the data file goes offline because of any kind of damage either over the disk, or also with the file itself. For example, there may be a chance that the file is not available anymore , may be because someone has deletedthe file. So most of the times, Oracle would offline the file by itself. In any case, if the file doesn't go offline by itself than you need to take it offline yourself as the recovery for the file can't happen if its online. You would get the error messages that oracle can't acquire the exclusive access over the file.
    Why is tablespace offline but data file is still online?Not sure that I did see this ever myself. Tablespace is nothing but the container for the datafile so if there is a file of it which is offline and that's the only file in the tablespace, the tablespace can't stay as online. Show us an output for the same.
    The links given by Anantha are must for you to read the cocept of offlining a tablespace. Make sure you read them for sure.
    HTH
    Aman....

  • File Adapter Read, Translation error

    Hi,
    I got problem with BPEL File Adapter. When i'm reading file i got scheme - termineted with semicol and without quoting.
    In uploaded test file there was line - first colums;second colums;""third"" column";forth column;
    And the problem are five quotes and adapter fails with translation error. Can i somehow ignore them?

    Hi,
    You have use the attribute quotedBy... By default, the specified string is &amp;quot; ("). If your data includes this character, you must override this default even if the field is not quoted.
    For example:
    {code}
    <element name="field" type="string" nxsd:style="terminated" nxsd:terminatedBy=";" nxsd:quotedBy="&lt;"/>
    {code}
    Read this...
    http://docs.oracle.com/cd/E23943_01/integration.1111/e10231/nfb.htm#CHDBBECG
    Cheers,
    Vlad

  • Translation Error not happening in File Adapter Inbound

    I have been working on a requirement where the file adapter picks up the file and gives to the BPEL for processing the data.
    So, in order to do Error Handling for file adapter, i have been reading the Oracle documentation for technology adapters. In that, some of the
    points which i felt valuable for my error handling when using file adapter are the uniqueMessageSeparator property, fault-policies for rejected messages,
    and one of the action for rejected messages could be writing the payload to a file or invoking any other webservice. .
    In my composite, i have configured the file adapter as a service for reading files, and i defined the xsd using the native format builder....If i put some data like 'aaaa' in the attribute of type integer , the .csv file is being picked up and its being passed to the BPEL without any error ? I believe there should be a translation error...right...
    But i am not getting the error, please tell me if anything is wrong in my xsd....
    Here is my xsd....
    <?xml version="1.0" encoding="UTF-8" ?>
    <xsd:schema xmlns:xsd="http://www.w3.org/2001/XMLSchema"
    xmlns:nxsd="http://xmlns.oracle.com/pcbpel/nxsd"
    xmlns:tns="http://TargetNamespace.com/InboundService"
    targetNamespace="http://TargetNamespace.com/InboundService"
    elementFormDefault="qualified"
    attributeFormDefault="unqualified"
    nxsd:version="NXSD"
    nxsd:stream="chars"
    nxsd:encoding="US-ASCII"
    nxsd:hasHeader="true"
    nxsd:headerLines="1"
    nxsd:headerLinesTerminatedBy="${eol}"
    <xsd:element name="names">
    <xsd:complexType>
    <xsd:sequence>
    <xsd:element name="name" minOccurs="1" maxOccurs="unbounded">
    <xsd:complexType>
    <xsd:sequence>
    <xsd:element name="FirstName" type="xsd:string" nxsd:style="" />
    <xsd:element name="LastName" type="xsd:string" nxsd:style="" />
    <xsd:element name="Number" type="xsd:integer" nxsd:style="" />
    </xsd:sequence>
    </xsd:complexType>
    </xsd:element>
    </xsd:sequence>
    </xsd:complexType>
    </xsd:element>
    </xsd:schema>
    In another composite, my file adapter is picking up the .xml file and if i put any wrong data over there, that is working fine, i am getting the trasnlation error, but only here in the case of csv file, i am not getting the error. Experts, please help me in this regard...I am using SOA 11G
    Thanks,
    Naresh

    pls help....

  • OraRRP Error with "Unable to copy data file;Error code 2, check disk space"

    Hi,
    Some users get this message -"Unable to copy data file;Error code 2, check disk space" when run report with orarrp, but most users do not get it.
    I check free space at both server and client side, they are very sufficient.
    I also checked directory exists for REPORTXX_TMP variable.
    My user call reports via URL (rwservlet) and it occur for all reports.
    How I can solve this problem?
    Thanks in advance.
    Tawatchai R.

    Hi,
    have the same problem now. One user has temporarily problems to download .rrpa files via URL (rwservlet) request. Error code: -"Unable to copy data file;Error code 2, check disk space". Did you get a solution??
    Thanks in advance. Axel

  • Error Message: The data provider required to connect to the local data file could not be found. The file will be added to the project but the typed DataSet associated with the file will ont be generated.

    I am currently taking the course "C# Fundamentals" from Bob Tabors site:  LearnVisualStudio.NET .  I'm on day 8 where we are integrating SQL Server with C#.  In Lesson 6 and 7, (havent' gotten past these two) I am having an
    issue.  In Lesson 6, "Retrieving Data with ADONET 2 in a Connected Scenario (SQL Server Compact Edition) when I go to create the data file I get the error message I typed into the title.  I cannot get the db to appear in the Solution
    Explorer, or open up so I can create a table.  When I try to skip by that lesson, and complete lesson7, I'm running into the same basic problem, where the error messages are different but I get an exception when I attempt to run the program.  Even
    though I have been able to create the table for this lesson, the Exception states I have an invalid object name 'Customers' which is the name of the table. I have tried reading several different suggestions for correcting this issue, but have not understood
    exactly what I need to do.  Since I work from home on my computer, I do not want to do anything with files or downloads unless I understand what I need to do, cause I do not want to lose my job.  Any help would be appreciated. 
    JennyBarrett7

    Hi
    According to your error message, we need to verify if Microsoft SQL Server Compact appears in the
    Change Data Source dialog. If not, you need to install
    SQL Server Compact components for Visual Studio firstly, and if you choose to install SQL Server Compact 4.0 , you should note that SQL Server Compact 4.0 supports in Visual Studio 2010 Service Pack 1 or later versions. I recommend you to install the latest
    Service Pack (SP) of SQL Server Compact, and latest SP of Visual Studio, then check if the error still occurs. For more information, see:
    http://blogs.msdn.com/b/sqlservercompact/archive/2011/03/15/sql-server-compact-4-0-tooling-support-in-visual-studio-2010-sp1-and-visual-web-developer-express-2010-sp1.aspx
    However if there is no problem with the installation of SQL Server Compact, it will be an issue that regards ASP.NET and website deployment. I suggest you to post the question in the ASP.NET forums at
    http://forums.asp.net/ . It is appropriate and more experts will assist you.
    In addition, you can review the following link:
    Working with SQL Server Compact in Visual Studio:http://msdn.microsoft.com/en-us/library/gg606540(v=vs.100).aspx
    Thanks
    Lydia Zhang

  • Error in date format when  I load a CSV file

    I am using Oracle G10 XE and I am trying to load data into my database from comma separated files.
    When I load the data from a CSV file which has the date with the following format "DD/MM/YYYY", I received the following error "ORA-01843: not a valid month".
    I have the NSL_LANG set to AMERICAN. I have tried the following command: "ALTER SESSION SET NLS DATE FORMAT="DD/MM/YYYY" and this does nothing. When I try to run "SELECT SYSDATE "NOW" FROM DUAL;" I get the date in this format "10-NOV-06".
    I will appreciate any help about migrating my data with date fields in format DD//MM/YYYY.
    Sincerely,
    Polonio

    See Re: Get error in date when I load a CSV file

  • Error in running Extract Definition Upload from Data File concurrent.

    Hi all,
    Am trying to upload the 834 Extract Layout from the data file by running the concurrent program, Extract Definition Upload from Data File.
    After running this concurrent program am getting the Extract Layout definition with Record layout and Data elements within it.
    But some record layouts has the changes at their repeating level in that.
    Please suggest me how do i get the same repeating levels for the record layouts when i move the 834 benefit extract layout definition
    from one instance to another instance.

    Hi,
    We have exactly the same error in IBolt. This error happens sometime and we don't find the reason.
    IBolt has been upgraded to Version 3.1 SP1 and a fix has been developed by the support. After installation, we've work hardly during 2 days for testing all the flows by the customer without having this error.
    Ibolt works as a service and when this error occurs, the service stopped and must be restarted.
    In our flows, we have put a delete of the observer.dll file just before the rest of the flow but it doesn't solve the problem.
    Deleting the directory can be temporarly a solution but the error come back another day.
    Each time the SBO client or DTW is started, the %temp%\SMS_OBJ_DLL directory is created...
    When you just delete the observer.dll file and leave the rest of the directory like it is, the file is re-created when you start and log to SBO. I would say that this file is a copy of another file observer_800178.dll (800178 depending of the version of SBO yoi have). 800178 corresponds to SBO 2007 A PL 42.
    Do you have more information since your post has been send ?
    Thanks in advance for your help.
    Best regards.

Maybe you are looking for