Exported data files change clarification

We are migrating a client from Oracle CRM On Demand to another CRM product. A couple of months ago we performed a full export and mapped all the needed fields from the current Oracle system to the new CRM system. This morning I went in and performed another full export to start the process of the actual migration process based on all the mappings we had performed previously. This time there were quite a few new data files that were exported and I believe this may be due to the "Release 21" that was just implemented. It appears initially that these new files are duplicate records that are associated to the main record. That is now there is an "Account Opportunity" data file where before it didn't exist. I looked online but cannot find any specifics on this. Can someone please let me know if this is the case or is there a document online somewhere that has this information? Thanks in advance for the help.

Hi Dave,
I've got this working on the copy of the rulebase I had, it's because your entity / relationship public names don't match the .xds file, possibly as a result of the upgrade. I've sent you an email about this.

Similar Messages

  • Need to rebuild Outlook 2011 for mac identity for exporting data file.

    I was exporting outlook for mac data file (to import it into apple mail), and suddenly Outlook email client crashed and when I tried again it said that I need to rebuild the Outlook for mac database using rebuild utility. I took help from some professional technicians but it was a dead end. They said that my outlook for mac database was corrupt and they could not help me any further. I had years of emails in outlook for mac that I needed to move to apple mail as I was fed up of these frequently occuring outlook 2011 errors. Please if anyone could help me in this case then it will be very helpful.
    Thanks

    Outlook for mac stores archives its data in .olm format and it originally stores its mails in .OLK format. If everything has failed then you can seek help of a tool that can recover these .olk14 messages. One tool I know that is capable of doing so is http://www.outlookmacdatabaserecovery.com/.
    Look for the trial version first and see of it works and recovers you emails.

  • Export data File Browse dialog box does not appear

    I am trying to export data to CSV and when I click on Browse to locate proper directory, nothing happens (the dialog doesn't show up).
    I normally right-click on any cell in "Result" TAB, choose Export->CSV.
    On the same note CSV format is a workaround to get to Excel. Is there an extension or option to save directly in Excel format. This would avoid having the client to open the file in Excel and then do a Save As with the proper XLS format instead of CSV.
    Using Win XP client connecting to 10.2 database. Using latest SQL Developer version (downloaded on Friday may 5th version 1.0.0.15.27).
    Thanks
    Robert
    Message was edited by:
    user509017

    I haven't seen the browse button problem that you describe. Is it still occurring? On all exports? There have been issues with export rerunning the query a few times which makes the application appear to lock up. Could that be the issue?
    There is an excel add-in. I assume that it isn't included for licensing issues, but it comes from one of the core developers of this app.
    esdev.sf.net.
    Regards,
    Eric

  • Export data file directory

    what's the default directory when exporting data from the application manager? why can't i find the file i just export or choose the directory when i try to reload the data?

    The default directory is $ARBORPATh/app.But you can export to database directory by$ARBORPATh/app/appname/dbname/filename.txtso that you can load directly this from App Manager under server file.

  • Peak bitrate of the exported VOB files change radically

    I have tested with Bitrate Viewer my timeline's .m2v file, its peak bitrate is a bit under 8.7Mbits. The project has one subtitle track, and two ac3 audio tracks (2.0 track is 320kbps (Encore encodes the 2.0) and 5.1 is 448kbps). So, as making peak bitrate calculations the maximum bitrate should not exceed 9.8Mbits, right? Still, when I'm viewing the exported DVD Folder with Bitrate Viewer, VTS_01_1.VOB has a peak bitrate of 12723 bits. In that VOB the average bitrate is around 7500bits as it should and mostly max. 8.2Mbits. How is this possible that there is this single super peak in the VOB?
    PS. Did another test, I changed the 2.0 AC3 track (which Encore encodes) to a previously AC3 encoded track my client provided me, now two of the VOBs have these super peaks as:
    VTS_01_1.VOB - 11010 bits
    VTS_01_5.VOB - 13486 bits
    What is the logic?

    Mick.
    Whilst there is nothing to actually prevent you from using a 320kbps stereo AC3 stream at all, I was only referring to
    i Encore's
    encoding to AC3 stereo being 192kbps. AFAIK (I do not have a copy of the specs in front of me so cannot be 100% certain here) the specs merely state that the maximum bitrate for AC3 is 448kbps. The general consensus rate for stereo is 192, and Encore's Automatic transcoding uses this value. It is certainly possible to use the whole 448 in stereo, assuming your encoder is capable of this. I have even seen surround 5.1 streams authored to DVD-Video at 640kbps in clear violation of the specs (Pink Floyd's "Pulse" Live album did this) which is complete stupidity. WHy not simply add a half-bitrate DTS stream? Ask Sony, I guess. But again I digress, and wander OT. Sorry about that.
    If your client is providing you with 320kbps streams for stereo, this ought to work just fine. But your original post did heavily imply that Encore was doing the encoding for these streams, hence my comment.
    >(2.0 track is 320kbps (Encore encodes the 2.0)
    As far as Replication houses rejecting masters goes, this is a whole can of small, pink wriggly things all of it's own, and the blunt fact of the matter is that replication houses
    i almost never bother to correctly verify the actual streams in your Video_TS.
    At best they will run the DDP files through the Eclipse system, and as all the various warnings are user-settable, what they will flag as an error is very much a case of pot luck. About the only way I know of to absolutely validate & verify your streams is to use the MEI DVD-Video Verification tool in combination with the full DVD-Video specifications, or else use a service such as Trai Forresters to get the job done for you. See http://www.dvdverification.com/public/157.cfm for a complete description of these services as well as a very good explanation as to why this is a good idea.
    To summarize, Replication companies do not verify your data. They do NOT, repeat NOT check your data for DVD-Video spec compliance. What they will do is run the
    i image
    through Eclipse, and may or may not act on the warnings & errors it flags. Odds are very high that unless you are using
    A - A very reputable replicator, or
    B - One that actually gives a damn (rare these days) then whatever happens - unless the eclipse warnings are flagged as FAIL errors, with the ABORT recommendation (and even this is user definable too) then they will most likely still press whatever you send them, and when it all goes pear shaped they will blame you. An image that passes Eclipse does not necessarily comply to the DVD-Video specifications. That is your job as the authoring house.
    Having said that, I would be extremely wary of attaching total reliance to any freeware utility in this matter. However, it would seem that there is a doubt over the true maximum bitrate of your stream, and this may well cause playability problems in some players. Not all, as not all DVD players are created equally, sad to say.
    I strongly suspect I have just scared the willies out of you now. That was not my intention, and I apologise if this is what has happened.
    Let me try to go into some more details here without repeating verbatim what you can go & read for yourself on Trai's website about this. In the earlier days of this game, replicators were a lot fewer in number, Eastern Europe, Asia & China had not got into the game, margins had not been cut to the bone to stay open & operational, and discs got checked & verified a hell of a lot better than is common these days. Problem is it can get bloody expensive. Deadlines loom, clients start getting twitchy, but the accountants keep wanting more & more for less & less and threaten to move contracts around of they do not get "discounted services". So companies start to cut corners.
    But again I digress and wander, and again I apologise. Most of this information in this post will be irrelevant to your project, and will not be anything to worry about.
    The point of making this post was not to scare the bejesus out of you, but to point out that you simply cannot rely on the replication house to tell you if your streams are out of spec - they simply will not do that any more than they will flag VOB errors in the navigation code either. This is just not something they ever check.
    The problem is that all authoring packages - there are no exceptions here at all and some are worse than others (with Apple's DVDSP being one of the worst offenders) - interpret the DVD specifications in their own manner, and can easily produce non compliant streams. If you are worried about a video file, you need to either triple check that file, or else re-encode it.
    Let's look at the numbers on your files again, just to be certain I have got all this correct.
    2 Audio Streams. 320 & 448 = 768kbps.
    1 Video Stream, VBR, Peak to 8700kbps.
    If this is correct, IMHO you are pushing your luck somewhat. That adds to 9468kbps, leaving zip overshoot room for any unexpected spikes in the bitstream caused by sudden shifts in the material - and in 2-pass VBR this
    i will
    happen more often than you think.
    What are you writing your master to, please? (I really want to hear "DLT Tape" and not "DVD-R Disc" here as well)
    How long is the video stream?
    What is the average bitrate of your video file as set? If the target is as stated, 7500, I would forget all about using VBR here and encode to a fixed bitrate (assuming you have the space to do this, I would set the CBR to around 7,000 maybe 7500 and master to DLT, as at that bitrate a written disc may choke in players anyway, regardless of stream spec compliance.
    One final thing - and this comes from personal experience too - is to use the best MPEG-2 encoder you can get your hands on. MainConcept is okay for a lot of things, but it struggles in certain circumstances.
    It
    i can
    also give peaks that are well outside what you
    i think
    you have set as your upper limit.
    You mention that you used MPEG Validator. BUT - and this is a biggie - if one says PASS and the other (Bitrate Viewer, which is no longer being developed either, so is way out of date) says FAIL, I would look for confirmation somewhere down the line. MPEG Validator will - and I am sure you already know this - only validate the streams in PAL, not NTSC. There are varying reports as to how accurate it really is.

  • Version 2.1.0.63 - export data file is locked

    I followed the tutorial for Getting Started with Oracle SQL Developer - Adding Data to a Table
    After adding a total of 5 rows to the newly created table: DEPENDENTS I followed the instructions to export to a .csv by right clicking on a cell and exporting to a file e.g. C:\DEPENDENTS.csv
    Although I can find the file, I can't open it with Excel 2003 - I tried disconnecting from the User HR and SQLDeveloper appears to hold a lock on the file.
    If I quit SQLDeveloper I can open the file in Excel - otherwise I get a message file C:\DEPENDENTS.csv is locked for editing, by 'another user' e.g. SQLDeveloper.
    I would expect the file to be created and be available for immediate use.
    Kind regards
    Simon
    Edited by: user626089 on 18-Feb-2010 05:15

    Has been reported several times and will hopefully be fixed in the upcoming patch.
    Regards,
    K.

  • How to create the Export Data and Import Data using flat file interface

    Hi,
    Request to let me know based on the requirement below on how to export and import data using flat file interface.....
    Please provide the steps involved for the same.......
    BW/BI - Recovery Process for SNP data. 
    For each SNP InfoProvider,
    create:
    1) Export Data:
    1.a)  Create an export data source, InfoPackage, comm structure, etc. necessary to create an ASCII fixed length flat file on the XI
    ctnhsappdata\iface\SCPI063\Out folder for each SNP InfoProvider. 
    1.b)  All fields in each InfoProvider should be exported and included in the flat file. 
    1.c)  A process chain should be created for each InfoProvider with a start event. 
    1.d)  If the file exists on the target drive it should be overwritten. 
    1.e)  The exported data file name should include the InfoProvider technical name.
    1.f)  Include APO Planning Version, Date of Planning Run, APO Location, Calendar Year/Month, Material and BW Plant as selection criteria.
    2) Import Data:
    2.a) Create a flat file source system InfoPackage, comm structure, etc. necessary to import ASCII fixed length flat files from the XI
    ctnhsappdata\iface\SCPI063\Out folder for each SNP InfoProvider.
    2.b)  All fields for each InfoProvider should be mapped and imported from the flat file.
    2.c)  A process chain should be created for each InfoProvider with a start event. 
    2.d)  The file should be archived in the
    ctnhsappdata\iface\SCPI063\Archive directory.  Each file name should have the date appended in YYYYMMDD format.  Each file should be deleted from the \Out directory after it is archived. 
    Thanks in advance.
    Tyson

    Here's some info on working with plists:
    http://developer.apple.com/documentation/Cocoa/Conceptual/PropertyLists/Introduc tion/chapter1_section1.html
    They can be edited with any text editor. Xcode provides a graphical editor for them - make sure to use the .plist extension so Xcode will recognize it.

  • Removing unwanted control characters in exported text files

    I am currently evaluating Crystal Reports 2008 to determine applicability to our requirements. I need to export data files to continuous text to be read by other application software. I have successfully created the files but have what I believe to be page feed or end-of-page control characters (small rectangles) in the output. Can someone enlighten me as to how I can suppress or remove these control characters?

    In the export to text options enter 0 for the number of lines per page. This will produce an unpaginated text document without the page control markers.

  • Exporting catalog, making changes, then importing catalog does not update file movements and deletions

    I export part of my master catalog to a laptop.  I include the image files so I can edit the files at full res.  After managing and editing files I import the catalog back into the master catalog. When importing I check to replace metadata, develop settings and negative files. I have been running some tests to be sure all my work is being captured in the master catalog by this process.  I find that if I change develop settings or photo ratings this is detected when the file is imported and the data in the master catalog is updated properly.  The surprise is that if I delete files or move files among folders in the exported catalog these changes are not detected and these changes are not updated inthe master catalog.  It seems bizarre to me that such changes are not detected but I do not see how to get LR to recognize these changes and include them in the catalog update.  Without this capability I don't see how to use catalogs to move part of a catalog to another computer for edits and file management and then move back to the master catalog.

    There are various reasons why Lightroom works this way, but you'll need use pick flags to indicate photos to be deleted, and other metadata like colours or collections.

  • Essbase Data Export not Overwriting existing data file

    We have an ODI interface in our environment which is used to export the data from Essbase apps to text files using Data export calc scripts and then we load those text files in a relational database. Laetely we are seeing some issue where the Data Export calc script is not overwriting the file and is just appending the new data to the existing file.
    The OverWriteFile option is set to ON.
    SET DATAEXPORTOPTIONS {
         DataExportLevel "Level0";
         DataExportOverWriteFile ON;     
    DataExportDimHeader ON;
         DataExportColHeader "Period";
         DataExportDynamicCalc ON;
    The "Scenario" variable is a substitution variable which is set during the runtime. We are trying to extract "Budget" but the calc script is not clearing the "Actual" scenario from the text file which was the scenario that was extracted earlier. Its like after the execution of the calc script, the file contains both "Actual" and "Budget" data. We are not able to find the root cause as in why this might be happening and why OVERWRITEFILE command is not being taken into account by the data export calc script.
    We have also deleted the text data file to make sure there are no temporary files on the server or anything. But when we ran the data export directly from Essbase again, then again the file contained both "Actual" as well as "Budget" data which really strange. We have never encountered an issue like this before.
    Any suggestions regarding this issue?

    Did some more testing and pretty much zoomed on the issue. Our Scenario is actually something like this "Q1FCST-Budget", "Q2FCST-Budget" etc
    This is the reason why we need to use a member function because Calc Script reads "&ODI_SCENARIO" (which is set to Q2FCST-Budget) as a number and gives an error. To convert this value to a string we are using @member function. And, this seems the root cause of the issue. The ODI_Scenario variable is set to "Q2FCST-Budget", but when we run the script with this calculation function @member("&ODI_SCENARIO"), the data file brings back the values for "Q1FCST-Budget" out of nowhere in addition to "Q2FCST-Budget" data which we are trying to extract.
    Successful Test Case 1:
    1) Put Scenario "Q2FCST-Budget" in hard coded letters in Script and ran the script
    e.g "Q2FCST-Phased"
    2) Ran the Script
    3) Result Ok.Script overwrote the file with Q2FCST-Budget data
    Successful Case 2:
    1) Put scenario in @member function
    e.g. @member("Q2FCST-Budget")
    2) Results again ok
    Failed Case:
    1) Deleted the file
    2) Put scenario in a substitution variable and used the member function "@member("&ODI_Scenario") and Ran the script . *ODI_SCENARIO is set to Q@FCST-Budget in Essbase variables.
    e.g. @member("&ODI_SCENARIO")
    3) Result : Script contained both "Q1FCST-Budget" as well as "Q2FCST-Budget" data values in the text file.
    We are still not close to the root cause and why is this issue happening. Putting the sub var in the member function changes the complete picture and gives us inaccurate results.
    Any clues anyone?

  • How to export data from a Dynpro table to Excel file?

    Hi
    Here I go again. I read the post <b>Looking for example to export data from a DynPro table to Excel file</b> and put the code lines into a Web Dynpro Project where we need to export a dynpro table to Excel file but exactly at line 23 it doesn't recognize <b>workBook = new HSSFWorkbook();</b>
    1     //Declare this in the end between the Begin others block.
    2     
    3     private FileOutputStream out = null;
    4     private HSSFWorkbook workBook = null;
    5     private HSSFSheet hsSheet = null;
    6     private HSSFRow row = null;
    7     private HSSFCell cell = null;
    8     private HSSFCellStyle cs = null;
    9     private HSSFCellStyle cs1 = null;
    10     private HSSFCellStyle cs2 = null;
    11     private HSSFDataFormat dataFormat = null;
    12     private HSSFFont f = null;
    13     private HSSFFont f1 = null;
    14     
    15     //Code to create the Excel.
    16     
    17     public void onActionExportToExcel(com.sap.tc.webdynpro.progmodel.api.IWDCustomEvent wdEvent )
    18     {
    19     //@@begin onActionExportToExcel(ServerEvent)
    20     try
    21     {
    22     out = new FileOutputStream("C:/mydirectory/myfiles/testexcel.xls");
    23     workBook = new HSSFWorkbook();
    24     hsSheet = workBook.createSheet("My Sheet");
    25     cs = workBook.createCellStyle();
    26     cs1 = workBook.createCellStyle();
    27     cs2 = workBook.createCellStyle();
    28     dataFormat = workBook.createDataFormat();
    29     f = workBook.createFont();
    30     f1 = workBook.createFont();
    31     f.setFontHeightInPoints((short) 12);
    32     // make it blue
    33     f.setColor( (short)HSSFFont.COLOR_NORMAL );
    34     // make it bold
    35     // arial is the default font
    36     f.setBoldweight(HSSFFont.BOLDWEIGHT_BOLD);
    37     
    38     // set font 2 to 10 point type
    39     f1.setFontHeightInPoints((short) 10);
    40     // make it red
    41     f1.setColor( (short)HSSFFont.COLOR_RED );
    42     // make it bold
    43     f1.setBoldweight(HSSFFont.BOLDWEIGHT_BOLD);
    44     f1.setStrikeout(true);
    45     cs.setFont(f);
    46     cs.setDataFormat(dataFormat.getFormat("#,##0.0"));
    47     
    48     // set a thick border
    49     cs2.setBorderBottom(cs2.BORDER_THICK);
    50     
    51     // fill w fg fill color
    52     cs2.setFillPattern((short) HSSFCellStyle.SOLID_FOREGROUND);
    53     cs2.setFillBackgroundColor((short)HSSFCellStyle.SOLID_FOREGROUND);
    54     // set the cell format to text see HSSFDataFormat for a full list
    55     cs2.setDataFormat(HSSFDataFormat.getBuiltinFormat("text"));
    56     cs2.setFont(f1);
    57     cs2.setLocked(true);
    58     cs2.setWrapText(true);
    59     row = hsSheet.createRow(0);
    60     hsSheet.createFreezePane(0,1,1,1);
    61     for(int i=1; i<10;i++)
    62     {
    63     cell = row.createCell((short)i);
    64     cell.setCellValue("Excel Column "+i);
    65     cell.setCellStyle(cs2);
    66     }
    67     workBook.write(out);
    68     out.close();
    69     
    70     //Read the file that was created.
    71     
    72     FileInputStream fin = new FileInputStream("C:/mydirectory/myfiles/testexcel.xls");
    73     byte b[] = new byte[fin.available()];
    74     fin.read(b,0,b.length);
    75     fin.close();
    76     
    77     wdContext.currentContextElement().setDataContent(b);
    78     }
    79     catch(Exception e)
    80     {
    81     wdComponentAPI.getComponent().getMessageManager().reportException("Exception while reading file "+e,true);
    82     }
    83     //@@end
    84     }
    I don't know why this happen? Any information I will appreciate it.
    Thanks in advance!!!
    Tokio Franco Chang

    After test the code lines appears this error stacktrace:
    [code]
    java.lang.NoClassDefFoundError: org/apache/poi/hssf/usermodel/HSSFWorkbook
         at com.sap.tc.webdynpro.progmodel.api.iwdcustomevent.ExportToExcel.onActionAct1(ExportToExcel.java:232)
         at com.sap.tc.webdynpro.progmodel.api.iwdcustomevent.wdp.InternalExportToExcel.wdInvokeEventHandler(InternalExportToExcel.java:147)
         at com.sap.tc.webdynpro.progmodel.generation.DelegatingView.invokeEventHandler(DelegatingView.java:87)
         at com.sap.tc.webdynpro.progmodel.controller.Action.fire(Action.java:67)
         at com.sap.tc.webdynpro.clientserver.task.WebDynproMainTask.handleAction(WebDynproMainTask.java:101)
         at com.sap.tc.webdynpro.clientserver.task.WebDynproMainTask.handleActionEvent(WebDynproMainTask.java:304)
         at com.sap.tc.webdynpro.clientserver.task.WebDynproMainTask.execute(WebDynproMainTask.java:649)
         at com.sap.tc.webdynpro.clientserver.cal.AbstractClient.executeTasks(AbstractClient.java:59)
         at com.sap.tc.webdynpro.clientserver.cal.ClientManager.doProcessing(ClientManager.java:252)
         at com.sap.tc.webdynpro.serverimpl.defaultimpl.DispatcherServlet.doWebDynproProcessing(DispatcherServlet.java:154)
         at com.sap.tc.webdynpro.serverimpl.defaultimpl.DispatcherServlet.doContent(DispatcherServlet.java:116)
         at com.sap.tc.webdynpro.serverimpl.defaultimpl.DispatcherServlet.doPost(DispatcherServlet.java:55)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:392)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:266)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:345)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:323)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:865)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:240)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:92)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:148)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:37)
         at com.sap.engine.core.cluster.impl6.session.UnorderedChannel$MessageRunner.run(UnorderedChannel.java:71)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(Native Method)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:95)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:159)
    Thanks in advance!!!
    Tokio Franco Chang
    [/code]

  • LKKBLF99   runtime error while export data to the local file

    Hi experts,
    i got this error while i export data to the local file.
    Runtime Errors         OBJECTS_NOT_CHARLIKE
    Date and Time          10.09.2011 13:32:43
    Short text
         The current statement requires a character-type data object.
    Error analysis
        At the statment
             "STRLEN( obj )..."
        , only character-type data objects are supported at the argument
        position "obj".
        In this particular case, the operand "obj" has the non-charcter-type type
        "P".
    "OBJECTS_NOT_CHARLIKE"
    "SAPLKKBL" or "LKKBLF9
    "GEN_FIELD_OUT2"
    Include               LKKBLF99
      Überführung int. in ext. Format via Konvertierungsexit
        if not gs_fc-convexit is initial or                     "INS TECH11
           not gs_fc-inttype ca 'CN'.
      elseif not gs_fc-convexit is initial.            "DEL TECH11
          write  to gs_out-hlpchar left-justified.
          gs_out-hlplen = strlen( gs_out-hlpchar ).
      ansonsten entspricht sich min. Ausgabelänge und Ausprägungslänge
        else.
          if gs_fc-tech_form ne 99.
            gs_out-hlplen = strlen(  ).
          endif.
        endif.
      endif.
    the error says the operand "obj" has the non-charcter-type type " P" .
    i do debug on this and find the LS_FIELDCAT-INTTYPE = " P". so i changed it the QUAN type with length 13 decimal places 2. i can't change it in "C" char type or "NUMC" type  because i need the output in decimals. the output is for amount and quantity so need this in decimal places up to 2 or 3.
    but still find the same, i do debug on it after change it into QUAN type. find the same value for LS_FIELDCAT-INTTYPE = " P".
    let me know how can i define this in char type with decimal places or what.
    Please give some suggestion, if you need code, i will post the code.
    Thanks in advance
    Sachin

    Hi Sachin,
    Check this links.
    [Run time error|OBJECTS_NOT_CHARLIKE : ALV export to local file;
    [Error|Runtime error : OBJECTS_NOT_CHARLIKE;
    Regards,
    Madhu.

  • Unable to export data from Web Access Data Sheet in Sharepoint to local excel or access file

    Greetings and good morning.
    I'm going to start off in broad terms with this question because I'm not 100 percent sure what information to provide.
    Long story short, we've got a Web Access Data Sheet list hosted in a Sharepoint 2010 environment. It is accessed and used by multiple people throughout the day. It contains several thousand line item entries. I'd call it a large data sheet.
    I think the size of the data sheet is casuign some instability in the list. I'd like to be able to export a defined range of data from the list into a local excel or access file. After that, I'd delete the stuff on the Access list to improve performance.
    But...when I attempt to use the Sharepoint Action bar to export - Excel locks up/crashes. If I try to export to Access, I get a similar issue.
    Any ideas? Could anyone begin by telling me what other information is required?

    Hi,
    If you would like to export data from Access Web Database in SharePoint 2010, you could go to Design With Access page in Settings. The url in my environment is http://sp/tt/_layouts/accsrv/ModifyApplication.aspx . Then choose the Table and export it to Excel
    or Modify it in Access.
    Regards,
    Rebecca Tu
    TechNet Community Support

  • Is it possible to export the "Mediathek" in a data file (excel or word) ?

    Is it possible to export the "Mediathek" in a data file (excel or word) ?

    see my old post:
    https://discussions.apple.com/message/6272619?messageID=6272619#6272619?messageI D=6272619

  • Export data from a table to text file using srcipt task

    Hi
    i am new to SSIS
    i have to export data from a table and append it into a existing file through SSIS script task
    please help
    Thanks
    Umesh

    Hi Umesh,
    The data structure of the source table and the structure of the destination file are the same, right? Is the destination file a flat file? Do you have to do it through Script Task? If the destination file is a flat file, this can be done easily by using
    the stock tasks/components other than .NET code. In the Data Flow Task, we choose the appropriate source adapter (such as OLE DB Source or ADO.NET Source) to extract data from the source table, perform transformation if necessary, and then load to the destination
    file via a Flat File Destination. When setting up the Flat File Destination, uncheck the “Overwrite data in the file” option so that the extracted data will be appended to the existing file.
    If you need to implement it through Script Task/Component indeed, you may benefit from the following code examples:
    http://stackoverflow.com/questions/8070163/how-to-add-custom-footer-to-an-ssis-flat-file-seperate-component-or-script-tas 
    http://stackoverflow.com/questions/8467326/add-header-and-footer-row-flat-file-ssis 
    If you need further help about the script, I suggest that you ask a new question in .NET forums where you can get more dedicated support:
    http://social.msdn.microsoft.com/Forums/vstudio/en-US/home?category=netdevelopment 
    Regards,
    Mike Yin
    TechNet Community Support

Maybe you are looking for

  • Use one agent for two different managment servers

    Hi , Can one agent be connected to two different managment server in different domains , or do i have to install two agents in the same server to connect to each managment server . thanks

  • JVM help

    hello, I am trying to get the hang of this JNI and JVM stuff...killing me. I am trying to get a C++ application to interrupt a java program that is in idle or just have it interrupt a certain thread then I can have the java application do something.

  • RMAN failure for "archive-log ... not found in controlfile"

    Recovery Manager: Release 8.1.6.0.0 - Production RMAN> connect target sys/XXXX@cldb; 2> connect catalog rman/XXXX@rmandb1; 3> 4> resync catalog; 5> 6> # Backup the database 7> run { 8> allocate channel ch1 type disk; 9> backup incremental level 0 for

  • Unable to connect to imap.aol.c​om please verify your server name entry.

    When I try to get or send mail with AOL on my Palm Treo 700p I get this VersaMail error: Unable to connect to imap.aol.com please verify your server name entry. All I have done today is have COX cable install a new modem for my desktop.  Any ideas wh

  • Using aliases in iTunes (with a twist)

    Okay, here is a whopper for everyone!  I have a Drobo attached to my Mac Mini which allows for my tremendous amount of data to be accessed at any given moment.  Among this data is my PDF library that is just over 20 GB.  My primary iTunes library is