Recalling Spreadsheets and appending results into 1 file.

Hi,
i am in a stage of my program where i have to do numerical analysis on a set of data collected by an Optical spectrum analyser... the data has been saved in multiple spreadsheets... (each with a row for the wavelengths, x-values, and power, y-values)...i need to get the power values and put them all in 1 spreadsheet file, with the wavelngth values in the first row.
so multiple 2D spreadsheets: Wavelength:  x1 x2 x3 .....
     Power 1     : P1 P2 P3....
     Wavelngth: x1 x2 x3...
     Power 2    : p1 p2 p3...
Will turn to: Wavelength:  x1 x2 x3...
Power11 P2 P3...
Power21 p2 p3...
attached below is my attempt at doing this, which doesnt seem to work. The loops are present so that 'read from spreadsheet' vi recalls all the spreadhseets...which it does do... but i don't know how to improve it so that all the files succesfully append into 1 2D array (as shown above).
Please can someone help me figure out how to do this (recall multiple spreadsheets and then save into 1). right now only the files being recalled in the last loop iteration are saved into the single spreadsheet. 
the file paths of the spreadsheets are all dependent on the number (N) so the file paths are such that 00.txt,01.txt,02.txt,03.txt...0N.txt, 10.txt,11.txt...1N.txt,20.txt,21.txt... etc ...NN.txt and teh final file will be RESULTS.txt (as seen in the pic)
Any input will be greatly appreciated, it seems like an easy concept, but the answer totally evades me.  
thanks a lot,
Asiri  
Attachments:
Savespreadsheet.jpg ‏121 KB

as you can see in the image, i have used the required VIs to recall and save the spreadsheets, what i am having problems with is appending ALL of the spreadsheets, my loops are not working properly, i only append a select number of spreadsheets/arrays together dependent on the loop iterations,
perhaps its something wrong with the way i got everything wired? i index at the outer loop, however this information gets lots in subsequent iterations... what i need ideally is some sort of VI which inserts sub arrays into consequative rows in an array though just one input (such as in my loops).
any suggestions are greatly appreciated.  
thank you!
asiri  
Attachments:
Savespreadsheet.jpg ‏121 KB

Similar Messages

  • Hello Anybody, I have a question. Can any of you please suggest me how to make an xml file from the database table with all the rows? Note:- I am having the XSD Schema file and the resulted XML file should be in that XSD format only.

    Hello Anybody, I have a question. Can any of you please suggest me how to make an xml file from the database table with all the records?
    Note:- I am having the XSD Schema file and the resulted XML file should be in that XSD format only.

    The Oracle documentation has a good overview of the options available
    Generating XML Data from the Database
    Without knowing your version, I just picked 11.2, so you made need to look for that chapter in the documentation for your version to find applicable information.
    You can also find some information in XML DB FAQ

  • How to call a SP with dynamic columns and output results into a .csv file via SSIS

    hi Folks, I have a challenging question here. I've created a SP called dbo.ResultsWithDynamicColumns and take one parameter of CONVERT(DATE,GETDATE()), the uniqueness of this SP is that the result does not have fixed columns as it's based on sales from previous
    days. For example, Previous day, customers have purchased 20 products but today , 30 products have been purchased.
    Right now, on SSMS, I am able to execute this SP when supplying  a parameter.  What I want to achieve here is to automate this process and send the result as a .csv file and SFTP to a server. 
    SFTP part is kinda easy as I can call WinSCP with proper script to handle it.  How to export the result of a dynamic SP to a .CSV file? 
    I've tried
    EXEC xp_cmdshell ' BCP " EXEC xxxx.[dbo].[ResultsWithDynamicColumns ]  @dateFrom = ''2014-01-21''"   queryout  "c:\path\xxxx.dat" -T -c'
    SSMS gives the following error as Error = [Microsoft][SQL Server Native Client 10.0]BCP host-files must contain at least one column
    any ideas?
    thanks
    Hui
    --Currently using Reporting Service 2000; Visual Studio .NET 2003; Visual Source Safe SSIS 2008 SSAS 2008, SVN --

    Hey Jakub, thanks and I did see the #temp table issue in our 2008R2.  I finally figured it out in a different way... I manage to modify this dynamic SP to output results into
    a physical table. This table will be dropped and recreated everytime when SP gets executed... After that, I used a SSIS pkg to output this table
    to a file destination which is .csv.  
     The downside is that if this table structure ever gets changed, this SSIS pkg will fail or not fully reflecting the whole table. However, this won't happen often
    and I can live with that at this moment. 
    Thanks
    --Currently using Reporting Service 2000; Visual Studio .NET 2003; Visual Source Safe SSIS 2008 SSAS 2008, SVN --

  • File Server - File size\type search and save results to file

    I already have a vb script to do what I want on our file server, but it is very inefficient and slow.  I was thinking that a powershell script may be more suitable now but I don't know anything about scripting in PS.  So far the vb code that I
    have works, and I am not the one who wrote it but I can manipulate it to do what I want it to.  The only problem is, when I scan the shared network locations it stops on some files that are password protected and I don't know how to get around it.  If
    someone else knows of a PS script to go through the file system and get all files of a certain type or size (right now, preferably size) and save the file name, size, path, owner and dates created\modified please point me to it and I can work with that.  If
    not, could I get some help with the current script that I have to somehow get around the password protected files?  They belong in a users' HOME directory so I can't do anything with them.  Here is my code:   
    'Script for scanning file folders for certain types of files and those of a certain size of larger'
    'Note: Script must be placed locally on whichever machine the script is running on'
    '***********VARIABLES FOR USE IN SCRIPT***********'
    'objStartFolder - notes the location of the folder you wish to begin your scan in'
    objStartFolder = "\\FileServer\DriveLetter\SharedFolder"
    'excelFileName - notes the location where you want the output spreadsheet to be saved to'
    excelFileName = "c:\temp\Results_Shared.xls"
    '**********END OF VARIABLES**********'
    Set objFSO = CreateObject("Scripting.FileSystemObject")
    fileName = objFSO.GetFileName(path)
    'beginning row and column for actual data (not headers)'
    excelRow = 3
    excelCol = 1
    'Create Excel Spreadsheet'
    Set objExcel = CreateObject("Excel.Application")
    Set objWorkbook = objExcel.Workbooks.Add()
    CreateExcelHeaders()
    'Loop to go through original folder'
    Set objFolder = objFSO.GetFolder(objStartFolder)
    Set colFiles = objFolder.Files
    For Each objFile in colFiles
    Call Output(excelRow) 'If a subfolder is met, output procedure recursively called'
    Next
    ShowSubfolders objFSO.GetFolder(objStartFolder)
    'Autofit the spreadsheet columns'
    ExcelAutofit()
    'Save Spreadsheet'
    objWorkbook.SaveAs(excelFileName)
    objExcel.Quit
    '*****END OF MAIN SCRIPT*****'
    '*****BEGIN PROCEDURES*****'
    Sub ShowSubFolders(Folder)
    'Loop to go through each subfolder'
    For Each Subfolder in Folder.SubFolders
    Set objFolder = objFSO.GetFolder(Subfolder.Path)
    Set colFiles = objFolder.Files
    For Each objFile in colFiles
    Call Output(excelRow)
    Next
    ShowSubFolders Subfolder
    Next
    End Sub
    Sub Output(excelRow)
    'convert filesize to readable format (MB)'
    fileSize = objFile.Size/1048576
    fileSize = FormatNumber(fileSize, 2)
    'list of file extensions currently automatically included in spreadsheet report:'
    '.wav, .mp3, .mpeg, .avi, .aac, .m4a, .m4p, .mov, .qt, .qtm'
    If fileSize > 100 then'OR objFile.Type="Movie Clip" OR objFile.Type="MP3 Format Sound" _ '
    'OR objFile.Type="MOV File" OR objFile.Type="M4P File" _'
    'OR objFile.Type="M4A File" OR objFile.Type="Video Clip" _'
    'OR objFile.Type="AAC File" OR objFile.Type="Wave Sound" _'
    'OR objFile.Type="QT File" OR objFile.Type="QTM File"'
    'export data to Excel'
    objExcel.Visible = True
    objExcel.Cells(excelRow,1).Value = objFile.Name
    objExcel.Cells(excelRow,2).Value = objFile.Type
    objExcel.Cells(excelRow,3).Value = fileSize & " MB"
    objExcel.Cells(excelRow,4).Value = FindOwner(objFile.Path)
    objExcel.Cells(excelRow,5).Value = objFile.Path
    objExcel.Cells(excelRow,6).Value = objFile.DateCreated
    objExcel.Cells(excelRow,7).Value = objFile.DateLastAccessed
    excelRow = excelRow + 1 'Used to move active cell for data input'
    end if
    End Sub
    'Procedure used to find the owner of a file'
    Function FindOwner(FName)
    On Error Resume Next
    strComputer = "."
    Set objWMIService = GetObject("winmgmts:" _
    & "{impersonationLevel=impersonate}!\\" & strComputer & "\root\cimv2")
    Set colItems = objWMIService.ExecQuery _
    ("ASSOCIATORS OF {Win32_LogicalFileSecuritySetting='" & FName & "'}" _
    & " WHERE AssocClass=Win32_LogicalFileOwner ResultRole=Owner")
    For Each objItem in colItems
    FindOwner = objItem.AccountName
    Next
    End Function
    Sub CreateExcelHeaders
    'create headers for spreadsheet'
    Set objRange = objExcel.Range("A1","G1")
    objRange.Font.Bold = true
    objExcel.Cells(1, 1).Value = "File Name"
    objExcel.Cells(1, 2).Value = "File Type"
    objExcel.Cells(1, 3).Value = "Size"
    objExcel.Cells(1, 4).Value = "Owner"
    objExcel.Cells(1, 5).Value = "Path"
    objExcel.Cells(1, 6).Value = "Date Created"
    objExcel.Cells(1, 7).Value = "Date Modified"
    End Sub
    Sub ExcelAutofit
    'autofit cells'
    Set objRange = objExcel.Range("A1")
    objRange.Activate
    Set objRange = objExcel.ActiveCell.EntireColumn
    objRange.Autofit()
    Set objRange = objExcel.Range("B1")
    objRange.Activate
    Set objRange = objExcel.ActiveCell.EntireColumn
    objRange.Autofit()
    Set objRange = objExcel.Range("C1")
    objRange.Activate
    Set objRange = objExcel.ActiveCell.EntireColumn
    objRange.Autofit()
    Set objRange = objExcel.Range("D1")
    objRange.Activate
    Set objRange = objExcel.ActiveCell.EntireColumn
    objRange.Autofit()
    Set objRange = objExcel.Range("E1")
    objRange.Activate
    Set objRange = objExcel.ActiveCell.EntireColumn
    objRange.Autofit()
    Set objRange = objExcel.Range("F1")
    objRange.Activate
    Set objRange = objExcel.ActiveCell.EntireColumn
    objRange.Autofit()
    Set objRange = objExcel.Range("G1")
    objRange.Activate
    Set objRange = objExcel.ActiveCell.EntireColumn
    objRange.Autofit()
    End Sub
    David Hood

    Accessing Excel through automation is bvery slow no matter what tool you use.  Scanning a disk is very slow for all tools.
    Since Vista all system have a search service that catalogues all major file itmes like size, extension, name and other attributes.  A search of a 1+Tb  volume can return in less that a second if you query the search service.
    You can easily batch the result into Excel by writ4ing to a CSV and opening in Excel. Use a template to apply formats.
    Example.  See how fast this returns results.
    #The following will find all log files in a system that are larger than 10Mb
    $query="SELECT System.ItemName, system.ItemPathDisplay, System.ItemTypeText,System.Size,System.ItemType FROM SystemIndex where system.itemtype='.log' AND system.size > $(10Mb)"
    $conn=New-Object -ComObject adodb.connection
    $conn.open('Provider=Search.CollatorDSO;Extended Properties="Application=Windows";')
    $rs=New-Object -ComObject adodb.recordset
    $rs.open($query, $conn)
    do{
    $p=[ordered]@{
    Name = $rs.Fields.Item('System.ItemName').Value
    Type = $rs.Fields.Item('System.ITemType').Value
    Size = $rs.Fields.Item('System.Size').Value
    New-Object PsObject -Property $p
    $rs.MoveNext()
    }Until($rs.EOF)
    ¯\_(ツ)_/¯

  • Very slow painting while reading and writing doubles into file

    for 15MB length file i = 7662080
    for 50MB length file i = 12414368
    Part of Code for writing into file follows like this:
    try{
    fos = new FileOutputStream("Angel.txt");
    File f = new File("Angel.txt");
         if(f.length() >=4)
         f.delete();
    fos = new FileOutputStream("Angel.txt");     
    dos = new DataOutputStream(new BufferedOutputStream(fos,1000000));
    int x=0;
    double y_last, y_new;
    for(int j=0 ;j<i ;j++)
    if(some condition)
    y_new = ....;
    try{
    //previously in vectors
    y_last = y_new;
    vect.add(new Line2D.Double(x, y_last, x, y_new)_;
    dos.writeDouble(y_new);
         }catch(Exception e){System.out.println(e);}
    dos.close();
    fos.close();
    x++;
    }catch(Exception excp){System.out.println(excp);}
    part of code for reading from file follows like this:
    public void paint(Graphics g)
    try{
         double y1, y2 =0;               
         Line2D.Double doub;
         raf = new RandomAccessFile("Angel.txt","r");
         dis = new DataInputStream(new BufferedInputStream(new FileInputStream(raf.getFD(),1000000)));
         raf.seek((rect.x*8));
         for (int i = 0/any value; (i < value as per choice); i++)
              g2.setStroke(new BasicStroke(0)); //2
              y1 = y2;
         y2 =dis.readDouble();
              doub=new Line2D.Double(i,y1,i,y2);
              g2.draw(doub);
    dis.close();
         raf.close();
    }catch(Exception excp){System.out.println(excp);}
    I tried using Object Streams but NotSerializable Exception is thrown as Line2D.Double objects
    are not serialized.
    Any idea to make reading and writing into file specially from MB files faster is appreciated.

    Why are you reading in the file in the paint method ?
    Create your data once before painting.
    I think you should explain what is your goal and what behavior you want.
    Denis

  • Insert a spreadsheet and word docs into a PDF with hyperlinks

    I have a spreadsheet index with hyperlinks to 455 single page word docs. I want to make a single pdf by inserting the spreadsheet followed by the word docs.
    Is there a way of having the hyperlinks recognize the inserted word docs pages inside the pdf via the original hyperlinks, or do I need to remove the hyperlinks before inserting the spreadsheet and then do bookmarks for all 455 pages? Or some other method?

    James,
    I doubt that will work. But it should be easy enough to find out. If you
    print to pdf from ID, I guarantee is will not work. If you export from
    ID, I doubt it will work, but ask in the ID forum, since the creation
    does not depend upon Acrobat.
    Mike

  • 8.1 EE can't find page and appends the welcome-file name to the url

    I have a problem where you enter a url that points to a static file in the xxxEar-WebModule_war directory. That is the "docroot" directory for the application. Somewhere along the line the welcome-file (index.html) is appended to the url. For example I enter http://server.com/app/file.xml and it gets changed to http://server.com/app/file.xml/index.html, which is not found.
    This was not a problem in 8.1 platform. Now that we upgraded to 8.1 enterprise it is a problem.
    All urls that are mapped to servlets work fine. I just can't seem to access static content.
    Thanks!

    select l.group#, f.member, l.sequence#, l.status, f.status
    from v$log l, v$logfile f
    where l.group#=f.group#
    order by 1;
    GROUP#
    MEMBER
    SEQUENCE# STATUS STATUS
    1
    D:\ORACLE\ORADATA\DBMAG98\LOG1.RDO
    41002 INACTIVE STALE
    2
    D:\ORACLE\ORADATA\DBMAG98\LOG2.RDO
    41003 INACTIVE
    GROUP#
    MEMBER
    SEQUENCE# STATUS STATUS
    3
    D:\ORACLE\ORADATA\DBMAG98\LOG3.RDO
    41004 CURRENT STALE
    sorry for inconvenient view

  • Check group membership and export result to file

    Hi everyone,
    Could you please assist?  I need a script that get's a samAccountName from a text or csv file, then checks if the user belongs to Group A or Group B or Group C.  If user belongs to group A, then output "Group A" and so on.  Sample
    output file:
    JDoe Group A
    JSmith Group B
    KSmith  Group C
    DDoe Group A Group B Group C
    Thanks in advance,

    Hi Phathuynh,
    We always recommend you can start your own script, and post the current script and the issue for a better response.
    In addition, you can also refer to the cmdlet "Get-ADGroupMember" to list the members in group, and the script below is for your reference:
    $output = @()
    $groups = "user1","user2"
    $array1 = Get-ADGroupMember Group1|select -ExpandProperty samaccountname
    $array2 = Get-ADGroupMember Group2|select -ExpandProperty samaccountname
    Foreach($u in $groups){
    $user = Get-ADUser $u|select -ExpandProperty samaccountname
    $param = @{"user" = $user
    "group A" = ""
    "group B" = ""}
    $p = New-Object Psobject -Property $param
    if ($array1 -contains $user){
    $p."group A"= "YES"
    if ($array2 -contains $user){
    $p."group B" = "YES"
    $output += $p
    $output
    The screenshot of result:
    If there is anything else regarding this issue, please feel free to post back.
    Best Regards,
    Anna Wang
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Selecting multiple records from the list and append it into internal table.

    hi
    there is a list of 10 records in the output screen with check boxes. if the user select any 3 records then those records should get appended in some other internal table of same structure.
    this is the requirement. can anybody help me it is very urgent.
    regards
    sami

    Hi,
    Loop at the internal table which contains the value in table control and move it to a work area. apend the work area to the next internal table. Like:
    loop at it1 into wa1 where sel = 'X'.
    append wa1 to it2.
    endloop.
    or
    loop at it1 into wa1.
    if wa1-sel = 'X'.
    append wa1 to it2.
    endif.
    endloop.
    Regards,
    RM

  • Submit report in background and get result into calling program

    Hi,
    I want to call the standard SAP program from Zprogram to create invoice using Submit. here the code
    SUBMIT RV60SBT1  TO SAP-SPOOL
                        SPOOL PARAMETERS print_parameters
                        WITHOUT SPOOL DYNPRO
                        VIA JOB name NUMBER number
                        WITH VKOR1 eq TVKO-VKORG
                        with X_VBELN eq gv_vbeln_so
                        with ALLEA eq 'X'
                        with ALLEL eq ''
                        AND RETURN.
    the standard SAP program creates invoice (billing document number). I want to get the billing document number which created using the above statement into the calling program.  ......could anyone pls tell me how to do this? ie how to read the billing document number in the calling program.
    Regrds
    shan
    Edited by: Shankar Raju Devadoss on Mar 21, 2011 9:00 AM

    Hi Shankar,
    You can try it in both ways.
    1.   By using the parameter ID VF. (GET PARAMETE ID)
    2.   Select the invoice numbers from the table VBRK for the current date (SY-DATUM) and
           sort it by time and get the latest invoice.
    Regards
    Hareesh Menon

  • Linking three databases and combing results into one resultset

    Hi Folks,
    I would like to receive input of respectable members of this forum for this problem. Please see details below and let me know what would be the best approach to implement it.
    Problem statement:I need to present data from three Oracle databases in the form of a report based upon parameter values entered by users.
    Problem Details:I have three tables which have same structure but different data. Each table is in a different database, i-e:
    DB1.TableX (say: DB1 holds data for Europe)
    DB2.TableX (DB2 holds data for Asia)
    DB3.TableX (DB3 holds data for North America)
    User Parameters:If p_country = 'ENGLAND' then go and fetch data from DB1.TableX only
    If p_country in ('ENGLAND', 'U.S.A') then go and fetch data from DB1.TableX and DB3.TableX etc.
    Users:Users of DB1, DB2 ,and DB3 will be running this report from their respective dbs so users of DB1 will be able to see data of DB2 and DB3 as well and the vice versa.

    dreporter wrote:
    User Parameters:If p_country = 'ENGLAND' then go and fetch data from DB1.TableX only
    If p_country in ('ENGLAND', 'U.S.A') then go and fetch data from DB1.TableX and DB3.TableX etc.You do not want to hit and process remote tables when they do not have the required rows.
    There's an old feature, dating back to Oracle v7 and still around, called partitioned views. Simply put, you create a view on the tables by using union all.
    A constraint per table specify the expected rows in that table. Oracle's CBO uses this constraint to determine which tables in the partition view to query (assuming of course the constraint column is used).
    For example, you have 5 organisations (same table structure) and a table per organisation. You add a organisation column to each of the tables, with a constraint on the column specifying which organisation's data is in that table. So table1 will have a check constraint on ORG_CODE that says 'BBC', and table2 will have the same constraint with value 'CNN', and table3 will have constraint with value 'SKY', etc.
    The partition view is a plain vanilla view. E.g.
    <i>create or replace view organisations as
    select * from table1 union all
    select * from table2 union all
    select * from table3 union all
    select * from table4 union all
    select * from table5</i>
    If you now query the partition view and use ORG_CODE = 'SKY', then only table3 will be queried and the other tables will not (the CBO prunes the other tables from the view using a special predicate).
    This works pretty well - especially when you only have Oracle Standard Edition. The question though is whether this will work across db links. But it is worth finding out and having a look at this option.

  • Logo in Excel Sheet and Appending Data under it

    Dear All,
    I have a requirement to insert Company Logo in excel sheet.
    Is it possible?
    Currently we have decided to have the logo already in the spreadsheet and append the data in excel under the logo.
    Any directions towards to development would be helpful. Thankyou.
    Regards,
    Durvi.

    Hi Guys.
    Sorry just answering my own question.
    The requirement was acheived by having logo already in the file.
    But Vijay's answer was good one, archived it already. I did not find the thread you provided. Already awarded you points.
    Thanks for your replies.
    Cheers.

  • Append an existing file?

    Hello world,
    I'm stuck in a predicament: The storage guys have asked me to reduce the frequency of backups on three of our instances to save bandwidth on the WAN when the backup files are copied across to another datacenter. The application owners do not want to backup
    any less frequently. 
    We do daily full backups and Tr-log backups every 15 minutes throughout the day.
    Storage have advised that the SAN only copies changes (ie new files or the changes to existing ones)
    Is it possible/ would it be of any benefit, to backup and append an existing file? Can I do this with the logs as with the databases?
    Thanks in advance

    Storage have advised that the SAN only copies changes (ie new files or the changes to existing ones)
    I am not very sure that it will pick changes from existing ones, I think if it sees the file is changed it will start coping as 1 big file and will not solve your purpose. You might get it reconfirm with storage on this. Also even if it does like you said
    I wonder if next backup runs and file is still getting copied over then backup will start failing saying "file is in use".
    If application owner is saying 15 min log backup then we cant reduce the frequency, but if you have some other drive/disk where you can divert tlog backup say for 2 hours then you can run differential backup every 2 hour which goes to SAN. You will be safe
    even if something happen you have tlog on 1 drive and
    differential get copied to other datacenter. (But point in time recovery will be issue as you will not have any tlog backups for restore)
    I cant think of anything else to reduce frequency, but for bandwidth see if you can use compress(enterprise edition) or zip backup file to save bandwidth.

  • Save digitizer and DIO waveforms into one spreadsheet file

    I am using attached mixedsignalscopeusingtclk.vi for data acquisition and would like to acquire 4-channel of data. My PXI-1042 system has PXI-5122 and PXI-6552, and I am using 2 channels of digitizer and 2 channels of digital I/O. I am using Labview 2010 SP1. The VI generated 1-D array of cluster for analog waveform and 1-D array of digital waveform. How do I save both waveforms to one spreadsheet file?
    Solved!
    Go to Solution.
    Attachments:
    mixedsignalscopeusingtclk.vi ‏258 KB

    Hi,
    You can use a Write to Measurement File express VI. Note that this is the easiest way however not the most efficient way to write to a file.
    Since you have the fetch VIs inside for loops, your output is a 1D array of waveforms. You will need to obtain the waveform element and convert it into a type that the express VI can read. If you unbundle the waveform and grab only the data and connect it straight to the express VI, LabVIEW will automatically place a "Convert to Dynamic Data" block in line to make it compatible. Same applies when you connect the boolean array to the Signals input of the express VI. And if you connect more than one signal to this input, LabVIEW will automatically place a "Merge Signals" block.
    If you reproduce the below code, you should be able to write the two channels to the same measurement file. If you have more than one channel of each type, you can grab multiple elements using the same Index Array block and take them through the same process shown below.
    Hope this helps.
    Tarek B
    Applications Engineer
    National Instruments

  • Reading data in XML files and outputting information into xl spreadsheet

    Can anyone help??? i want to read an XML file, and after reading the file, store certain data into a XL speadsheet. i am new to all this and dont know where to start. Can anyone provide me with coded examples or links to certain sites that can provide me with examples. I would be very grateful, thanks!!!

    you will either have to use a third party toolkit or develop your own for this purpose. In case you develop your own, please be kind to share it with me ;)
    I am not sure about the free ones. But Sitraka is know for developing such kits. For instance you may use sitraka's JClass which will be more then what you need. I guess JClass is free for evaluation, but you have to pay them after that.
    all the best, and update this thread if you find anything for free !!
    Here are some usefull inks(It seems the very first link provide some free and useful apis. Let me know, if you are able to use them)
    http://www.andykhan.com/jexcelapi/
    http://www.sitraka.com/software/jclass/
    http://www.grapecity.com/india/Product/Database/ (From the JAva drop down, select Formula One for Java)
    Rakesh

Maybe you are looking for

  • Problem with javascript in narrative view

    Hi gurús! I'm trying to put some conditional logic in a narrative view of my analysis. I found some tips, and wrote that: <script> if('@{FIN01TIME}{ }'=='Period') document.write("[b]Fiscal Year : [/b]@1       [b]Fiscal Quarter : [/b]@2       [b]Fisca

  • SOLVED: The file or folder is corrupted and unreadable.

    On a Windows 7 Pro workstation, whenever a user logs in, they receive an error that the file C:\Windows\System32\en-us\winlogon.exe.mui is missing or corrupted. Every normal restart of Windows automatically invokes chkdsk c: /f. I've tried the follow

  • 2nd display/monitor not working (strange problem)

    hello, I have a mac pro with this specification... Model Name:          Mac Pro   Model Identifier:          MacPro1,1   Processor Name:          Dual-Core Intel Xeon   Processor Speed:          2.66 GHz   Number Of Processors:          2   Total Num

  • How to use MacBook in the Arctic?

    I will need to use my MacBook in -30 degree Celcius temperatures. Is it safe to do so? Will my MacBook function? Do I need to keep it heated? What recommendations do you have for using the computer in extreme temperatures?

  • EJB Question Please help me

    Hi I am preparing for certification exam. I don't know the correct answer of these question. Please help me. Regards Arghya Banerjee ======================================================================== 1. While testing a BMP Entity Bean, a develo