Proper format of cells after Writing to csv file?

I have been searching, yet struggling with what is probably pretty simple to do.  I am writing data to a csv but my data is not coming out the way I need it to look in the spreadsheet once its been written.  Each of the channel's data from the front panel need to be in their own Column.  Please see attached csv for what I'm aiming to accomplish in a finished program.  As of right now, I can write to the csv file but the data and fields are not in their proper columns/rows.  Can someone get me jump started on this please?  I'm very new and green in Labview.  This is my first program.  Thank you. 
Solved!
Go to Solution.
Attachments:
Front_Panel.vi ‏23 KB
CSVdata.csv ‏1 KB

It is recommended to turn on context help (ctrl+H) so you can get a brief description of stuff you mouse over.  You can get detailed information by clicking on "details".
For example, you would have seen the terminal for "transpose" in the context window.  Curious to see what it did, you would have clicked on "details".  That would have shown you what "transpose" does.
I hate it that Microsoft trains you to ignore help files because they write them so poorly.
Bill
(Mid-Level minion.)
My support system ensures that I don't look totally incompetent.
Proud to say that I've progressed beyond knowing just enough to be dangerous. I now know enough to know that I have no clue about anything at all.

Similar Messages

  • How to update a specific cell value in a CSV file.

    Hi
    I would like to know whether it is possible in openscript to update a cell value of the CSV file.
    For example if the test script requires to update the cell of 6th row and 8th column, then how to achieve through openscript API?
    Please kindly let me know.

    Hi,
    Hope this helps!
    Table table;
    String[] rowvalus;
    int targetRow=5,targetColumn=7;
    * Note: Rows and columns are 0-based index.
    table =utilities.loadCSV("C:\OracleATS\OFT\DataBank\Test.csv");
    rowvalus=table.getRows().get(targetRow).getAll();
    info("Before change :"+rowvalus[targetColumn]);
    rowvalus[targetColumn]="NewValue";
    utilities.saveCSV(table, "C:\OracleATS\OFT\DataBank\Test.csv", true);
    table =utilities.loadCSV("C:\OracleATS\OFT\DataBank\Test.csv");
    rowvalus=table.getRows().get(targetRow).getAll();
    info("After change :"+rowvalus[targetColumn]);
    Note:- Please first take the backup of your CSV file before using this code snippet
    Thanks
    -POPS

  • Problem in writing to CSV file

    Hi all,
    i am retreiving a long String which is not in a proper format (ie) Lot of Spaces in between and my String contains comma(,) seperated and I am trying to put that string in to a CSV file. I am able to write the data in the CSV file but not in a proper format.i need to write the data in the below format.
    For Ex:
    Emp-No,Address,Name,Manager -------> this is Heading
    1000,abcd,name,Man-name -----> After i have to put the data in next line
    My String will be like this,
    abcd , , , jsjsj, , ,saalkj,jfjfjfjf,mmm, nnn, , ,
    I am able to get the data properly in the CSV format.
    Please help me to solve this. Thanks in Advance!!!!

    Are you using any reporting frameworks like Jasper
    Reports or you using simple select queries in MySQL
    to export the data to CSV format.Detail more clearly
    to get assist from others.
    Are you using Session beans or Entity beans on Weblogic?no no no, he can create the csv file, whether its in the correct format is the problem of the consumer of this csv file, isnt'it?
    are you using a microsoft product?

  • Format check for each column in csv file

    i want to find out how to check each column value is correct or not while uploading flat file(csv).
    after uploading to internal table,
    it should display error record and line no of the record.
    ex matnr(18) type c,
        werks(4) type c.
    if we input matnr value more than 18 characters in csv file,
    the entire record of internal table should be displayed in log file.
    it should validate before sap database validations.

    Hi,
    Since the file format is csv, while uplaoding the data to the program using the function module, the importing FILE FORMAT is 'ASC' and not 'DAT'.
    For this the Internal table should contain only one filed of type C length (255 or 1024, upto you).
    Then after this using SPLIT command you will push into another Internal table.
    This is where you have to perform the validations.
    In my opinion instead of checking the number of characters, chcek whether the material or plant exists in SAP or not?
    This is because you have already fixed the filed length of material to 18 charecters, obvious when you try to move a variable with more than 18 charecters to this filed, it will automatically truncate and take only 18 charecters..
    Hope I haev answered you. If still have any questions. shoot.
    BR
    Sree

  • To show zeroes after decimal in CSV file

    Hi all,
    A program generates a .CSV file , downloaded to local PC.
    The values should be shown as 2.0000(upto four decimals) and 0 should be 0.0000.
    I tried appending ' - single quote to char field , it's working fine but it's shown
    as '2.0000 in CSV file.
    Is there any other way to show the values with four decimals as 2.0000 ?
    Thanks in advance
    Kaavya.

    Hi,
    Once your value with 4 decimals is ready in variable with type P, move it to anew 'character' variable & use this variable for downloading.
    I hope this helps,
    Regards
    RBC

  • How to Format DataGridView Cells after user input, also find average of Timespans

    I'm a rather new to programming and slowly learning as best I can, and feel that I'm probably just getting hung up on some things because of my lack of overall understanding.  I've searched plenty online and these forums to figure out my issues, but
    somethings just go over my head or I can't figure out how to make it work for what I'm trying to do.
    On to what I'm trying to do.
    I'm working on building an app for scoring swim meets for our conference.  I'm using a lcal SQL DB for all of the data and have a DGV on my form.  We don't have timing systems so we take 2 times from stop watches and then need to do an average
    of the times to determine a swimmers time.  I have 3 fields for times, Time1, Time2 and AvgTime.
    What I'm needing help with is how do I allow the user to not worry about formatting the time they enter but have it automatically format what they input after they exit the cell on Time1 and Time2.  Then I need to have it figure out the average of those
    two times in AvgTime cell.  I'm able to get the averaging to work if I have the datatype set to decimal or int, but I can't get it to work if I have them set has TimeSpan.
    Below is the code I have currently.  As you can see I've got things commented out that I found online but couldn't make work.
    Thanks for taking the time to review this and help me out.
    Public Class EventScoring
    Private Sub EventScoring_Load(sender As Object, e As EventArgs) Handles MyBase.Load
    'TODO: This line of code loads data into the 'MeetDataSet.DualLineup' table. You can move, or remove it, as needed.
    Me.DualLineupTableAdapter.Fill(Me.MeetDataSet.DualLineup)
    'TODO: This line of code loads data into the 'MeetDataSet.EventList' table. You can move, or remove it, as needed.
    Me.EventListTableAdapter.Fill(Me.MeetDataSet.EventList)
    'DualLineupDataGridView.Columns(5).DefaultCellStyle.Format = ("mm\:ss\.ff")
    MeetDataSet.DualLineup.Columns("AvgTime").Expression = "(([Time1] + [Time2]) /2)"
    End Sub
    Private Sub Sub_Btn_Click(ByVal sender As Object, ByVal e As EventArgs) Handles Sub_Btn.Click
    Try
    Me.Validate()
    Me.DualLineupBindingSource.EndEdit()
    Me.DualLineupTableAdapter.Update(Me.MeetDataSet.DualLineup)
    MsgBox("Update successful")
    Catch ex As Exception
    MsgBox("Update failed")
    End Try
    End Sub
    'Private Sub DualLineupDataGridView_CellFormatting(ByVal sender As Object, ByVal e As DataGridViewCellFormattingEventArgs) Handles DualLineupDataGridView.CellFormatting
    ' If ((e.ColumnIndex = 5) AndAlso (Not IsDBNull(e.Value))) Then
    ' Dim tp As TimeSpan = CType(e.Value, TimeSpan)
    ' Dim dt As DateTime = New DateTime(tp.Ticks)
    ' e.Value = dt.ToString("mm\:ss\.ff")
    ' End If
    'End Sub
    'Private Sub DualLineupDataGridView_CurrentCellDirtyStateChanged(ByVal sender As Object, ByVal e As DataGridViewCellFormattingEventArgs) Handles DualLineupDataGridView.CurrentCellDirtyStateChanged
    ' If ((e.ColumnIndex = 5) AndAlso (Not IsDBNull(e.Value))) Then
    ' Dim tp As TimeSpan = CType(e.Value, TimeSpan)
    ' Dim dt As DateTime = New DateTime(tp.Ticks)
    ' e.Value = dt.ToString("mm\:ss\.ff")
    ' End If
    'End Sub
    End Class

    AB,
    If you're ok with your database other than working out this issue with time, you might want to give the following a try. It's pretty flexible in that you can give it a string or a decimal, and I'll explain:
    The string can be in the format of "mm:ss" (minutes and seconds) or in the format of "mm:ss:f~".
    The reason that I put the tilda there is because you're not limited on the number of decimal places - it will work out the math to figure out what you meant.
    Also though, you can give it a decimal value representing the total number of seconds. Do be sure to clearly denote that it's a decimal type (otherwise it will assume it's a double and will fail). Here's the class:
    Public Class SwimmerTime
    Private _totalSeconds As Decimal
    Public Sub New(ByVal value As Object)
    Try
    Dim tSeconds As Decimal = 0
    If TypeOf value Is String Then
    Dim s As String = CType(value, String)
    If Not s.Contains(":"c) Then
    Throw New ArgumentException("The string is malformed and cannot be used.")
    Else
    Dim elements() As String = s.Split(":"c)
    For Each element As String In elements
    If Not Integer.TryParse(element, New Integer) Then
    Throw New ArgumentException("The string is malformed and cannot be used.")
    End If
    Next
    If elements.Length = 2 Then
    tSeconds = (60 * CInt(elements(0)) + CInt(elements(1)))
    ElseIf elements.Length = 3 Then
    tSeconds = (60 * CInt(elements(0)) + CInt(elements(1)))
    Dim divideByString As String = "1"
    For Each c As Char In elements(2)
    divideByString &= "0"
    Next
    tSeconds += CDec(elements(2)) / CInt(divideByString)
    Else
    Throw New ArgumentException("The string is malformed and cannot be used.")
    End If
    End If
    ElseIf TypeOf value Is Decimal Then
    Dim d As Decimal = DirectCast(value, Decimal)
    tSeconds = d
    Else
    Throw New ArgumentException("The type was not recognizable and cannot be used.")
    End If
    If tSeconds = 0 Then
    Throw New ArgumentOutOfRangeException("Total Seconds", "Must be greater than zero.")
    Else
    _totalSeconds = tSeconds
    End If
    Catch ex As Exception
    Throw
    End Try
    End Sub
    Public Shared Function GetAverage(ByVal value1 As Object, _
    ByVal value2 As Object) As SwimmerTime
    Dim retVal As SwimmerTime = Nothing
    Try
    Dim st1 As SwimmerTime = New SwimmerTime(value1)
    Dim st2 As SwimmerTime = New SwimmerTime(value2)
    If st1 IsNot Nothing AndAlso st2 IsNot Nothing Then
    Dim tempList As New List(Of Decimal)
    With tempList
    .Add(st1.TotalSeconds)
    .Add(st2.TotalSeconds)
    End With
    Dim averageSeconds As Decimal = tempList.Average
    retVal = New SwimmerTime(averageSeconds)
    End If
    Catch ex As Exception
    Throw
    End Try
    Return retVal
    End Function
    Public ReadOnly Property FormattedString As String
    Get
    Dim ts As TimeSpan = TimeSpan.FromSeconds(_totalSeconds)
    Return String.Format("{0:00}:{1:00}:{2:000}", _
    ts.Minutes, _
    ts.Seconds, _
    ts.Milliseconds)
    End Get
    End Property
    Public ReadOnly Property TotalSeconds As Decimal
    Get
    Return _totalSeconds
    End Get
    End Property
    End Class
    I'll show how to use it by example:
    Option Strict On
    Option Explicit On
    Option Infer Off
    Public Class Form1
    Private Sub Form1_Load(ByVal sender As System.Object, _
    ByVal e As System.EventArgs) _
    Handles MyBase.Load
    Try
    Dim averageSwimmerTime As SwimmerTime = _
    SwimmerTime.GetAverage("02:24:05", 145.3274D)
    MessageBox.Show(String.Format("Formatted String: {0}{1}Actual Value: {2}", _
    averageSwimmerTime.FormattedString, vbCrLf, _
    averageSwimmerTime.TotalSeconds), _
    "Average Swimmer Time")
    Stop
    Catch ex As Exception
    MessageBox.Show(String.Format("An error occurred:{0}{0}{1}", _
    vbCrLf, ex.Message), "Program Error", _
    MessageBoxButtons.OK, MessageBoxIcon.Warning)
    End Try
    Stop
    End Sub
    End Class
    An advantage here is that since it returns an instance of the class, you have access to the formatted string, and the actual decimal value.
    I hope you find this helpful. :)
    Still lost in code, just at a little higher level.

  • Format before OR after writing the paper

    Is it better to format your paper as you are writing or do it after you are finished? I'm writing papers for college that require either APA or MLA style - depending on the class - and I'm not sure if I should spend time before I start setting up the style or after I've written it.
    Any help would be appreciated.
    Thanks.
    Chip

    I agree that this is a personal preference. With APA, however, you can always insert your cover page first, headers, and adjust your type styles before you begin writing. Then at least you'll have the basic formatting finished. Once completed with your paper, go back and make any necessary changes.

  • Format XML data when writing to a file

    Hi,
    I am trying to write runtime data (XML) to a file. I am doign the following to write to the file:
    public void saveFile() throws Exception {
    try {
    TransformerFactory transformerFactory =
    TransformerFactory.newInstance();
    Transformer transformer = transformerFactory.newTransformer();
    DOMSource source = new DOMSource(doc);
    OutputStream output = new FileOutputStream(XMLFile);
    StreamResult result = new StreamResult();
    transformer.transform(source, result);
    output.flush();
    output.close();
    } catch (Exception x) {
    throw x;
    The output from the above code is a clutter of elements. All the elements are on the same line some times.
    1) How can I format the output? (Indent the elements in the file)
    Is there a way to achieve this?
    Sometimes, I see a square character at the end of some lines. I thought the output file is corrupt because of this character. But it does not seem to be a problem.
    2) How do I avoid this character?

    Also, I am planning on using xpath in conjunction with DOM. However I read on one of the websites not to use too many xpath queries. I am using jdk 1.5.
    Is it tight. My code might have to be run many times. Is the following ok?
    or should I eliminate xpath and do my own traversing. I am concerned about runtime performance if I have to continue to update the xml file frequently.
    public Element getMatchingNode(MyAttributes attr) throws Exception {
    try {
    File myFile = new File(XMLFile);
    if (myFile.exists()) {
    System.out.println(XMLFile + " File exists");
    XPathFactory factory = XPathFactory.newInstance();
    XPath xPath = factory.newXPath();
    InputSource inputSource =
    new InputSource(new FileInputStream(myFile));
    String matchingString = this.getMatchingString(attr);
    Element module =
    (Element)xPath.evaluate("/" + TEST_TAG +
    "/" +
    TEST_TAG2 +
    "/" +
    DETAIL +
    matchingString,
    inputSource,
    XPathConstants.NODE);
    return module;
    } else
    return null;
    } catch (Exception x) {
    throw x;
    private String getMatchingString(MyAttributes attrs){
    StringBuffer buffer = new StringBuffer();
    for (Enumeration e=attrs.keys(); e.hasMoreElements();){
    String attribute = (String)e.nextElement();
    String value = attrs.get(attribute);
    buffer.append("[@" + attribute + "=" + "'" + value + "']");
    return buffer.toString();
    }

  • .csv files exported not converting to Tabular format jdev patch 6908968

    After jdeveloper patch 6908968 on 12.0.6, the data is copied into an Excel .csv file but the entire data is copied to the single cell/column.
    Developed an OAF page which has the standard Oracle Export Functionality for one of our requirements. When exporting from the OAF page, it is creating a .csv file(as expected) , but the data is not copied to individual cells. The whole data is going to one cell rather than into individual columns.
    According to Bug 4699555 you can use one of the following workarounds :-
    a. change "List separator" value from ";" to "," in the Regional Options,
    b. in Excel, after opening the .csv file, go to "Data" -> "Text to Columns ...",
    choose "comma" as the delimiter in the wizard.
    c. instead of Opening the file, Save it with .txt extension open MS Excel and open the saved file and specify delimited + comma as the delimiter during the text import process.
    The workarounds work, however this is not acceptable as there are 380,000 users. This worked as expected with 11.5.10 and not with upgraded R12. 12.0.6 with jdeveloper patch '6908968' for developing the OAF page with jdeveloper export button option.
    Is there a setup step missing or that changed for R12 and the column delimiter?

    You MAY want to think about a product that will read the output from your DVD player and convert to DV AVI
    Matt with Grass Valley Canopus in their tech support department stated that the model
    110 will suffice for most hobbyist. If a person has a lot of tapes that were played
    often the tape stretches and the magnetic coding diminishes. If your goal is to encode
    tapes in good shape buy the 110, if you will be encoding old tapes of poor quality buy
    the model 300
    Both the 110 and 300 are two way devices so you may output back to tape... if you don't
    need that, look at the model 55
    http://www.grassvalley.com/products/advc55 One Way Only to Computer
    http://www.grassvalley.com/products/advc110 for good tapes, or
    http://www.grassvalley.com/products/advc300 better with OLD tapes

  • Writing ouput to the next column in a .csv file

    I am Writing to .CSV file with PL/SQL code:
    CREATE OR REPLACE PROCEDURE Write_File(p_name IN VARCHAR)
    AS
    BEGIN
    DECLARE
    CURSOR emp_sel IS SELECT ename,hiredate FROM EMP;
    file_name VARCHAR2(20);
    output_file utl_file.file_type;
    BEGIN
    file_name:= (p_name||'.csv');
    output_file:=utl_file.fopen('d:\test_dir',p_name||'.csv','W');
    FOR i IN emp_sel LOOP
    utl_file.put_line(output_file,'NAME:'||i.ename);
    utl_file.put_line(output_file,'hire Date:'||i.hiredate);
    END LOOP;
    utl_file.fclose(output_file);
    EXCEPTION
    WHEN OTHERS THEN
    dbms_output.put_line(SQLCODE||SQLERRM);
    END;
    END;
    Just wanted to know how Move to the next Column in the .csv file.
    at present the ename and hiredate r written in the 1st Column C1 column.
    I want to write the hiredate data in C2 column.
    how to?

    user545846,
    I answered you in the other thread:
    Re: Calling utl_file package from a java program
    Good Luck,
    Avi.

  • Power Query Doesn't Load Complete CSV FIle After Pivot Transform

    Hello, After transforming a CSV file using the following code I can load the entire file (45,000 rows) into the Excel Data Model.
    let
    Source = Csv.Document(File.Contents("C:\Users\jd\Desktop\2014 JW Annual Review\ALL STORES 01-07-15 AR OPEN.csv"),null,",",null,1252),
    #"First Row as Header" = Table.PromoteHeaders(Source),
    #"Changed Type" = Table.TransformColumnTypes(#"First Row as Header",{{"AR-OPEN", type text}, {"SALE CO", type text}, {"NAME", type text}, {"N-CD", Int64.Type}, {"REFER", type text}, {"JRNL CO", type text}, {"SCHED", type text}, {"AGME.CO.ID", Int64.Type}, {"SO", type text}, {"ODATE", type text}, {"JRNL TYPE", type text}, {"CONTROL", type text}, {"DATE", type text}, {"A-ACCT", type text}, {"CONTROL2", type text}}),
    #"Removed Columns" = Table.RemoveColumns(#"Changed Type",{"AR-OPEN"}),
    #"Split Column by Delimiter" = Table.SplitColumn(#"Removed Columns","SALE CO",Splitter.SplitTextByDelimiter(" "),{"SALE CO.1", "SALE CO.2", "SALE CO.3", "SALE CO.4", "SALE CO.5", "SALE CO.6", "SALE CO.7", "SALE CO.8", "SALE CO.9", "SALE CO.10", "SALE CO.11", "SALE CO.12", "SALE CO.13", "SALE CO.14", "SALE CO.15", "SALE CO.16", "SALE CO.17", "SALE CO.18", "SALE CO.19", "SALE CO.20", "SALE CO.21", "SALE CO.22", "SALE CO.23", "SALE CO.24", "SALE CO.25"}),
    #"Changed Type1" = Table.TransformColumnTypes(#"Split Column by Delimiter",{{"SALE CO.1", Int64.Type}, {"SALE CO.2", type text}, {"SALE CO.3", type text}, {"SALE CO.4", type text}, {"SALE CO.5", type text}, {"SALE CO.6", type text}, {"SALE CO.7", type text}, {"SALE CO.8", type text}, {"SALE CO.9", type text}, {"SALE CO.10", type text}, {"SALE CO.11", type text}, {"SALE CO.12", type text}, {"SALE CO.13", type text}, {"SALE CO.14", type text}, {"SALE CO.15", type text}, {"SALE CO.16", type text}, {"SALE CO.17", type text}, {"SALE CO.18", type text}, {"SALE CO.19", type text}, {"SALE CO.20", type text}, {"SALE CO.21", type text}, {"SALE CO.22", type text}, {"SALE CO.23", type text}, {"SALE CO.24", type text}, {"SALE CO.25", type text}, {"NAME_1", type text}, {"DATE_2", type text}, {"NAME_3", type text}, {"CONTROL_4", type text}}),
    #"Split Column by Delimiter1" = Table.SplitColumn(#"Changed Type1","REFER",Splitter.SplitTextByDelimiter(" "),{"REFER.1", "REFER.2", "REFER.3", "REFER.4", "REFER.5", "REFER.6", "REFER.7", "REFER.8", "REFER.9", "REFER.10", "REFER.11", "REFER.12", "REFER.13", "REFER.14", "REFER.15", "REFER.16", "REFER.17", "REFER.18", "REFER.19", "REFER.20", "REFER.21", "REFER.22", "REFER.23", "REFER.24", "REFER.25"}),
    #"Changed Type2" = Table.TransformColumnTypes(#"Split Column by Delimiter1",{{"REFER.1", type text}, {"REFER.2", type text}, {"REFER.3", type text}, {"REFER.4", type text}, {"REFER.5", type text}, {"REFER.6", type text}, {"REFER.7", type text}, {"REFER.8", type text}, {"REFER.9", type text}, {"REFER.10", type text}, {"REFER.11", type text}, {"REFER.12", type text}, {"REFER.13", type text}, {"REFER.14", type text}, {"REFER.15", type text}, {"REFER.16", type text}, {"REFER.17", type text}, {"REFER.18", type text}, {"REFER.19", type text}, {"REFER.20", type text}, {"REFER.21", type text}, {"REFER.22", type text}, {"REFER.23", type text}, {"REFER.24", type text}, {"REFER.25", type text}}),
    #"Split Column by Delimiter2" = Table.SplitColumn(#"Changed Type2","JRNL CO",Splitter.SplitTextByDelimiter(" "),{"JRNL CO.1", "JRNL CO.2", "JRNL CO.3", "JRNL CO.4", "JRNL CO.5", "JRNL CO.6", "JRNL CO.7", "JRNL CO.8", "JRNL CO.9", "JRNL CO.10", "JRNL CO.11", "JRNL CO.12", "JRNL CO.13", "JRNL CO.14", "JRNL CO.15", "JRNL CO.16", "JRNL CO.17", "JRNL CO.18", "JRNL CO.19", "JRNL CO.20", "JRNL CO.21", "JRNL CO.22", "JRNL CO.23", "JRNL CO.24", "JRNL CO.25"}),
    #"Changed Type3" = Table.TransformColumnTypes(#"Split Column by Delimiter2",{{"JRNL CO.1", Int64.Type}, {"JRNL CO.2", type text}, {"JRNL CO.3", type text}, {"JRNL CO.4", type text}, {"JRNL CO.5", type text}, {"JRNL CO.6", type text}, {"JRNL CO.7", type text}, {"JRNL CO.8", type text}, {"JRNL CO.9", type text}, {"JRNL CO.10", type text}, {"JRNL CO.11", type text}, {"JRNL CO.12", type text}, {"JRNL CO.13", type text}, {"JRNL CO.14", type text}, {"JRNL CO.15", type text}, {"JRNL CO.16", type text}, {"JRNL CO.17", type text}, {"JRNL CO.18", type text}, {"JRNL CO.19", type text}, {"JRNL CO.20", type text}, {"JRNL CO.21", type text}, {"JRNL CO.22", type text}, {"JRNL CO.23", type text}, {"JRNL CO.24", type text}, {"JRNL CO.25", type text}}),
    #"Split Column by Delimiter3" = Table.SplitColumn(#"Changed Type3","SO",Splitter.SplitTextByDelimiter(" "),{"SO.1", "SO.2", "SO.3", "SO.4", "SO.5", "SO.6", "SO.7", "SO.8", "SO.9", "SO.10", "SO.11", "SO.12", "SO.13", "SO.14", "SO.15", "SO.16", "SO.17", "SO.18", "SO.19", "SO.20", "SO.21", "SO.22", "SO.23", "SO.24", "SO.25"}),
    #"Changed Type4" = Table.TransformColumnTypes(#"Split Column by Delimiter3",{{"SO.1", Int64.Type}, {"SO.2", type text}, {"SO.3", type text}, {"SO.4", type text}, {"SO.5", type text}, {"SO.6", type text}, {"SO.7", type text}, {"SO.8", type text}, {"SO.9", type text}, {"SO.10", type text}, {"SO.11", type text}, {"SO.12", type text}, {"SO.13", type text}, {"SO.14", type text}, {"SO.15", type text}, {"SO.16", type text}, {"SO.17", type text}, {"SO.18", type text}, {"SO.19", type text}, {"SO.20", type text}, {"SO.21", type text}, {"SO.22", type text}, {"SO.23", type text}, {"SO.24", type text}, {"SO.25", type text}}),
    #"Split Column by Delimiter4" = Table.SplitColumn(#"Changed Type4","ODATE",Splitter.SplitTextByDelimiter(" "),{"ODATE.1", "ODATE.2", "ODATE.3", "ODATE.4", "ODATE.5", "ODATE.6", "ODATE.7", "ODATE.8", "ODATE.9", "ODATE.10", "ODATE.11", "ODATE.12", "ODATE.13", "ODATE.14", "ODATE.15", "ODATE.16", "ODATE.17", "ODATE.18", "ODATE.19", "ODATE.20", "ODATE.21", "ODATE.22", "ODATE.23", "ODATE.24", "ODATE.25"}),
    #"Changed Type5" = Table.TransformColumnTypes(#"Split Column by Delimiter4",{{"ODATE.1", type date}, {"ODATE.2", type text}, {"ODATE.3", type text}, {"ODATE.4", type text}, {"ODATE.5", type text}, {"ODATE.6", type text}, {"ODATE.7", type text}, {"ODATE.8", type text}, {"ODATE.9", type text}, {"ODATE.10", type text}, {"ODATE.11", type text}, {"ODATE.12", type text}, {"ODATE.13", type text}, {"ODATE.14", type text}, {"ODATE.15", type text}, {"ODATE.16", type text}, {"ODATE.17", type text}, {"ODATE.18", type text}, {"ODATE.19", type text}, {"ODATE.20", type text}, {"ODATE.21", type text}, {"ODATE.22", type text}, {"ODATE.23", type text}, {"ODATE.24", type text}, {"ODATE.25", type text}}),
    #"Split Column by Delimiter5" = Table.SplitColumn(#"Changed Type5","JRNL TYPE",Splitter.SplitTextByDelimiter(" "),{"JRNL TYPE.1", "JRNL TYPE.2", "JRNL TYPE.3", "JRNL TYPE.4", "JRNL TYPE.5", "JRNL TYPE.6", "JRNL TYPE.7", "JRNL TYPE.8", "JRNL TYPE.9", "JRNL TYPE.10", "JRNL TYPE.11", "JRNL TYPE.12", "JRNL TYPE.13", "JRNL TYPE.14", "JRNL TYPE.15", "JRNL TYPE.16", "JRNL TYPE.17", "JRNL TYPE.18", "JRNL TYPE.19", "JRNL TYPE.20", "JRNL TYPE.21", "JRNL TYPE.22", "JRNL TYPE.23", "JRNL TYPE.24", "JRNL TYPE.25"}),
    #"Changed Type6" = Table.TransformColumnTypes(#"Split Column by Delimiter5",{{"JRNL TYPE.1", type text}, {"JRNL TYPE.2", type text}, {"JRNL TYPE.3", type text}, {"JRNL TYPE.4", type text}, {"JRNL TYPE.5", type text}, {"JRNL TYPE.6", type text}, {"JRNL TYPE.7", type text}, {"JRNL TYPE.8", type text}, {"JRNL TYPE.9", type text}, {"JRNL TYPE.10", type text}, {"JRNL TYPE.11", type text}, {"JRNL TYPE.12", type text}, {"JRNL TYPE.13", type text}, {"JRNL TYPE.14", type text}, {"JRNL TYPE.15", type text}, {"JRNL TYPE.16", type text}, {"JRNL TYPE.17", type text}, {"JRNL TYPE.18", type text}, {"JRNL TYPE.19", type text}, {"JRNL TYPE.20", type text}, {"JRNL TYPE.21", type text}, {"JRNL TYPE.22", type text}, {"JRNL TYPE.23", type text}, {"JRNL TYPE.24", type text}, {"JRNL TYPE.25", type text}}),
    #"Split Column by Delimiter6" = Table.SplitColumn(#"Changed Type6","DATE",Splitter.SplitTextByDelimiter(" "),{"DATE.1", "DATE.2", "DATE.3", "DATE.4", "DATE.5", "DATE.6", "DATE.7", "DATE.8", "DATE.9", "DATE.10", "DATE.11", "DATE.12", "DATE.13", "DATE.14", "DATE.15", "DATE.16", "DATE.17", "DATE.18", "DATE.19", "DATE.20", "DATE.21", "DATE.22", "DATE.23", "DATE.24", "DATE.25"}),
    #"Changed Type7" = Table.TransformColumnTypes(#"Split Column by Delimiter6",{{"DATE.1", type date}, {"DATE.2", type text}, {"DATE.3", type text}, {"DATE.4", type text}, {"DATE.5", type text}, {"DATE.6", type text}, {"DATE.7", type text}, {"DATE.8", type text}, {"DATE.9", type text}, {"DATE.10", type text}, {"DATE.11", type text}, {"DATE.12", type text}, {"DATE.13", type text}, {"DATE.14", type text}, {"DATE.15", type text}, {"DATE.16", type text}, {"DATE.17", type text}, {"DATE.18", type text}, {"DATE.19", type text}, {"DATE.20", type text}, {"DATE.21", type text}, {"DATE.22", type text}, {"DATE.23", type text}, {"DATE.24", type text}, {"DATE.25", type text}}),
    #"Reordered Columns" = Table.ReorderColumns(#"Changed Type7",{"AGME.CO.ID", "SCHED", "NAME", "NAME_1", "N-CD", "SALE CO.1", "SALE CO.2", "SALE CO.3", "SALE CO.4", "SALE CO.5", "SALE CO.6", "SALE CO.7", "SALE CO.8", "SALE CO.9", "SALE CO.10", "SALE CO.11", "SALE CO.12", "SALE CO.13", "SALE CO.14", "SALE CO.15", "SALE CO.16", "SALE CO.17", "SALE CO.18", "SALE CO.19", "SALE CO.20", "SALE CO.21", "SALE CO.22", "SALE CO.23", "SALE CO.24", "SALE CO.25", "REFER.1", "REFER.2", "REFER.3", "REFER.4", "REFER.5", "REFER.6", "REFER.7", "REFER.8", "REFER.9", "REFER.10", "REFER.11", "REFER.12", "REFER.13", "REFER.14", "REFER.15", "REFER.16", "REFER.17", "REFER.18", "REFER.19", "REFER.20", "REFER.21", "REFER.22", "REFER.23", "REFER.24", "REFER.25", "JRNL CO.1", "JRNL CO.2", "JRNL CO.3", "JRNL CO.4", "JRNL CO.5", "JRNL CO.6", "JRNL CO.7", "JRNL CO.8", "JRNL CO.9", "JRNL CO.10", "JRNL CO.11", "JRNL CO.12", "JRNL CO.13", "JRNL CO.14", "JRNL CO.15", "JRNL CO.16", "JRNL CO.17", "JRNL CO.18", "JRNL CO.19", "JRNL CO.20", "JRNL CO.21", "JRNL CO.22", "JRNL CO.23", "JRNL CO.24", "JRNL CO.25", "SO.1", "SO.2", "SO.3", "SO.4", "SO.5", "SO.6", "SO.7", "SO.8", "SO.9", "SO.10", "SO.11", "SO.12", "SO.13", "SO.14", "SO.15", "SO.16", "SO.17", "SO.18", "SO.19", "SO.20", "SO.21", "SO.22", "SO.23", "SO.24", "SO.25", "ODATE.1", "ODATE.2", "ODATE.3", "ODATE.4", "ODATE.5", "ODATE.6", "ODATE.7", "ODATE.8", "ODATE.9", "ODATE.10", "ODATE.11", "ODATE.12", "ODATE.13", "ODATE.14", "ODATE.15", "ODATE.16", "ODATE.17", "ODATE.18", "ODATE.19", "ODATE.20", "ODATE.21", "ODATE.22", "ODATE.23", "ODATE.24", "ODATE.25", "JRNL TYPE.1", "JRNL TYPE.2", "JRNL TYPE.3", "JRNL TYPE.4", "JRNL TYPE.5", "JRNL TYPE.6", "JRNL TYPE.7", "JRNL TYPE.8", "JRNL TYPE.9", "JRNL TYPE.10", "JRNL TYPE.11", "JRNL TYPE.12", "JRNL TYPE.13", "JRNL TYPE.14", "JRNL TYPE.15", "JRNL TYPE.16", "JRNL TYPE.17", "JRNL TYPE.18", "JRNL TYPE.19", "JRNL TYPE.20", "JRNL TYPE.21", "JRNL TYPE.22", "JRNL TYPE.23", "JRNL TYPE.24", "JRNL TYPE.25", "CONTROL", "DATE.1", "DATE.2", "DATE.3", "DATE.4", "DATE.5", "DATE.6", "DATE.7", "DATE.8", "DATE.9", "DATE.10", "DATE.11", "DATE.12", "DATE.13", "DATE.14", "DATE.15", "DATE.16", "DATE.17", "DATE.18", "DATE.19", "DATE.20", "DATE.21", "DATE.22", "DATE.23", "DATE.24", "DATE.25", "DATE_2", "A-ACCT", "NAME_3", "CONTROL_4", "CONTROL2"}),
    #"Removed Columns1" = Table.RemoveColumns(#"Reordered Columns",{"DATE_2", "A-ACCT", "NAME_3", "CONTROL_4", "CONTROL2"}),
    #"Reordered Columns1" = Table.ReorderColumns(#"Removed Columns1",{"AGME.CO.ID", "SCHED", "NAME", "NAME_1", "N-CD", "CONTROL", "SALE CO.1", "SALE CO.2", "SALE CO.3", "SALE CO.4", "SALE CO.5", "SALE CO.6", "SALE CO.7", "SALE CO.8", "SALE CO.9", "SALE CO.10", "SALE CO.11", "SALE CO.12", "SALE CO.13", "SALE CO.14", "SALE CO.15", "SALE CO.16", "SALE CO.17", "SALE CO.18", "SALE CO.19", "SALE CO.20", "SALE CO.21", "SALE CO.22", "SALE CO.23", "SALE CO.24", "SALE CO.25", "REFER.1", "REFER.2", "REFER.3", "REFER.4", "REFER.5", "REFER.6", "REFER.7", "REFER.8", "REFER.9", "REFER.10", "REFER.11", "REFER.12", "REFER.13", "REFER.14", "REFER.15", "REFER.16", "REFER.17", "REFER.18", "REFER.19", "REFER.20", "REFER.21", "REFER.22", "REFER.23", "REFER.24", "REFER.25", "JRNL CO.1", "JRNL CO.2", "JRNL CO.3", "JRNL CO.4", "JRNL CO.5", "JRNL CO.6", "JRNL CO.7", "JRNL CO.8", "JRNL CO.9", "JRNL CO.10", "JRNL CO.11", "JRNL CO.12", "JRNL CO.13", "JRNL CO.14", "JRNL CO.15", "JRNL CO.16", "JRNL CO.17", "JRNL CO.18", "JRNL CO.19", "JRNL CO.20", "JRNL CO.21", "JRNL CO.22", "JRNL CO.23", "JRNL CO.24", "JRNL CO.25", "SO.1", "SO.2", "SO.3", "SO.4", "SO.5", "SO.6", "SO.7", "SO.8", "SO.9", "SO.10", "SO.11", "SO.12", "SO.13", "SO.14", "SO.15", "SO.16", "SO.17", "SO.18", "SO.19", "SO.20", "SO.21", "SO.22", "SO.23", "SO.24", "SO.25", "ODATE.1", "ODATE.2", "ODATE.3", "ODATE.4", "ODATE.5", "ODATE.6", "ODATE.7", "ODATE.8", "ODATE.9", "ODATE.10", "ODATE.11", "ODATE.12", "ODATE.13", "ODATE.14", "ODATE.15", "ODATE.16", "ODATE.17", "ODATE.18", "ODATE.19", "ODATE.20", "ODATE.21", "ODATE.22", "ODATE.23", "ODATE.24", "ODATE.25", "JRNL TYPE.1", "JRNL TYPE.2", "JRNL TYPE.3", "JRNL TYPE.4", "JRNL TYPE.5", "JRNL TYPE.6", "JRNL TYPE.7", "JRNL TYPE.8", "JRNL TYPE.9", "JRNL TYPE.10", "JRNL TYPE.11", "JRNL TYPE.12", "JRNL TYPE.13", "JRNL TYPE.14", "JRNL TYPE.15", "JRNL TYPE.16", "JRNL TYPE.17", "JRNL TYPE.18", "JRNL TYPE.19", "JRNL TYPE.20", "JRNL TYPE.21", "JRNL TYPE.22", "JRNL TYPE.23", "JRNL TYPE.24", "JRNL TYPE.25", "DATE.1", "DATE.2", "DATE.3", "DATE.4", "DATE.5", "DATE.6", "DATE.7", "DATE.8", "DATE.9", "DATE.10", "DATE.11", "DATE.12", "DATE.13", "DATE.14", "DATE.15", "DATE.16", "DATE.17", "DATE.18", "DATE.19", "DATE.20", "DATE.21", "DATE.22", "DATE.23", "DATE.24", "DATE.25"}),
    #"Unpivoted Columns" = Table.UnpivotOtherColumns(#"Reordered Columns1", {"AGME.CO.ID", "SCHED", "NAME", "NAME_1", "N-CD", "CONTROL"}, "Attribute", "Value"),
    #"Replaced Value" = Table.ReplaceValue(#"Unpivoted Columns",".*","",Replacer.ReplaceText,{"Attribute"}),
    #"Split Column by Delimiter7" = Table.SplitColumn(#"Replaced Value","Attribute",Splitter.SplitTextByDelimiter("."),{"Attribute.1", "Attribute.2"}),
    #"Changed Type8" = Table.TransformColumnTypes(#"Split Column by Delimiter7",{{"Attribute.1", type text}, {"Attribute.2", Int64.Type}}),
    #"Removed Columns2" = Table.RemoveColumns(#"Changed Type8",{"Attribute.2"}),
    #"Renamed Columns" = Table.RenameColumns(#"Removed Columns2",{{"Attribute.1", "Type"}}),
    #"Removed Columns3" = Table.RemoveColumns(#"Renamed Columns",{"SCHED", "NAME", "NAME_1", "N-CD"}),
    #"Renamed Columns1" = Table.RenameColumns(#"Removed Columns3",{{"AGME.CO.ID", "companyid"}, {"CONTROL", "control"}}),
    #"Reordered Columns2" = Table.ReorderColumns(#"Renamed Columns1",{"companyid", "control", "Type", "Value"})
    in
    #"Reordered Columns2"
    I need to do one more transform which is a Pivot like so:
    #"Pivoted Column" = Table.Pivot(#"Reordered Columns2", List.Distinct(#"Reordered Columns2"[Type]), "Type", "Value")
    in
    #"Pivoted Column"
    and I only get 326 rows. Because it is a pivot without aggregation I get errors but even if I remove errors I still don't get the complete result set.
    Any ideas?

    Hello again and thanks for your help. I used List.First because the value column is text however I only got one row returned so I thought I would give you all my code (because I changed some things to make it simpler) and some pictures. Here's my code:
    let
    Source = Csv.Document(File.Contents("C:\Users\jdonnelly\Desktop\2014 JW Annual Review\ALL STORES 01-07-15 AR OPEN.csv"),null,",",null,1252),
    #"First Row as Header" = Table.PromoteHeaders(Source),
    #"Changed Type" = Table.TransformColumnTypes(#"First Row as Header",{{"AR-OPEN", type text}, {"SALE CO", type text}, {"NAME", type text}, {"N-CD", Int64.Type}, {"REFER", type text}, {"JRNL CO", type text}, {"SCHED", type text}, {"AGME.CO.ID", Int64.Type}, {"SO", type text}, {"ODATE", type text}, {"JRNL TYPE", type text}, {"CONTROL", type text}, {"DATE", type text}, {"A-ACCT", type text}, {"CONTROL2", type text}}),
    #"Removed Columns" = Table.RemoveColumns(#"Changed Type",{"AR-OPEN", "NAME", "NAME_1", "N-CD", "JRNL CO", "SCHED", "AGME.CO.ID", "JRNL TYPE", "DATE", "DATE_2", "A-ACCT", "NAME_3", "CONTROL_4", "CONTROL2"}),
    #"Reordered Columns" = Table.ReorderColumns(#"Removed Columns",{"CONTROL", "SALE CO", "REFER", "SO", "ODATE"}),
    #"Split Column by Delimiter" = Table.SplitColumn(#"Reordered Columns","ODATE",Splitter.SplitTextByDelimiter(" "),{"ODATE.1", "ODATE.2", "ODATE.3", "ODATE.4", "ODATE.5", "ODATE.6", "ODATE.7", "ODATE.8", "ODATE.9", "ODATE.10", "ODATE.11", "ODATE.12", "ODATE.13", "ODATE.14", "ODATE.15", "ODATE.16", "ODATE.17", "ODATE.18", "ODATE.19"}),
    #"Changed Type1" = Table.TransformColumnTypes(#"Split Column by Delimiter",{{"ODATE.1", type date}, {"ODATE.2", type text}, {"ODATE.3", type text}, {"ODATE.4", type text}, {"ODATE.5", type text}, {"ODATE.6", type text}, {"ODATE.7", type text}, {"ODATE.8", type text}, {"ODATE.9", type text}, {"ODATE.10", type text}, {"ODATE.11", type text}, {"ODATE.12", type text}, {"ODATE.13", type text}, {"ODATE.14", type text}, {"ODATE.15", type text}, {"ODATE.16", type text}, {"ODATE.17", type text}, {"ODATE.18", type text}, {"ODATE.19", type text}}),
    #"Removed Columns1" = Table.RemoveColumns(#"Changed Type1",{"ODATE.2", "ODATE.3", "ODATE.5", "ODATE.6", "ODATE.8", "ODATE.9", "ODATE.11", "ODATE.12", "ODATE.14", "ODATE.15", "ODATE.17", "ODATE.18"}),
    #"Changed Type2" = Table.TransformColumnTypes(#"Removed Columns1",{{"ODATE.1", type date}, {"ODATE.4", type date}, {"ODATE.7", type date}, {"ODATE.10", type date}, {"ODATE.13", type date}, {"ODATE.16", type date}, {"ODATE.19", type date}}),
    #"Split Column by Delimiter1" = Table.SplitColumn(#"Changed Type2","SO",Splitter.SplitTextByDelimiter(" "),{"SO.1", "SO.2", "SO.3", "SO.4", "SO.5", "SO.6", "SO.7", "SO.8", "SO.9", "SO.10", "SO.11", "SO.12", "SO.13", "SO.14", "SO.15", "SO.16", "SO.17", "SO.18", "SO.19"}),
    #"Changed Type3" = Table.TransformColumnTypes(#"Split Column by Delimiter1",{{"SO.1", Int64.Type}, {"SO.2", type text}, {"SO.3", type text}, {"SO.4", type text}, {"SO.5", type text}, {"SO.6", type text}, {"SO.7", type text}, {"SO.8", type text}, {"SO.9", type text}, {"SO.10", type text}, {"SO.11", type text}, {"SO.12", type text}, {"SO.13", type text}, {"SO.14", type text}, {"SO.15", type text}, {"SO.16", type text}, {"SO.17", type text}, {"SO.18", type text}, {"SO.19", type text}}),
    #"Removed Columns2" = Table.RemoveColumns(#"Changed Type3",{"SO.2", "SO.3", "SO.5", "SO.6", "SO.8", "SO.9", "SO.11", "SO.12", "SO.14", "SO.15", "SO.17", "SO.18"}),
    #"Split Column by Delimiter2" = Table.SplitColumn(#"Removed Columns2","REFER",Splitter.SplitTextByDelimiter(" "),{"REFER.1", "REFER.2", "REFER.3", "REFER.4", "REFER.5", "REFER.6", "REFER.7", "REFER.8", "REFER.9", "REFER.10", "REFER.11", "REFER.12", "REFER.13", "REFER.14", "REFER.15", "REFER.16", "REFER.17", "REFER.18", "REFER.19"}),
    #"Changed Type4" = Table.TransformColumnTypes(#"Split Column by Delimiter2",{{"REFER.1", type text}, {"REFER.2", type text}, {"REFER.3", type text}, {"REFER.4", type text}, {"REFER.5", type text}, {"REFER.6", type text}, {"REFER.7", type text}, {"REFER.8", type text}, {"REFER.9", type text}, {"REFER.10", type text}, {"REFER.11", type text}, {"REFER.12", type text}, {"REFER.13", type text}, {"REFER.14", type text}, {"REFER.15", type text}, {"REFER.16", type text}, {"REFER.17", type text}, {"REFER.18", type text}, {"REFER.19", type text}}),
    #"Removed Columns3" = Table.RemoveColumns(#"Changed Type4",{"REFER.2", "REFER.3", "REFER.5", "REFER.6", "REFER.8", "REFER.9", "REFER.11", "REFER.12", "REFER.14", "REFER.15", "REFER.17", "REFER.18"}),
    #"Split Column by Delimiter3" = Table.SplitColumn(#"Removed Columns3","SALE CO",Splitter.SplitTextByDelimiter(" "),{"SALE CO.1", "SALE CO.2", "SALE CO.3", "SALE CO.4", "SALE CO.5", "SALE CO.6", "SALE CO.7", "SALE CO.8", "SALE CO.9", "SALE CO.10", "SALE CO.11", "SALE CO.12", "SALE CO.13", "SALE CO.14", "SALE CO.15", "SALE CO.16", "SALE CO.17", "SALE CO.18", "SALE CO.19"}),
    #"Changed Type5" = Table.TransformColumnTypes(#"Split Column by Delimiter3",{{"SALE CO.1", Int64.Type}, {"SALE CO.2", type text}, {"SALE CO.3", type text}, {"SALE CO.4", type text}, {"SALE CO.5", type text}, {"SALE CO.6", type text}, {"SALE CO.7", type text}, {"SALE CO.8", type text}, {"SALE CO.9", type text}, {"SALE CO.10", type text}, {"SALE CO.11", type text}, {"SALE CO.12", type text}, {"SALE CO.13", type text}, {"SALE CO.14", type text}, {"SALE CO.15", type text}, {"SALE CO.16", type text}, {"SALE CO.17", type text}, {"SALE CO.18", type text}, {"SALE CO.19", type text}}),
    #"Removed Columns4" = Table.RemoveColumns(#"Changed Type5",{"SALE CO.2", "SALE CO.3", "SALE CO.5", "SALE CO.6", "SALE CO.8", "SALE CO.9", "SALE CO.11", "SALE CO.12", "SALE CO.14", "SALE CO.15", "SALE CO.17", "SALE CO.18"}),
    #"Unpivoted Columns" = Table.UnpivotOtherColumns(#"Removed Columns4", {}, "Attribute", "Value"),
    #"Changed Type6" = Table.TransformColumnTypes(#"Unpivoted Columns",{{"Value", type text}}),
    #"Split Column by Delimiter4" = Table.SplitColumn(#"Changed Type6","Attribute",Splitter.SplitTextByDelimiter("."),{"Attribute.1", "Attribute.2"}),
    #"Changed Type7" = Table.TransformColumnTypes(#"Split Column by Delimiter4",{{"Attribute.1", type text}, {"Attribute.2", type text}}),
    #"Removed Columns5" = Table.RemoveColumns(#"Changed Type7",{"Attribute.2"}),
    #"Renamed Columns" = Table.RenameColumns(#"Removed Columns5",{{"Attribute.1", "Attribute"}})
    #"Pivoted Column" = Table.Pivot(#"Renamed Columns", List.Distinct(#"Renamed Columns"[Attribute]), "Attribute", "Value", List.First)
    in
    #"Pivoted Column"
    Before the Table.Pivot my View is as follows:
    It runs down to row 18,200 or so. I chose a smaller file to make things simpler.
    After the Table.Pivot as written above this is what my view shows:
    So I received only one row. Is there a way to make it all rows?
    (For background: the CSV file is from a PICK database that has multi-variate fields where all but one of the fields has multiple values correlated with multiple values in other fields.)
    Thanks again
    John Donnelly

  • Data set created out of csv file: columns not recognized, too many rows

    After uploading a csv file I got out of the OPP the columns don't get recognized in the dataset viewer. Perhaps this happens because the delimiter is a semicolon instead of a comma? Unfortunately, I have no influence how the data gets exported out of the OPP. Is there any option to configure which delimiter to recognize, like it's possible in excel?
    Further, is it possible to configure from which row on the data should be taken into account, similar to excel? With the csv file I get out of the OPP, the data only starts with row 7, while row 6 has the headings.

    After uploading a csv file I got out of the OPP the columns don't get recognized in the dataset viewer. Perhaps this happens because the delimiter is a semicolon instead of a comma? Unfortunately, I have no influence how the data gets exported out of the OPP. Is there any option to configure which delimiter to recognize, like it's possible in excel?
    The delimiter cannot be configured. Comma is the delimiter. I would suggest that you open this file in text editor then replace semicolon with comma. After that you can reupload your dataset.
    Further, is it possible to configure from which row on the data should be taken into account, similar to excel? With the csv file I get out of the OPP, the data only starts with row 7, while row 6 has the headings.
    This is not configurable as well.

  • Gmail Contacts Import Refuses My .CSV File

    I have created a file in Numbers of contacts for gmail.
    I have exported it as a .csv file
    and I selected Western Mac OS Roman
    (When this failed I tried Unicode and Windows Latin1)
    Both of those failed to be accepted by gmail too.
    Each time I try and import this file into gmail
    to the Contacts section it says...
    Sorry we cannot recognize the file format.
    When I made my .csv file...
    I started with a pile of email addresses with names.
    I made sure that each line in the document read:
    name(,) last name(,) <email address> (I hit return here)
    Can somebody please tell me what I am doing wrong???
    thanks.

    Which is the decimal separator in your system.
    If it is the comma, Numbers will not use the comma as values separator but semi-colons.
    This may explain your problem.
    Yvan KOENIG (VALLAURIS, France) vendredi 5 mars 2010 21:09:17

  • Can we strore .CSV file into CLOB datatype

    hi
    can we strore .CSV file into CLOB datatype
    its giving me error ot hexa coonversion?
    can anyone provide sample code...
    when i m sending mail from oracle database when i send as blob object then nope but when i send attachment stored in table with clob column it gives me hex to raw conversion error
    folllowing is my code--
    CREATE OR REPLACE PROCEDURE com_maildata_prc1( pi_sender IN com_batch_contact_dtl.strreceivername%TYPE,
    pi_recipients IN com_batch_contact_dtl.stremailaddr%TYPE,
    pi_subject IN VARCHAR2 ,
    pi_text IN VARCHAR2 ,
    pi_filename IN VARCHAR2 ,
    pi_blob IN cLOB
    IS
    conn utl_smtp.connection;
    i NUMBER;
    len NUMBER;
    BEGIN
    ('com_send_email_prc',Vn_Process_Id);
    conn := demo_mail.begin_mail( sender => pi_sender,
    recipients => pi_recipients,
    subject => pi_subject,
    mime_type => demo_mail.MULTIPART_MIME_TYPE
    demo_mail.begin_attachment(conn => conn,
    mime_type => 'application/csv',
    inline => TRUE,
    filename => pi_filename,
    transfer_enc => 'base64'
    -- split the Base64 encoded attachment into multiple lines
    i := 1;
    len := DBMS_LOB.getLength(pi_blob);
    WHILE (i < len) LOOP
    IF(i + demo_mail.MAX_BASE64_LINE_WIDTH < len)THEN
    UTL_SMTP.Write_raw_Data (conn
    , UTL_ENCODE.Base64_Encode(
    DBMS_LOB.Substr(pi_blob, demo_mail.MAX_BASE64_LINE_WIDTH, i)));
    ELSE
    UTL_SMTP.Write_raw_Data (conn
    , UTL_ENCODE.Base64_Encode(
    DBMS_LOB.Substr(pi_blob, (len - i)+1, i)));
    END IF;
    UTL_SMTP.Write_Data(conn, UTL_TCP.CRLF);
    i := i + demo_mail.MAX_BASE64_LINE_WIDTH;
    END LOOP;
    demo_mail.end_attachment(conn => conn);
    demo_mail.attach_text(
    conn => conn,
    data => pi_text,
    mime_type => 'text/csv');
    demo_mail.end_mail( conn => conn );
    END;
    Thanx in advance...
    Message was edited by:
    user549162
    Message was edited by:
    user549162

    Hi
    Ignore "SQL*Loader-292: ROWS parameter ignored when an XML, LOB or VARRAY column is loaded" error
    after importing your csv file just change length CHAR(100000).
    ex: your column col1 CHAR(1000) to change CHAR(100000).
    and deploy your mapping and execute
    Regards,
    Venkat

  • Remove column name in generated CSV file through ODI

    Hi,
    I want to remove the column names in CSV file which I am creating through database.
    I have tried putting "heading to zero in the the datastore", which I created for csv file, but it is not working.
    Please assist.
    Thanks,
    Sudeep Naik

    Okay let me explain.After creating your csv file you open it.If 1st row contains column name then during reverse you use heading as 1. Regarding c1,c2 these are the column name defined by odi.Its not going to come as your data.The way i told you just make it by giving header 1 then reverse it.After that just drag it to your interface.Then on interface right click on source table and click on view data.Here you can find if your column name is coming as data or not.Hope you are clear now.
    Thanks.

Maybe you are looking for

  • ITunes store wont loads, just comes up with a blank screen

    After installing the lastest update for iTunes i can now not open the iTunes store. The screen just stays white no matter what i try to do. Anyone got any ideas how i can fix this problem?

  • Acrobat Standard 9.1 - Forms Distribution

    I am trying to distribute a form using Acrobat Standard 9.1.  When I try to use either to use the adobe.com or "Save a local copy and manually send it later" option, I am getting this error: "The filename you specified is not valid because it does no

  • Web Developer Needed for a Hawaii Real Estate Agency

    Web Developer Needed for a Hawaii Real Estate Agency (Telecommuting OK. No Relocation) We are an innovative, technology oriented, real estate agency searching for an excellent web developer to continue to improve and grow our 40 agents business. This

  • Planning scenarios

    Friends,     In  SAP documentation, i remember having read that the way sales orders or independent customer requirements consume PIR can be controlled by requirement class and type. then is it possible that even if i choose strategy 10, i can achiev

  • Adobe Photoshop Imageでの画像の削除ができません

    Adobe Photoshop Image の画像が削除できるものとできないものがあります.画像を削除しようとすると「○○○を削除できません.ほかの人またはプログラムによって使用されています.ファイルを使用している可能性が あるプログラムを全て閉じてからやり直してください.」と言うエラーが出てしまいます.プログラムは全部閉じてるし.再起動や一度電源を切ってもだめです.どうしたら削除できるのでしょうか? わかる人がいましたら教えてください.