Export-csv output maximum row size

Hi all. May I know the maximum row size that can output to .csv using 'export-csv'??

There's no limit on writing to a CSV file.
You can try:
# max rows to export to CSV:
$i=0
While ($true) { #Forever
$i++
$i
"Some text" | Export-Csv -Path .\test12.csv -Append -NoTypeInformation
until you run out of disk space, which may take a while. A million lines using the above script makes a 5 BM file..
Now reading it back is an entirely different story. Try
Delimit..
Sam Boutros, Senior Consultant, Software Logic, KOP, PA http://superwidgets.wordpress.com (Please take a moment to Vote as Helpful and/or Mark as Answer, where applicable)

Similar Messages

  • Error when trying to export CSV output.

    When I try to export a regular query to csv I get the following error. Any help would be appreciated.
    ORA-06502: PL/SQL: numeric or value error: character string buffer too small
    ORA-06512: at "FLOWS_030100.F", line 279
    ORA-06512: at line 22
    DAD name: apex
    PROCEDURE : f
    URL : http://adssys2-dc-aus:7777/pls/apex/f?p=107:26:2190330892121429:FLOW_EXCEL_OUTPUT_R4268816727314684_en-us
    PARAMETERS :
    ===========
    p:
    107:26:2190330892121429:FLOW_EXCEL_OUTPUT_R4268816727314684_en-us

    I had this problem. I was going to ask oracle about it.
    I had a plsql function body returning query. I also had a funtion in this plsql function body that return a string. After I removed the funtion, I was able to export.
    declare
    qsql varchar2(4000);
    lqsql varchar2(4000);
    begin
    qsql := 'select * from emp ';
    qsql := qsql || ' where dept_no = 3 ';
    select group_of_emp( :Pxx_START, :Pxx_END ) into l_qsql from dual;
    /* The function returns: and ( emp_no >0 or emp_no < 100 ) */
    qsql := qsql || lqsql;
    qsql := qsql || ' group by emp_no ';
    return qsql;
    end;
    "select group_of_emp( :Pxx_START, :Pxx_END ) into l_qsql from dual; " was the cause of my error.

  • What is the maximum file size for CSV that Excel can open ? (Excel 2013 64bit)

    Hello,
    Before anyone jumps in, I am not talking about the maximum worksheet size of 1048576 rows by 16384 columns.
    I have  client whom has a 1.5 Gb CSV file, 1.9, 2.6, 5, 17 and 89 Gb file (Huge).
    If I open the 1.5 Gb, the file opens (After waiting 5 minutes) and then a warning pops up that only the first 1048576 rows have loaded. That is fair enough.
    If I try and open any of the others, Excel comes up to a blank worksheet. No errors. It just seems to ignore the file I tried to open. This happens from within Excel (File - open) or from double clicking the file in explorer.
    Excel goes to this blank page almost imeadiatly. It does not even try to open the file.
    If I try with Ms Access, I get a size warning and it refuses to load the file. (At least I get a warning)
    I would have expected Excel to load at least the first 1048576 rows  (If that is what there are in the file), and give an error.
    The computer is more than capable (Xeon processors, 16 Gb ram, SSD hard disks top of the line HP Z820 power workstation).
    With the 1.5 Gb file loaded to 1048576 rows, it uses 15% ram/pagefile. CPU's hit about 5%.
    I have confirmed it is Win 7 64bit, Excel 64bit. I am fairly confident we are over the file size but without an error message, I don't know what to tell my client whom is looking to me for answers.
    I have already discussed that the 89gb file in Excel is unreasonable and they are looking at a stat's package but I need an answer on these smaller files.
    Anyone got any ides ?
    Michael Jenkin (Mickyj) www.mickyj.com (Community website) - SBS MVP (2004 - 2008) *5 times Microsoft MVP award winner *Previously MacWorld Australia contributer *Previously APAC Vice Chairman Culminis (Pro IT User group support system)* APAC chairman GITCA
    *Director Business Technology Partners, Microsoft Small Business Specialist, SMB150 2012 Member

    Hi,
    The 1,048,576 rows & 16,384 columns is the
    workbook size limitation in Excel 2013. Thus, I recommend we try the Mr. Bernie's suggestions to import the large CSV file.
    1. Use VBA to read the file line by line and split/examine the import file in sections. If you have further question about the VBA, please post your question to the MSDN forum for Excel
    http://social.msdn.microsoft.com/Forums/en-US/home?forum=exceldev&filter=alltypes&sort=lastpostdesc
    2. Use Excel 2013 add-ins. Power Pivot and Power Query. For more detailed information, please see the below articles: 
    http://social.technet.microsoft.com/Forums/en-US/9243a533-4575-4fd6-b93a-4b95d21d9b10/table-with-more-than-1-048-576-rows-in-power-query-excel-2013?fo
    http://www.microsofttrends.com/2014/02/09/how-much-data-can-powerpivot-really-manage-how-about-122-million-records/
    Please Note: Since the web site is not hosted by Microsoft, the link may change without notice. Microsoft does not guarantee the accuracy of this information.
    Thanks
    George Zhao
    Forum Support
    Come back and mark the replies as answers if they help and unmark them if they provide no help.
    If you have any feedback on our support, please click "[email protected]"

  • Scheduled Excel/CSV output with 65k rows

    Hi Experts,
    I just wanted to confirm if in 3.1, can the scheduled excel/csv output handle >65 rows.  From what we've read, it is possible and we assumed that's thru manual run.  However, we aren't too sure if it is the same when scheduled.
    Kind Regards,
    Mark

    Hi,
    When the Webi excel output is > 65K rows, the 1st tab contains the 1st 65K rows and the rest goes to 2nd tab and forth and each tab has maximum 65K rows.
    However, there is a limitation on the total number of tabs can be created in a excel file during this conversion.  You will find out when you get there.
    Hope this helps,
    Jin-Chong

  • CSV output (export to excel) don't use right code table

    CSV output (export to excel) don't use right code table. We use code table
    8859-2 (reports and forms are OK). Why export to excel don't use this code table.
    Thanks for answers

    You could also take the example from my blob that Denes mentioned and instrument it to output in either CSV or HTML. This way, you could have a single procedure that worked for both formats.
    All you would need to do is handle each field differently - pad a TD and /TD (with proper brackets, of course) if its HTML or just append a comma if it's CSV.
    Thanks,
    &#150; Scott &#150;
    http://spendolini.blogspot.com/
    http://sumnertech.com/

  • Maximum file size for export into MP4?

    Hello,
    I am not able to export 2 hour HD video into standard MP4 file. It seems that reaching 100% export algorithm gets into loop. I was waiting for hours and still had seen progress at exactly 100% with final file size on hard disk to be 0 bytes. I am using CS5 on MAC OS X. I had to split my timeline to 2 parts and to export them separately (which is embarrasing). Is there something like maximum file size for export? I guess that 2h video would have about 25-35GB.
    Thank you
    jiri

    You are right.
    So I am running AP Pro 5.0.4, Adobe Media Encoder 5.0.1.0 (64bit). Operating system Mac OS X ver 10.7.3. All applications are up to date.
    MacBook Pro Intel i5 2.53GB, 8GB RAM, nVidia GT 330M 256 MB, 500GB HDD
    Video is 1920x1080 (AVCHD) 25fps in a .MTS container  (major part of timeline), 1280x720 30fps in .MOV container (2mins), Still images 4000x3000 in .JPG
    No error message is generated during export - everything finishes without any problem...just file created has 0 byte size (as described above).
    This is my largest video project (1h 54min) I dont have any other problem with other projects.
    I dont run any other special software, at the moment of export all usual applications are closed so that MacBook "power" can go to Media Encoder. No codecs installed, using VLC Player or Quick Time.
    Attached please find printscreen from Export settings (AP Pro). Writing this ppost I tried to export only the first 4mins from timeline where all kind of media is used...and it was OK.
    As a next step I will try to export (same settings) 1h 30mins as I still believe problem comes with length of video exported.
    Let me know your opinion

  • Export-csv only generating the output only for single server

    Hi Team,
    From the below script I'm unable to generate the single output file for all the condition. The script only giving the last server output, It's skipping the all the other servers in the file.
    $ScriptBlock = {
        param
            $Server,
            $ExportCSV
    $Counters= import-csv "G:\testcounter.csv"
    foreach($Counter in $Counters)
    $ObjectName=$Counter.ObjectName
    $CounterName=$Counter.CounterName
    $InstanceName=$Counter.InstanceName
    $Result=Get-Counter -Counter "\\$server\$ObjectName($InstanceName)\$CounterName"
        $CounterSamples = $Result | % {$_.CounterSamples}
        $MasterArray = @()
        foreach ($CounterSample in $CounterSamples)
            $TempArray = @()
            $TempArray = "" | Select Server, ObjectName, CounterName, InstanceName, SampleValue, DateTime
            $Split = $CounterSample.Path.Remove(0,2)
            $Split = $Split.Split("\")
            $TempArray.Server = $Split[0]
            $TempArray.ObjectName = $Split[1].Split("(")[0]
            $TempArray.CounterName = $Split[2]
            $TempArray.InstanceName = $CounterSample.InstanceName
            $TempArray.SampleValue = $CounterSample.CookedValue
            $TempArray.DateTime = $CounterSample.TimeStamp.ToString("yyyy-MM-dd HH:mm:ss")
            $MasterArray += $TempArray
        $MasterArray | Export-Csv $ExportCSV -NoType
    $Servers = import-csv "G:\testcounter.csv"
    foreach ($Server in $Servers)
        $server=$server.server
        If (Test-Connection -quiet -computer $server){
            $ExportCSV = "G:\PerformaneData.csv"
            Start-Job -ScriptBlock $ScriptBlock -ArgumentList @($server, $ExportCSV)

    Hi RatheeshAV,
    In addition, to export the result to csv file, please also try to wait for all the jobs to complete then retrieve all the data and write it to a file in one step:
    $ScriptBlock = {
    param ($Server)
    #SCRIPT
    $MasterArray }
    $Servers = import-csv "G:\testcounter.csv"
    Get-Job | Remove-Job
    $jobs=@()
    foreach ($Server in $Servers){
    $server=$server.server
    If(Test-Connection -quiet -computer $server){
    Write-Host $Server -ForegroundColor green
    $ExportCSV = "G:\PerformaneData.csv"
    $job+=Start-Job -ScriptBlock $ScriptBlock -ArgumentList $server
    $jobs | Wait-Job
    $jobs | Receive-job | Export-CSV 'd:\temp.csv' -NoTypeInformation
    Refer to:
    PS3 Export-CSV -Append from
    multiple instances to the same csv file
    If there is anything else regarding this issue, please feel free to post back.
    Best Regards,
    Anna Wang

  • Creating a CSV output of just the rows of data.

    Dear All,
    I am trying to create a CSV output from SQL*Plus using the following method... All I need is the pure data, and nothing else.
    set verify off
    set feedback off
    set termout off
    set heading off
    set echo off
    spool <filename>
    <my sql statement>
    spool <off>
    The problem is that the SQL statement is displayed in the output file.
    Does anyone know how to switch this off, before I spool the file, or can I spool the file immediately after execution. I.e. Build this into the SQL statement (somehow)???
    Thanks
    David.

    If you want to handle 100% the output file you can use the package UTL_FILE but the file is going to be store in the server.
    UTL_FILE
    With the UTL_FILE package, your PL/SQL programs can read and write operating system text files. UTL_FILE provides a restricted version of operating system stream file I/O.
    http://download-west.oracle.com/docs/cd/B10501_01/appdev.920/a96612/u_file.htm#ARPLS069
    Joel Pérez
    http://otn.oracle.com/experts

  • Public Folder To Mailbox Map Generator maximum mailbox size

    Hello,
    I am in the planning stages of a 2010 to 2013 public folder migration.
    Upon running the following script to generate the mailbox mappings in 2013 I receive the following error regarding the maximum mailbox size requested.
    [PS] C:\Program Files\Microsoft\Exchange Server\V15\scripts>.\PublicFolderToMailboxMapGenerator.ps1 5120MB E:\Migration\
    PublicFolders\PFMapNameToSize.csv E:\Migration\PublicFolders\PFMapFolderToMailbox.csv
    [25/11/2014 11:32:08] Reading public folder list...
    [25/11/2014 11:32:08] Loading folder hierarchy...
    [25/11/2014 11:32:08] The size of the folder @{FolderName=\IPM_SUBTREE\Company\Admin Team\Info; FolderSize=174
    62349238} (17462349238) is greater than the mailbox size 5368709120
    [25/11/2014 11:32:08] Unable to load public folders...
    I want to set the maximum mailbox size to 5GB, however it seems that this is not possible given that some of the existing folders exceed the maximum mailbox size specified.
    I have increased the mailbox size gradually until there are no errors reported, but this leaves me with a maximum mailbox size of 50GB which is not what I require.
    There is circa 45 GB of data in the PF hierarcy currently, with the largest folder containing around 42 GB of data.
    I thought the idea was that the mapping generator would split the public folder data up and distribute it across multiple mailboxes as required?
    Can anyone assist with achieving this please?
    Regards
    Matt

    Hi Belinda,
    Thanks for the reply. I've done some more investigating, as I thought the public folders sizes being reported were larger than expected.
    Upon running the Get-PublicFolderStatistics cmdlet across the public folder hierarchy I can indeed see that the largest public folder we have is only 4.46GB in size, not the 42GB being reported by the generator.
    PublicFolderToMailboxMapGenerator
    [PS] C:\Program Files\Microsoft\Exchange Server\V15\scripts>.\PublicFolderToMailboxMapGenerator.ps1 25600MB E:\Migration
    \PublicFolders\PFMapNameToSize.csv E:\Migration\PublicFolders\PFMapFolderToMailbox.csv
    [25/11/2014 11:40:10] Reading public folder list...
    [25/11/2014 11:40:10] Loading folder hierarchy...
    [25/11/2014 11:40:10] The size of the folder @{FolderName=\IPM_SUBTREE\Company\Dept\SubDept Inbox\Offers; FolderSize=45160928457} (45160928457) is greater than the mailbox size 26843545600
    [25/11/2014 11:40:10] Unable to load public folders...
    Get-PublicFolderStatistics
    Name ItemCount TotalItemSize LastUserAccessTime LastAccessTime FolderPath Offers 4314 4.46 GB (4,788,856,098 bytes) 25/11/2014 11:41 26/11/2014 08:14 Company\Dept\SubDept Inbox\Offers
    Going back a step - before running the PublicFolderToMailboxMapGenerator.ps1
    , the issue is initially occurs when running the Export-PublicFolderStatistics.ps1 as this is where the folder sizes are calculated and output to the PFMapNameToSize.csv. This is where the folder size is inflated by 10 times!
    Is this a bug?
    Does anyone know how to change this behaviour?
    Regards
    Matt

  • Create export-csv with dynamic (unknown amount) columns

    Hi,
    This is my first post, hopefully I include everything you need.  My code block is at the bottom of my post.....
    I have written a PS script that gives me the info I need, but I would like to format it differently.  Currently I import a CSV with 2 columns - username and print queue name.  The script then takes each username and looks it up in SCCM
    to find the workstations the username has logged into.  I then export to csv - this is where I would like to format it differently.  The export csv has 4 columns username, print queue name, workstation name, and details. 
    username1,printqueue,workstation1,details
    username1,printqueue,workstation2,details
    username2,printqueue,workstation1,details
    username2,printqueue,workstation2,details
    My problem is that if the user logs into 5 workstations I have 5 rows with duplicated username and print queue name.  If my next username logs into 8 workstations I have 8 rows, etc.  I would like to format the export dynamically using as many
    columns (not rows) as need.  For example
    username1, printqueue,workstation1,workstaion2, workstation3
    username2,printqueue,workstation1,workstation2, workstation3,workstation4
    I expect where I have my pscustomobject would be where I could change this - I have tried setting up a count variable and using that in my loop to be able to assign dynamic column name, I have tried using an array but still new to setting that up, maybe
    I just wasn't doing it correctly.  If anyone has any ideas that would be great.
    Thanks, Kevin
    Here is my code:
    #Declare variables
    $now = Get-date
    $date = get-date -uformat "%Y_%m_%d_%I%M%p"
    $Queues = Import-Csv "c:\Users.csv"
    $SiteName="XXX"
    $SCCMServer="your.sccm.server.com"
    $SCCMNameSpace="root\sms\site_$SiteName"
    $CSVPath1 = "c:\PrinterUsersByComputer_$date.CSV"
    #Notify user that script is starting
    Write-Host -ForegroundColor Green "Starting script"
    #Loop through each account and find any workstations from SCCM that the user has logged into
    foreach ($Queue in $Queues){
    $Queue.Username
    $WSs = Get-WmiObject -namespace $SCCMNameSpace -computer $SCCMServer -query "select Name from sms_r_system where LastLogonUserName = '$($queue.username)'" | Select-Object -ExpandProperty Name
    #Check if the workstation variable has data
    if ($WSs){
    foreach ($WS in $WSs) {
    #Check if workstation is ALIVE and if we have access to it to get OS version
    $rtn = (Test-Connection -Cn $WS -BufferSize 16 -Count 1 -Quiet)
    IF ($rtn -match 'True') {
    $OS = Get-WmiObject Win32_OperatingSystem -ComputerName $WS | Select-Object -ExpandProperty Caption -ErrorAction Stop
    #Write to screen device to show progress
    Write-Host -ForegroundColor Green "Computer information found: $WS"
    #Write to log file the username, workstation(s) logged into, and OS version
    [PSCustomObject] [Ordered] @{
    'UserName' = $Queue.Username
    'Print Queue' = $Queue.PrintQueue
    'ComputerName' = $WS
    'Operating System' = $OS
    } | Export-Csv $CSVpath1 -Append -NoTypeInformation

    Ok, just a sec, I'm not sure we are on the same page.  I don't think that isn't what I'm look for. Unless I'm missing something, your code above assume 4 workstation columns for every row.  What if there is 5 or 8.  My current output
    from the full script is this (based on checking if PC is pingable, whether I'm denied the WMI query, etc)
    UserName
    Print Queue
    ComputerName
    Operating System
    User1
    PQ1
    Computer1
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer2
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer3
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer4
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer5
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer6
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer7
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer8
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer9
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer10
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer11
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer12
    Microsoft Windows 7 Enterprise 
    User1
    PQ1
    Computer13
    No device name to query
    User1
    PQ1
    Computer14
    Microsoft Windows 7 Enterprise 
    User2
    PQ2
    Computer15
    Microsoft Windows 7 Enterprise 
    User3
    PQ3
    Computer1
    Microsoft Windows 7 Enterprise 
    User4
    PQ4
    Computer2
    Access denied connecting to device.    No access to it: Computer2
    What I would really like is this, getting rid of the OS name for now:
    User1
    PQ1
    Computer1
    Computer2
    Computer3
    Computer4
    Computer5
    User2
    PQ3
    Computer1
    User3
    PQ13
    Computer1
    Computer2
    I would just like to condense the workstations to columns instead of 1 workstation per row and be able to expand the # of columns required based on the amount of workstations in $WSs -

  • CSV output includes div info in first column

    All,
    I have a report on a page that allows for CSV output. For some reason when I look at the output the first column in the header row includes div information and appears as follows:
    &#60;div id="report_1585218597659393_catch"&#62;"Title"Then in the last column there is an "&#60;/div&#62;". Does anyone know what would cause this? I've tried changing some column names and even deleting and recreating the report but neither helped.
    Dan

    All,
    As was pointed out in earlier replies, this is being caused by enabling PPR pagination. If PPR and CSV export are both needed, I would suggest using the pre-3.1 style PPR report templates. Those have the PPR stuff burned in directly.
    I did fix the underlying bug though, yet this is not in the latest 3.1.1 patch release. So I don't know when this fix will become available.
    Regards,
    Marc

  • Csv output w/o quotes

    I have seen a lot of good ideas regarding csv output (such as the trick to output more than 10000 rows), and I was hoping someone could steer me in the right direction with this question.
    How would I go about generating csv ouput that does not contain quotes around the field data... just commas. Thanks
    Kevin

    Kevin,
    Essentially this is working as any other report, the only difference is that instead of using a template with HTML tags, the column values are enclosed in double quotes and separated by comma. This is hard-wired into the engine. So this can't be changed in the current version. You could write your own PL/SQL code to print out whatever you like, for this you might find the following posting helpful:
    Re: Tab delimited export
    Regards,
    Marc

  • Maximum file size of 2 GB exceeded please choose a shorter bounce time

    I have done a thorough search online (w/google) trying several combinations of words but I seem to be the only person on the planet with this error. I guess I will try to remake the project but I don't think it will fix the problem. I also will update my OS to 10.4.10, anyways here is the the error message.
    "Maximum file size of 2 GB exceeded please choose a shorter bounce time"
    I just bought and installed iLife 08 and the 8.1 update for Garageband. I have a 3 hr and 50 track on the time line and I have added chapter marks with 16k pics. The original combined track size (wav files) did exceed 2GB but I edited the files in itunes (to mono wav files) and swapped in the new files so that the max file size should only be 1.5GB. I still get the error message above when I try to export the podcast to disk using AAC w/Mono Podcast setting. Your help is appreciated.
    Message was edited by: Thad

    I am now the third person in the world to get this error message, except my GB project is a mere 1 hour and 45 minutes long. I've outputted longer projects before and never got this error before. It's a new one to me, and frustrating. How exactly am I supposed to choose a "shorter bounce time" if there is no explanation anywhere in Appleworld of what a bounce time is or how to set it shorter?
    Again, I'm not doing anything different with this project than others that shared successfully, and the dialog box estimates my outputted file size to be approx. 100 MB.
    Can anybody help me get through this error blockage? Please?

  • What is the maximum file size allowed by Yosemite's Archive Utility?

    I've seen radically different answers to this question, ranging from ".zip files have a 4GB size limit" to "no limit other than the OS maximum file size". It's been very difficult to find current info on this.
    I have lots of large folders full of recording projects --- I'm talking up to 45GB or so --- and I've been experimenting with different methods of zipping them in preparation for cloud backup. Several applications have been willing to zip my enormous folders into archives, but then choke when unzipping them, telling me that the files cannot be unzipped or are corrupted. Using the "Compress" shortcut in the finder has so far yielded the best results, keeping all necessary resource forks, etc., and unzipping properly, even when other apps have insisted the files are corrupted. However, I'm afraid to continue with this procedure until I can get a straight answer on this issue: have I just been lucky so far? IS there an actual size limit?
    ...It only seems to work when I both zip and unzip with the Archive Utility. Other archiving apps will either render the included projects unreadable by my recording software when zipping (I suspect because of alteration or stripping of resource forks), or will refuse to read the giant .zip archives generated by Archive Utility.
    I've set up the Archive Utility to produce .zip files, as I feel these are the most likely to be recognized by a variety of cloud storage services.ny
    ...Any guidance would be greatly appreciated!

    Hi,
    The 1,048,576 rows & 16,384 columns is the
    workbook size limitation in Excel 2013. Thus, I recommend we try the Mr. Bernie's suggestions to import the large CSV file.
    1. Use VBA to read the file line by line and split/examine the import file in sections. If you have further question about the VBA, please post your question to the MSDN forum for Excel
    http://social.msdn.microsoft.com/Forums/en-US/home?forum=exceldev&filter=alltypes&sort=lastpostdesc
    2. Use Excel 2013 add-ins. Power Pivot and Power Query. For more detailed information, please see the below articles: 
    http://social.technet.microsoft.com/Forums/en-US/9243a533-4575-4fd6-b93a-4b95d21d9b10/table-with-more-than-1-048-576-rows-in-power-query-excel-2013?fo
    http://www.microsofttrends.com/2014/02/09/how-much-data-can-powerpivot-really-manage-how-about-122-million-records/
    Please Note: Since the web site is not hosted by Microsoft, the link may change without notice. Microsoft does not guarantee the accuracy of this information.
    Thanks
    George Zhao
    Forum Support
    Come back and mark the replies as answers if they help and unmark them if they provide no help.
    If you have any feedback on our support, please click "[email protected]"

  • How to add SaveFileDialog to PowerShell Get-ADUser Export-CSV

    Hi,
    I am having a bit of difficulties with getting the SaveFileDialog to work when I use the Get-ADUser export-CSV.
    Current code:
    $handler_Output_Click=
    $User = Get-Aduser $textBox1.Text -Properties DisplayName,sAMAccountName,EmailAddress,Mobile,Company,Title,Enabled,LockedOut,Description,Created,Modified,LastLogonDate,AccountExpirationDate,AccountLockoutTime,BadLogonCount,CannotChangePassword,LastBadPasswordAttempt,PasswordLastSet,PasswordExpired,LogonWorkstations,CanonicalName | Select DisplayName,sAMAccountName,EmailAddress,Mobile,Company,Title,Enabled,LockedOut,Description,Created,Modified,LastLogonDate,AccountExpirationDate,AccountLockoutTime,BadLogonCount,CannotChangePassword,LastBadPasswordAttempt,PasswordLastSet,PasswordExpired,LogonWorkstations,CanonicalName | Export-CSV C:\temp\test.csv -NoTypeInformation ';' -encoding utf8
    $richTextBox1.Text = "A file 'test.csv' has been created to C:\temp\ based on the user: $($textBox1.Text)"
    Here a specific filename is already defined and I have to edit the code each time I want a different filename.
    It would be perfect if I could implement the SaveFileDialog box so I have the ability to name the file before saving and possibly even have the option to select the file type (among .CSV and All files).
    This it the export/output button itself:
    $System_Drawing_Point = New-Object System.Drawing.Point
    $System_Drawing_Point.X = 502
    $System_Drawing_Point.Y = 38
    $Output.Location = $System_Drawing_Point
    $Output.Name = "Output"
    $System_Drawing_Size = New-Object System.Drawing.Size
    $System_Drawing_Size.Height = 23
    $System_Drawing_Size.Width = 85
    $Output.Size = $System_Drawing_Size
    $Output.TabIndex = 2
    $Output.Text = "Export as file"
    $Output.UseVisualStyleBackColor = $True
    $Output.add_Click($handler_Output_Click)
    # $form1.AcceptButton = $Output
    $Output.DataBindings.DefaultDataSourceUpdateMode = 0
    $form1.Controls.Add($Output)
    And in the beginning of my script I also have defined the following:
    [System.Windows.Forms.Application]::EnableVisualStyles();
    [reflection.assembly]::loadwithpartialname("System.Windows.Forms") | Out-Null
    [reflection.assembly]::loadwithpartialname("System.Drawing") | Out-Null
    [reflection.assembly]::loadwithpartialname("System.Windows.Forms.SaveFileDialog") | Out-Null
    And also:
    $Output = New-Object System.Windows.Forms.Button
    Any ideas how can I implement the SaveFileDialog so when I press the "Export as file" button the PowerShell command "Get-Aduser $textBox1.Text -Properties DisplayName,sAMAc..." is ran and I can choose the file name from a pop-up
    dialog box where to save the file and also put a filename? Currently I have to edit the code in order to assign a new file name (or go rename the file from that location).
    Thank you in advance,
    Henri
    EDIT:
    I know that the below is the answer to the SaveFileDialog, however I cannot imagine how I could implement it to my script into the "Get-Aduser $textBox1.Text -Properties a,b,c,d | select a,b,c,d | Export-CSV C:\temp\test.csv" cmdlet.
    Function Get-SaveFile($initialDirectory)
    [System.Reflection.Assembly]::LoadWithPartialName("System.windows.forms") |
    Out-Null
    $SaveFileDialog = New-Object System.Windows.Forms.SaveFileDialog
    $SaveFileDialog.initialDirectory = $initialDirectory
    $SaveFileDialog.filter = "All files (*.*)| *.*"
    $SaveFileDialog.ShowDialog() | Out-Null
    $SaveFileDialog.filename

    Just run the dialog before exporting the file.  Why is that a problem?
    \_(ツ)_/
    It works now! I made some modifications and it works. Thank you very much for the advice.
    $handler_Output_Click={
    Add-Type -AssemblyName System.Windows.Forms
    $SaveAs1 = New-Object System.Windows.Forms.SaveFileDialog
    $SaveAs1.Filter = "CSV Files (*.csv)|*.csv|Text Files (*.txt)|*.txt|Excel Worksheet (*.xls)|*.xls|All Files (*.*)|*.*"
    $SaveAs1.SupportMultiDottedExtensions = $true;
    $SaveAs1.InitialDirectory = "C:\temp\"
    if($SaveAs1.ShowDialog() -eq 'Ok'){
    $User = Get-Aduser $textBox1.Text -Properties DisplayName,sAMAccountName,EmailAddress,Mobile,Company,Title,Enabled,LockedOut,Description,Created,Modified,LastLogonDate,AccountExpirationDate,AccountLockoutTime,BadLogonCount,CannotChangePassword,LastBadPasswordAttempt,PasswordLastSet,PasswordExpired,LogonWorkstations,CanonicalName | Select DisplayName,sAMAccountName,EmailAddress,Mobile,Company,Title,Enabled,LockedOut,Description,Created,Modified,LastLogonDate,AccountExpirationDate,AccountLockoutTime,BadLogonCount,CannotChangePassword,LastBadPasswordAttempt,PasswordLastSet,PasswordExpired,LogonWorkstations,CanonicalName | Export-CSV $($SaveAs1.filename) -NoTypeInformation ';' -Encoding UTF8
    $richTextBox1.Text = "A file $($SaveAs1.filename) has been created based on the user: $($textBox1.Text)"

Maybe you are looking for

  • Memory leak in JSpinner implementation (maybe others?)

    Hi, I am developing an application using Java and Swing, and have run into some problems with memory leaks. After examining the source code and making an example program (provided below), I can only come to the conclusion that there is a bug in the i

  • Show/Hide Divs Problem in IE6

    Hello, and thanks for taking a look at this. I have inherited a web site which isn't performing very well in IE6, more details later. The setup is as follows: The website delivers training content. The design is such that there are no scrollbars, all

  • Multiple query in print layout

    I have a problem. I wanna select with apex both two region that i have in my page to compose a document that is made by this two report. The problem is that when i define my report queries in Shared components it is possible to define only one query

  • Some Albums on iTunes separate into one album for each song on iPod

    Some of my albums, imported from CD's, show up properly on iTunes. After the sync, each song in the album shows up on the iPod Touch as a separate album with the one song. All the songs on the album are "ticked" in the iTunes list before the sync.

  • Having issue with update Adobe 11.0.06.  Error 1603.

    Having issue with update Adobe 11.0.06.  Error 1603.