How to deal with Upload limit of Document Library

Hi
We are working on SharePoint Online (Office365). We have a requirement to upload files of 20GB+ size. I have come across a limitation that we can't upload files to document library whose size is more than 2GB. Please provide an clarification on this
limitation.
We want our user to upload files up to 20GB. Can we give him an option to upload file using SP Designer. If we can create a folder an user can open site in SPD then they can upload files. Is there any limitation of file size in folder.
If this is not the feasible option kindly suggest us any other way by which we can fulfill this requirement.
Regards
Vikrant Kaushal

Vikrant,
Please see the SharePoint Online Boundaries and Limits document
http://office.microsoft.com/en-gb/office365-sharepoint-online-enterprise-help/sharepoint-online-software-boundaries-and-limits-HA102694293.aspx#_SharePoint_Online_for_2
The 2Gb file limit is a hard limit that cannot be exceeded in either on-premises or cloud based versions of SharePoint due to a SQL/File system limitation.
The best bet that you can do is look at using a storage location based in perhaps Azure and then linking to this location from Office 365. Unfortunately you can't use search crawl for none Office 365 locations, so I would think about using a links list to
surface the large files in SharePoint as a link item, thus making them available to search etc.
Paul 
Please ensure that you mark a question as Answered once you receive a satisfactory response. This helps people in future when searching and helps prevent the same questions being asked multiple times.

Similar Messages

  • How external users can upload files in document library hosted in onpremise SharePoint?

    Hi all,
    We have setup a document library(DL) in SharePoint site hosed on-premise with no direct access from public zone/internet being an intranet solution. We have some external users who need to upload some files that needs to go in this document
    library. These files are quite confidential in nature so security is a prime requirement. Following are the couple of options I have in mind:
    1. External users upload their file in their designated folder to secure FTP using their own credential set. And also there is a background process (kind of directory watcher) that keep looking at the all the folders
    and moves the files to document library
    2. We host a single page application which has One time password (OTP) kind of feature. External user comes to the secure site to specify Vendor team name and mobile number. OTP is generated and user is asked to specify the password along with file
    they would like to upload. This page also maps the vendor name with respective folder in DL and moves the files to SharePoint there on the fly.
    Pls. share your opinion about these approaches. Also welcome your thoughts if there is any other better option.
    Thanks.

    Hi,
    I agree with the Margriet, the first option is simple and safe, the users could not access the document library directly.
    For security, it is an importance for users not access the document library directly, it will reduce the risk of the documents in the library.
    Thanks,
    Jason
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Jason Guo
    TechNet Community Support

  • How to deal with 4GB limit of the mail box

    Thunderbird says there is not enough space, even after archiving too. please guide to increase the limit from 4 GB to higher

    Is that message regarding Thunderbird folders or the server?
    Background info:
    Pop mail account folders and synchronised subscribed imap folders download emails and store them in mbox files.
    You will see Folders in your Folder Pane, but in reality the emails are stored one after the other, in the order they were downloaded in a file. So they are not stored as separate emails.
    If you have them stored in eg: Inbox, these emails are actually stored in one file; a single document.
    So it is a good idea to organise them into separate folder in the Folder Pane to make the files smaller, This means therre is less risk and it uses less memory to open the files. Archiving by year and month and keeping folder structure will organise the emails into small files.
    See Archiving options:
    https://support.mozilla.org/en-US/kb/archived-messages
    Emails that are deleted are 'marked as deleted' and from your point of view they are moved to the 'Deleted' folder. In reality they are still in the Inbox (or what ever folder) just hidden, marked as deleted and still taking up space.
    It is possible that you have old deleted mail taking up space.
    So to remove all of those 'marked as deleted' emails you would need to right click on folder and select 'compact' or you can compact all folders using 'File > Compact folders.
    Please note that if the folder has any corruption, you may lose emails.
    As a precaution especially on the Inbox as this folder sees the most downloading, moving and deleting; I would move all good emails into suitable folders; delete what you do not want and then when the Inbox is empty compact the folder. As general useage, the Inbox should ideally be kept low in email numbers and not used as a general storage.
    The Sent folder can also get a lot of emails, so make sure you Archive that folder as well.
    Junk and Deleted folders need to be emptied and compacted too.
    '''If you got that message from the server'''
    Check how many emails are kept on the server.
    '''If you have Pop mail accounts''' then the server would expect you to download and delete frequently off the server.
    You can configure this in your Account Settings.
    right click on mail account name in Folder Pane and select 'Settings', then choose 'Server Settings for the mail account
    or
    Tools > Account Settings > Server Settings for the mail account
    or
    Menu icon > options > Account Settings > Server Settings for the mail account
    You can select to 'Leave messages on server'
    You can also select 'for at most eg: 32 days'
    which will only keep the more recent email
    You can also select 'Until I delete them'
    This will delete anything you have deleted if less than eg: 32 days.
    Click on Ok to save changes.
    '''If you have an IMAP mail account''' and you have reached you quota limits, I would suggest you synchronise folders and then in offline mode move some emails into suitably named folders in Local Folders. These emails will be kept on your computer and not the server.
    Remember any emails only stored on your computer will need to be backed up.
    Backup info:
    * http://kb.mozillazine.org/Profile_backup
    * http://kb.mozillazine.org/IMAP_backup

  • How to deal with file(for example .xml)? what format of dir should be?

    I'd like to operate the file in disk, and want to use relative directory?
    How to deal with file dir? what format of dir should be?

    Hi Kamlesh,
    Thanks for your response.
    Actually, In the "Process External Bank Statement" window, i see that there are few entries which is for the previous year and which has not been reconciled. I have never worked practically on BRS and hence, i am scared to make any changes in the clients database without being confident on what i am doing. I need to reconcile for one of their Bank a/c for the month of April '08. I have the copy of the statements for the month ending 31st Mar 08 and 30th Apr 08. The closing balances are as below:
    31/03/08 - 2300000.00
    30/04/08 - 3100000.00
    Now my OB for Bank a/c for April '08 in SAP is 2300000.00 Dr.
    When i go to External Bank Reconciliation - Selection Criteria Screen (Manual Reconciliation), here are the detail that i enter:
    Last Balance: INR -7,000,000.00000 (Grayed out by the system)
    Ending Balance: INR -3,100,000.00000 (Entered by me)
    End Date: 30/04/08 (Entered by me)
    "Reconciliation Bank Statement" Screen opens up and shows the below balances in the screen:
    Cleared Book Balance: INR -7,000,000.00000
    Statement Ending Balance: INR -3,100,000.00000
    Difference: INR 3,800,000.00000
    As per the Bank statement, i have found all the transactions listed out here for the month of Apr '08 but, i also found that the open transactions for the previous month from April '08 have been lying in "Process External Bank Statement" window.
    Could you please help me solve my issue as to what needs to be done or could you also get me some links from where i can get few documents for processing External Bank Reconciliations?
    That will be of a great help for me. I need steps as to what needs to be done first and then the next so that i can arrive at the correct closing balance for the month April '08.
    Thanks in Advance....
    Regards,
    Kaushal

  • How to deal with special character in source file

    Hi experts,
                      i am doing a file to file scenario in which my source file contains many special characters when i am puting this file into moni its going with the special character .My source file is a fixed length file so in content conversion i have specified the file length but due to these special charcters these field lenght is also varing.So please guide me how to deal with these special characters in sender adapter
    regards,
    Saurabh

    you could try using a Java Mapping to change the encoding manually. For that, set the encoding of the OutputFormat of the XML you'll serialize. Try the following code piece for the mapping (inside a try/catch declaration):
    DocumentBuilderFactory factory = DocumentBuilderFactory.newInstance();
    DocumentBuilder documentBuilder = factory.newDocumentBuilder();
    Document input = documentBuilder.parse(in);
    OutputFormat format = new OutputFormat(XML, "ISO-8859-1", false);
    XMLSerializer serializer = new XMLSerializer(out, format);
    For more details check this guide:
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/uuid/502991a2-45d9-2910-d99f-8aba5d79fb42

  • How to deal with BP inconsistencies in SRM?

    Dear Experts,
    How to deal with Business Partner inconsistency in SRM?
    A BP is damaged beyond repair?
    Should we delete the BP of the user (SU01 user)? Are we allowed to delete a BP? or do we need to archive the BP?
    What will happen to the follow-on documents of the particular BP?
    Should we close all the transactional documents or we should simply the change the BP of the documents manaully?
    Please, share your expereince.
    Thanks and regards,
    Ranjan

    Hi
    What kind of Inconsistencies are there...Deletion of BP is last resort in such cases. Although its only possible to put a deletion/blockecd indicator for the BP. BP is never deleted completely from SRM unless you archive it.
    Regards
    Virender Singh

  • Move of mailbox with over 30K of folders fails, how to deal with this.

    Hello all,
    I'm in the process of moving all Exchange 2010 mailboxes to newly created databases located on new storage.
    Every mailbox except one moved successfully, the "problem" mailbox keeps failing and I need some advise on how to deal with this.
    The mailbox that keeps failing is only 4 GB but has over 30K of folders that the users created manually over the years.
    They are unable to easelly delete folders as they have created many sub folders within the folders.
    I posted a part of the failed move request log below.
    Can anyone advise on how to solve this?
    Thanks in advance.
    21-2-2015 18:15:13 [EXHUB1] Fatal error MapiExceptionMaxSubmissionExceeded has occurred.
    Error details: MapiExceptionMaxSubmissionExceeded: Unable to save changes. (hr=0x80004005, ec=1242)
    Diagnostic context:
        Lid: 55847   EMSMDBPOOL.EcPoolSessionDoRpc called [length=7574]
        Lid: 43559   EMSMDBPOOL.EcPoolSessionDoRpc returned [ec=0x0][length=360][latency=0]
        Lid: 23226   --- ROP Parse Start ---
        Lid: 27962   ROP: ropWriteStream [45]
        Lid: 27962   ROP: ropSetProps [10]
        Lid: 27962   ROP: ropSaveChangesMessage [12]
        Lid: 17082   ROP Error: 0x4DA     
        Lid: 18273  
        Lid: 21921   StoreEc: 0x4DA     
        Lid: 27962   ROP: ropExtendedError [250]
        Lid: 1494    ---- Remote Context Beg ----
        Lid: 26426   ROP: ropSaveChangesMessage [12]
        Lid: 27420   dwParam: 0xE1F000B
        Lid: 36537   dwParam: 0xE1F000B
        Lid: 5587    StoreEc: 0x8004010F PropTag: 0xE1F000B 
        Lid: 33375   StoreEc: 0x8004010F
        Lid: 27420   dwParam: 0x806684B0
        Lid: 36537   dwParam: 0x806684B0
        Lid: 5587    StoreEc: 0x8004010F PropTag: 0x806684B0
        Lid: 64931   StoreEc: 0x8004010F
        Lid: 46509  
        Lid: 40193   StoreEc: 0x8004010F
        Lid: 5587    StoreEc: 0x8004010F PropTag: 0x30130102
        Lid: 21970   StoreEc: 0x8004010F PropTag: 0x668F0040
        Lid: 64319   dwParam: 0x5000
        Lid: 39743   dwParam: 0x6730
        Lid: 45016   StoreEc: 0x4DA     
        Lid: 57304  
        Lid: 5041    StoreEc: 0x4DA     
        Lid: 4465    StoreEc: 0x4DA     
        Lid: 6926    StoreEc: 0x4DA     
        Lid: 1750    ---- Remote Context End ----
        Lid: 31418   --- ROP Parse Done ---
        Lid: 21457  
        Lid: 19665   StoreEc: 0x4DA     
       at Microsoft.Mapi.MapiExceptionHelper.ThrowIfError(String message, Int32 hresult, SafeExInterfaceHandle iUnknown, Exception innerException)
       at Microsoft.Mapi.MapiProp.SaveChanges(SaveChangesFlags flags)
       at Microsoft.Exchange.MailboxReplicationService.MoveObjectInfo`1.CreateMessage(T obj, GetAdditionalProperties getAdditionalPropertiesCallback)
       at Microsoft.Exchange.MailboxReplicationService.MoveObjectInfo`1.CreateMessage(T obj)
       at Microsoft.Exchange.MailboxReplicationService.LocalMailbox.Microsoft.Exchange.MailboxReplicationService.IMailbox.SaveSyncState(Byte[] key, String syncStateStr)
       at Microsoft.Exchange.MailboxReplicationService.MailboxWrapper.<>c__DisplayClass57.<Microsoft.Exchange.MailboxReplicationService.IMailbox.SaveSyncState>b__56()
       at Microsoft.Exchange.MailboxReplicationService.ExecutionContext.Execute(GenericCallDelegate operation)
       at Microsoft.Exchange.MailboxReplicationService.MailboxWrapper.Microsoft.Exchange.MailboxReplicationService.IMailbox.SaveSyncState(Byte[] key, String syncState)
       at Microsoft.Exchange.MailboxReplicationService.MailboxCopierBase.SaveSyncState(SaveStateFlags flags)
       at Microsoft.Exchange.MailboxReplicationService.BaseJob.SaveSyncState(SaveStateFlags flags)
       at Microsoft.Exchange.MailboxReplicationService.BaseJob.SaveState(SaveStateFlags flags, UpdateRequestDelegate updateRequestDel)
       at Microsoft.Exchange.MailboxReplicationService.MoveBaseJob.SortMessages(Object[] wiParams)
       at Microsoft.Exchange.MailboxReplicationService.CommonUtils.CatchKnownExceptions(GenericCallDelegate del, FailureDelegate failureDelegate)
    Error context: --------
    Operation: IMailbox.SaveSyncState
    OperationSide: Target
    Primary (a4966c13-778a-4d70-9087-f077e2523f66)
    Key: FEF1D04E070AB74A99B809B7797674AE000000000000000000000000000000009B5540B45B37774CAB84806C1964B598
    SyncStateLength: 27050362
    21-2-2015 18:15:13 [EXHUB1] Relinquishing job.

    Hi,
    Based on the description, please run the following commands to check if there are any message limit size. Try to increase the limit size then move this mailbox.
    Get-mailbox –identity “user name” | fl MaxSendSize,MaxReceiveSize
    Get-transportconfig | fl MaxSendSize,MaxReceiveSize
    Get-transportrule | fl MessageSizeOver
    Get-sendconnector | fl MaxMessageSize
    Get-receiveconnector | fl MaxMessageSize
    Note: Once we increased the value we need to wait for the AD replication and cache to be cleared and then test the move mailbox.
    If this doesn’t work, please try to move mailbox with the skip corrupted items. This will skip the corrupted items and move only the good items to the target mailbox.
    Similar thread:
    https://social.technet.microsoft.com/Forums/en-US/5383a826-4383-4b26-a503-2c3bb635f9d2/mailbox-move-to-exchange-2010-fails-for-mailboxes-over-23-gb?forum=exchangesvrdeploylegacy
    Best Regards.

  • How to deal with CGEHS (central goverment employees health scheme0

    HI!
    All!
    Can any one please tel me as to how to  deal with CGEHS --- Central Government Employees Health Scheme in indian pay roll, which is not available in the Standard Indian Pay roll.
    How much development is needed?
    My understanding is that the following are required to be done..
    A) we need to create a cumulative wage type in It 14 whcih applies to all the employees then accordingly a pcr
    b) we need to create a unique ID, i mean CGEHS id and assign to every employee in IT 0105
    c)  creation of various clinics/hospitals.
    d) provision for reimbursement... etc...
    My doubts are..
    a) how to deal with the retired employees CGEHS contribution and reimbursement?
    how much developement is needed interms of user exits or abap developement.?
    Can any one please give some idea or some kind of document may be proto kind of doc will also do...
    Its little urgent..
    answer would be awarded the points.
    thanking you in advance..
    regards
    rao

    Further to Madhu,I feel you can redefine your feature IGMOD in such a way so that the same set of fields become applicable under different module pools.
    V_t588D and Feature IGMOD needs to be redefined in such cases.

  • Batch uploading to a document library with content types

    Hi,
    I have a document library with two content types. Now I need to upload documents to this document library. However, if I have large number of files, then uploading through web UI is a pain as metadata will not be attached. How should I automate it  from
    my computer using powershell without relying on Administrator to run the powershell script. Or is there any other way to do it without using powershell? or  can powershell be run remotely from an end user computer?
    Note I am using SharePoint 2010.

    Hi,
    According to your post, my understanding is that you want to butch upload files to document library with content types.
    There are two good tools in CodePlex which can batch upload files with content types, you can have a look at them.
    DocMetaMan : Bulk document Upload and MetaData (Taxonomy) Setter
    SharePoint Multiple File Upload Script
    There are some third-party tools which can achieve the same scenario, you can refer to it.
    http://sharepoint.stackexchange.com/questions/74694/uploading-multiple-files-with-metadata
    Thanks,
    Jason
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Jason Guo
    TechNet Community Support

  • How to deal with deadlock on wwv_flow_data table when http server times out

    There are some threads about a deadlock on the wwv_flow_data table. None of them contain a real explanation for this behaviour. In my case I will try to explain what I think is happening. Maybe it helps somebody who is hitting the same matter.
    In my case with APEX 3.2.1 I am navigating from one page to another. Doing this APEX will lock the table wwv_flow_data. As soon as the other page is shown the lock will be released. But now this other page contains a bad performing query (standaard report region). After 5 minutes the http server (modplsql) will time out and present the message "No response from the application server" on the screen. In the meanwhile the query is still running on the database server and the lock stays on the wwv_flow_data table.
    Normal user behaviour will be that the user will use the back button to return to the previous page and tries it again to navigate to the other page or
    the user will try to refresh the page with the bad performing query.
    And voila now you will have a deadlock on the wwv_flow_data table since a second session is trying to do the same thing while the first hasn't finished yet.
    How to deal with it?
    First of all. Have a good look at the bad performing query. Maybe you can improve it that it will succeed before the http server will timeout.
    In my case the 11gr1 optimizer couldn't handle a subquery factoring clause in the best way. After changing it back to a classical inline query the problem was solved.
    Secondly you could increase the timeout parameter of the http server. Although this not the best way.
    Maybe it would better if APEX in a next version would release the lock on the table wwv_flow_date earlier or do a rollback just before the moment that the http server is timing out.
    regards,
    Mathieu Meeuwissen

    Hello Shmoove,
    I saw your reply here and you probably understand the problems the HTTP 100 response may cause.
    I am trying to send image that was taken by getSnapshot. The problem is that the server respond with this HTTP 100 message.
    I suspect that the reason that my server doesn't recognize the file that I'm sending from J2me is that the "server to client" response to the 100 message comes after the second message of (see what the TCPIP viewer shows down here):
    POST /up01/up02.aspx HTTP/1.1
    Content-Type: multipart/form-data; boundary=xxxxyyyyzzz
    Connection: Keep-Alive
    Content-length: 6294
    User-Agent: UNTRUSTED/1.0
    Host: szekely.dnsalias.com:80
    Transfer-Encoding: chunked
    400: Client to Server (126 bytes)
    78
    --xxxxyyyyzzz
    Content-Disposition: form-data; name="pic"; filename="david.jpg"
    Content-Type: application/octet-stream
    400: Connected to Server
    400: Server to Client (112 bytes)
    HTTP/1.1 100 Continue
    Server: Microsoft-IIS/5.1
    Date: Wed, 23 Mar 2005 00:47:02 GMT
    X-Powered-By: ASP.NET
    Any help will be appreciated,
    David

  • Error while uploading document with same name in Document library

    Hi,
    Whenever I try to upload document with same name in Document library it throws an error. I know its a default behavior but is there any way so that I can upload documents with same name or can we append some unique id with title of the document?
    Please suggest.
    Thanks

    No, unfortunately this is not possible. SharePoint treats the file name as the primary key identifier for a document so if you try to upload a second document with the same name it will assume this is a new version of the file. If you have versioning enabled
    on the document library, it will add this as a new version. If not, it will overwrite the original file with the new one.
    If you need two documents with the same name, you will have to either place them in different folders within a document library or in different document libraries.
    http://stackoverflow.com/questions/11894968/uploading-documents-with-same-name-to-sharepoint-2010
    Regards,
    Rajendra Singh
    If a post answers your question, please click Mark As Answer on that post and Vote as Helpful
    http://sharepointundefind.wordpress.com/

  • How to deal with OpenCL grey out issue.

    First off, I'm not native english speaker.
    I have Nvidia GeForce 550 Ti video card, and its support OpenCL v1.1.
    I'm also face a "random" OpenCL grey out issue like other people here, but I found the "issue" and how to deal with.
    open a command windows (using WinKey + R, and type cmd <enter>)
    type "cd C:\Program Files\Adobe\Adobe Photoshop CS6 (64 Bit)" <enter>
    type "sniffer_gpu" <enter>  -->> to check your GPU has OpenCL in Photoshop?
    When GPU is in power-saving mode (core running in 51Mhz, using MSI Afterburner monitor), I running sniffer_gpu, got this report
    C:\Program Files\Adobe\Adobe Photoshop CS6 (64 Bit)>sniffer_gpu
    Device: 0000000000328D68 has video RAM(MB): 1023
    Vendor string:    NVIDIA Corporation
    Renderer string:  GeForce GTX 550 Ti/PCIe/SSE2
    Version string:   3.0.0
    OpenGL version as determined by Extensionator...
    OpenGL Version 3.0
    Has NPOT support: TRUE
    Has Framebuffer Object Extension support: TRUE
    OpenGL ok
    Return code: 1
    C:\Program Files\Adobe\Adobe Photoshop CS6 (64 Bit)>
    No OpenCL support
    Then, I switch GPU back to normal status (core running at 900Mhz)
    C:\Program Files\Adobe\Adobe Photoshop CS6 (64 Bit)>sniffer_gpu
    Device: 00000000001F8D68 has video RAM(MB): 1023
    Vendor string:    NVIDIA Corporation
    Renderer string:  GeForce GTX 550 Ti/PCIe/SSE2
    Version string:   3.0.0
    OpenGL version as determined by Extensionator...
    OpenGL Version 3.0
    Has NPOT support: TRUE
    Has Framebuffer Object Extension support: TRUE
    OpenGL ok
    OpenCL ok, version=1.1 CUDA 4.2.1
    Return code: 3
    C:\Program Files\Adobe\Adobe Photoshop CS6 (64 Bit)>
    OpenCL support !!
    Nowadays, most of new GPUs were support power-saving, but when sniffer_gpu running at GPU power-saving state, the OpenCL computing report maybe "time-out", cause the detector "guess" the GPU does not support OpenCL. This is why many people says "OpenCL" sometimes work, sometimes grey-out.
    I hope this bug can be fixed, or at least people can understand how to "enable OpenCL everytime when you launch photoshop".
    Bye.

    Do check the following link
    http://blogs.oracle.com/soabpm/2010/01/soa_suite_11g_weblogictransact.html

  • How to deal with "Error 1001. The specified service already exists" when install a service using installer package?

    Hi everybody,
    I wrote a "Class Library" project which is a service using Visual Stodio 2008 recently, then tried to use a Visual Studio 2008
    Setup Project to install it.
    Here is what I did for the "Class Library":
    1. Finish the program.cs, Service.cs
    2. Add Installer
    3. Change the serviceInstaller so that "StartType" to be Aotumatic
    4. Change the ServiceProcessInstaller2 so that "Account" to be LocalSystem
    5.
    6. Click in F5 (Start Debugging)
    Here is what I did for the Setup Project:
    1. Add the exe file built from the "Class Library" project to the Application Folder
    2. On the Custom Action Editor, add the exe file from 1 to Install and Commit
    3. Change the property of the project so that "RemovePreviousVersion" to be true
    4. Click on F6(Build Solution)
    Then I tried to run the msi file from the built of the Setup Project. Because I modified the two projects serveral times, I uninstalled the Class Library using "Control Panel->Add or Remove Programs" before I reinstall. Two things I notived:
    1. After unstall, the registry was not cleaned up about the installed program
    2. After several rounds install/uninstall, I got "Error 1001. The specified service already exists"
    My questions are:
    1. How to cleanup the registry when uninstall a program?
    2. How to deal with the "Error 1001. The specified service already exists"?
    3. Did I do anytbing wrong with the "Class Library" or the "Setup Project"?
    Thanks a lot!
    Helen

    Hi Simon, not a problem!
    I spent some more time on this and here are few more notes:
    it is called Major Upgrade, when you are installing new version of the product upon a previous one and
    MSI supports 2 strategies:
    Strategy 1. Install a new version and uninstall previous one. (Install a new version right upon previously installed version (file merging is performed based on dll version number) and the delete previously
    installed files)
    Strategy 2. Uninstall previous version and install a new one (Delete all previous files and install from scratch new files.)
    From the first look it seems that 1st strategy is weird and buggy. But, remember, MSI is great because it's transactional!!! That means that if once some of the phases (Installation, Uninstallation, Rollback, Comit) fails, your machine
    will be reverted to the previous state and it'll be still functional. 
    Let's consider both strategies:
    Consider you have installed product_v1.msi and you want to install product_v2.msi.
    Strategy 1
    1. MSI engine copies files from Product_v1 directory to TEMP directory
    2. MSI engine merges files based on the assembly version (between v1 and v2)
    3. Once merging is completed successfully it removes files in TEMP (RemoveExistingProducts  action triggers it) and you got product_v2 installed, otherwise if it fails MSI engine revert machine to V1 and copies previous files from TEMP.
    Strategy 2
    1. MSI engine tottaly removes all files from v1.
    2. MSI engine installs v2 files and if something goes wrong you cannot revert back, because RemoveExistingProducts  allready worked out and MSI doesn't have files to revert machine back
    I recommend to everybody to use Strategy 1 and leverage MSI transaction functionality. And you can set this strategies by defining sequence of RemoveExistingProducts action. See more info
    here.  So, I think it's not even a bug in VS as I said in the upper post it is default recommened behaviour.
    AND, you got "Error 1001. The specified service already exists"
    because if we follow Strategy 1 MSI engine tries to install Windows Service on top of the existing service and OF COURSE it fails MSI engine (StopServices, DeleteServices actions are executed before actual
    installation and  they look at ServiceControl table). In order to stop service first and delete them you have to fill ServiceContol table of the MSI (and then StopServices, DeleteServices actions will recognize what to they have to stop
    and delete), like this:
    *clip*clip*clip*
    ' see http://msdn.microsoft.com/en-us/library/windows/desktop/aa371634(v=vs.85).aspx for more info
    ' Update the Service Entry to stop and delete service while uninstalling
    query = "INSERT INTO ServiceControl (ServiceControl, Name, Event, Arguments, Wait, Component_) VALUES ('MAD_Service', 'Service name', '160', '', '1', '"
    + componentName + "')"
    Set view = database.OpenView(query)
    : CheckError
    view.Execute : CheckError
    ' Update the Service Entry to stop and delete service while installing
    query = "INSERT INTO ServiceControl (ServiceControl, Name, Event, Arguments, Wait, Component_) VALUES ('MAD2_Service', 'Service name', '10', '', '1', '"
    + componentName + "')"
    Set view = database.OpenView(query)
    : CheckError
    view.Execute : CheckError
    *clip*clip*clip*
    We can uninstall service first by following Strategy 2, but then we lose transactional support.
    So, Simon did I encourage you to change your code a bit?:)
    And, btw, if you don't want to change the strategy, please don't rely on SequenceID in MSI table, it can be change, you have to get the at the runtime.
    Hope it will help to everybody!
    See also more advanced explanation of how MSI works
    here.
    Truly yours, Marat

  • How to deal with credentials for external applications using a Java Client/

    Hi Guys,
    This is the case. I am integrating an external application with an ADF Application. I have implemented some programmatic ViewObjects that are being filled up by a REST Java Client Wrapper. Everything is working fine but the issue is that the credentials the wrapper is using are hard coded inside the java class. I am thinking to ask for the credentials at the beginning of my taskflow and then store them somewhere and use them then to create my client wrapper (passing them in the constructor).
    However, I don't know if my approach is good and I would like you to share your experiences or how to deal with this.
    Regards

    You can use Credential Store Framework to store the credentials securely in the weblogic server instead of hardcoding in the java class.
    The Credential Store Framework:
    - enables you to manage credentials securely
    - provides an API for storage, retrieval, and maintenance of credentials in different back-end repositories
    Check the documentation on CSF API -
    http://docs.oracle.com/cd/E29505_01/core.1111/e10043/devcsf.htm
    Major Steps -
    1. Create a credential map and key in em console to store the password (http://docs.oracle.com/cd/E25054_01/core.1111/e10043/csfadmin.htm)
    2. Use CSF API to retrieve the stored password
    3. In jazn-data.xml give permissions to access CSF key and map

  • How to deal with payment of vendor to us (credit notes I guess)

    Normally I use F07 for outgoing payments, it's linked to vendors, but how to deal with incoming payments of vendors?
    We got paid two invoices(credit notes I guess) and I want to put them under the right vendor, but how do I do it?
    I can't use F07 I think because I need to use positive amounts (for normal purchase invoices) and can't use negative amounts in this transaction. F08 is for ingoing payment but that session is linked to customers I thought..not for vendors?
    How to deal with this kind ot transactions? What transaction can I use the best for payments like this
    (we got paid an amount , no invoices yet, but I know the vendor).

    whith F-02 it is possible I noticed.

Maybe you are looking for

  • Ipod nano after 1 minute automatically in standby

    From the replacement pogram I received the Ipod nano 8GB. When I play music, after about 1 minute it seems to get in a standby modus automatically and stops playing. When I push the button on the top, the screen starts and I have to start the music a

  • Shameful support

    Been a customer for over 10 years. Been running a web hosting company also for over ten years. Verizon has blacklisted my company's mail server IP address. So, it says to visit this page to request delisting. http://my.verizon.com/micro/whitelist/ I

  • Please help me un-drop this float!

    This newbie is sure it's something simple. The template for the home page that's still very much under construction came from DW CS3. All I wanted to do is to swap the order of the #menu div with the #mainContent div, so the flow of html is: the head

  • LInk Inspector Doesn't HIghlight

    What am I doing wrong? I can launch the Hyperlink Inspector fine, but nothing works except for the "make hyperlinks active" button. So I can't add a URL or do anything in this window. Ideas? Thanks!

  • CS6 Windows not downloading Premire, After Effects, or Photoshop because the product isn't genuine. What does that mean?

    I have Production Premium CS6, I got the serial number and started downloading all the components until it stopped and said that it wasn't a Adobe Genuine Software. I don't know why this happened. I downloaded all of the things from my downloads page