Vector class size limitation

Hi guys,
Does anyone know how big a Vector class can be?
I'm planning to have a Vector object of size ~200,000 to 500,000 elements.
I would really want to use the vector in my design and benefit from its sync feature. The only concern is how reliable it will be..
Has anyone have an idea?
Thanks!

Shouldn't that response be qualified? Somethinglike,
where your datastructure doesn't need to be thread
safe, use an ArrayList? I thought that was theonly
real reason to use ArrayList over Vector. Arethere
other reasons? Anyhow, his answer to your questionis
correct, afaik.If you need synchronization... use ArrayList and the
Collections.synchronizedList() method. Vector has too
many non-List methods.
I still say that is wrong, if your data structure needs to be thread safe, then you should use the synchronize on it where ever you use it.
Is this code thread safe?
Vector memberVector = new Vector();
public void addToVector(Object c, Class cType ) {
memberVector.add( c );
memberVector.add(cType);
}This should produce a vector with alternating objects and then a class that is associated with that object. However, will this code always work?
The answer is no, even though the vector is 'thread safe' this operation is not. If 2 threads call this function on the same instance of the object, then you can get a situation where 2 objects appear in the list one after another. or 2 classes, it really depends on the thread task scheduler. The only difference between this and an ArrayList is that if you had used an ArrayList the internal structure might get messed up and you might get only one of the objects, and 2 classes or some other combination.
So the answer to this problem is this
Vector memberVector = new Vector();
public void addToVector(Object c, Class cType ) {
synchronized ( memberVector ) {
memberVector.add( c );
memberVector.add(cType);
}You must sync on the member vector, but in this case, syncing on 'this' would work as well assuming that you used data encapsulation.
I can't 'prove' it, but if you use proper syncing, then Vector, or any 'thread safe' data structure is no better then a non-thread safe version. It may be that it's easier to debug (sometimes) when dealing with vectors, since no elements will accidently be overwritten ,etc. But it's still unsafe.

Similar Messages

  • The get() of Vector class, is it thread safe??

    Hi all:
    the get() of Vector class, is it thread safe?? I looked into the API, but no info is available, so any help is appreciated

    You just have to look out when you perform two or more methods on the Vector in sucession. Then it's better to lock on the vector itself. If you for instance access the vector to query whether the size is greater than zero and immediately remove an item while another thread is removing elements from the same vector, you might get a dirty read and call get() on the vector while there are no elements left.

  • Can someone explain to me Vector Class !!

    Can someone explain to me, or tell me where I can find a good tutorial about Vector class.....in simple terms....any good simple, easy to understand tutorials???
    Thank you.
    Rod

    Start with this:
    http://java.sun.com/j2se/1.4/docs/api/java/util/Vector.html
    Essentially, a vector is a single Java object that allows you to store a group of similar things. Each item, or element, inside the Vector must be of the same type, or Java class. The important thing to remember about Vectors is that their size is not fixed. If you don't know how many items you need to store in advance, a Vector is a useful datastore, because you can simply add items into it without worrying about reaching the maximum size. The Vector object will grow itself as you add elements.
    The important methods are described in the above API. Read through it carefully.

  • Size limitation?

    Hi,
    ignoring the device's physical mem, are there any JAR file max size limitation to take note of? I'm worried bcos i hav alot of graphical content in my midlet.
    For JSR184 developers:
    BTW any1 know wat is the size limit for each M3G file?
    Thanks!

    Well - most Nokia phones currently in use have 64kb size limit (Series-40, version 1.0), the next generation of non-handheld-phones of nokia (Series-40; 2.0) has 128kb.
    Not to mention the object-size limit of some siemens phones - nothing may be bigger (uncrompressed!) than 16kb - no classes - no images (uncrompressed not png file size).
    No classes, no images, no arrays.
    Most symbian based devices (e.g series 60) do not have such size limits.
    lg Clemens

  • Running out of heap with size-limited cache

    I'm experimenting with creating a size-limited cache using Coherence 3.6.1 and not having any luck. I'm creating four 1GB storage-enabled cache nodes and then a non-storage-enabled process to populate them with test data. When I monitor the cache nodes w/ visualvm and run my test case to populate the cache, I just see the heaps get larger and larger until the processes either just stop responding, or I get an OutOfMemoryError. I've tried slowing down and even stopping, waiting for awhile, and then restarting the test case populating the caches to see if that helps, but it doesn't seem to make a difference.
    A related question, since Coherence doesn't seem to be able to recover from the OOME, is there anyway to configure it to end the process when that happens instead of just sitting there?
    Thanks.
    Stack Trace:
    SRVCoherence[93191:9953 0] 2011/06/29 00:39:04 1,015.5 MB/191.38 KB ERROR Coherence   -2011-06-29 00:39:04.679/449.883 Oracle Coherence GE 3.6.1.0 <Error> (thread=DistributedCache:PartitionedPofCache, member=4):
    java.lang.OutOfMemoryError: Java heap space
         at com.tangosol.util.SegmentedHashMap.grow(SegmentedHashMap.java:893)
         at com.tangosol.util.SegmentedHashMap.grow(SegmentedHashMap.java:840)
         at com.tangosol.util.SegmentedHashMap.ensureLoadFactor(SegmentedHashMap.java:809)
         at com.tangosol.util.SegmentedHashMap.putInternal(SegmentedHashMap.java:724)
         at com.tangosol.util.SegmentedHashMap.put(SegmentedHashMap.java:418)
         at com.tangosol.util.SimpleMapIndex.insertInternal(SimpleMapIndex.java:236)
         at com.tangosol.util.SimpleMapIndex.insert(SimpleMapIndex.java:133)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$Storage.updateIndex(PartitionedCache.CDB:20)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$ResourceCoordinator.processEvent(PartitionedCache.CDB:74)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$ResourceCoordinator.finalizeInvokeSingleThreaded(PartitionedCache.CDB:56)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$ResourceCoordinator.finalizeInvoke(PartitionedCache.CDB:9)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache.processChanges(PartitionedCache.CDB:3)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache.onPutAllRequest(PartitionedCache.CDB:68)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$PutAllRequest.onReceived(PartitionedCache.CDB:85)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid.onMessage(Grid.CDB:11)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid.onNotify(Grid.CDB:33)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.PartitionedService.onNotify(PartitionedService.CDB:3)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache.onNotify(PartitionedCache.CDB:3)
         at com.tangosol.coherence.component.util.Daemon.run(Daemon.CDB:42)
         at java.lang.Thread.run(Thread.java:680)Cache Config:
    <?xml version="1.0"?>
    <!DOCTYPE cache-config SYSTEM "cache-config.dtd">
    <cache-config>
        <caching-scheme-mapping>
            <cache-mapping>
                <cache-name>Event</cache-name>
                <scheme-name>EventScheme</scheme-name>
            </cache-mapping>
        </caching-scheme-mapping>
        <caching-schemes>
            <distributed-scheme>
                <scheme-name>EventScheme</scheme-name>
                <service-name>PartitionedPofCache</service-name>
                <partition-count>16381</partition-count>
                <serializer>
                    <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
                </serializer>
                <backing-map-scheme>
                    <local-scheme>
                        <high-units>300m</high-units>
                        <low-units>200m</low-units>
                        <unit-calculator>BINARY</unit-calculator>
                        <eviction-policy>LRU</eviction-policy>
                    </local-scheme>
                </backing-map-scheme>
                <autostart>true</autostart>
            </distributed-scheme>
            <invocation-scheme>
                <service-name>InvocationService</service-name>
                <thread-count>5</thread-count>
                <autostart>true</autostart>
            </invocation-scheme>
        </caching-schemes>
    </cache-config>

    Hi Timberwolf,
    Some tips to keep in mind when cnofiguring a cache to be size limited:
    - High units are per storage node. For instance, 5 storage nodes with a high-units configuration of 10mb can theoretically store up to 50mb. However this implies a perfect distribution which is normally not the case.
    - High units are per named cache. If two caches are mapped to a scheme and the scheme specifies a high units of 10mb, then each cache will have a high units of 10mb, meaning 20mb of storage total. This is especially important to keep in mind when using a wildcard (*) naming convention.
    - For backing maps, high units only take primary storage into account. With a backup count of 1 (the default) a high-units setting of 10mb really means that up to 20mb of data will be stored on that node for that cache.
    - The entire heap of a dedicated storage node cannot be used for storage. In other words, a 1gb heap cannot hold 1gb of cache data. At the very most, ~75% can be used for cache storage (around 750mb.) If the storage node is performing grid functions (such as entry processors, filters, or aggregations) then the total high-units setting should be less than 75%.
    A sample configuration could look like this: 1gb JVM with 700mb dedicated storage. Half of the storage is primary and half is backup. Therefore, the total high-units for that node should not exceed 350mb.
    Hope it helps.
    Thanks,
    Cris

  • DLP Attachment Scanning - Size Limitations

    Is it documented anywhere what the attachment file size limitations are for DLP scanning?  In the ESA Configuration documentation I read:
    "To scan attachments, the content scanning engine extracts the attachment for the RSA Email DLP scanning engine to scan." 
    Can you identify what scanning engine is referenced by "content scanning engine" and what is the maximum attachment size it can process?  Also are those settings modifiable and some indcation of performance impact if they are increased to a maximum of 50 MB per attachment?
    I know you can make some modifications in the DLP policy, however it is our desire to DLP scan every document sent up to our allowable maximum email size.
    If large attachments cannot be scanned we may be forced to reduce our maximum message/attachment file size. 
    We are currently using Async O/S 7.5.1-102 and will be moving to 7.6.0 when it comes GA.

    Hello David,
    The content scanning engine in reference is the same AsyncOS scanning engine responsible or Message and Content Filter scanning. The maximum size of attachment to scan for this scanning engine is controlled by your 'scanconfig' setttings, as configured in the IronPort CLI. The default 'maximum size of attachment to scan' is 5MB.
    IronPort1.example.com>scanconfig
    There are currently 6 attachment type mappings configured to be SKIPPED.
    Choose the operation you want to perform:
    - NEW - Add a new entry.
    - DELETE - Remove an entry.
    - SETUP - Configure scanning behavior.
    - IMPORT - Load mappings from a file.
    - EXPORT - Save mappings to a file.
    - PRINT - Display the list.
    - CLEAR - Remove all entries.
    - SMIME - Configure S/MIME unpacking.
    []> setup
    1. Scan only attachments with MIME types or fingerprints in the list.
    2. Skip attachments with MIME types or fingerprints in the list.
    Choose one:
    [2]>
    Enter the maximum depth of attachment recursion to scan:
    [5]>
    Enter the maximum size of attachment to scan:
    [5242880]>
    <...>
    Any message that is larger than this limit will be skipped by the scanning engine. This would mean that pertinent DLP policys and filters would not match that same message. Naturally, allowing larger messages to be scanned will result in performance risks, as more system resources would be required to complete the content scanning.
    Regards,
    -Jerry

  • Exchange 2013 Mail Size Limits

    I am having an issue with setting the max send and receive size on Exchange 2013.  I keep getting the following error when I attempt to send a 20 meg file server to an internal exchange account OR if I attempt to send a 20 meg file from the exchange
    server to an external account: 
    #550 5.3.4
    ROUTING.SizeLimit; message size exceeds fixed maximum size for route ##
    I have checked the mail sizes and below is the report.  I currently have both send and receive set to 100MB.  Is there some other setting in 2013 that I am not aware of?
    AnonymousSenderToRecipientRatePerHour                       : 1800
    ClearCategories                                            
    : True
    ConvertDisclaimerWrapperToEml                               : False
    DSNConversionMode                                          
    : UseExchangeDSNs
    ExternalDelayDsnEnabled                                     : True
    ExternalDsnDefaultLanguage                                  :
    ExternalDsnLanguageDetectionEnabled                         : True
    ExternalDsnMaxMessageAttachSize                             : 100 MB (104,857,600 bytes)
    ExternalDsnReportingAuthority                               :
    ExternalDsnSendHtml                                        
    : True
    ExternalPostmasterAddress                                   :
    GenerateCopyOfDSNFor                                        :
    HygieneSuite                                               
    : Standard
    InternalDelayDsnEnabled                                     : True
    InternalDsnDefaultLanguage                                  :
    InternalDsnLanguageDetectionEnabled                         : True
    InternalDsnMaxMessageAttachSize                             : 100 MB (104,857,600 bytes)
    InternalDsnReportingAuthority                               :
    InternalDsnSendHtml                                        
    : True
    InternalSMTPServers                                        
    JournalingReportNdrTo                                       : <>
    LegacyJournalingMigrationEnabled                            : False
    LegacyArchiveJournalingEnabled                              : False
    LegacyArchiveLiveJournalingEnabled                          : False
    RedirectUnprovisionedUserMessagesForLegacyArchiveJournaling : False
    RedirectDLMessagesForLegacyArchiveJournaling                : False
    MaxDumpsterSizePerDatabase                                  : 18 MB (18,874,368 bytes)
    MaxDumpsterTime                                            
    : 7.00:00:00
    MaxReceiveSize                                             
    : 100 MB (104,857,600 bytes)
    MaxRecipientEnvelopeLimit                                   : 500
    MaxRetriesForLocalSiteShadow                                : 2
    MaxRetriesForRemoteSiteShadow                               : 4
    MaxSendSize                                                
    : 100 MB (104,857,600 bytes)
    MigrationEnabled                                           
    : False
    OpenDomainRoutingEnabled                                    : False
    RejectMessageOnShadowFailure                                : False
    Rfc2231EncodingEnabled                                      : False
    SafetyNetHoldTime                                          
    : 2.00:00:00
    ShadowHeartbeatFrequency                                    : 00:02:00
    ShadowMessageAutoDiscardInterval                            : 2.00:00:00
    ShadowMessagePreferenceSetting                              : PreferRemote
    ShadowRedundancyEnabled                                     : True
    ShadowResubmitTimeSpan                                      : 03:00:00
    SupervisionTags                                            
    : {Reject, Allow}
    TLSReceiveDomainSecureList                                  : {}
    TLSSendDomainSecureList                                     : {}
    VerifySecureSubmitEnabled                                   : False
    VoicemailJournalingEnabled                                  : True
    HeaderPromotionModeSetting                                  : NoCreate
    Xexch50Enabled                                             
    : True

    Hello Landfish,
    Good Day...
    The output gives the information that Size limit set for Receive and Send is 100 mb, but setting could have changed. So you can follow the below steps to resolve the issue. 
    There are basically three places where you can configure default message size limits on Exchange:
    Organization transport settings
    Send/receive connector settings
    User mailbox settings.
    To check your server’s current limit you can open Exchange Management Shell
    Try the below commands to check the Message quota size limit
    get-transportconfig | ft maxsendsize, maxreceivesize
    get-receiveconnector | ft name, maxmessagesize
    get-sendconnector | ft name, maxmessagesize
    get-mailbox Administrator |ft Name, Maxsendsize, maxreceivesize
    To change the above size limits based on your requirement.
    Set-TransportConfig -MaxSendSize 200MB -MaxReceiveSize 500MB (Size is based on your requirement)
    Attachment size limit
    To set up the rule you can use the below PowerShell cmdlet, as the method is quite simple
    New-TransportRule -Name LargeAttach -AttachmentSizeOver 20MB -RejectMessageReasonText "Message attachment size over 20MB - email rejected."
    For More info
    https://technet.microsoft.com/en-us/library/bb124708(v=exchg.150).aspx
    Remember to mark as helpful if you find my contribution useful or as an answer if it does answer your question.That will encourage me - and others - to take time out to help you Check out my latest blog posts @ Techrid.com

  • How to Get Around the Memo Size Limitations in CR ?

    I am Using Crystal Reports 2008, SQL Database and ASP .Net Visual Studio 2010 for Team Foundation with Crystal Viewer embedded in a web page.  All current update and patches area installed.
    Database has Memo Fields up to 164000 characters in length. Viewer show fine, but with the reports that have been designed to print this information, only seeing part of the Memo field.
    This happens with both RTF, Text and HTML formatted data from within the database field.
    I have read that there is a limitation on the size of a Memo field that Crystal Reports will print (65,534).
    I actually received an Crystal Reports error box when i try to concatenate multiple substring fields as a formula.
    Does anyone have any suggestions or ideas on a work-around ? 
    Due to legal considerations, this data has to be output as it was input, so it can't be hacked. It can be parsed and again merged  but I really donu2019t want to try and write SQL procedures to parse HTML code into readable multiple pieces based on variable length tags with large memo fields.
    Please offer any and every suggestion,
    Thanks to all  ! !
    Edited by: Ludek Uher on Oct 21, 2010 1:31 PM

    yes sir,
    already did but i didn't receive any answers. . .   Memo Field Size Limitations with Crystal Reports 2008 ?
    Thanks for your help.

  • Where is the Vector class? Can't use it.

    Hello!
    I've downloaded Flex Builder 3 Trial from Adobe's website.
    Now I want to use the Vector class, but this code for example:
      var v:Vector.<String>;
      v = new Vector.<String>();
    Gives me this error:
    1046: Type was not found or was not a compile-time constant: Vector.
    At the IDE, going to
      Window -> Preferences -> Flex -> Installed Flex SDKs
    I can see that I have Flex 3.2 SDK.
    Even more than that: When I search for "Vector" in the help ("Adobe® Flex™ 3 Language Reference") I can see all the details about Vector.
    What can I do?
    I'd like to avoid installing any newer SDKs due to all the bugs I see listed here in the forums.
    Thank you!

    It looks like you are trying to type Java code in Flex Builder.  Flex applications are generally built using ActionScript.
    These docs might help you learn ActionScript:
    http://www.flexafterdark.com/docs/ActionScript-Language
    http://www.flexafterdark.com/docs/ActionScript-vs-Java
    I hope that helps...
    Ben Edwards

  • How do I modify the size limitation on images in PSE 12?

      I frequently use panorama tools in Photoshop CS6 & CC and PSE 12 to generate large images (several sunsets/sunrises reach as much as 370 MB or more --- PNG or TIF without layers).  It apears that PSE 12 (and previous) have a limit less than a tenth of that.  I also understand that there is a registry fix which would allow for larger images to be displayed. 
      A) Is that so?
       B)  What is the patch?
       C)  What are the (unintended) consequences?
      I use Win 7 Pro.  One machine is 32 bit; the primary production machine, 64 bit.  Thanks

    Everything you need to know about PSE maximum sizes is in http://helpx.adobe.com/photoshop-elements/kb/maximum-image-size-limits-photoshop.html.
    The registry fix you are thinking of affects the Organizer only. The Editor has an absolute limit of 30,000 x 30,000 pixels. File size itself is not the constraining factor.
    Cheers,
    Neale
    Insanity is hereditary, you get it from your children
    If this post or another user's post resolves the original issue, please mark the posts as correct and/or helpful accordingly. This helps other users with similar trouble get answers to their questions quicker. Thanks.

  • Nio ByteBuffer and memory-mapped file size limitation

    I have a question/issue regarding ByteBuffer and memory-mapped file size limitations. I recently started using NIO FileChannels and ByteBuffers to store and process buffers of binary data. Until now, the maximum individual ByteBuffer/memory-mapped file size I have needed to process was around 80MB.
    However, I need to now begin processing larger buffers of binary data from a new source. Initial testing with buffer sizes above 100MB result in IOExceptions (java.lang.OutOfMemoryError: Map failed).
    I am using 32bit Windows XP; 2GB of memory (typically 1.3 to 1.5GB free); Java version 1.6.0_03; with -Xmx set to 1280m. Decreasing the Java heap max size down 768m does result in the ability to memory map larger buffers to files, but never bigger than roughly 500MB. However, the application that uses this code contains other components that require the -xMx option to be set to 1280.
    The following simple code segment executed by itself will produce the IOException for me when executed using -Xmx1280m. If I use -Xmx768m, I can increase the buffer size up to around 300MB, but never to a size that I would think I could map.
    try
    String mapFile = "C:/temp/" + UUID.randomUUID().toString() + ".tmp";
    FileChannel rwChan = new RandomAccessFile( mapFile, "rw").getChannel();
    ByteBuffer byteBuffer = rwChan.map( FileChannel.MapMode.READ_WRITE,
    0, 100000000 );
    rwChan.close();
    catch( Exception e )
    e.printStackTrace();
    I am hoping that someone can shed some light on the factors that affect the amount of data that may be memory mapped to/in a file at one time. I have investigated this for some time now and based on my understanding of how memory mapped files are supposed to work, I would think that I could map ByteBuffers to files larger than 500MB. I believe that address space plays a role, but I admittedly am no OS address space expert.
    Thanks in advance for any input.
    Regards- KJ

    See the workaround in http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=4724038

  • Is there a file size limitation to using this service?

    I am working on a large PDF file (26 mb) and I need to re-size the original and mess with the margins. I don't beleive there is an easy way to do this in Adobe Acrobat 6.0 Pro. It sounds like I have to convert the file back to a Word document, do the adjustments there and then produce a new PDF. I have two questions:
    Is there a file size limitation to using this service?
    Will a PDF to Word doc. conversion mantain the format of the orginal PDF?
    Thanks
    Tim

    Good day Tim,
    There is a 100MB file size limitation for submitting files to the ExportPDF service.  As for the quality of the conversion, from our FAQ entitled Will Adobe ExportPDF convert both text and formatting information?:
    Adobe ExportPDF is capable of exporting high quality information, but the quality of your Word or Excel document depends on the quality of the PDF file you start with. For instance, if your PDF file was originally authored in Microsoft Word or Excel and converted to PDF using the PDFMaker functionality of Adobe Acrobat®, your PDF file contains a rich set of information that can be captured by Adobe ExportPDF. This includes relative positioning of tables, images, and even multi-column text, as well as page, paragraph, and font attributes.
    If your PDF file was originally authored using simpler PDF generation methods, such as “print to PDF” or “scan to PDF” options, Adobe ExportPDF will convert any recognizable text and then use sophisticated conversion intelligence to preserve as much of the page layout as possible.
    Please let us know if you have any other questions!
    Kind regards,
    David

  • Input file size limition in File Reader Adapter

    Hi,
    We have a File Read Activity(using File Reader Adapter) which works on a Polling model. Whenever a input file is found in a designated directory it invokes our BPEL web service. During load testing we found out that if input file size exceeds 7 MB then we get "insufficient heap space" error. We found a reference to this size limitation in http://download.oracle.com/docs/cd/E11036_01/integrate.1013/b28994.pdf which has below lines..
    "Files that are 7MB or larger in size cannot be delivered. As an alternative, you can debatch large files (if they have multiple messages), and publish these files in messages of size less than 7 MB. This alternative is applicable only to structured files (comma-delimited or fixed position), which contain more than one message. It is not applicable to binary or XML files."
    Our customer will have the data in XML format and the size of the files will be pretty huge(running into 49, 69 MBs). Do we have any workaround here or we are left with no option but to ask the customer to split the files and load it? The later option is currently difficult one for the customer since when split there is a possibility of XML becoming invalid. Any pointers in this issue is appreciated.
    Thanks!
    Balaji

    Hi,
    BPEL can not process more than 7 MB size xml structred file. but,debatching will work for xml files also provided if the file has repeated multiple structures (means multiple repeated root elements).
    Regards,
    Dhanumjay

  • Increasing Email Size Limits for some users

    Hi All,
    For some users (management) i needed to arrange that they are able to send bigger files then the default 10MB limit to each other.
    I used this procedure to do so,
    https://www.simple-talk.com/sysadmin/exchange/increasing-email-size-limits-for-your-high-profile-users-in-exchange-2010/
    In a nutshell:
    Set Global Transport Org. Config to 200MB
    Set Transport Send and Receive connector to 200MB
    Created a Distribution group adding the MT members and me (for testing)
    Created a transport Rule for the 10MB limit for all users and exception for the distribution group
    (Believe me, i am not happy with the 200MB size limit)
    All seems to work but i have one issue that i cannot tackel or maybe it is by design.
    -Managemant can also send big files to all
    users who do not have this privelige, this is not someting i want, sending 100MB+ attachments to over 150 users.
    Is this by design or do i mis something?
    Thanks,
    Ron

    Hi RonGielgens,
    Attachment size in exchange has to be set on
    Global settings
    Connectors (send and receive connectors)
    Mailbox
    The global settings will affect all the others. of course connectors will affect specific servers and mailbox one user. You have to set the global to the absolute max size you want to allow in your organization.
    after you do this you will need either to set the 10MB on each mailbox
    Get-Mailbox | Set-Mailbox -MaxSendSize 10MB -MaxReceiveSize 10MB
    Then change the limit on specific mailbox
    or another way is to create a transport rule that will reject the email if the attachment size it > 10MB unless the sender/recipient is the required user or group.
    so basically create 2 transport rules
    1. reject messages sent to anyone with attachment > 10mb unless the recipient is userX
    2. reject messages sent from anyone with attachment > 10mb unless the sender is userX
    Normal case it should not allow.
    Thanks, MAS
    Please mark as helpful if you find my comment helpful or as an answer if it does answer your question. That will encourage me - and others - to take time out to help you.

  • PDF file size limited to graphics memory in Reader?

    I've created a form (in LiveCycleDS) that allows for an unlimited number of photos to be loaded into it. I put an image into a subform that is duplicated every time a user clicks a button thus creating an unlimited number of images that can hold photos. I then extended it with Reader Extensions.
    I'm running into a problem when I try to load a large number of photos into the form.  It gets to about 47MB of images when it locks-up Adobe reader.  I've been able to bring up other applications after the lock-up and when I switch back to Reader the artifacts of the other application are then displayed within Reader.
    What is the practical limit to the size of a file created with LiveCycle?  Is it tied to the amount of graphics memory a computer has?  My machine has 2GB of RAM while my video card has only 384MB.  I haven't been able to figure out if there is a file size limitation.

    It's entirely normal for file size to increase. PDFs are much more compressed than print streams. For some printers the print size is almost constant (a huge bitmap), for others, it's a collection of graphical items and the size varies enormously. Rarely anything you can do.

Maybe you are looking for

  • I just noticed that my ethernet port stopped working.

    I looked, and it seems to be a bit damaged. I also noticed that the power supply chord tends to come loose. How much do you think it will cost to have this fixed at the Apple Store? Will it be worth it?

  • File Sharing passwords

    I am trying to share files between iBook and iMac. Both show up in the other computer's network, and printing wirelessly works as well. However, it is looking for a password for registered user in each computer, and any of the passwords I know to hav

  • Update pricing procedure in PO

    Hi, I am having a problem in updating the pricing procedure in PO. Earlier in inforecord, we had a couple of condition types which has been removed now. For the current validity period of the inforecords only the basic price condition is there. But t

  • Xdb.jar and ojdbc14.jar incompatibility ( XMLType.createXML() ) ??

    I am using the latest xdb.jar (from xdk_java_9_2_0_5_0.zip) and ojdbc14.jar and it appears that xdb.jar calls a method that does not exist in ojdbc14.jar. It happens when executing OPAQUE opaque = rset.getOPAQUE("value"); XMLType xt = XMLType.createX

  • Transfer of condition records

    hi guys, I have  couple of questions regardign the transfer of the conidtiopn records from R/3 to CRM IF we trying to create a new condition table ex A922 in R/3 do we need to create a new adpater object ZDNL_COND_A922 or we can download using the ad