Max datafile size limit?

Hi,
Is there any maximum datafile size limit? May I resize my datafiles to, let's say, more than 4GB size?
(OS: RHEL4 32bit, DB: 10.2.0.3)
Thanks!

http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/limits002.htm#i287915
<quote>
Database file size      Maximum      Operating system dependent. Limited by maximum operating system file size; typically 2^22 or 4 MB blocks
<quote>
So give it a try.

Similar Messages

  • Max Datafile Size Oracle 9i

    Hi There,
    We have multiple tablespaces in our database and the recently we got a message that the size of a particular tables has reached 97% and it needs to be increased.
    Now currently the size of that tablespace is 29GB and before increasing the size of this tablespace we would like to know if there is a limitation for an Oracle data file size or are there any performance issues or implications that we may potentially face with a single data file of 30GB in size.
    This is the biggest tablespace in our system other tablespaces are max 12 or 17 GB.
    I am not a DBA just working in the support role as a developer. Could you please provide some insight on above query.
    Many thanks in advance.
    Regards
    Sam

    Aman, Please go through below MOS notes.
    Oracle9i Database Limits [ID 217143.1]
    Database file size
    + Maximum
    Operating system dependent. Limited by maximum operating system file size;
    typically 222 or 4M blocks
    What is The Maximum Datafile Size Limit In Oracle Database 10gR2 [ID 804733.1]
    Maximum number of Database blocks allowed in a single datafile in 10gR2 are as follows:
    Small File Tablespace (Normal Tablespace) : 4194303 (2^22 -1)
    Big File Tablespace (New in 10gR2) : 4294967295 (2^32 -1)
    Max datafile size for SMALL FILE NORMAL TABLESPACE would be:
    Database Block Size     Maximum Datafile File Size
    2k     4194303 * 2k = 8 GB
    4k     4194303 * 4k = 16 GB
    8k     4194303 * 8k = 32 GB
    16k     4194303 * 16k = 64 GB
    32k     4194303 * 32k = 128 GB
    Max datafile size for BIG FILE TABLESPACE would be:
    Database Block Size     Maximum Datafile Size
    2k     4294967295 * 2k = 8 TB
    4k     4294967295 * 4k = 16 TB
    8k     4294967295 * 8k = 32 TB
    16k     4294967295 * 16k = 64 TB
    32k     4294967295 * 32k = 128 TB

  • Max datafiles size

    hi there,
    i want to know how much is the maximum size of a datafile?
    I'm using oracle 8.1.7.4 on aix 4.3.3
    db_block_size=8192
    i have a datafile of 2GB and i need to expand it.
    I was wondering if the maximum datafile size is 2 gb, so i do not need to increase this but create e new one
    thanks

    Without any reference at hand, the AIX (4.3.3, JFS) limits as I can recall
    File size: 2GB
    File size if large files enabled: near 64GB
    File system size: 64GB with std fragment size.
    Also observe the ulimit of the user who is using the file system.

  • What would be the maximum datafile size that can support sql*loader

    Hi,
    I would like to load datafile from xls file which nearly 5 gb into oracle table by using sql*loader. Could you please tell me how much is max datafile size we can load by using sql*loader?
    Thanks
    VAMSHI

    Hello,
    The Size limit is mainly given by the OS. So you should care about what the OS could support as SQL*Loader files are unlimited on *64 Bit* but limited to *2GB* in *32 Bit* OS:
    http://download.oracle.com/docs/cd/E11882_01/server.112/e10839/appg_db_lmts.htm#UNXAR382
    Else, you should be able to load these data into the Table. So you must check that you have enough place inside the Tablespace and/or the Disk (if the Tablespace has to be extended).
    Please find enclosed a link about SQL*Loader and scroll down to Limits / Defaults:
    http://www.ordba.net/Tutorials/OracleUtilities~SQLLoader.htm
    Hope this help.
    Best regards,
    Jean-Valentin

  • Max file size in OSX Tiger?

    I regularly backup my hard drive to my USB 2.x 80GB drive with Retrospect. As of the moment my backup file is 20GB's. Is there a max file size limit like there was in the old Mac os?
    Thanks,
    John

    This article may answer your question:
    http://docs.info.apple.com/article.html?artnum=25557
    However, why are you creating a 20 GB file? Why not do a duplicate from one volume to another?
    Mind you, you won't get booting as USB can't boot Mac OS X, but at least all the files will be there, and you won't suffer from an all your eggs in one basket issue, where if one file is corrupted the entire file needs to be recreated from scratch.

  • Solaris-x86 mount fat32 partition, the partition max size limit?

    solaris10 x86, laptop, 10G FAT32 partition for windows & x86 exchange data.
    the fat32 partition mount as normal, can be read fine.
    but write some file by x86, that can not find by windows.
    anyboy know did the solaris-x86 mount fat32 partition, the partition max size limit? or no limit, why this problem occur?

    Mounting Windows partition in Solaris
    The easiest way to share data now is to do it through a FAT32 partition. Solaris
    recognises it as partition of type pcfs. It is specified as device:drive where drive is
    either the DOS logical drive letter (c through z) or a drive number (1 through 24).
    Drive letter c is equivalent to drive number 1 and represents the Primary DOS partition
    on the disk; drive letters d through z are equivalent to drive numbers 2 through 24,
    and represent DOS drives within the Extended DOS partition.Syntax is
    mount -F pcfs device:drive /directroy-name
    where directory name specifies the location where the file system is mounted.
    To mount the first logical drive (d:) in the Extended DOS partition from an IDE hard
    disk in the directory /d use
    mount -F pcfs /dev/dsk/c0d0p0:d /d
    You can use mount directory-name after appending following line is in
    /etc/vfstab file
    device:drive directory-name pcfs no rw
    for example
    c0d0s0:c /c pcfs no rw
    If your windows partition like the following means
    C: - NTFS, D:-FAT32, E:-NTFS, F:-FAT32
    Then you can only mount D, F not C & E.
    Mounting D Drive:
    mount -F pcfs /dev/disk/c0d0p0:c /mountpoint
    Mounting F Drive
    mount -F pcfs /dev/disk/c0d0p0:d /mountpoint
    The driveletter only for fat not including other file systems (ntfs or any linux filesystems).

  • SGA Max Size limit?

    Hi,
    I have Fujitsu mid range Server with 16gb RAM and 64 bit Windows Server 2003,10g R2 db installed, current i have SGA size 4gb..
    What is SGA max size limit????
    One of my report runs in 24 seconds...*will this issue b solved increasing the SGA size upto 10,12 gb?*

    Yes,
    You can also go for a 10046 event tracing...
    ACCEPT sid PROMPT 'Enter SID: '
    ACCEPT serial PROMPT 'Enter SERIAL#: '
    ACCEPT action PROMPT 'Enter TRUE or FALSE: '
    EXEC sys.DBMS_SYSTEM.SET_SQL_TRACE_IN_SESSION(&sid,&serial,&action);
    prompt Trace &action for &sid,&serial
    exec DBMS_SYSTEM.SET_EV(10,20,10046,12,”);
    Then you can check your dump file and see whcih events are higher......
    For Eg. content could be like:
    =====================
    PARSING IN CURSOR #6 len=107 dep=1 uid=44 oct=6 lid=44 tim=1621758552415 hv=3988607735 ad='902c07a8'
    UPDATE rn_lu_lastname_loca set entr_loca_id_plz14 = translate(entr_loca_id_plz14,'_','-') where rowid = :b1
    END OF STMT
    PARSE #6:c=0,e=981,p=0,cr=0,cu=0,mis=1,r=0,dep=1,og=0,tim=1621758552403
    BINDS #6:
    bind 0: dty=1 mxl=32(18) mal=00 scl=00 pre=00 oacflg=13 oacfl2=1 size=32 offset=0
    bfp=10331d748 bln=32 avl=18 flg=09
    value="AAAHINAATAAAwTTABV"
    WAIT #6: nam='db file sequential read' ela= 12170 p1=6 p2=197843 p3=1
    WAIT #6: nam='db file sequential read' ela= 8051 p1=14 p2=261084 p3=1
    WAIT #6: nam='db file sequential read' ela= 7165 p1=19 p2=147722 p3=1
    WAIT #6: nam='db file sequential read' ela= 9604 p1=19 p2=133999 p3=1
    WAIT #6: nam='db file sequential read' ela= 6381 p1=19 p2=133801 p3=1
    EXEC #6:c=10000,e=45750,p=5,cr=1,cu=10,mis=0,r=1,dep=1,og=4,tim=1621758598343
    FETCH #5:c=0,e=357,p=0,cr=5,cu=0,mis=0,r=0,dep=1,og=4,tim=1621758598896
    EXEC #1:c=30000,e=116691,p=36,cr=35,cu=10,mis=0,r=1,dep=0,og=4,tim=1621758599043
    WAIT #1: nam='SQL*Net message to client' ela= 5 p1=1413697536 p2=1 p3=0
    WAIT #1: nam='SQL*Net message from client' ela= 2283 p1=1413697536 p2=1 p3=0
    Lines that start with WAIT
    len Length of SQL statement.
    dep Recursive depth of the cursor.
    uid Schema user id of parsing user.
    oct Oracle command type.
    lid Privilege user id.
    ela Elapsed time. 8i: in 1/1000th of a second, 9i: 1/1'000'000th of a second 
    tim Timestamp. Pre-Oracle9i, the times recorded by Oracle only have a resolution of 1/100th of a second (10mS). As of Oracle9i some times are available to microsecond accuracy (1/1,000,000th of a second). The timestamp can be used to determine times between points in the trace file. The value is the value in v$timer when the line was written. If there are TIMESTAMPS in the file you can use the difference between 'tim' values to determine an absolute time. 
    hv Hash id.
    ad SQLTEXT address (see v$sqlarea and v$sqltext).
    Lines that start with PARSE, EXEC or FETCH
    #n  n = number of cursor 
    c  cpu time 
    e  elapsed time 
    p  physical reads 
    cr  consistant reads 
    cu  current mode reads 
    mis miss in cache (?) 
    r  rows processed 
    dep recursive depth 
    og  optimizer goal 
    tim time  Content

  • How do define the limit of the max heap size?

    Hi All,
    I would like to know what should be the limit of the JVM max heap size.
    What will happen if we will not define it?
    What is the purpose of defining it from the technical point of view?
    Thanks
    Edited by: Anna78 on Jul 31, 2008 12:36 PM

    Defining a max heap space too large can have the following effect:
    If you create new objects, the VM may decide it is not worth getting rid of garbage-collectable ones, as there
    is still plenty of space between the current heap size and the max allowed. The result will be that the
    application will run faster and will consume more memory than it really needs.
    If the heap size is too small, but still sufficient, the application will do a lot of garbage-collection and therefore
    run slower. On the other hand, it will stay inside the tight space it has been allowed to use.
    The speed difference may or may not be noticeable, while the difference between 256M and 512M may
    or may not matter on today's computers.

  • HotSpot(TM)64-BitServer VM warning: (benign)Hit CMSMarkStack max size limit

    Hi I am getting the CMSMarkStack max size limit. Could anyone explain why i am getting that.
    61083.003: [GC 61083.003: [ParNew: 523392K->0K(523840K), 0.1802670 secs] 2866168K->2364464K(4193856K), 0.1804250 secs]
    61087.107: [GC 61087.107: [ParNew: 523392K->0K(523840K), 0.1970010 secs] 2887856K->2396761K(4193856K), 0.1971990 secs]
    61087.349: [GC [1 CMS-initial-mark: 2396761K(3670016K)] 2408426K(4193856K), 0.0330660 secs]
    61087.382: [CMS-concurrent-mark-start]
    61089.382: [CMS-concurrent-mark: 2.000/2.000 secs]
    61089.382: [CMS-concurrent-preclean-start]
    61089.637: [CMS-concurrent-preclean: 0.253/0.255 secs]
    61089.637: [CMS-concurrent-abortable-preclean-start]
    CMS: abort preclean due to time 61090.703: [CMS-concurrent-abortable-preclean: 0.224/1.067 secs]
    61090.721: [GC[YG occupancy: 336074 K (523840 K)]61090.721: [Rescan (parallel) , 0.4475020 secs]61091.169: [weak refs processing, 1.8464740 secs]Java HotSpot(TM) 64-Bit Server VM warning: (benign) Hit CMSMarkStack max size limit
    [1 CMS-remark: 2396761K(3670016K)] 2732836K(4193856K), 2.5285500 secs]
    61095.521: [CMS-concurrent-sweep-start]
    61104.793: [CMS-concurrent-sweep: 9.271/9.271 secs]
    61104.793: [CMS-concurrent-reset-start]
    61104.821: [CMS-concurrent-reset: 0.029/0.029 secs]
    61110.338: [GC 61110.338: [ParNew: 523392K->0K(523840K), 0.6183310 secs] 2133391K->1628588K(4193856K), 0.6184950 secs]
    61162.032: [GC 61162.032: [ParNew: 523392K->0K(523840K), 0.2259220 secs] 2151980K->1662904K(4193856K), 0.2261040 secs]
    61171.154: [GC 61171.155: [ParNew: 523392K->0K(523840K), 0.1890640 secs] 2186296K->1686907K(4193856K), 0.1892200 secs]
    regards
    R.Sriram

    "errno = 28" is an error code from the OS which means "No space left on device" this could indicate you don't have enough swap space.
    I suspect you are using too much memory even when you don't think you are.
    I would start with a simple "hello world" program and increase the memory until you get an error. If you can't run even a hello world program you have a serious system error.

  • Max Character File Size Limit exceeded.The document is too large to process

    Hi
    I have  made set on section on products in report.
    the Reports contains around 50 products when im opening the report
    in draft mode the following error is displayed
    "Max Character File Size Limit exceeded.The document is too large to be proceesed by the server .Contact your business Objects Administrator."
    can some body help me out how to increase my report character file size

    Hi,
      If you are using Business Objects XIR2, there is a performance parameter in the Web_Intelligence Report Server where you can increase the size of that file. The parameter is Maximun Character File Size. Go to the CMC>Server>server.Web_IntelligenceReportServer in the Properties tab you will see it.
    Cheers,
    Luigi

  • Connection pool size limit error

    Hi all,
    I am trying to execute a BAPI function from MII, execution fails with the following message;
    [ERROR] Unable to make RFC call Exception: [Problem retrieving JCO.Function object: Connection pool <ECC_Server>:800:02:EN:ECCUser is exhausted. The current pool size limit (max connections) is 1 connections.]
    [WARN] [SAP_JCo_Function_0] Skipping execution of output links due to action failure.
    [ERROR] Uncaught exception from SAP_JCo_Function_0, Problem retrieving JCO.Function object: Connection pool <ECC_Server>:800:02:EN:ECCUser is exhausted. The current pool size limit (max connections) is 1 connections.
    Config:
    1. In 'SAP MII: Connections' of type JCO and have given pool size to 100.
    2. In 'SAP MII: Credential Stores' store is created and same is being used in Start Session.
    3. In  JCO_Function block, we can search for the Function Module and set it.
    MII Version:
    14.0.2 Build(82)
    Am I missing something?
    Has any one seen this? please advise.
    Thanks,
    Message was edited by: Shridhar N

    Check if there is another JCo connection configured with the same IP and User. I have found in the past that even though there are two connections configured because they have the same ip and user they are put into one pool with the lowest max pool of the two connections.

  • LabView RT FTP file size limit

    I have created a few very large AVI video clips on my PXIe-8135RT (LabView RT 2014).  When i try to download these from the controller's drive to a host laptop (Windows 7) with FileZilla, the transfer stops at 1GB (The file size is actually 10GB).
    What's going on?  The file appears to be created correctly and I can even use AVI2 Open and AVI2 Get Info to see that the video file contains the frames I stored.  Reading up about LVRT, there is nothing but older information which claim the file size limit is 4GB, yet the file was created at 10GB using the AVI2 VIs.
    Thanks,
    Robert

    As usual, the answer was staring me right in the face.  FileZilla was reporting the size in an odd manner and the file was actually 1GB.  The vi I used was failing.  After fixing it, it failed at 2GB with error -1074395965 (AVI max file size reached).

  • [AS3 AIR] 2880x2880 Size Limit, Camera, Filter, CameraRoll Issues & Finger Friendly Components

    AS3 AIR ANDROID
    I started playing with this Adobe AIR for Adroid by building an app that would let the user take a picture using the mobile device's native camera app.  Then, I'm applying filter effects to make the image look cool.  Then, I'm allowing the user to save the image back to the camera roll.
    Here are some questions that I have:
    KEEPING UP WITH CURRENT TECHNOLOGY
    Are we limited to the 2880x2880 stage size limit?  Although, this dimension does yield 8+megapixels, it's not in the ratio that most camera sensors are built (widescreen).  Plus, you can bet that newer cameras will have even higher dimensions.  Will this be updated to keep up with current technology requirements?
    IMPORTING & MANIPULATING CAMERA DATA
    Code
    var bmpData:BitmapData = new BitmapData($loader.width, $loader.height);
    bmpData.draw(DisplayObject($loader));
    bmp = new Bitmap(bmpData);
    bmp.width = Capabilities.screenResolutionX;
    bmp.height = Capabilities.screenResolutionY;
    if (CameraRoll.supportsAddBitmapData) {
        var cameraRoll:CameraRoll = new CameraRoll();              
        cameraRoll.addEventListener(ErrorEvent.ERROR, onCrError);
        cameraRoll.addEventListener(Event.COMPLETE, onCrComplete);
        var savedBmpData:BitmapData = new BitmapData (bmp.width, bmp.height);
        savedBmpData.draw(DisplayObject(bmp));
        cameraRoll.addBitmapData(savedBmpData);
    } else {
        trace("~" + "Camera Roll not supported for this device.");
    addChild(bmp);
    When you capture an image using the mobile device's camera app, you have to use the Loader object.
    So, here, I am doing just that with these steps:
    First, I'm creating a BitmapData object and sizing it the same as the camera image.
    Pass the camera image into the BitmapData object.
    Create a Bitmap object and pass in the BitmapData.
    Resize it to fit on the stage.
    Check for Camera Roll and then create a savedBmpData BitmapData object at the size of the screen.
    Pass in the bmp.
    Save it to the Camera Roll.
    The problem is that when the image is displayed on the phone, it shows THE ENTIRE (uncropped) image.  However, the image that is saved to the phone is only the top 800x480 corner of the image.  What is wrong?  How do we save an image to the Camera Roll that is larger than the display area of the phone.  It seems like the only way to save the entire camera image is to resize it to fit the stage and then capture the stage into a bitmapData object.
    FILTERS
    If you apply any filters to the bitmapData object, the filter effects will display on the phone, but if the image is saved to the Camera Roll, then all the filters are lost and it only saves the original (unfiltered) image.
    FINGER FRIENDLY UI COMPONENTS
    Do they exist?
    ADDITIONAL NOTES
    The max image size that can be saved is 2039x2039 pixels on the HTC Evo.  Anything bigger than this resulted in a CameraRoll Error #1 (which there is no documentation for what that means ANYWHERE on the web).

  • Is there a size limit on internal HD? iBook G4 1Ghz

    Recently bought a used 12" iBook G4. Of course no installation discs came with it.
    When received it had a 30 Gb hard drive, 528 mb RAM, and 10.3.9.
    I wanted to max out RAM, increase HD size, and move to 10.4.x.
    (Additional background - I have various other Macs and can use Fire Wire Target Mode from any of them to the "new" iBook.)
    RAM upgrade successful. I borrowed install discs from a friend with a similar though not identical machine; his is a 1.33 Ghz. When I removed the old HD and replaced with a new 160 Gb drive, the Install discs did not see the HD. So I took the iBook apart again, put the new drive in an OWC external enclosure connected to my Powerbook G4. Saw that the drive needed formatting, duh, I did so. Put it back in the iBook. Started up from the install disc, it still did not see the HD. Started the iBook in Fire Wire Target mode; could not see the newly installed HD from the other machine either; all I could see on the iBook end of the Firewire was the CD/DVD. I have looked through various threads here related both to iBook HD replacement and to system upgrade concerns - nothing I find matches my issue.
    I have since reinstalled the original 30 Gb HD, have used the borrowed installation discs to do a clean install of 10.3.x (so the similar but not identical was close enough) and a retail 10.4 Install to upgrade the system. So, 2 out of 3; I have my expanded memory, a clean upgraded system. But I hate to give up on a bigger internal HD!!!
    1. Is there a size limit on the internal HD on the iBook? Is 160 Gb too large, and should I try again with a less aggressive increase? I do have a 60 Gb I could pull out of a PB G4 I will probably sell...
    2. If size is not the problem, it seems that maybe it would help if I somehow pre-install the system onto the 160Gb before I physically put it into the iBook. But I do not understand the whole process of making bootable backups etc. Is there a way I can copy my current installation, or do a reinstallation of 10.3.x or 10.4.x on the 160 Gb while is is external, connected via Firewire, and then swap the 160 Gb into the iBook and have it just boot up? I see hard drives for sale in eBay "with OS X 10.x.x already installed..."? I tried, both from my Powerbook and from the iBook to do an Install onto the 160 Gb while it was attached via Firewire. The PB didn't want to have anything to do with those iBook Install discs. The iBook wouldn't allow me the option of doing an Installation onto the External drive. And I assume that a simple Select-All-and-Copy isn't going to do it for me. So how do I install onto the drive while it is external, and will it work once I move it into the iBook?
    3. Probably not important, but out of curiosity... if I could use my Powerbook G4 Installation discs to burn a new system onto my 160 Gb drive (when it was connected as an external to the PB), and then put that drive into the iBook, would that work? Or would iBook vs. Powerbook differences throw off the installation?
    Thanks!
    Stan

    Shiftless:
    Is there a size limit on the internal HD on the iBook?
    The only limitation of HDD size is due to availability. The largest capacity ATA/IDE HDD available is 250 GB. Theoretically there is no limit to the capacity your computer will support, if larger capacity HDDs were available.
    If size is not the problem, it seems that maybe it would help if I somehow pre-install the system onto the 160Gb before I physically put it into the iBook.
    This is not necessary, although you could do it this way. I note in your later responses that you have already attempted this process and have some difficulties. Here is the procedure I would recommend, after the HDD is installed.
    • Clone the old internal HDD to an external firewire HDD. I gather you have done this.
    • Format and erase the newly installed HDD (directions follow).
    • Install new OS from disk or clone back from external HDD. (Post back for directions if you choose to install from this and then restore from backup).
    Formatting, Partitioning Erasing a Hard Disk Drive
    Warning! This procedure will destroy all data on your Hard Disk Drive. Be sure you have an up-to-date, tested backup of at least your Users folder and any third party applications you do not want to re-install before attempting this procedure.
    • With computer shut down insert install disk in optical drive.
    • Hit Power button and immediately after chime hold down the "C" key.
    • Select language
    • Go to the Utilities menu (Tiger) Installer menu (Panther & earlier) and launch Disk Utility.
    • Select your HDD (manufacturer ID) in left side bar.
    • Select Partition tab in main panel. (You are about to create a single partition volume.)
    • Click on Options button
    • Select Apple Partition Map (PPC Macs) or GUID Partition Table (Intel Macs)
    • Click OK
    • Select number of partition in pull-down menu above Volume diagram.
    (Note 1: One partition is normally preferable for an internal HDD.)
    • Type in name in Name field (usually Macintosh HD)
    • Select Volume Format as Mac OS Extended (Journaled)
    • Click Partition button at bottom of panel.
    • Select Erase tab
    • Select the sub-volume (indented) under Manufacturer ID (usually Macintosh HD).
    • Check to be sure your Volume Name and Volume Format are correct.
    • Click Erase button
    • Quit Disk Utility.
    cornelius

  • Is there a size limit for the txt files to be viewed in the ipod ?

    when I pass a text file (.txt) in the notes folder, then I try to see it on my ipod nano, I can see just a part of it, the beginning. Is there a size limit for the .txt files to be viewed in the ipod nano? How to increase it?

    4kByte file size, 1000 notes max.

Maybe you are looking for

  • Can I authorize two computers withe the same iTunes

    Can I authorize two computers withe the same iTunes

  • How to edit and update table control into database?

    I am doing table control. Here are my codes: *& Report  ZHERA_TABLE2 REPORT  ZHERA_TABLE2. ***&SPWIZARD: DATA DECLARATION FOR TABLECONTROL 'ZTABLE_CONTROL' *&SPWIZARD: DEFINITION OF DDIC-TABLE TABLES:   ZHERA. *&SPWIZARD: TYPE FOR THE DATA OF TABLECO

  • A lot of information is wrong on the maps

    Hello, i have tried to use my GSM  as a GPS and I have seen some problems (nokia 5800 Xpress Music). To go from Alzate in Brianza  (Italia) to Sirmione (Italia) and a lot of information were wrong : something like 10 roundabout (I am not sure for th

  • Question on BPM 11g Workflow Services

    Hi, There was no out of the box support feature which would enable the end users ( agents ) to fetch a work item one at a time from some queue . This allocation/fetch would be done on the basis of some algorithm ( SLA, Priority etc ) . We developed t

  • Dotted line

    This may sound pretty basic, but have got a sheet of design drawings and need to annotate using dotted lines to indicate measurements etc. Can someone tell me how to do this please?