Weird viewing (one spread per pasteboard)

Okay I might just be losing it, but...
I'm looking at an old archived InDesign magazine issue. It's set up in spreads just like I'd expect. But I can't scroll through the spreads. When I click on a page in the pages dialog box, I can jump to it. But then the entire pasteboard is empty except for that one spread (where normally I could scroll up for the previous spread or down for the next spread). Is this some weird viewing preference? How do I get a normal view?
Thanks, Phyllis

"Are you saying there is no vertical scroll available at all? What happens when you zoom out to entire pasteboard?"
No matter how far out I zoom, I only see one spread. It's just then surrounded by a very large amount of white space, until I can see all the way to the edge of the pasteboard in all directions. It's as if the other pages don't exist until I double-click in the Pages panel and switch to one of them.
I'll try the INX option when I'm back at work again. I fear what would happen to this long document from switching orientations. I would guess this was done in an earlier version of InDesign (don't know which one as I only have CS3). Is there a way to tell which version it was created in? It's difficult going through someone else's files (especially older ones), but I occasionally have to go through archives.
Thanks, Phyllis

Similar Messages

  • Only one spread visible on pasteboard (CS4)

    Something weird happened to my indd file (a 128-page book) and although all the pages are all accounted for, only one spread is visible at a time: if I scroll up or down in the document window, there is nothing but blank space. I can access any other spread by double clicking in the Pages panel, and I can go forward or backward to the next spread using a keyboard shortcut I set up, but it is disturbing not to be able to scroll up or down to other spreads. I found a reference to this problem on another site (http://www.techplex.net/postst15582_weird-viewing-one-spread-per-pasteboard.html) where it was suggested that exporting as INX would fix the problem. I tried that and although opening an INX version did solve the scrolling problem, the text fell differently on the pages, so this was not a usable solution. Same thing for exporting as IDML. I don't know what accounts for the changes.
    I'm hoping there's a simple way to get my scrolling back.

    You've obviously got some corruption in your file.  That is illustrated by the fact you are getting this problem, and also by the fact an inx or idml export fixes it.  I would recommend biting the bullet and doing it otherwise something else could go wrong and your whole file might become unusable.
    The text falling on different pages will be because of text reflow, sounds like the file was created in an earlier version and updated to CS4?  The text wouldn't reflow according to CS4's different composer until you edited it, or did a inx/idml conversion like this.

  • My keyboard on macbook pro (laptop) is acting weird. One key is not responding at all. Have verified using Keyboard viewer and some other keys are printing the unresponsive character at random.

    my keyboard on macbook pro (laptop) is acting weird. One key is not responding at all. Have verified using Keyboard viewer and some other keys are printing the unresponsive character at random. "z" is the unresponsive character.
    Is it a damaged keyboard ?
    The laptop is just 2 months old, will Apple replace it with a new one if its indeed a damaged keyboard or just repair, I use it for official purposes so being without a laptop is not much of an option.

    No one here works for Apple, so we don't know what Apple might or might not do.  If it's a genuine defect, they will of course repair it under warranty.  It is not their responsibility if it effects your ability to work or not, so that's on you.
    If, however, they determine that the key is problematic as a result of your misuse of the laptop, then everything is on you.  And trust me, if they find a glob of dried up beer or coffee there, they will charge you.
    Your only choice is to take it in for repair.

  • How can I view just one spread at a time of my interactive PDF?

    Hi,
    Apologies of this has been asked before.
    I am creating an interactive PDF of a magazine, to be viewed online.
    I would like to be able to 'flip' through the spreads one spread at a time. I do not want to be able to scroll down.
    I do not want to use SWF for various reasons (viewing on ipad etc)
    When I open the magazine online it displays like this (one spread below the other) Is there a way for the spreads to be 'stacked' so I can only see one at any one time?
    Thanks
    Ben

    Hi Ben,
    You might want to create a pdf portfolio and check.
    Regards,
    Rave

  • How do I create one page per spread?

    I want to create a digital booklet that is only one page per spread.  Right now I have a spread with two pages, one is filled and the other is blank.  How do I delete the blank page w/o moving the next page after the blank into its place?  Your help is appreciated, thanks!

    Just uncheck Facing Pages from New Document settings window while you create your document, or do it later from File > Document Setup.
    You have to remove your empty pages manually after that. One easy way to do that is use Pages panel (Window>Pages), just activate thumbnails of empty pages by clicking them control button pushed down (in mac it´s command). After you have activated all your empty page thumbnails, just hit small carbage can icon at the bottom of pages panel...
    Or use Layout > Pages > Delete Pages tool, where you can give page numbers of all the pages you want to delete.

  • How can I view more than one result on the screen when I search for apps? eg I search for my hobby and there are over 1500 results - one result per screen - do I have to swipe 1500 times to see them all?

    I search for apps about my hobby. There are over 1500 results. Each single result takes up the whole screen. I will have to swipe the screen 1500 times to see all the results. How can I see 8 or 10 results and scroll through fast? Thanks very much for any suggestions!

    Bump:)
    But I've found some online reviews of this and I think we
    are stuck with this amazingly awful search system until
    they realize people like me won't bother searching for or buying
    Apps.
    Does a novel display one sentence per page?
    Do you listen to a radio documentary that transmits one word a minute?
    Or how about a song that is broken into four parts because the singers ran out of breath?
    Searching app products has never been easy or a pleasure, and now it's tougher than ever. Who has the time to swipe 1,600 times? It HURTS - and hurts my brain just thinking about it.
    Woe is me!
    Woe are we!

  • Delivery from sales scheduling agreement with one item per schedule line

    Hi all,
    I have a sales scheduling agreement where the same part number has 3 different JIT schedule lines and I need to create a delivery with 3 item lines: every item line of the delivery has to be referred to one schedule line of the scheduling agreement.
    When I create the delivery with transaction VL10E it has only one item line and its quantity is the sum of the 3 schedule lines of the sales scheduling agreement.
    I've tried to change the customizing of the User Role, choosing the Split Schedule Line = 2 ("One item per schedule line") but the delivery is created again with only one item line.
    How can I solve this issue?
    Thanks in advance for reply

    Dear Alex
    The SAP standard behaviour is that all schedule items with delivery date in the past have to be rescheduled anyway to one date and are combined therefore.
    Partially this behaviour can be influenced in VL10* by 2 parameters in
    the user role : Split per sched.line and Rule (delivery qty).
    Please find below informations about these different parameters
    "Split per sched.line propose you 3 Rule for creating one delivery item
    per schedule line. In a delivery list in item view with multiple
    schedule lines, you can decide for the same document item whether you
    want to generate one delivery item per schedule line or whether you want
    to deliver the cumulated quantity of the last schedule lines selected.
    About rule for delivery quantity, the following rules are currently
    implemented:
    1. No cumulation
    The open schedule line quantity to be delivered is the quantity at which
    the list line is delivered.
    2. Sum of all open schedule line quantities to be delivered that fall
    before the chronologically latest date in the to field (following the
    Deliv. creation date field) of the selection criteria.
    3. Sum of all open schedule line quantities to be delivered whose
    delivery creation date falls within the selection time frame (between
    the dates in the Deliv. creation date and to fields, respectively),
    making the calculation of the quantity to be delivered consistent with
    selection criteria. However, calculationof the quantities delivered with
    the order's schedule line quantities is not affected."
    Hope this helps.
    Regards
    Tonia

  • Question: Viewing one page at a time

    How do I change from viewing a spread as two pages to just viewing one page at a time?

    If the doc is set up with spreads then turning spreads off may change the structure of a document. Or, at least, it did back in CS4 when I turned spreads off to look at a document in single-page layout because I was working on a laptop. When I turned spreads back on, lots of features (like anchored objects or text wraps set up to align to spine or "outside edge" for example) were broken.
    I never researched any further - I simply got in the habit of exporting PDFs and viewing those in single-page mode, instead of trying to affect layout in InDesign. Maybe this isn't broken anymore, or perhaps there's a non-destructive method of getting a single-page view. Or, more importantly, maybe the document you're looking at doesn't have any of these spread-dependent features. Or maybe you dont need it to keep the spreads features. Anyway, be aware that you might lose something if you untick that box.

  • Actual and Plan Versions annual view - Multiple versions per company code

    Good Day,
    Iu2019m creating an FI-CO query with an annual view of both Actuals and Plan Values, which uses period and version as input values. The dilemma is around the plan versions which are hardcoded per company codes (eg. 1 Company Code = versions 501 u2013 512)
    Can anyone suggest a possible solutions as normally there is only one version used for planning but in this instance there are multiple versions per Company Code.
    Currently the report is hardcoded for one Company Code and going forward it should cater for more than one(Company Company* 12 Version)
    Kind Regards

    Remove restriction on comp code. There could be 2 models.
    1. Include comp code, value type and version in the rows. This way you get multiple records for a comp code. One record for actual data and one record each for plan versions. To have this model at the query, the data model in the cube also should be in this way (what we call as the account model), multiple records differentiated into actual/plan by valuetype and different versions by version.
    2. Include comp code in rows. Create selections, one for actuals (restrict with key fig and value type 10, actuals), one for each plan versions (restrict key fig with value type 20, version 501. Create one selection for each version this way). This way we have one record per comp code with diff key fig for actual and plan versions. We call this as key fig model. To have this model at the query, the data model in the cube could be account model or key fig model. If it is key fig model then no need of selections. The key fig could be taken directly from the cube.

  • Can see only one photo per page. How to revert to multiple images per page?

    First noticed on my iMac and now on my Macbook (10.4.9)--I'm no longer seeing a month's worth of photos on a single page. Only one photo per page, sometimes with miniature ones in the top border. I thought it was simply an inadvertent click on by part and easily corrected. But all attempts on both computers and reading of Help index haven't led to restoration of an entire page of photos and elimination of the annoyance. I'm wondering now if something I downloaded might have altered iPhoto program itself, making it impossible to view large numbers of photos at the same time. Any suggestions appreciated.
    Cap

    Cap,
    This view, with the one photo big and the thumbnail strip across the top, is what happens when you double-click a photo with "Edit in: Main Window" set in the Preferences. There should be a button visible that says either Done or Return, to exit edit mode and return you to the library. If you still see only one photo, use the size slider in the lower right corner to reduce the photo size and get more thumbnails on the page.
    If, for some reason, you can't exit the edit mode, you may have a corrupt preference file. Quit iPhoto. Use the Finder to navigate to Macintosh HD/ Users/ (your user name)/ Library/ Preferences/ com.apple.iPhoto.plist and move that file to the Trash. Launch iPhoto, and you should be returned to the main viewing mode. You'll have to go in the Preferences and reset them to your liking.
    Regards.

  • Is it possible to datamerge more than one image per page?

    I am building a product catalogue using datamerge.  Is it possible to data merge more than one image per page? (ie I have a logo, product image and dimensioned drawing to go with each product page)
    What would the coding be for the Data Field Name?
    Would it be
    @Logo
    @Product Image
    @Dimension Drawing
    Add image fields in the data source file
    By adding image fields to the data source file, you can allow a different image to appear on each merged record. For example, when you merge documents that include information from various companies, you may want to include an image of each company’s logo as part of the merge.
    Open the data source file.
    At the beginning of the data field name, type an “at” symbol (@) to insert text or paths that refer to image files.The @ symbol is required only in the first line; subsequent lines should include the image paths. Paths, which are case-sensitive, must follow the naming conventions of the operating system in which they’re stored.
    If you get an error message when you type the @ symbol at the beginning of the field, type an apostrophe (') before the @ symbol (such as '@Photos) to validate the function. Some applications, such as Microsoft Excel, reserve the @ symbol for functions. 
    (Windows) Example of image references in data source file
    Bill Tucker
    36
    C:\Photos\BillTucker.jpg
    Dat Nguyen
    53
    C:\MyDocuments\dat.gif
    Maria Ruiz
    26
    C:\Photos\Ruiz.psd
    (Mac OS) Example of image references in data source file
    Bill Tucker
    36
    Mac HD:Photos:BillTucker.jpg
    Dat Nguyen
    53
    Desktop:Family:dat.gif
    Maria Ruiz
    26
    Mac HD:Photos:Ruiz.psd
    You can use InDesign to view the path of an image on your operating system. Insert an image in an InDesign document, and then use the Links panel to view the image’s location. With the image selected, choose Copy Info > Copy Full Path from the Links panel menu. You may need to edit the path after you paste it in your data source. This technique is especially useful for images on a server. 

    anna-rees wrote:
    I am building a product catalogue using datamerge.  Is it possible to data merge more than one image per page? (ie I have a logo, product image and dimensioned drawing to go with each product page)
    What would the coding be for the Data Field Name?
    Would it be
    @Logo
    @Product Image
    @Dimension Drawing
    Yes, pretty much exactly as you describe. You would then just add those fields to the boxes in your page. I tend to add an apostrophe before the @ symbol in Excel so it doesn't think it's a function - as per text you quoted above.
    If you get an error message when you type the @ symbol at the beginning of the field, type an apostrophe (') before the @ symbol (such as '@Photos) to validate the function. Some applications, such as Microsoft Excel, reserve the @ symbol for functions.

  • How to improve speed of queries that use ORM one table per concrete class

    Hi,
    Many tools that make ORM (Object Relational Mapping) like Castor, Hibernate, Toplink, JPOX, etc.., have the one table per concrete class feature that maps objects to follow structure:
    CREATE TABLE ABSTRACTPRODUCT (
        ID VARCHAR(8) NOT NULL,
        DESCRIPTION VARCHAR(60) NOT NULL,
        PRIMARY KEY(ID)
    CREATE TABLE PRODUCT (
        ID VARCHAR(8) NOT NULL REFERENCES ABSTRACTPRODUCT(ID),
        CODE VARCHAR(10) NOT NULL,
        PRICE DECIMAL(12,2),
        PRIMARY KEY(ID)
    CREATE UNIQUE INDEX iProduct ON Product(code)
    CREATE TABLE BOOK (
        ID VARCHAR(8) NOT NULL REFERENCES PRODUCT(ID),
        AUTHOR VARCHAR(60) NOT NULL,
        PRIMARY KEY (ID)
    CREATE TABLE COMPACTDISK (
        ID VARCHAR(8) NOT NULL REFERENCES PRODUCT(ID),
        ARTIST VARCHAR(60) NOT NULL,
        PRIMARY KEY(ID)
    there is a way to improve queries like
    SELECT
        pd.code CODE,   
        abpd.description DESCRIPTION,
        DECODE(bk.id,NULL,cd.artist,bk.author) PERSON
    FROM
        ABSTRACTPRODUCT abpd,
        PRODUCT pd,
        BOOK bk,
        COMPACTDISK cd
    WHERE
        pd.id = abpd.id AND
        bk.id(+) = abpd.id AND
        cd.id(+) = abpd.id AND
        pd.code like '101%'
    or like this:
    SELECT
        pd.code CODE,   
        abpd.description DESCRIPTION,
        DECODE(bk.id,NULL,cd.artist,bk.author) PERSON
    FROM
        ABSTRACTPRODUCT abpd,
        PRODUCT pd,
        BOOK bk,
        COMPACTDISK cd
    WHERE
        pd.id = abpd.id AND
        bk.id(+) = abpd.id AND
        cd.id(+) = abpd.id AND
        abpd.description like '%STARS%' AND
        pd.price BETWEEN 1 AND 10
    think in a table with many rows, then exists something inside MaxDB to improve this type of queries? like some anotations on SQL? or declare tables that extends another by PK? on other databases i managed this using Materialized Views, but i think that this can be faster just using PK, i'm wrong? the better is to consolidate all tables in one table? what is the impact on database size with this consolidation?
    note: with consolidation i will miss NOT NULL constraint at database side.
    thanks for any insight.
    Clóvis

    Hi Lars,
    i dont understand because the optimizer get that Index for TM at execution plan, and because dont use the join via KEY column, note the WHERE clause is "TM.OID = MF.MY_TIPO_MOVIMENTO" by the key column, and the optimizer uses an INDEX that the indexed column is ID_SYS, that isnt and cant be a primary key, because its not UNIQUE, follow the index columns:
    indexes of TipoMovimento
    INDEXNAME     COLUMNNAME          SORT     COLUMNNO     DATATYPE     LEN     INDEX_USED     FILESTATE     DISABLED
    ITIPOMOVIMENTO     TIPO               ASC     1          VARCHAR          2     220546          OK          NO
    ITIPOMOVIMENTO     ID_SYS               ASC     2          CHAR          6     220546          OK          NO
    ITIPOMOVIMENTO     MY_CONTA_DEBITO          ASC     3          CHAR          8     220546          OK          NO
    ITIPOMOVIMENTO     MY_CONTA_CREDITO     ASC     4          CHAR          8     220546          OK          NO
    ITIPOMOVIMENTO1     ID_SYS               ASC     1          CHAR          6     567358          OK          NO
    ITIPOMOVIMENTO2     DESCRICAO          ASC     1          VARCHAR          60     94692          OK          NO
    after i create the index iTituloCobrancaX7 on TituloCobranca(OID,DATA_VENCIMENTO) in a backup instance and get surprised with the follow explain:
    OWNER     TABLENAME     COLUMN_OR_INDEX          STRATEGY                    PAGECOUNT     
         TC          ITITULOCOBRANCA1     RANGE CONDITION FOR INDEX          5368     
                   DATA_VENCIMENTO               (USED INDEX COLUMN)          
         MF          OID               JOIN VIA KEY COLUMN               9427     
         TM          OID               JOIN VIA KEY COLUMN               22     
                                  TABLE HASHED          
         PS          OID               JOIN VIA KEY COLUMN               1350     
         BOL          OID               JOIN VIA KEY COLUMN               497     
                                       NO TEMPORARY RESULTS CREATED          
         JDBC_CURSOR_19                    RESULT IS COPIED   , COSTVALUE IS     988
    note that now the optimizer gets the index ITITULOCOBRANCA1 as i expected, if i drop the new index iTituloCobrancaX7 the optimizer still getting this execution plan, with this the query executes at 110 ms, with that great news i do same thing in the production system, but the execution plan dont changes, and i still getting a long execution time this time at 413516 ms. maybe the problem is how optimizer measure my tables.
    i checked in DBAnalyser that the problem is catalog cache hit rate (we discussed this at [catalog cache hit rate, how to increase?|;
    ) and the low selectivity of this SQL command, then its because of this that to achieve a better selectivity i must have an index with, MF.MY_SACADO, MF.TIPO and TC.DATA_VENCIMENTO, as explained in previous posts, since this type of index inside MaxDB isnt possible, i have no choice to speed this type of query without changing tables structure.
    MaxDB developers can develop this type of index? or a feature like this dont have any plans to be made?
    if no, i must create another schema, to consolidate tables to speed queries on my system, but with this consolidation i will get more overhead, i must solve the less selectivity because i think if the data on tables increase, the query becomes impossible, i see that CREATE INDEX supports FUNCTION, maybe a   FUNCTION that join data of two tables can solve this?
    about instance configuration it is:
    Machine:
    Version:       '64BIT Kernel'
    Version:       'X64/LIX86 7.6.03   Build 007-123-157-515'
    Version:       'FAST'
    Machine:       'x86_64'
    Processors:    2 ( logical: 8, cores: 8 )
    data volumes:
    ID     MODE     CONFIGUREDSIZE     USABLESIZE     USEDSIZE     USEDSIZEPERCENTAGE     DROPVOLUME     TOTALCLUSTERAREASIZE     RESERVEDCLUSTERAREASIZE     USEDCLUSTERAREASIZE     PATH     
    1     NORMAL     4194304          4194288          379464          9               NO          0               0               0               /db/SPDT/data/data01.dat     
    2     NORMAL     4194304          4194288          380432          9               NO          0               0               0               /db/SPDT/data/data02.dat     
    3     NORMAL     4194304          4194288          379184          9               NO          0               0               0               /db/SPDT/data/data03.dat     
    4     NORMAL     4194304          4194288          379624          9               NO          0               0               0               /db/SPDT/data/data04.dat     
    5     NORMAL     4194304          4194288          380024          9               NO          0               0               0               /db/SPDT/data/data05.dat
    log volumes:
    ID     CONFIGUREDSIZE     USABLESIZE     PATH               MIRRORPATH
    1     51200          51176          /db/SPDT/log/log01.dat     ?
    parameters:
    KERNELVERSION                         KERNEL    7.6.03   BUILD 007-123-157-515
    INSTANCE_TYPE                         OLTP
    MCOD                                  NO
    _SERVERDB_FOR_SAP                     YES
    _UNICODE                              NO
    DEFAULT_CODE                          ASCII
    DATE_TIME_FORMAT                      ISO
    CONTROLUSERID                         DBM
    CONTROLPASSWORD                       
    MAXLOGVOLUMES                         2
    MAXDATAVOLUMES                        11
    LOG_VOLUME_NAME_001                   /db/SPDT/log/log01.dat
    LOG_VOLUME_TYPE_001                   F
    LOG_VOLUME_SIZE_001                   6400
    DATA_VOLUME_NAME_0005                 /db/SPDT/data/data05.dat
    DATA_VOLUME_NAME_0004                 /db/SPDT/data/data04.dat
    DATA_VOLUME_NAME_0003                 /db/SPDT/data/data03.dat
    DATA_VOLUME_NAME_0002                 /db/SPDT/data/data02.dat
    DATA_VOLUME_NAME_0001                 /db/SPDT/data/data01.dat
    DATA_VOLUME_TYPE_0005                 F
    DATA_VOLUME_TYPE_0004                 F
    DATA_VOLUME_TYPE_0003                 F
    DATA_VOLUME_TYPE_0002                 F
    DATA_VOLUME_TYPE_0001                 F
    DATA_VOLUME_SIZE_0005                 524288
    DATA_VOLUME_SIZE_0004                 524288
    DATA_VOLUME_SIZE_0003                 524288
    DATA_VOLUME_SIZE_0002                 524288
    DATA_VOLUME_SIZE_0001                 524288
    DATA_VOLUME_MODE_0005                 NORMAL
    DATA_VOLUME_MODE_0004                 NORMAL
    DATA_VOLUME_MODE_0003                 NORMAL
    DATA_VOLUME_MODE_0002                 NORMAL
    DATA_VOLUME_MODE_0001                 NORMAL
    DATA_VOLUME_GROUPS                    1
    LOG_BACKUP_TO_PIPE                    NO
    MAXBACKUPDEVS                         2
    LOG_MIRRORED                          NO
    MAXVOLUMES                            14
    LOG_IO_BLOCK_COUNT                    8
    DATA_IO_BLOCK_COUNT                   64
    BACKUP_BLOCK_CNT                      64
    _DELAY_LOGWRITER                      0
    LOG_IO_QUEUE                          50
    _RESTART_TIME                         600
    MAXCPU                                8
    MAX_LOG_QUEUE_COUNT                   0
    USED_MAX_LOG_QUEUE_COUNT              8
    LOG_QUEUE_COUNT                       1
    MAXUSERTASKS                          500
    _TRANS_RGNS                           8
    _TAB_RGNS                             8
    _OMS_REGIONS                          0
    _OMS_RGNS                             7
    OMS_HEAP_LIMIT                        0
    OMS_HEAP_COUNT                        8
    OMS_HEAP_BLOCKSIZE                    10000
    OMS_HEAP_THRESHOLD                    100
    OMS_VERS_THRESHOLD                    2097152
    HEAP_CHECK_LEVEL                      0
    _ROW_RGNS                             8
    RESERVEDSERVERTASKS                   16
    MINSERVERTASKS                        28
    MAXSERVERTASKS                        28
    _MAXGARBAGE_COLL                      1
    _MAXTRANS                             4008
    MAXLOCKS                              120080
    _LOCK_SUPPLY_BLOCK                    100
    DEADLOCK_DETECTION                    4
    SESSION_TIMEOUT                       180
    OMS_STREAM_TIMEOUT                    30
    REQUEST_TIMEOUT                       5000
    _IOPROCS_PER_DEV                      2
    _IOPROCS_FOR_PRIO                     0
    _IOPROCS_FOR_READER                   0
    _USE_IOPROCS_ONLY                     NO
    _IOPROCS_SWITCH                       2
    LRU_FOR_SCAN                          NO
    _PAGE_SIZE                            8192
    _PACKET_SIZE                          131072
    _MINREPLY_SIZE                        4096
    _MBLOCK_DATA_SIZE                     32768
    _MBLOCK_QUAL_SIZE                     32768
    _MBLOCK_STACK_SIZE                    32768
    _MBLOCK_STRAT_SIZE                    16384
    _WORKSTACK_SIZE                       8192
    _WORKDATA_SIZE                        8192
    _CAT_CACHE_MINSIZE                    262144
    CAT_CACHE_SUPPLY                      131072
    INIT_ALLOCATORSIZE                    262144
    ALLOW_MULTIPLE_SERVERTASK_UKTS        NO
    _TASKCLUSTER_01                       tw;al;ut;2000*sv,100*bup;10*ev,10*gc;
    _TASKCLUSTER_02                       ti,100*dw;63*us;
    _TASKCLUSTER_03                       equalize
    _DYN_TASK_STACK                       NO
    _MP_RGN_QUEUE                         YES
    _MP_RGN_DIRTY_READ                    DEFAULT
    _MP_RGN_BUSY_WAIT                     DEFAULT
    _MP_DISP_LOOPS                        2
    _MP_DISP_PRIO                         DEFAULT
    MP_RGN_LOOP                           -1
    _MP_RGN_PRIO                          DEFAULT
    MAXRGN_REQUEST                        -1
    _PRIO_BASE_U2U                        100
    _PRIO_BASE_IOC                        80
    _PRIO_BASE_RAV                        80
    _PRIO_BASE_REX                        40
    _PRIO_BASE_COM                        10
    _PRIO_FACTOR                          80
    _DELAY_COMMIT                         NO
    _MAXTASK_STACK                        512
    MAX_SERVERTASK_STACK                  500
    MAX_SPECIALTASK_STACK                 500
    _DW_IO_AREA_SIZE                      50
    _DW_IO_AREA_FLUSH                     50
    FBM_VOLUME_COMPRESSION                50
    FBM_VOLUME_BALANCE                    10
    _FBM_LOW_IO_RATE                      10
    CACHE_SIZE                            262144
    _DW_LRU_TAIL_FLUSH                    25
    XP_DATA_CACHE_RGNS                    0
    _DATA_CACHE_RGNS                      64
    XP_CONVERTER_REGIONS                  0
    CONVERTER_REGIONS                     8
    XP_MAXPAGER                           0
    MAXPAGER                              64
    SEQUENCE_CACHE                        1
    _IDXFILE_LIST_SIZE                    2048
    VOLUMENO_BIT_COUNT                    8
    OPTIM_MAX_MERGE                       500
    OPTIM_INV_ONLY                        YES
    OPTIM_CACHE                           NO
    OPTIM_JOIN_FETCH                      0
    JOIN_SEARCH_LEVEL                     0
    JOIN_MAXTAB_LEVEL4                    16
    JOIN_MAXTAB_LEVEL9                    5
    _READAHEAD_BLOBS                      32
    CLUSTER_WRITE_THRESHOLD               80
    CLUSTERED_LOBS                        NO
    RUNDIRECTORY                          /var/opt/sdb/data/wrk/SPDT
    OPMSG1                                /dev/console
    OPMSG2                                /dev/null
    _KERNELDIAGFILE                       knldiag
    KERNELDIAGSIZE                        800
    _EVENTFILE                            knldiag.evt
    _EVENTSIZE                            0
    _MAXEVENTTASKS                        2
    _MAXEVENTS                            100
    _KERNELTRACEFILE                      knltrace
    TRACE_PAGES_TI                        2
    TRACE_PAGES_GC                        20
    TRACE_PAGES_LW                        5
    TRACE_PAGES_PG                        3
    TRACE_PAGES_US                        10
    TRACE_PAGES_UT                        5
    TRACE_PAGES_SV                        5
    TRACE_PAGES_EV                        2
    TRACE_PAGES_BUP                       0
    KERNELTRACESIZE                       5369
    EXTERNAL_DUMP_REQUEST                 NO
    _AK_DUMP_ALLOWED                      YES
    _KERNELDUMPFILE                       knldump
    _RTEDUMPFILE                          rtedump
    _UTILITY_PROTFILE                     dbm.utl
    UTILITY_PROTSIZE                      100
    _BACKUP_HISTFILE                      dbm.knl
    _BACKUP_MED_DEF                       dbm.mdf
    _MAX_MESSAGE_FILES                    0
    _SHMKERNEL                            44601
    __PARAM_CHANGED___                    0
    __PARAM_VERIFIED__                    2008-05-03 23:12:55
    DIAG_HISTORY_NUM                      2
    DIAG_HISTORY_PATH                     /var/opt/sdb/data/wrk/SPDT/DIAGHISTORY
    _DIAG_SEM                             1
    SHOW_MAX_STACK_USE                    NO
    SHOW_MAX_KB_STACK_USE                 NO
    LOG_SEGMENT_SIZE                      2133
    _COMMENT                              
    SUPPRESS_CORE                         YES
    FORMATTING_MODE                       PARALLEL
    FORMAT_DATAVOLUME                     YES
    OFFICIAL_NODE                         
    UKT_CPU_RELATIONSHIP                  NONE
    HIRES_TIMER_TYPE                      CPU
    LOAD_BALANCING_CHK                    30
    LOAD_BALANCING_DIF                    10
    LOAD_BALANCING_EQ                     5
    HS_STORAGE_DLL                        libhsscopy
    HS_SYNC_INTERVAL                      50
    USE_OPEN_DIRECT                       YES
    USE_OPEN_DIRECT_FOR_BACKUP            NO
    SYMBOL_DEMANGLING                     NO
    EXPAND_COM_TRACE                      NO
    JOIN_TABLEBUFFER                      128
    SET_VOLUME_LOCK                       YES
    SHAREDSQL                             YES
    SHAREDSQL_CLEANUPTHRESHOLD            25
    SHAREDSQL_COMMANDCACHESIZE            262144
    MEMORY_ALLOCATION_LIMIT               0
    USE_SYSTEM_PAGE_CACHE                 YES
    USE_COROUTINES                        YES
    FORBID_LOAD_BALANCING                 YES
    MIN_RETENTION_TIME                    60
    MAX_RETENTION_TIME                    480
    MAX_SINGLE_HASHTABLE_SIZE             512
    MAX_HASHTABLE_MEMORY                  5120
    ENABLE_CHECK_INSTANCE                 YES
    RTE_TEST_REGIONS                      0
    HASHED_RESULTSET                      YES
    HASHED_RESULTSET_CACHESIZE            262144
    CHECK_HASHED_RESULTSET                0
    AUTO_RECREATE_BAD_INDEXES             NO
    AUTHENTICATION_ALLOW                  
    AUTHENTICATION_DENY                   
    TRACE_AK                              NO
    TRACE_DEFAULT                         NO
    TRACE_DELETE                          NO
    TRACE_INDEX                           NO
    TRACE_INSERT                          NO
    TRACE_LOCK                            NO
    TRACE_LONG                            NO
    TRACE_OBJECT                          NO
    TRACE_OBJECT_ADD                      NO
    TRACE_OBJECT_ALTER                    NO
    TRACE_OBJECT_FREE                     NO
    TRACE_OBJECT_GET                      NO
    TRACE_OPTIMIZE                        NO
    TRACE_ORDER                           NO
    TRACE_ORDER_STANDARD                  NO
    TRACE_PAGES                           NO
    TRACE_PRIMARY_TREE                    NO
    TRACE_SELECT                          NO
    TRACE_TIME                            NO
    TRACE_UPDATE                          NO
    TRACE_STOP_ERRORCODE                  0
    TRACE_ALLOCATOR                       0
    TRACE_CATALOG                         0
    TRACE_CLIENTKERNELCOM                 0
    TRACE_COMMON                          0
    TRACE_COMMUNICATION                   0
    TRACE_CONVERTER                       0
    TRACE_DATACHAIN                       0
    TRACE_DATACACHE                       0
    TRACE_DATAPAM                         0
    TRACE_DATATREE                        0
    TRACE_DATAINDEX                       0
    TRACE_DBPROC                          0
    TRACE_FBM                             0
    TRACE_FILEDIR                         0
    TRACE_FRAMECTRL                       0
    TRACE_IOMAN                           0
    TRACE_IPC                             0
    TRACE_JOIN                            0
    TRACE_KSQL                            0
    TRACE_LOGACTION                       0
    TRACE_LOGHISTORY                      0
    TRACE_LOGPAGE                         0
    TRACE_LOGTRANS                        0
    TRACE_LOGVOLUME                       0
    TRACE_MEMORY                          0
    TRACE_MESSAGES                        0
    TRACE_OBJECTCONTAINER                 0
    TRACE_OMS_CONTAINERDIR                0
    TRACE_OMS_CONTEXT                     0
    TRACE_OMS_ERROR                       0
    TRACE_OMS_FLUSHCACHE                  0
    TRACE_OMS_INTERFACE                   0
    TRACE_OMS_KEY                         0
    TRACE_OMS_KEYRANGE                    0
    TRACE_OMS_LOCK                        0
    TRACE_OMS_MEMORY                      0
    TRACE_OMS_NEWOBJ                      0
    TRACE_OMS_SESSION                     0
    TRACE_OMS_STREAM                      0
    TRACE_OMS_VAROBJECT                   0
    TRACE_OMS_VERSION                     0
    TRACE_PAGER                           0
    TRACE_RUNTIME                         0
    TRACE_SHAREDSQL                       0
    TRACE_SQLMANAGER                      0
    TRACE_SRVTASKS                        0
    TRACE_SYNCHRONISATION                 0
    TRACE_SYSVIEW                         0
    TRACE_TABLE                           0
    TRACE_VOLUME                          0
    CHECK_BACKUP                          NO
    CHECK_DATACACHE                       NO
    CHECK_KB_REGIONS                      NO
    CHECK_LOCK                            NO
    CHECK_LOCK_SUPPLY                     NO
    CHECK_REGIONS                         NO
    CHECK_TASK_SPECIFIC_CATALOGCACHE      NO
    CHECK_TRANSLIST                       NO
    CHECK_TREE                            NO
    CHECK_TREE_LOCKS                      NO
    CHECK_COMMON                          0
    CHECK_CONVERTER                       0
    CHECK_DATAPAGELOG                     0
    CHECK_DATAINDEX                       0
    CHECK_FBM                             0
    CHECK_IOMAN                           0
    CHECK_LOGHISTORY                      0
    CHECK_LOGPAGE                         0
    CHECK_LOGTRANS                        0
    CHECK_LOGVOLUME                       0
    CHECK_SRVTASKS                        0
    OPTIMIZE_AGGREGATION                  YES
    OPTIMIZE_FETCH_REVERSE                YES
    OPTIMIZE_STAR_JOIN                    YES
    OPTIMIZE_JOIN_ONEPHASE                YES
    OPTIMIZE_JOIN_OUTER                   YES
    OPTIMIZE_MIN_MAX                      YES
    OPTIMIZE_FIRST_ROWS                   YES
    OPTIMIZE_OPERATOR_JOIN                YES
    OPTIMIZE_JOIN_HASHTABLE               YES
    OPTIMIZE_JOIN_HASH_MINIMAL_RATIO      1
    OPTIMIZE_OPERATOR_JOIN_COSTFUNC       YES
    OPTIMIZE_JOIN_PARALLEL_MINSIZE        1000000
    OPTIMIZE_JOIN_PARALLEL_SERVERS        0
    OPTIMIZE_JOIN_OPERATOR_SORT           YES
    OPTIMIZE_QUAL_ON_INDEX                YES
    DDLTRIGGER                            YES
    SUBTREE_LOCKS                         NO
    MONITOR_READ                          2147483647
    MONITOR_TIME                          2147483647
    MONITOR_SELECTIVITY                   0
    MONITOR_ROWNO                         0
    CALLSTACKLEVEL                        0
    OMS_RUN_IN_UDE_SERVER                 NO
    OPTIMIZE_QUERYREWRITE                 OPERATOR
    TRACE_QUERYREWRITE                    0
    CHECK_QUERYREWRITE                    0
    PROTECT_DATACACHE_MEMORY              NO
    LOCAL_REDO_LOG_BUFFER_SIZE            0
    FILEDIR_SPINLOCKPOOL_SIZE             10
    TRANS_HISTORY_SIZE                    0
    TRANS_THRESHOLD_VALUE                 60
    ENABLE_SYSTEM_TRIGGERS                YES
    DBFILLINGABOVELIMIT                   70L80M85M90H95H96H97H98H99H
    DBFILLINGBELOWLIMIT                   70L80L85L90L95L
    LOGABOVELIMIT                         50L75L90M95M96H97H98H99H
    AUTOSAVE                              1
    BACKUPRESULT                          1
    CHECKDATA                             1
    EVENT                                 1
    ADMIN                                 1
    ONLINE                                1
    UPDSTATWANTED                         1
    OUTOFSESSIONS                         3
    ERROR                                 3
    SYSTEMERROR                           3
    DATABASEFULL                          1
    LOGFULL                               1
    LOGSEGMENTFULL                        1
    STANDBY                               1
    USESELECTFETCH                        YES
    USEVARIABLEINPUT                      NO
    UPDATESTAT_PARALLEL_SERVERS           0
    UPDATESTAT_SAMPLE_ALGO                1
    SIMULATE_VECTORIO                     IF_OPEN_DIRECT_OR_RAW_DEVICE
    COLUMNCOMPRESSION                     YES
    TIME_MEASUREMENT                      NO
    CHECK_TABLE_WIDTH                     NO
    MAX_MESSAGE_LIST_LENGTH               100
    SYMBOL_RESOLUTION                     YES
    PREALLOCATE_IOWORKER                  NO
    CACHE_IN_SHARED_MEMORY                NO
    INDEX_LEAF_CACHING                    2
    NO_SYNC_TO_DISK_WANTED                NO
    SPINLOCK_LOOP_COUNT                   30000
    SPINLOCK_BACKOFF_BASE                 1
    SPINLOCK_BACKOFF_FACTOR               2
    SPINLOCK_BACKOFF_MAXIMUM              64
    ROW_LOCKS_PER_TRANSACTION             50
    USEUNICODECOLUMNCOMPRESSION           NO
    about send you the data from tables, i dont have permission to do that, since all data is in a production system, the customer dont give me the rights to send any information. sorry about that.
    best regards
    Clóvis

  • Amazon print guidelines specify one page per sheet

    Amazon print guidelines specify one page per sheet
    I am setting up a new book and not sure I understand Amazon guidlines "one page per sheet".
    I am pretty sure this setup is one page per sheet when Exported to PDF -- can any one please confirm this?

    Thanks, you mean be sure "Spreads" is unchecked as in this screenshot (I don't know what I am doing here)...?

  • How do you get table column names to show one time per page

    I have a report where I want the column names to show one time per page.
    I have the table with the column names set below the header and above the
    'for-each Row'.
    When I view the report in PDF format the colunm names show only one time on the first page and on the other pages the column names do not show at all. If I view the report in RTF, the same thing happens, shows column names on the first page, but not on the other pages after that.
    I highlighted the table and went to table properties, clicked on ROW tab and then checked (Repeat as header row at the top of each page) and it still does not work for PDF or RTF format. I have looked at other forums with this issue and it doesn't work for me in PDF format.
    Could someone tell me what I am doing wrong.....
    I appreciate anyone that can help me.
    Thank you,
    Susie

    I figured out what I was doing wrong.
    I was creating a seperate table for the column names and not using the table where the column names where created originally. I deleted the column names from the original table and created a seperate table above that table with the column names only.
    So I went back to the original table and highlighted the row with the column names and went to table properties/ROW tab/ and clicked (Repeat as header row at the top of each page) and now it works perfectly in PDF format where the column names show one time for each page. Deleted the other table where I had the column names only.
    thanks
    Susie

  • What if i need more than one prepareStatements per connection

    hi all!!
    Please take some time to look into this scenario.
    On the front end: User enters/ updates his/her records and submits. An ArrayList of object goes to backend.
    On back end: i have to loop for the size of the ArrayList submitted and see if some record is to be updated or inserted.
    Now at back end i do following steps:
    I have only one Connection object named connection.
    Step 1:
    query (say query1 )to see if the record in DB exists for the primary key as extracted from the element from ArrayList. So one prepareStatment ( say prepareStatement1) for this. I write a query and execute :
    prepareStatment1= connection.prepareStatement(query1);
    Step2 :
    A query ( say query2) to update Data in DB. And i have a separtae prepareStatement for this and i execute
    prepareStatment2 = connection.prepareStatement(query2);
    Sep3
    A query (say query3 )to insert Data in DB. And i have a separtae prepareStatement for this and i execute
    prepareStatment3 = connection.prepareStatement(query3);
    Step4:
    for: size of arraylist:
    1. take out the element at ith index from the list.
    2. see if the record exists i.e execute step1 and see if resultSet.next() is true.
    3. if record exists step2 applies so execute update query
    4. if record doesnot exist then go to step3 and execute insert query
    A weird thing happened .. insert was running fine but update was not. May be i was not checking appropriately. BUT the error that occurrd was that
    [Microsoft][ODBC SQL Server Driver]Connection is busy with
    results for another hstmt.
    So i changed my path and in loop i prepareStatement and execute query to see if record exists or not and store the result in some boolean variable and then close this prepareStatement. and if the result existed again prepare a statement for update or else prepare a statement for insert and execute the respective.
    and then again close this prepareStatement and go for next iteration of loop
    so each iteration i am preparing statement and then closing it BECAUSE Connection doesnot allow more than one statement per connection at one time. Am i right.??
    what should i do.? Although i have solved problem by closing and preparing in loop but that is so inefficient. Is there some elegant way???
    please reply.
    PS: I am using ms sql server 8.0 as my DBMS. I am developing web application using Sun Creator.

    No, you can use as many PreparedStatements as you like, all from one Connection, and you can keep them open as long as you like. (I do this with no problems.) Where you will run into problems is when you are trying to use two ResultSets at the same time. That's what the error message means, it just doesn't use JDBC terminology.

Maybe you are looking for