Missing configuration entry for data - index relationship

IN ST04 -> configuration -> Data Class
all data classes are in yellow I'm trying to activate them and getting the following error
"Missing configuration entry for data - index relationship"
This is the result of  a system copy
Thanks
Kkhaldoun

Hi All,
Just an update, I was able to get around my problems by patching the 2004s media from sp6 to sp8, uninstalling my scs, and database and re-installing from the patched media.

Similar Messages

  • WD4VC Callable Object: No parser configuration entry for the tag Invoke

    I am trying to use a Visual Composer model as a callable object. I can deploy OK (as a Web Dynpro model, having checked the copmiler options "Deploy Model with PAR" and "Add Guided Procedure information").
    When I try to test the Callable Object, I get the following error:
    com.sap.tc.wd4vc.intapi.info.exception.WD4VCRuntimeException: No parser configuration entry for the tag Invoke.
    No idea how to get around this.
    Any help greatly appreciated.
    Kind Regards,
    Tony.

    Hi Mario,
    We have only three runtime in VC 7.0
    1. XGraph
    2. Flash
    3. Webdynpro.
    In your blog on wiki Flex runtime is mentioned  as given below:
    Webdynpro runtime is not fully supported in NW04s and some tags may not be supported as stated in the error. The reccomendation is to use the flex runtime.
    Is Flex runtime and  Flash runtime is same ,
    I have tried to deploy the model in Flash runtime and  Portal runtime error ocurs by diasply renderning JSP Component,
    But when i deploy it in Webdynpro runtime , it works fine for displaying data in Table.But gives error :No parser configuration entry for the tag ChartView , when i used chart in place of table .
    Thanks.

  • No parser configuration entry for the tag ChartView

    Hi All,
    We have created chartview in VC 7.0 but getting below error while deploying in webdynpro runtime.
    <b>com.sap.tc.wd4vc.intapi.info.exception.WD4VCRuntimeException: No parser configuration entry for the tag ChartView</b>
    Rohit

    Hi Mario,
    We have only three runtime in VC 7.0
    1. XGraph
    2. Flash
    3. Webdynpro.
    In your blog on wiki Flex runtime is mentioned  as given below:
    Webdynpro runtime is not fully supported in NW04s and some tags may not be supported as stated in the error. The reccomendation is to use the flex runtime.
    Is Flex runtime and  Flash runtime is same ,
    I have tried to deploy the model in Flash runtime and  Portal runtime error ocurs by diasply renderning JSP Component,
    But when i deploy it in Webdynpro runtime , it works fine for displaying data in Table.But gives error :No parser configuration entry for the tag ChartView , when i used chart in place of table .
    Thanks.

  • Configurational settings for data source in BI server

    Hi Sdns,
    Where can we make  configurational settings for data source in BI server.
    Advance thanks to replies.
    karun

    Go to T-code RSA13.
    Search you desired BI system.
    Right click on it and Display Data Source Tree.
    There you can see all your datasource with BI as source system.
    You can do the same for other source systems.
    Double click on the datasource to see the fields and configuration.

  • PKG-Building: WARNING: missing directory entry for /tmp

    Hi,
    After I ran the 'pkgmk -o -r / -d /tmp/pkg -f my.prototype' I got followings:
    ## Building pkgmap from package prototype file.
    ## Processing pkginfo file.
    WARNING: missing directory entry for </tmp>
    WARNING: missing directory entry for </tmp/test>
    ## Attempting to volumize 4 entries in pkgmap.
    part  1 -- 167 blocks, 16 entries
    ## Packaging one part.
    /tmp/pkg/WChen/pkgmap
    /tmp/pkg/WChen/pkginfo
    /tmp/pkg/WChen/root/tmp/test/test1.txt
    /tmp/pkg/WChen/root/tmp/test/test2.txt
    /tmp/pkg/WChen/root/tmp/test/test3.txt
    ## Validating control scripts.
    ## Packaging complete.
    What does here "WARNING: missing directory entry for </tmp>" mean? Though this is not critical, I want to know if it's possible to get rid of it.

    you want add this into pacmans path
    Server = ftp://ftp.archlinux.org/tur/hapy
    then
    pacman -S gnomad2
    its so much easier
    but if your current like if you ran pacman -Suy after oct 1 i belive
    then it probably wont work
    mine did till the new upgrades
    ive not been able to build it from source like your doing ive tried many times
    i read some where in the read me files bout jsut adding  modprobe njbfs then mount it like
    a mass storage device im gonna try looking into that today
    i also posted ? to the forum on this

  • Configure listener for data guard

    HI everyone,
    I am currently setting data guard (Physical standby database) for my database. But I have problem to configure the listener on both servers. Can anyone provide me some example?
    Oracle: 10g R2
    O/S: Windows
    Primary database ken10g
    standby database: ken10gbk
    Following is the content of my current listener files on both of servers:
    Primary server:
    # listener.ora Network Configuration File: C:\oracle\product\10.2.0\db_1\network\admin\listener.ora
    # Generated by Oracle configuration tools.
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = C:\oracle\product\10.2.0\db_1)
    (PROGRAM = extproc)
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))
    (ADDRESS = (PROTOCOL = TCP)(HOST = Primary_server)(PORT = 1521))
    Standby Server:
    # listener.ora Network Configuration File: C:\oracle\product\10.2.0\db_1\NETWORK\ADMIN\listener.ora
    # Generated by Oracle configuration tools.
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = C:\oracle\product\10.2.0\db_1)
    (PROGRAM = extproc)
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1))
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = standby_server)(PORT = 1521))
    Thanks in advance.
    Ken

    Hi Ken,
    You need to configure this on both primary and standby, I would have kept different listener name on primary and standby. Also if you are going to use dataguard broker you would need to set GLOBAL_DBNAME in your listener.ora file
    I have give a sample entry for tnsnames.ora and listener.ora
    TNSNAMES.ORA on primary
    STNDBY =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = TCP)(HOST = node2-prv)(PORT = 10521))
    (CONNECT_DATA =
    (SERVICE_NAME = STNDBY)
    PRIM =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = node1-prv)(PORT = 10521)))
    PRIMARY =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = TCP)(HOST = node1-prv)(PORT = 10521))
    (CONNECT_DATA =
    (SERVICE_NAME = PRIMARY)
    EXTPROC_CONNECTION_DATA =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC0))
    (CONNECT_DATA =
    (SID = PLSExtProc)
    (PRESENTATION = RO)
    Copy the same file to the standby server and adjust it based on the listener.ora file. Also update the listener.ora file so that it listen the SIDs mentioned in the tnsnames.ora file.
    Listener.ora
    LISTENER_STBY =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = node2-prv)(PORT = 10521))
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC0))
    SID_LIST_LISTENER_STBY =
    (SID_LIST =
    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = /u01/app/oracle/product/10.2.0/db10g)
    (PROGRAM = extproc)
    (SID_DESC =
    (SID_NAME = stndby)
    (GLOBAL_DBNAME = stndby_DGMGRL)
    (ORACLE_HOME = /u01/app/oracle/product/10.2.0/db10g)
    )

  • Issue with table ROOSPRMSF entries for data source 0FI_AP_4

    Hi Experts,
    I am facing with an issue where we found incosistencies with table ROOSPRMSF in R/3 system.
    In BW , we have done initializations based on fiscal period selections (none of the selections overlap) for data source 0FI_AP_4.
    We have done in total 7 initializations. So in BW system in table RSSDLINITSEL we have 7 initialization requests.
    But in R/3 system we have 49 records for data source 0FI_AP_4 in ROOSPRMSF table out of which 42 are invalid records.
    I suspect that these 42 invalid records are created due to the execution of program RSSM_OLTP_INIT_DELTA_UPDATE when the tables ROOSPRMSF are actually holding the 7 initialization request entries.   Due to this each and every initialization request is linked to rest of the other intialization requests and ended with 49 records in table ROOSPRMSF table.
    Now our data loads are running fine but daily a short dump is raised . In the daily loads, BW init records in RSSDLINITSEL are compared with ROOSPRMSF entries and all the other 42 records which are invalid are written into system log and a short dump is raised.
    In order to fix these inconsistencies i checked for OSS note 852443. (Point 3 in OSS note)
    But it is specified to delete the delta queue for data source 0FI_AP_4 in RSA7 and instructed to execute the program RSSM_OLTP_INIT_DELTA_UPDATE so that the ROOSPRMSF table will be reconstructed with valid records available in RSSDLINITSEL. 
    From OSS note 852443 point 3
    "3. If the RSSDLINIT table in the BW system already contains entries, check the requests listed there in the RNR column in the monitor (transaction RSRQ). Compare these entries with the entries in the ROOSPRMSF and ROOSPRMSC tables with the INITRNR field. If, in the ROOSPRMSF and ROOSPRMSC tables for your DataSource source system combination, there are more entries with different INITRNR numbers, use transaction RSA7 in an OLTP source system to delete all entries and then use the RSSM_OLTP_INIT_DELTA_UPDATE report mentioned in the next section. For a DataMart source system, delete the entries that you cannot find in the RSSDLINIT table using the procedure described above."
    My question is if we delete the delta queue in RSA7 then all the tables in R/3 (ROOSPRMSF, ROOSPRMSC, Time stamp table) and BW (RSSDLINITSEL, initialization requests will be deleted) will be cleared. Then how will the program RSSM_OLTP_INIT_DELTA_UPDATE  copy entries into ROOSPRMSF table in R/3 ?
    Could any one please clarify this ?
    Thanks
    Regards,
    Jeswanth

    Hi Amarnath,
    Did you unhide the new field in RSA6 and regenerated the DataSource?
    Often SAP will populate newly added fields (belonging to the same (set) of table(s) used for extraction) automatically (e.g. SAP uses 'move-corresponding' in it's extractor-code, or, in this case, reading all fields from the DD, FM BWFIU_TRANSFORM_FIELDLIST).
    If the DataSource looks fine to you and the field is still not populated in RSA3 you can't go without a user-exit.
    Grtx,
    Marco

  • Missing registry entries for burning CDs

    I used iTunes to burn soe playlists. This worked well. However, I also installed Nero Exprss after that. After this when  I brought up iTunes it gave the message "MIssing registry entries - cannot import or burn CDs or DVDs - reinstall iTunes" Though I was still able to import CDs after that, I cannot burn CDs. The error does not go away even after re-installing iTunes several times (either usiing an Update or after unistalling and doing a fresh install).
    How can I fix this problem ?

    I'd start with the following document, with one modification. At step 12 after typing GEARAspiWDM press the Enter/Return key once prior to clicking OK. (Pressing Return adds a carriage return in the field and is important.)
    iTunes for Windows: "Registry settings" warning when opening iTunes

  • BPC NW 7.0: Data Load: rejected entries for ENTITY member

    Hi,
    when trying to load data from a BW info provider into BPC (using UJD_TEST_PACKAGE & process chain scheduling), a number of records is being rejected due to missing member entries for the ENTITY dimension in the application.
    However, the ENTITY member actually do exist in the application. Also, the dimension is processed with no errors. The dimension member are also visible usnig the Excel Client naviagtion pane for selecting members.
    The error also appears when kicking of the data load from the Excel Client for BPC. Any ideas how to analyze this further or resolve this?
    Thanks,
    Claudia Elsner

    Jeffrey,
    this question is closely related to the issue, because there is also a short dump when trying to load the data into BPC. I am not sure whether both problems are directly related though:
    Short dump with UJD_TEST_PACKAGE
    Problem desription of the post:
    When running UJD_TEST_PACKAGE, I get a short dump.
    TSV_TNEW_PAGE_ALLOC_FAILED
    No more storage space available for extending an internal table.
    Other keywords are CL_SHM_AREA and ATTACHUPDATE70.
    When I looked at NOTES, I found this Note 928044 - BI lock server". Looking at the note and debugging UJD_TEST_PACKAGE leaves me some questions:
    1. Do I need a BI lock server?
    2. Should I change the enque/table_size setting be increased on the central instance from 10000 to 25000 or larger?
    Claudia

  • How to add in ODBC connection and configuration for data connection

    Hello expert,
    on my computor, there is a oracle client for oracle 10g. now there is new oracle 11g installed , I want to create a ODBC connection to this new database. but when I create ODBC , I can't find server type for oracle 11g , there only has driver for oracle 10g. will you please tell me how to create ODBC for this new DB? will you please tell me where is the configuration file for data source on my computor?
    Many Thanks,

    918440 wrote:
    Hello expert,
    on my computor, there is a oracle client for oracle 10g. now there is new oracle 11g installed , I want to create a ODBC connection to this new database. but when I create ODBC , I can't find server type for oracle 11g , there only has driver for oracle 10g. will you please tell me how to create ODBC for this new DB? will you please tell me where is the configuration file for data source on my computor?
    Many Thanks,So, on this machine you have
    - Oracle 10g client (only), and in a seperate ORACLE_HOME
    - Oracle 11g rdbms
    First, the 10g client will connect quite well to the 11g database
    Second, the 11g database includes the basic client admin software, but the odbc drivers are not installed by default. You have to run OUI and choose 'custom' installation, then select 'windows components' from the list of available components.
    I've never wanted to install the odbc driver as part of the database installation. Only on remote client installations.
    What are you really trying to achieve?

  • Missing default flag for communication type in crm/business partner

    Hello!
    We've connected a crm 5.0 with an R/3-system (4.6c).
    The crm-system is the leading system...
    While inserting a new customer in crm with phone- and fax-number.
    After "save" we've got red bdocs, telling us that the standard entry for communication type is missing.
    In crm 5.0 there's no standard flag for the phone number but a fixed range with from-date and to-date.
    The R/3-system needs the default flag (an "X") as a sign that the number is the standard number.
    Anywhere there should be the problem...
    Is here anybody with some advices for us?
    Thanks a lot in advance.
    Best regards,
    Ingo

    Dear Ingo,
    I assume you talk about error message XR002
    You have 3 options to resolve the issue:
    1. In order to avoid error message XR002 you must set the Standard
    Indicator Validity in the maintained telephone in BP transaction.
    You can find this indicator in the yellow arrow near the telephone
    number. Here you must press the button "Standard Indicator Validity".
    2. In customizing you have the possibility to switch off the time
    dependency of the addresses causing the XR002.
    3. The error occurs due to missing ADRU entries for the communication
    data (Email, Phone, Fax etc).
    The function modules used in data exchange assumes that there is
    an ADRU entry for all addresses. This assumption is partially correct.
    If an address is created in Basis release 6.40 and above, then for all
    communication data of that address ADRU entries will be created.
    But if a Business Partner(with address) is created in release 6.20 or
    below and system is upgraded to 7.0, ADRU entries for the communication
    data of thoses addresses are not automatically created. When the user
    opens that BP in change mode & modifies the communication data and
    saves it then ADRU entries for that BP's communication data are created.
    There exist a correction report and several notes, so I think if above info
    is not able to help, it would be better to open a support message!
    Regards,Gerhard

  • What are the best solutions for data warehouse configuration in 10gR2

    I need help on solutions to be provided to my Client for upgrading the data warehouse.
    Current Configuration: Oracle database 9.2.0.8. This database contains the data warehouse and one more data mart on the same host.Sizes are respectively 6 Terabyte(retention policy of 3 years+current year) and 1 Terabyte. The ETL tool and BO Reporting tools are also hosted on the same host. This current configuration is really performing poor.
    Client cannot go for a major architectural or configuration changes to its existing environment now due to some constraints.
    However, they have agreed to separate out the databases on separate hosts from the ETL tools and BO objects. Also we are planning to upgrade the database to 10gR2 to attain stability, better performance and overcome current headaches.
    We cannot upgrade the database to 11g as the BO is at a version 6.5 which isn't compatible with Oracle 11g. And Client cannot afford to upgrade anything else other than the database.
    So, my role is very vital in providing a perfect solution towards better performance and take a successful migration of Oracle Database from one host to another (similar platform and OS) in addition to upgrade.
    I have till now thought of the following:
    Move the Oracle database and data mart to separate host.
    The host will be the same platform, that is, HP Superdome with HP-UX 32-bit OS (we cannot change to 64-bit as ETL tool doesn't support)
    Install new Oracle database 10g on the new host and move the data to it.
    Exploring all new features of 10gR2 to help data warehouse, that is, SQL MODEL Clause introduction, Parallel processing, Partitioning, Data Pump, SPA to study pre and post migrations.
    Also thinking of RAC to provide more better solution as our main motive is to show a tremendous performance enhancement.
    I need all your help to prepare a good road map for my assignment. Please suggest.
    Thanks,
    Tapan

    SGA=27.5 GB and PGA=50 MB
    Also I am pasting part of STATSPACK Report, eliminating the snaps of DB bounce. Please suggest the scope of improvement in this case.
    STATSPACK report for
    Snap Id Snap Time Sessions Curs/Sess Comment
    Begin Snap: 582946 11-Mar-13 20:02:16 46 12.8
    End Snap: 583036 12-Mar-13 18:24:24 60 118.9
    Elapsed: 1,342.13 (mins)
    Cache Sizes (end)
    ~~~~~~~~~~~~~~~~~
    Buffer Cache: 21,296M Std Block Size: 16K
    Shared Pool Size: 6,144M Log Buffer: 16,384K
    Load Profile
    ~~~~~~~~~~~~ Per Second Per Transaction
    Redo size: 1,343,739.01 139,883.39
    Logical reads: 100,102.54 10,420.69
    Block changes: 3,757.42 391.15
    Physical reads: 6,670.84 694.44
    Physical writes: 874.34 91.02
    User calls: 1,986.04 206.75
    Parses: 247.87 25.80
    Hard parses: 5.82 0.61
    Sorts: 1,566.76 163.10
    Logons: 10.99 1.14
    Executes: 1,309.79 136.35
    Transactions: 9.61
    % Blocks changed per Read: 3.75 Recursive Call %: 43.34
    Rollback per transaction %: 3.49 Rows per Sort: 190.61
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer Nowait %: 99.90 Redo NoWait %: 100.00
    Buffer Hit %: 96.97 In-memory Sort %: 100.00
    Library Hit %: 99.27 Soft Parse %: 97.65
    Execute to Parse %: 81.08 Latch Hit %: 99.58
    Parse CPU to Parse Elapsd %: 3.85 % Non-Parse CPU: 99.34
    Shared Pool Statistics Begin End
    Memory Usage %: 7.11 50.37
    % SQL with executions>1: 62.31 46.46
    % Memory for SQL w/exec>1: 26.75 13.47
    Top 5 Timed Events
    ~~~~~~~~~~~~~~~~~~ % Total
    Event Waits Time (s) Ela Time
    CPU time 492,062 43.66
    db file sequential read 157,418,414 343,549 30.49
    library cache pin 92,339 66,759 5.92
    PX qref latch 63,635 43,845 3.89
    db file scattered read 2,506,806 41,677 3.70
    Background Wait Events for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> ordered by wait time desc, waits desc (idle events last)
    Avg
    Total Wait wait Waits
    Event Waits Timeouts Time (s) (ms) /txn
    log file sequential read 176,386 0 3,793 22 0.2
    log file parallel write 2,685,833 0 1,813 1 3.5
    db file parallel write 239,166 0 1,350 6 0.3
    control file parallel write 33,432 0 79 2 0.0
    LGWR wait for redo copy 478,120 536 75 0 0.6
    rdbms ipc reply 10,027 0 47 5 0.0
    control file sequential read 32,414 0 40 1 0.0
    db file scattered read 4,101 0 30 7 0.0
    db file sequential read 13,946 0 29 2 0.0
    direct path read 203,694 0 14 0 0.3
    log buffer space 363 0 13 37 0.0
    latch free 3,766 0 9 2 0.0
    direct path write 80,491 0 6 0 0.1
    async disk IO 351,955 0 4 0 0.5
    enqueue 28 0 1 21 0.0
    buffer busy waits 1,281 0 1 0 0.0
    log file single write 172 0 0 1 0.0
    rdbms ipc message 10,563,204 251,286 992,837 94 13.7
    pmon timer 34,751 34,736 78,600 2262 0.0
    smon timer 7,462 113 76,463 10247 0.0
    Instance Activity Stats for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    Statistic Total per Second per Trans
    CPU used by this session 49,206,154 611.0 63.6
    CPU used when call started 49,435,735 613.9 63.9
    CR blocks created 6,740,777 83.7 8.7
    Cached Commit SCN referenced 423,253,503 5,256.0 547.2
    Commit SCN cached 19,165 0.2 0.0
    DBWR buffers scanned 48,276,489 599.5 62.4
    DBWR checkpoint buffers written 6,959,752 86.4 9.0
    DBWR checkpoints 454 0.0 0.0
    DBWR free buffers found 44,817,183 556.5 57.9
    DBWR lru scans 137,149 1.7 0.2
    DBWR make free requests 162,528 2.0 0.2
    DBWR revisited being-written buff 4,220 0.1 0.0
    DBWR summed scan depth 48,276,489 599.5 62.4
    DBWR transaction table writes 5,036 0.1 0.0
    DBWR undo block writes 2,989,436 37.1 3.9
    DDL statements parallelized 3,723 0.1 0.0
    DFO trees parallelized 4,157 0.1 0.0
    DML statements parallelized 3 0.0 0.0
    OS Block input operations 29,850 0.4 0.0
    OS Block output operations 1,591 0.0 0.0
    OS Characters read/written 182,109,814,791 2,261,447.1 235,416.9
    OS Integral unshared data size ################## 242,463,432.4 ############
    OS Involuntary context switches 188,257,786 2,337.8 243.4
    OS Maximum resident set size 43,518,730,619 540,417.4 56,257.5
    OS Page reclaims 159,430,953 1,979.8 206.1
    OS Signals received 5,260,938 65.3 6.8
    OS Socket messages received 79,438,383 986.5 102.7
    OS Socket messages sent 93,064,176 1,155.7 120.3
    OS System time used 10,936,430 135.8 14.1
    OS User time used 132,043,884 1,639.7 170.7
    OS Voluntary context switches 746,207,739 9,266.4 964.6
    PX local messages recv'd 55,120,663 684.5 71.3
    PX local messages sent 55,120,817 684.5 71.3
    Parallel operations downgraded 1 3 0.0 0.0
    Parallel operations not downgrade 4,154 0.1 0.0
    SQL*Net roundtrips to/from client 155,422,335 1,930.0 200.9
    SQL*Net roundtrips to/from dblink 18 0.0 0.0
    active txn count during cleanout 16,529,551 205.3 21.4
    background checkpoints completed 43 0.0 0.0
    background checkpoints started 43 0.0 0.0
    background timeouts 280,202 3.5 0.4
    branch node splits 4,428 0.1 0.0
    buffer is not pinned count 6,382,440,322 79,257.4 8,250.7
    buffer is pinned count 9,675,661,370 120,152.8 12,507.9
    bytes received via SQL*Net from c 67,384,496,376 836,783.4 87,109.3
    bytes received via SQL*Net from d 6,142 0.1 0.0
    bytes sent via SQL*Net to client 50,240,643,657 623,890.4 64,947.1
    bytes sent via SQL*Net to dblink 3,701 0.1 0.0
    calls to get snapshot scn: kcmgss 145,385,064 1,805.4 187.9
    calls to kcmgas 36,816,132 457.2 47.6
    calls to kcmgcs 3,514,770 43.7 4.5
    change write time 369,373 4.6 0.5
    cleanout - number of ktugct calls 20,954,488 260.2 27.1
    cleanouts and rollbacks - consist 6,357,174 78.9 8.2
    cleanouts only - consistent read 10,078,802 125.2 13.0
    cluster key scan block gets 69,403,565 861.9 89.7
    Instance Activity Stats for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    Statistic Total per Second per Trans
    cluster key scans 41,311,211 513.0 53.4
    commit cleanout failures: block l 413,776 5.1 0.5
    commit cleanout failures: buffer 414 0.0 0.0
    commit cleanout failures: callbac 41,194 0.5 0.1
    commit cleanout failures: cannot 174,382 2.2 0.2
    commit cleanouts 11,469,056 142.4 14.8
    commit cleanouts successfully com 10,839,290 134.6 14.0
    commit txn count during cleanout 17,155,424 213.0 22.2
    consistent changes 145,418,277 1,805.8 188.0
    consistent gets 8,043,252,188 99,881.4 10,397.7
    consistent gets - examination 3,180,028,047 39,489.7 4,110.9
    current blocks converted for CR 9 0.0 0.0
    cursor authentications 14,926 0.2 0.0
    data blocks consistent reads - un 143,706,500 1,784.6 185.8
    db block changes 302,577,666 3,757.4 391.2
    db block gets 336,562,217 4,179.4 435.1
    deferred (CURRENT) block cleanout 2,912,793 36.2 3.8
    dirty buffers inspected 627,174 7.8 0.8
    enqueue conversions 1,296,337 16.1 1.7
    enqueue releases 13,053,200 162.1 16.9
    enqueue requests 13,239,092 164.4 17.1
    enqueue timeouts 185,878 2.3 0.2
    enqueue waits 114,120 1.4 0.2
    exchange deadlocks 7,390 0.1 0.0
    execute count 105,475,101 1,309.8 136.4
    free buffer inspected 1,604,407 19.9 2.1
    free buffer requested 258,126,047 3,205.4 333.7
    hot buffers moved to head of LRU 22,793,576 283.1 29.5
    immediate (CR) block cleanout app 16,436,010 204.1 21.3
    immediate (CURRENT) block cleanou 2,860,013 35.5 3.7
    index fast full scans (direct rea 12,375 0.2 0.0
    index fast full scans (full) 3,733 0.1 0.0
    index fast full scans (rowid rang 192,148 2.4 0.3
    index fetch by key 1,321,024,486 16,404.5 1,707.7
    index scans kdiixs1 406,165,684 5,043.8 525.1
    leaf node 90-10 splits 50,373 0.6 0.1
    leaf node splits 697,235 8.7 0.9
    logons cumulative 884,756 11.0 1.1
    messages received 3,276,719 40.7 4.2
    messages sent 3,257,171 40.5 4.2
    no buffer to keep pinned count 569 0.0 0.0
    no work - consistent read gets 4,406,092,172 54,715.0 5,695.8
    opened cursors cumulative 20,527,704 254.9 26.5
    parse count (failures) 267,088 3.3 0.4
    parse count (hard) 468,996 5.8 0.6
    parse count (total) 19,960,548 247.9 25.8
    parse time cpu 323,024 4.0 0.4
    parse time elapsed 8,393,422 104.2 10.9
    physical reads 537,189,332 6,670.8 694.4
    physical reads direct 292,545,140 3,632.8 378.2
    physical writes 70,409,002 874.3 91.0
    physical writes direct 59,248,394 735.8 76.6
    physical writes non checkpoint 69,103,391 858.1 89.3
    pinned buffers inspected 11,893 0.2 0.0
    prefetched blocks 95,892,161 1,190.8 124.0
    prefetched blocks aged out before 1,495,883 18.6 1.9
    Instance Activity Stats for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    Statistic Total per Second per Trans
    process last non-idle time ################## ############## ############
    queries parallelized 417 0.0 0.0
    recursive calls 122,323,299 1,519.0 158.1
    recursive cpu usage 3,144,533 39.1 4.1
    redo blocks written 180,881,558 2,246.2 233.8
    redo buffer allocation retries 5,400 0.1 0.0
    redo entries 164,728,513 2,045.6 213.0
    redo log space requests 1,006 0.0 0.0
    redo log space wait time 2,230 0.0 0.0
    redo ordering marks 2,563 0.0 0.0
    redo size 108,208,614,904 1,343,739.0 139,883.4
    redo synch time 558,520 6.9 0.7
    redo synch writes 2,343,824 29.1 3.0
    redo wastage 1,126,585,600 13,990.0 1,456.4
    redo write time 718,655 8.9 0.9
    redo writer latching time 7,763 0.1 0.0
    redo writes 2,685,833 33.4 3.5
    rollback changes - undo records a 522,742 6.5 0.7
    rollbacks only - consistent read 335,177 4.2 0.4
    rows fetched via callback 1,100,990,382 13,672.1 1,423.3
    session connect time ################## ############## ############
    session cursor cache count 1,061 0.0 0.0
    session cursor cache hits 1,687,796 21.0 2.2
    session logical reads 8,061,057,193 100,102.5 10,420.7
    session pga memory 1,573,228,913,832 19,536,421.0 2,033,743.8
    session pga memory max 1,841,357,626,496 22,866,054.4 2,380,359.0
    session uga memory 1,074,114,630,336 13,338,399.4 1,388,529.0
    session uga memory max 386,645,043,296 4,801,374.0 499,823.6
    shared hash latch upgrades - no w 410,360,146 5,095.9 530.5
    sorts (disk) 2,657 0.0 0.0
    sorts (memory) 126,165,625 1,566.7 163.1
    sorts (rows) 24,048,783,304 298,638.8 31,088.3
    summed dirty queue length 5,438,201 67.5 7.0
    switch current to new buffer 1,302,798 16.2 1.7
    table fetch by rowid 6,201,503,534 77,010.5 8,016.8
    table fetch continued row 26,649,697 330.9 34.5
    table scan blocks gotten 1,864,435,032 23,152.6 2,410.2
    table scan rows gotten 43,639,997,280 541,923.3 56,414.3
    table scans (cache partitions) 26,112 0.3 0.0
    table scans (direct read) 246,243 3.1 0.3
    table scans (long tables) 340,200 4.2 0.4
    table scans (rowid ranges) 359,617 4.5 0.5
    table scans (short tables) 9,111,559 113.2 11.8
    transaction rollbacks 4,819 0.1 0.0
    transaction tables consistent rea 824 0.0 0.0
    transaction tables consistent rea 1,386,848 17.2 1.8
    user calls 159,931,913 1,986.0 206.8
    user commits 746,543 9.3 1.0
    user rollbacks 27,020 0.3 0.0
    write clones created in backgroun 7 0.0 0.0
    write clones created in foregroun 4,350 0.1 0.0
    Buffer Pool Statistics for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> Standard block size Pools D: default, K: keep, R: recycle
    -> Default Pools for other block sizes: 2k, 4k, 8k, 16k, 32k
    Free Write Buffer
    Number of Cache Buffer Physical Physical Buffer Complete Busy
    P Buffers Hit % Gets Reads Writes Waits Waits Waits
    D 774,144 95.6############ 233,869,082 10,089,734 0 0########
    K 504,000 99.9############ 3,260,227 1,070,338 0 0 65,898
    R 63,504 96.2 196,079,539 7,511,863 535 0 0 0
    Buffer wait Statistics for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> ordered by wait time desc, waits desc
    Tot Wait Avg
    Class Waits Time (s) Time (ms)
    data block 7,791,121 14,676 2
    file header block 587 101 172
    undo header 151,617 71 0
    segment header 299,312 58 0
    1st level bmb 45,235 7 0
    bitmap index block 392 1 3
    undo block 4,250 1 0
    2nd level bmb 14 0 0
    system undo header 2 0 0
    3rd level bmb 1 0 0
    Latch Activity for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    ->"Get Requests", "Pct Get Miss" and "Avg Slps/Miss" are statistics for
    willing-to-wait latch get requests
    ->"NoWait Requests", "Pct NoWait Miss" are for no-wait latch get requests
    ->"Pct Misses" for both should be very close to 0.0
    Pct Avg Wait Pct
    Get Get Slps Time NoWait NoWait
    Latch Requests Miss /Miss (s) Requests Miss
    Consistent RBA 2,686,230 0.0 0.2 0 0
    FAL request queue 86 0.0 0 0
    FAL subheap alocation 0 0 2 0.0
    FIB s.o chain latch 1,089 0.0 0 0
    FOB s.o list latch 4,589,986 0.5 0.0 2 0
    NLS data objects 1 0.0 0 0
    SQL memory manager worka 5,963 0.0 0 0
    Token Manager 0 0 2 0.0
    active checkpoint queue 719,439 0.3 0.1 0 1 0.0
    alert log latch 184 0.0 0 2 0.0
    archive control 4,365 0.0 0 0
    archive process latch 1,808 0.6 0.6 0 0
    begin backup scn array 3,387,572 0.0 0.0 0 0
    cache buffer handles 1,577,222 0.2 0.0 0 0
    cache buffers chains ############## 0.5 0.0 430 354,357,972 0.3
    cache buffers lru chain 17,153,023 0.1 0.0 1 385,505,654 0.5
    cas latch 538,804,153 0.3 0.0 7 0
    channel handle pool latc 1,776,950 0.5 0.0 0 0
    channel operations paren 2,901,371 0.3 0.0 0 0
    checkpoint queue latch 99,329,722 0.0 0.0 0 11,153,369 0.1
    child cursor hash table 3,927,427 0.0 0.0 0 0
    commit callback allocati 8,739 0.0 0 0
    dictionary lookup 7,980 0.0 0 0
    dml lock allocation 6,767,990 0.1 0.0 0 0
    dummy allocation 1,898,183 0.2 0.1 0 0
    enqueue hash chains 27,741,348 0.1 0.1 4 0
    enqueues 17,450,161 0.3 0.1 6 0
    error message lists 132,828 2.6 0.2 1 0
    event group latch 884,066 0.0 0.7 0 0
    event range base latch 1 0.0 0 0
    file number translation 34 38.2 0.9 0 0
    global tx hash mapping 577,859 0.0 0 0
    hash table column usage 4,062 0.0 0 8,757,234 0.0
    hash table modification 16 0.0 0 2 0.0
    i/o slave adaptor 0 0 2 0.0
    job workq parent latch 4 100.0 0.3 0 494 8.7
    job_queue_processes para 1,950 0.0 0 2 0.0
    ksfv messages 0 0 4 0.0
    ktm global data 8,219 0.0 0 0
    lgwr LWN SCN 2,687,862 0.0 0.0 0 0
    library cache 310,882,781 0.9 0.0 34 104,759 4.0
    library cache load lock 30,369 0.0 0.3 0 0
    library cache pin 153,821,358 0.1 0.0 2 0
    library cache pin alloca 126,316,296 0.1 0.0 4 0
    list of block allocation 2,730,808 0.3 0.0 0 0
    loader state object free 566,036 0.1 0.0 0 0
    longop free list parent 197,368 0.0 0 8,390 0.0
    message pool operations 14,424 0.0 0.0 0 0
    messages 25,931,764 0.1 0.0 1 0
    mostly latch-free SCN 40,124,948 0.3 0.0 5 0
    Latch Sleep breakdown for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> ordered by misses desc
    Get Spin &
    Latch Name Requests Misses Sleeps Sleeps 1->4
    cache buffers chains ############## 74,770,083 1,062,119 73803903/884
    159/71439/10
    582/0
    redo allocation 170,107,983 3,441,055 149,631 3292872/1467
    48/1426/9/0
    library cache 310,882,781 2,831,747 89,240 2754499/6780
    6/7405/2037/
    0
    shared pool 158,471,190 1,755,922 55,268 1704342/4836
    9/2826/385/0
    cas latch 538,804,153 1,553,992 6,927 1547125/6808
    /58/1/0
    row cache objects 161,142,207 1,176,998 27,658 1154070/1952
    0/2560/848/0
    process queue reference 1,893,917,184 1,119,215 106,454 78758/4351/1
    36/0/0
    Library Cache Activity for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    ->"Pct Misses" should be very low
    Get Pct Pin Pct Invali-
    Namespace Requests Miss Requests Miss Reloads dations
    BODY 3,137,721 0.0 3,137,722 0.0 0 0
    CLUSTER 6,741 0.1 4,420 0.2 0 0
    INDEX 353,708 0.8 361,065 1.2 0 0
    SQL AREA 17,052,073 0.3 54,615,678 0.9 410,682 19,628
    TABLE/PROCEDURE 3,521,884 0.2 12,922,737 0.1 619 0
    TRIGGER 1,975,977 0.0 1,975,977 0.0 1 0
    SGA Memory Summary for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    SGA regions Size in Bytes
    Database Buffers 22,330,474,496
    Fixed Size 779,288
    Redo Buffers 17,051,648
    Variable Size 7,180,648,448
    sum 29,528,953,880

  • Missing "dtend" entries on .ics files for a published calendar

    From the top... I am using iCal Version 2.0.5 (1069), Mac OS X 10.4.10, and The Missing Sync for Palm OS v6.0.1; also the MarkSpace Events and Tasks conduits to sync my Palm with iCal via iSync.
    My goal is to publish a grouped calendar to icalx.com (which I've used for years).
    When I publish the calendar, everything appears to transfer fine, and I can see the published calendars on icalx.com.
    HOWEVER, most events in the graphical views (such as day or week) appear to have lost their end times, and thus stretch to 11:59 PM that day (which would be the default).
    I have examined the .ics file that icalx.com is receiving from my local iCal app, and it appears that on many of the events there is no entry for DTEND (the date and time when an event ends). There is an entry for DTSTART. I experimented with changing the end time for an event in iCal, and then refreshing the published calendar. The result is that DTEND is now defined, and the event shows up properly. Simply entering the same event ending time in the end time field does not result in a proper ending time on the icalx.com calendar. Changing the end time in iCal by dragging the lower boundary of the event box DOES modify the time properly on the icalx.com calendar.
    What is perplexing is that the proper start and end times ARE shown in iCal, and on my Palm.
    Has this kind of behavior been seen by others? Any hints on how to correct it, or where to start?
    Thanks everyone for your help.

    OK, previously I went to the online version and clicked on the "Download" link in the left column, but I did not know if I re-published from the downloaded calender, would it replace/over-write the existing calendar or create a new.1.ics file.
    I think I had tried this before, and the version I was updating, did not seem to update the version others had subscribed to, even though, I thought, I had re-published the same calendar.
    Anyway, I have tested the "new" re-published calendars and they seem to be updating for the subscribers too.
    Is there a way to make sure that I don't have any orphaned calendars? Some of the calendars have changed names and I was wondering if I am able to what is exactly sitting on my ME account. I can't find any of the calendars when I log into the account.
    Any thoughts?
    Thanks
    Gary

  • Entry for host name is missing or empty (Secure Store key: admin/host/).

    Hi All,
    I am in the process of installing EP7 on SLES9 and MaxDB 7.6 on a clustered HA environment where the SCS, DB and JC components are installed under and referenced by their virtual hostname (eg. sapinst SAPINST_USE_HOSTNAME <virt name>).
    The SCS instance is installed, the Database Instance is installed and I'm currently in the process of Installing the JC and Deploying the EP packages. 
    During the installation, SDM is put into standalone mode and sapinst tries to deploy the package EPBC06_0.SCA where it fails and the sdm_server.err shows :
    "com.sap.sdm.serverext.servertype.inqmy.TargetEngineConfigurationException: Cannot retrieve Engine logon data from Secure Store: Entry for host name is missing or empty (Secure Store key: admin/host/).        at com.sap.sdm.serverext.servertype.inqmy.InQMyTargetSystemConfigurator.getLoginDataFromSecStoreInQMyTargetSystemConfigurator.java:286)        at com.sap.sdm.serverext.servertype.inqmy.InQMyTargetSystemConfigurator.getEngineLoginDataInQMyTargetSystemConfigurator.java:197)        at com.sap.sdm.serverext.servertype.inqmy.EngineStateServiceImpl.determineCurrentEngineStateEngineStateServiceImpl.java:53)        at com.sap.sdm.app.proc.deployment.states.State.determineJ2EEEngineState(State.java:97)        at com.sap.sdm.app.proc.deployment.states.StateBeforeFirstDeployment.getJ2EEEngineStateChangeDescriptionStateBeforeFirstDeployment.java:75)        at com.sap.sdm.app.view.proc.deployment.mapper.StateMapper.map(StateMapper.java:56)        at com.sap.sdm.app.view.proc.deployment.mapper.FlowMapper.map(FlowMapper.java:28)        at com.sap.sdm.app.view.session.AppViewUpdater.updateAppView(AppViewUpdater.java:22)        at com.sap.sdm.app.view.session.AppViewUpdater.sessionStateChanged(AppViewUpdater.java:18)        at com.sap.sdm.app.proc.deployment.impl.SessionStateObserversImpl.notifySessionStateChangedSessionStateObserversImpl.java:46)        at com.sap.sdm.app.proc.deployment.states.InstContext.processEventServerSide(InstContext.java:85)        at com.sap.sdm.app.proc.deployment.states.InstContext.processEvent(InstContext.java:59)        at com.sap.sdm.app.view.controllers.DeployEventProcessor.processEvent(DeployEventProcessor.java:11)        at com.sap.sdm.client_server.deployserver.DeployCmdProcessor.processEvent(DeployCmdProcessor.java:264)        at com.sap.sdm.client_server.deployserver.DeployCmdProcessor.process(DeployCmdProcessor.java:108)        at com.sap.sdm.gui.server.GuiAdminRoleCmdProcessor.processGuiAdminRoleCmdProcessor.java:72)        at com.sap.sdm.is.cs.session.server.SessionCmdProcessor.process(SessionCmdProcessor.java:67)        at com.sap.sdm.is.cs.cmd.server.CmdServer.execCommandCmdServer.java:76)        at com.sap.sdm.client_server.launch.ServerLauncher$ConnectionHandlerImpl.handle(ServerLauncher.java:286)        at com.sap.sdm.is.cs.ncserver.NetCommServer.serveNetCommServer.java:43)        at com.sap.sdm.is.cs.ncwrapper.impl.ServiceWrapper.serveServiceWrapper.java:39)        at com.sap.bc.cts.tp.net.Worker.run(Worker.java:50)       at java.lang.Thread.run(Thread.java:816)"
    The same problem occurs if I try to deploy this package or any other package using SDM in standalone or integrated mode. 
    configtool -> Secure Store -> admin/host/<SID> is the virtual hostname of the JC instance.
    I've tried the following to no avail:
    - Changed configtool -> Secure Store -> admin/host/<SID> to be the physical hostname of the box.
    - Changed configtool -> Server -> Dispatcher -> Service -> P4 -> bindhost (from 0.0.0.0 to JC virt IP).
    - Changed the <Host> entry in sdmrepository.sdc from <phys hostname> to <jc virt hostname>.
    The SAPLOCALHOST, SAPGLOBALHOST all seem to be set correctly in the SAP profiles.
    Has anyone had this issue?  If so, what do I need to check/change?  Any ideas?
    Notice the message does not reference the <SID> in "admin/host", could this be related to the issue?  If so, what can cause this?
    TIA
    Anthony

    Hi All,
    Just an update, I was able to get around my problems by patching the 2004s media from sp6 to sp8, uninstalling my scs, and database and re-installing from the patched media.

  • Pass data from configuration entry point

    I have a need to post data from a Configuration Entry Point during a Test UUTs execution. I would like to update a parameter in my test sequence, using the configuration entrypoint if the test sequence is actively running in Test UUts. If it's not active, I would update a .ini file which gets read by the process model when a new Test UUTs entrypoint is started.
    My first guess has been to determine if the TestUUTs sequence is active, and I have tried using the PropertyObject RunState.ProcessModelClient.IsExecuting, but I just realized the Configuration Entry Point is always running in the process model.
    Everything I have read indicates that most Configuration Entry Points just write data to a file. Am I doing something fundamentally wrong?
    How do i determine if my sequence file is being executed with TestUUts?
    Thanks for your help. Woefully confused,
    Brian

    hi bce,
    [ If I used a station global, how would my changes be saved? Can you do this programmatically? Is there a better way? ]
    When Teststand closes, it automatically saves the StationGlobals. You would use PropertyObjectFile.SaveFileIfModified if you needed to force the save. But you shouldn't really need to do this as the latest value is available to all sequences and sequence files.
    Regards
    Ray Farmer
    Regards
    Ray Farmer

Maybe you are looking for

  • Skype not load correctly in my win 8..

    hallo everyone ihave problem with my Skype application.. after installing in my win 8.. Skype not load correctly in my win 8.. it display  like this.. anyone know why this happens?? Itry to sign in in there, but it just not respond..

  • Posting in 13-16 period

    Hi, Can anyone please tell me with examples the entries which we do in 13-16 periods ?? Regards.

  • Add ADF UI component at runtime

    Hi , I have created Ui components in my page at runtime based on the input. I have a panel at design time. To that I am adding decoration boxes at run time, also, i am adding InputTextBoxes to the Decoration Box at runtime. for(int i=0;i<noOfDecorati

  • Roaming between mesh and non mesh and mesh AP design

    Hi There, I need some verification on the following: - Can client roam between Mesh AP 1550 series and Non-Mesh AP? - I have 1550 series AP to be installed and all the mounting area are with data point ready, which connects back to switch. Would conf

  • Sharing Views

    Hi, I am just sharing my view and want to know with your views is that possible or not. "I have two cases, if I want to delete ORGANIZATION_ID and ORG_ID based data from EBS R12" Will following queries work fine?. -- 1st DECLARE CURSOR c1 IS SELECT a