Solution on Data Handling

Hi All,
I am lookin for some idea of modelling on BW.
The requirement is to have the information on weekly basis (Current Week or History week)
Now, the solution was to copy unchanged data as well to the current week.
This was helping to have a week basis snapshot in the cube.
But off late we got to know that the data is huge and copying will eat time
Is there any way we could avoid copying?
Hope I am clear.
Thanks In Advance

I've managed to override the dispatchEvent method when
extending the HTTPService, and it looks like that could be the way
forward. At the moment I've got it checking to see if the event
it's dispatching is a result event and if so then I'm doing my
stuff and changing that event to my custom one.
Question now is, there will be many of these custom
HTTPServices allover my application and I want to be able to dump
any CakeDebug I get in the response to a central location in the
application --- what would be the best way to do that (without
passing round the CakeDebugManager -- or whatever I create). I
can't have the CakeDebugManager add a listener to every HTTPService
as I'll need a reference to the HTTPService and that will break my
encapsulation.
As I say, I'm still new to Flex.

Similar Messages

  • Plan data handling profile

    Dear Gurus,
    There is a customzing point under Project System -> Costs -> Planned Costs -> Maintain plan data handling profile.
    Does anyone have any idea about:
    - where can it be used?
    - what is it used for?
    I was looking for any tool which makes plan figures on WBS elements out of purchase orders and I found this but there is no info at all, no where.
    Thanks in advance,
    Jeno

    This profile is aimed to be used in transaction AD31 (and associated Tcode AD32). Those belongs to an add-on of IS A&D solution.
    The purpose is to provide a functionality to calculate the plan cost at completion by adding the actuals costs (including or not the commitments form PO and/or PR) to the cost to complete entered in a specific planning version.
    This functionality can also be used to calculate the cost to complete by calculating the difference between the plan costs at completion stored on a dedicated planning version, and actuals costs (including or not the commitments form PO and/or PR).
    As SAP provides, independently, the possibility to copy actual to plan, the profile gives the flexibility to calculate the cost to complete by adding 2 planning versions.

  • Data handling issue

    The problem is that my application receives sterilized java object in zip format from third party system then my application can update into DB2 tables in a given time frame (max 1 min). This results into a buffer overflows which implies that data are discarded due to buffer size limit. Every min, my application receives the data from third party.
    At present, application data handling is implemented in queuing mechanism. Application expected to handle 100000 java objects per min. Do you have any idea to implement the solution for above issue.
    At present I have an idea to implement virtual table (array list) and keep the data into memory..use it after 1 min. required your valuable input.

    831403 wrote:
    The problem is that my application receives sterilized java object in zip format from third party system then my application can update into DB2 tables in a given time frame (max 1 min). This results into a buffer overflows which implies that data are discarded due to buffer size limit. Every min, my application receives the data from third party.
    At present, application data handling is implemented in queuing mechanism. Application expected to handle 100000 java objects per min. Do you have any idea to implement the solution for above issue.
    At present I have an idea to implement virtual table (array list) and keep the data into memory..use it after 1 min. required your valuable input.My valuable input: Re-read your own question and try again. What do you actually want? What you've written is totally incomprehensible.
    Question: What is a "sterilized" object? I'm familiar with several euphemisms, but that's a new one on me.
    This for nothing:
    1. Simply reading 100,000 objects could take a reasonable percentage of your 1 minute, let alone marshalling and then sending them to a database for update. This suggests to me that your solution isn't scalable.
    2. Buffer overflows (if that's indeed what it is; you haven't supplied an exception) are a result of bad programming. Find the bug, and fix it.
    Winston

  • Data handling tool

    plz enlist  data handling tools in labview?

    What do you mean by data handling?  What exactly are you trying to do?
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Receiving a DNS Solution Yahoo Error Handler Page and some pages won't display

    I recently downloaded the Stuffit Expander and ever since then Firefox hasn't been the same. I receive a DNS Solution Yahoo Error Handler page when I do a search sometimes. And now some pages won't even display for me when they should. How do I get rid of this Yahoo thing (In simple terms please)? Thanks!

    Thank you jscher2000! This seems to have fixed it! I appreciate it :)

  • Mass data handling in an Enterprise SOA

    Hello together,
    i'm working on my diploma thesis with the title "Design of a service-oriented architecture based on SAP NetWeaver".
    For this work an enterprise service is designed to calculate a special price matrix (10x5) from a given price for a product.
    Additionally the matrix for a product can have four variations, depending from input values. I think for few products there's no problem.
    But it's also possible that someone wants to export the price matrices for all 15.000 products. So there are 60.000 (4 x 15.000) service requests at once.
    So my questions:
    Is it (without bigger problems) possible for a service to handle so many requests at one time? Or generally, how behaves enterprise SOA in cases of mass data requests like that (network traffic, workload, overhead for enterprise services...)?
    An alternative is to storage all matrices in a DB (would be much redundancy, because variations of the matrices are very similar).
    Thanks for any answers
    Joschi

    Having the same service to get one product price as well as all product prices would not be recommended. Two separate services would be more reasonable
    Mass data handling in SOA is not entirely different than let's say in BAPI/RFC environment. parallelism, multi-threading, locking, commit and many others are typical elements of considerations.
    From a scalability standpoint, you can always be interested in how optimal its performance is whether you call it once or a million times. If the service is a rather slim one (lightweight in terms of number and types of parameters as well as processing code), you'll surely have better chance than with a service that's heavyweight.

  • Screen data handling in Screen Enhancement with BAdI's ?

    Hi all,
    I'm trying to build a Screen Enhancement with (new) BAdI concept.
    But now I'm having a Problem. The new BAdI's don't allow variable Attributes  (instance/static attribute) in the BAdI Interface.
    If I try to define an instance or static Attribute for the BAdI-Interface I get the following error message:
    "BAdI interface <my_badi_if> contains non-constant attributes "
    In my Screen Enhancement I have build a Subscreen with new fields in it. These field data couldn't be saved in the BAdI class (e.g. via BAdI-methods my_badi->get_data_from_screen... and my_badi->put_data_on_screen...) because having no attributes.
    How these screen-data could be handled without BAdI-Interface-Attributes?
    Thanks in advance,
    Anja

    Having the same service to get one product price as well as all product prices would not be recommended. Two separate services would be more reasonable
    Mass data handling in SOA is not entirely different than let's say in BAPI/RFC environment. parallelism, multi-threading, locking, commit and many others are typical elements of considerations.
    From a scalability standpoint, you can always be interested in how optimal its performance is whether you call it once or a million times. If the service is a rather slim one (lightweight in terms of number and types of parameters as well as processing code), you'll surely have better chance than with a service that's heavyweight.

  • Error -2003: a data handler needed by the movie could not be found

    Hello.
    I just purchased QuickTime Pro and the MPEG-2 add-on for it, for the single purpose of transcoding some MPEG-2 recordings to H.264. However, when I try to open any of my MPEG-2 files in Quicktime, I get the message: "error -2003: a data handler needed by the movie could not be found (my-filename.mpg)"
    I can play these files in a number of different free players--VLC, Mplayer, Windows Media Player, RealPlayer... but not in the player I paid $50 for.
    Can someone please help me?

    That doesn't really help. Besides, I haven't even got to the point where I'm trying to export. I get this error message when I try to open the file. I can't even play back the clip.
    And the MPEG-2 page clearly states: "The QuickTime MPEG-2 Playback Component provides QuickTime users with the ability to import and play back MPEG-2 content, including both multiplexed (a.k.a. muxed, where the audio and video tracks are interleaved together into one track) and non-multiplexed (a.k.a. elementary) streams."
    Also, Streamclip crashes when I try to load my MPEG-2 clips, so that's not going to work (and even if it did, I am trying to find a simple way to take the HDTV content I've recorded off the air and get it onto my AppleTV--I want fewer steps, not more).
    Looks like I'll have to waste some time calling customer service to get a refund, since there's no way to do this via the website.

  • How is table control data handled in Session Method?

    Hi Friends,
    Can any1 plz tell me how is table control data handled in Session Method.
    In Call tran we do it by count. But how in Session Method?
    Thank you.
    Regards,
    Varun.

    Hi,
    u can do it by both methods. u can check out this program. u need to comment the performs of table control fields and write ur own perform statements.And u have to declare the table control fields as separate internal tables.
    report zcustomer_change_bict
           no standard page heading line-size 255.
    include bdcrecx1.
    data: begin of it_record occurs 0,
            kunnr like rf02d-kunnr,
            d0360 like rf02d-d0360,
            anred(30) type c,
            name1(35) type c,
            namev(35) type c,
            telf1(35) type c,
            abtnr(10) type c,
          end of it_record.
    data: begin of it_knvk1 occurs 0,
          anred like knvk-anred,
          end of it_knvk1.
    data: begin of it_knvk2 occurs 0,
          name1 like knvk-name1,
          end of it_knvk2.
    data: begin of it_knvk3 occurs 0,
          namev like knvk-namev,
          end of it_knvk3.
    data: begin of it_knvk4 occurs 0,
          telf1 like knvk-telf1,
          end of it_knvk4.
    data: begin of it_knvk5 occurs 0,
          abtnr like knvk-abtnr,
          end of it_knvk5.
    data : fld(20) type c,
           cnt(2) type n.
    start-of-selection.
      call function 'GUI_UPLOAD'
        exporting
          filename                      = 'C:\CUSTCH.TXT'
         filetype                      = 'ASC'
        has_field_separator            = 'X'
        tables
          data_tab                      = it_record
    exceptions
       file_open_error               = 1
       file_read_error               = 2
       no_batch                      = 3
       gui_refuse_filetransfer       = 4
       invalid_type                  = 5
       no_authority                  = 6
       unknown_error                 = 7
       bad_data_format               = 8
       header_not_allowed            = 9
       separator_not_allowed         = 10
       header_too_long               = 11
       unknown_dp_error              = 12
       access_denied                 = 13
       dp_out_of_memory              = 14
       disk_full                     = 15
       dp_timeout                    = 16
       others                        = 17
      if sy-subrc <> 0.
        message id sy-msgid type sy-msgty number sy-msgno
                with sy-msgv1 sy-msgv2 sy-msgv3 sy-msgv4.
      endif.
      perform open_group.
      loop at it_record.
        refresh bdcdata.
        refresh: it_knvk1,it_knvk2,it_knvk3,it_knvk4,it_knvk5.
        split it_record-anred at ',' into table it_knvk1.
        split it_record-name1 at ',' into table it_knvk2.
        split it_record-namev at ',' into table it_knvk3.
        split it_record-telf1 at ',' into table it_knvk4.
        split it_record-abtnr at ',' into table it_knvk5.
    screen 101
        perform bdc_dynpro      using 'SAPMF02D' '0101'.
        perform bdc_field       using 'BDC_CURSOR'
                                      'RF02D-D0360'.
        perform bdc_field       using 'BDC_OKCODE'
                                      '/00'.
        perform bdc_field       using 'RF02D-KUNNR'
                                      it_record-kunnr.
        perform bdc_field       using 'RF02D-D0360'
                                      it_record-d0360.
    screen 360
        perform bdc_dynpro      using 'SAPMF02D' '0360'.
        perform bdc_field       using 'BDC_CURSOR'
                                      'KNVK-ABTNR(01)'.
        perform bdc_field       using 'BDC_OKCODE'
                                      '=ENTR'.
    *perform bdc_field       using 'KNVK-ANRED(01)'
                                 record-ANRED_01_003.
    *perform bdc_field       using 'KNVK-NAMEV(01)'
                                 record-NAMEV_01_004.
    *perform bdc_field       using 'KNVK-NAME1(01)'
                                 record-NAME1_01_005.
    *perform bdc_field       using 'KNVK-TELF1(01)'
                                 record-TELF1_01_006.
    *perform bdc_field       using 'KNVK-ABTNR(01)'
                                 record-ABTNR_01_007.
        move 1 to cnt.
        loop at it_knvk1 .
          concatenate 'knvk-anred( ' cnt ' ) ' into fld.
          perform bdc_field using fld it_knvk1-anred.
          cnt = cnt + 1.
        endloop.
        move 1 to cnt.
        loop at it_knvk2 .
          concatenate 'knvk-name1( ' cnt ' ) ' into fld.
          perform bdc_field using fld it_knvk2-name1.
          cnt = cnt + 1.
        endloop.
        move 1 to cnt.
        loop at it_knvk3 .
          concatenate 'knvk-namev( ' cnt ' ) ' into fld.
          perform bdc_field using fld it_knvk3-namev.
          cnt = cnt + 1.
        endloop.
        move 1 to cnt.
        loop at it_knvk4 .
          concatenate 'knvk-telf1( ' cnt ' ) ' into fld.
          perform bdc_field using fld it_knvk4-telf1.
          cnt = cnt + 1.
        endloop.
        move 1 to cnt.
        loop at it_knvk5 .
          concatenate 'knvk-abtnr( ' cnt ' ) ' into fld.
          perform bdc_field using fld it_knvk5-abtnr.
          cnt = cnt + 1.
        endloop.
    screen 360
        perform bdc_dynpro      using 'SAPMF02D' '0360'.
        perform bdc_field       using 'BDC_CURSOR'
                                      'KNVK-NAMEV(01)'.
        perform bdc_field       using 'BDC_OKCODE'
                                      '=ENTR'.
        perform bdc_transaction using 'XD02'.
      endloop.
      perform close_group.

  • SRM7.0 Meta Data Handling or Component Configuration

    Hi Experts!
    Let's say you for some reaseon want to remove the a field in the Shopping Cart at item level. To achieve this you could e.g.
    1. Use the Component Configurator (SE80) - and use the browser-based Editor for the WDP ABAP Component Configurator
    2. Use the Meta Data Handling Concept and configure the field control there
    My guess is that one should use #2 if one need to control the behaviour of the field according to roles or authorities, but is this correct, or are there other considerations that should be taken into account when performing UI adjustments using the above mentioned methods?
    brgs Ziggy

    Check the below link, this should help you.
    http://wiki.sdn.sap.com/wiki/display/SRM/HowtoaddCustomerfieldsinSAP+SRM
    Regards,
    Jagadish

  • US to UK Date Handling

    Hi all,
    Is it possible to set up date handling in Project Siena so it recognises Date input as dd/mm/yy instead of the US mm/dd/yy.

    just format a date in whatever format you like with Text function
    Example
    Text(Today(),"dd/mm/yy")               03/11/14
    Text(Today(),"dd/mmmm/yy")          03/November/14
    Check this page out for all function
    http://technet.microsoft.com/library/dn690139
    And for date inputs is much more better to create own date picker. Here is a link to one of them
    https://onedrive.live.com/redir?resid=435EB1CB289C9DE5!21363&authkey=!AH6jXP3fUNUux6c&ithint=file%2csiena
     

  • What are the best solutions for data warehouse configuration in 10gR2

    I need help on solutions to be provided to my Client for upgrading the data warehouse.
    Current Configuration: Oracle database 9.2.0.8. This database contains the data warehouse and one more data mart on the same host.Sizes are respectively 6 Terabyte(retention policy of 3 years+current year) and 1 Terabyte. The ETL tool and BO Reporting tools are also hosted on the same host. This current configuration is really performing poor.
    Client cannot go for a major architectural or configuration changes to its existing environment now due to some constraints.
    However, they have agreed to separate out the databases on separate hosts from the ETL tools and BO objects. Also we are planning to upgrade the database to 10gR2 to attain stability, better performance and overcome current headaches.
    We cannot upgrade the database to 11g as the BO is at a version 6.5 which isn't compatible with Oracle 11g. And Client cannot afford to upgrade anything else other than the database.
    So, my role is very vital in providing a perfect solution towards better performance and take a successful migration of Oracle Database from one host to another (similar platform and OS) in addition to upgrade.
    I have till now thought of the following:
    Move the Oracle database and data mart to separate host.
    The host will be the same platform, that is, HP Superdome with HP-UX 32-bit OS (we cannot change to 64-bit as ETL tool doesn't support)
    Install new Oracle database 10g on the new host and move the data to it.
    Exploring all new features of 10gR2 to help data warehouse, that is, SQL MODEL Clause introduction, Parallel processing, Partitioning, Data Pump, SPA to study pre and post migrations.
    Also thinking of RAC to provide more better solution as our main motive is to show a tremendous performance enhancement.
    I need all your help to prepare a good road map for my assignment. Please suggest.
    Thanks,
    Tapan

    SGA=27.5 GB and PGA=50 MB
    Also I am pasting part of STATSPACK Report, eliminating the snaps of DB bounce. Please suggest the scope of improvement in this case.
    STATSPACK report for
    Snap Id Snap Time Sessions Curs/Sess Comment
    Begin Snap: 582946 11-Mar-13 20:02:16 46 12.8
    End Snap: 583036 12-Mar-13 18:24:24 60 118.9
    Elapsed: 1,342.13 (mins)
    Cache Sizes (end)
    ~~~~~~~~~~~~~~~~~
    Buffer Cache: 21,296M Std Block Size: 16K
    Shared Pool Size: 6,144M Log Buffer: 16,384K
    Load Profile
    ~~~~~~~~~~~~ Per Second Per Transaction
    Redo size: 1,343,739.01 139,883.39
    Logical reads: 100,102.54 10,420.69
    Block changes: 3,757.42 391.15
    Physical reads: 6,670.84 694.44
    Physical writes: 874.34 91.02
    User calls: 1,986.04 206.75
    Parses: 247.87 25.80
    Hard parses: 5.82 0.61
    Sorts: 1,566.76 163.10
    Logons: 10.99 1.14
    Executes: 1,309.79 136.35
    Transactions: 9.61
    % Blocks changed per Read: 3.75 Recursive Call %: 43.34
    Rollback per transaction %: 3.49 Rows per Sort: 190.61
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer Nowait %: 99.90 Redo NoWait %: 100.00
    Buffer Hit %: 96.97 In-memory Sort %: 100.00
    Library Hit %: 99.27 Soft Parse %: 97.65
    Execute to Parse %: 81.08 Latch Hit %: 99.58
    Parse CPU to Parse Elapsd %: 3.85 % Non-Parse CPU: 99.34
    Shared Pool Statistics Begin End
    Memory Usage %: 7.11 50.37
    % SQL with executions>1: 62.31 46.46
    % Memory for SQL w/exec>1: 26.75 13.47
    Top 5 Timed Events
    ~~~~~~~~~~~~~~~~~~ % Total
    Event Waits Time (s) Ela Time
    CPU time 492,062 43.66
    db file sequential read 157,418,414 343,549 30.49
    library cache pin 92,339 66,759 5.92
    PX qref latch 63,635 43,845 3.89
    db file scattered read 2,506,806 41,677 3.70
    Background Wait Events for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> ordered by wait time desc, waits desc (idle events last)
    Avg
    Total Wait wait Waits
    Event Waits Timeouts Time (s) (ms) /txn
    log file sequential read 176,386 0 3,793 22 0.2
    log file parallel write 2,685,833 0 1,813 1 3.5
    db file parallel write 239,166 0 1,350 6 0.3
    control file parallel write 33,432 0 79 2 0.0
    LGWR wait for redo copy 478,120 536 75 0 0.6
    rdbms ipc reply 10,027 0 47 5 0.0
    control file sequential read 32,414 0 40 1 0.0
    db file scattered read 4,101 0 30 7 0.0
    db file sequential read 13,946 0 29 2 0.0
    direct path read 203,694 0 14 0 0.3
    log buffer space 363 0 13 37 0.0
    latch free 3,766 0 9 2 0.0
    direct path write 80,491 0 6 0 0.1
    async disk IO 351,955 0 4 0 0.5
    enqueue 28 0 1 21 0.0
    buffer busy waits 1,281 0 1 0 0.0
    log file single write 172 0 0 1 0.0
    rdbms ipc message 10,563,204 251,286 992,837 94 13.7
    pmon timer 34,751 34,736 78,600 2262 0.0
    smon timer 7,462 113 76,463 10247 0.0
    Instance Activity Stats for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    Statistic Total per Second per Trans
    CPU used by this session 49,206,154 611.0 63.6
    CPU used when call started 49,435,735 613.9 63.9
    CR blocks created 6,740,777 83.7 8.7
    Cached Commit SCN referenced 423,253,503 5,256.0 547.2
    Commit SCN cached 19,165 0.2 0.0
    DBWR buffers scanned 48,276,489 599.5 62.4
    DBWR checkpoint buffers written 6,959,752 86.4 9.0
    DBWR checkpoints 454 0.0 0.0
    DBWR free buffers found 44,817,183 556.5 57.9
    DBWR lru scans 137,149 1.7 0.2
    DBWR make free requests 162,528 2.0 0.2
    DBWR revisited being-written buff 4,220 0.1 0.0
    DBWR summed scan depth 48,276,489 599.5 62.4
    DBWR transaction table writes 5,036 0.1 0.0
    DBWR undo block writes 2,989,436 37.1 3.9
    DDL statements parallelized 3,723 0.1 0.0
    DFO trees parallelized 4,157 0.1 0.0
    DML statements parallelized 3 0.0 0.0
    OS Block input operations 29,850 0.4 0.0
    OS Block output operations 1,591 0.0 0.0
    OS Characters read/written 182,109,814,791 2,261,447.1 235,416.9
    OS Integral unshared data size ################## 242,463,432.4 ############
    OS Involuntary context switches 188,257,786 2,337.8 243.4
    OS Maximum resident set size 43,518,730,619 540,417.4 56,257.5
    OS Page reclaims 159,430,953 1,979.8 206.1
    OS Signals received 5,260,938 65.3 6.8
    OS Socket messages received 79,438,383 986.5 102.7
    OS Socket messages sent 93,064,176 1,155.7 120.3
    OS System time used 10,936,430 135.8 14.1
    OS User time used 132,043,884 1,639.7 170.7
    OS Voluntary context switches 746,207,739 9,266.4 964.6
    PX local messages recv'd 55,120,663 684.5 71.3
    PX local messages sent 55,120,817 684.5 71.3
    Parallel operations downgraded 1 3 0.0 0.0
    Parallel operations not downgrade 4,154 0.1 0.0
    SQL*Net roundtrips to/from client 155,422,335 1,930.0 200.9
    SQL*Net roundtrips to/from dblink 18 0.0 0.0
    active txn count during cleanout 16,529,551 205.3 21.4
    background checkpoints completed 43 0.0 0.0
    background checkpoints started 43 0.0 0.0
    background timeouts 280,202 3.5 0.4
    branch node splits 4,428 0.1 0.0
    buffer is not pinned count 6,382,440,322 79,257.4 8,250.7
    buffer is pinned count 9,675,661,370 120,152.8 12,507.9
    bytes received via SQL*Net from c 67,384,496,376 836,783.4 87,109.3
    bytes received via SQL*Net from d 6,142 0.1 0.0
    bytes sent via SQL*Net to client 50,240,643,657 623,890.4 64,947.1
    bytes sent via SQL*Net to dblink 3,701 0.1 0.0
    calls to get snapshot scn: kcmgss 145,385,064 1,805.4 187.9
    calls to kcmgas 36,816,132 457.2 47.6
    calls to kcmgcs 3,514,770 43.7 4.5
    change write time 369,373 4.6 0.5
    cleanout - number of ktugct calls 20,954,488 260.2 27.1
    cleanouts and rollbacks - consist 6,357,174 78.9 8.2
    cleanouts only - consistent read 10,078,802 125.2 13.0
    cluster key scan block gets 69,403,565 861.9 89.7
    Instance Activity Stats for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    Statistic Total per Second per Trans
    cluster key scans 41,311,211 513.0 53.4
    commit cleanout failures: block l 413,776 5.1 0.5
    commit cleanout failures: buffer 414 0.0 0.0
    commit cleanout failures: callbac 41,194 0.5 0.1
    commit cleanout failures: cannot 174,382 2.2 0.2
    commit cleanouts 11,469,056 142.4 14.8
    commit cleanouts successfully com 10,839,290 134.6 14.0
    commit txn count during cleanout 17,155,424 213.0 22.2
    consistent changes 145,418,277 1,805.8 188.0
    consistent gets 8,043,252,188 99,881.4 10,397.7
    consistent gets - examination 3,180,028,047 39,489.7 4,110.9
    current blocks converted for CR 9 0.0 0.0
    cursor authentications 14,926 0.2 0.0
    data blocks consistent reads - un 143,706,500 1,784.6 185.8
    db block changes 302,577,666 3,757.4 391.2
    db block gets 336,562,217 4,179.4 435.1
    deferred (CURRENT) block cleanout 2,912,793 36.2 3.8
    dirty buffers inspected 627,174 7.8 0.8
    enqueue conversions 1,296,337 16.1 1.7
    enqueue releases 13,053,200 162.1 16.9
    enqueue requests 13,239,092 164.4 17.1
    enqueue timeouts 185,878 2.3 0.2
    enqueue waits 114,120 1.4 0.2
    exchange deadlocks 7,390 0.1 0.0
    execute count 105,475,101 1,309.8 136.4
    free buffer inspected 1,604,407 19.9 2.1
    free buffer requested 258,126,047 3,205.4 333.7
    hot buffers moved to head of LRU 22,793,576 283.1 29.5
    immediate (CR) block cleanout app 16,436,010 204.1 21.3
    immediate (CURRENT) block cleanou 2,860,013 35.5 3.7
    index fast full scans (direct rea 12,375 0.2 0.0
    index fast full scans (full) 3,733 0.1 0.0
    index fast full scans (rowid rang 192,148 2.4 0.3
    index fetch by key 1,321,024,486 16,404.5 1,707.7
    index scans kdiixs1 406,165,684 5,043.8 525.1
    leaf node 90-10 splits 50,373 0.6 0.1
    leaf node splits 697,235 8.7 0.9
    logons cumulative 884,756 11.0 1.1
    messages received 3,276,719 40.7 4.2
    messages sent 3,257,171 40.5 4.2
    no buffer to keep pinned count 569 0.0 0.0
    no work - consistent read gets 4,406,092,172 54,715.0 5,695.8
    opened cursors cumulative 20,527,704 254.9 26.5
    parse count (failures) 267,088 3.3 0.4
    parse count (hard) 468,996 5.8 0.6
    parse count (total) 19,960,548 247.9 25.8
    parse time cpu 323,024 4.0 0.4
    parse time elapsed 8,393,422 104.2 10.9
    physical reads 537,189,332 6,670.8 694.4
    physical reads direct 292,545,140 3,632.8 378.2
    physical writes 70,409,002 874.3 91.0
    physical writes direct 59,248,394 735.8 76.6
    physical writes non checkpoint 69,103,391 858.1 89.3
    pinned buffers inspected 11,893 0.2 0.0
    prefetched blocks 95,892,161 1,190.8 124.0
    prefetched blocks aged out before 1,495,883 18.6 1.9
    Instance Activity Stats for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    Statistic Total per Second per Trans
    process last non-idle time ################## ############## ############
    queries parallelized 417 0.0 0.0
    recursive calls 122,323,299 1,519.0 158.1
    recursive cpu usage 3,144,533 39.1 4.1
    redo blocks written 180,881,558 2,246.2 233.8
    redo buffer allocation retries 5,400 0.1 0.0
    redo entries 164,728,513 2,045.6 213.0
    redo log space requests 1,006 0.0 0.0
    redo log space wait time 2,230 0.0 0.0
    redo ordering marks 2,563 0.0 0.0
    redo size 108,208,614,904 1,343,739.0 139,883.4
    redo synch time 558,520 6.9 0.7
    redo synch writes 2,343,824 29.1 3.0
    redo wastage 1,126,585,600 13,990.0 1,456.4
    redo write time 718,655 8.9 0.9
    redo writer latching time 7,763 0.1 0.0
    redo writes 2,685,833 33.4 3.5
    rollback changes - undo records a 522,742 6.5 0.7
    rollbacks only - consistent read 335,177 4.2 0.4
    rows fetched via callback 1,100,990,382 13,672.1 1,423.3
    session connect time ################## ############## ############
    session cursor cache count 1,061 0.0 0.0
    session cursor cache hits 1,687,796 21.0 2.2
    session logical reads 8,061,057,193 100,102.5 10,420.7
    session pga memory 1,573,228,913,832 19,536,421.0 2,033,743.8
    session pga memory max 1,841,357,626,496 22,866,054.4 2,380,359.0
    session uga memory 1,074,114,630,336 13,338,399.4 1,388,529.0
    session uga memory max 386,645,043,296 4,801,374.0 499,823.6
    shared hash latch upgrades - no w 410,360,146 5,095.9 530.5
    sorts (disk) 2,657 0.0 0.0
    sorts (memory) 126,165,625 1,566.7 163.1
    sorts (rows) 24,048,783,304 298,638.8 31,088.3
    summed dirty queue length 5,438,201 67.5 7.0
    switch current to new buffer 1,302,798 16.2 1.7
    table fetch by rowid 6,201,503,534 77,010.5 8,016.8
    table fetch continued row 26,649,697 330.9 34.5
    table scan blocks gotten 1,864,435,032 23,152.6 2,410.2
    table scan rows gotten 43,639,997,280 541,923.3 56,414.3
    table scans (cache partitions) 26,112 0.3 0.0
    table scans (direct read) 246,243 3.1 0.3
    table scans (long tables) 340,200 4.2 0.4
    table scans (rowid ranges) 359,617 4.5 0.5
    table scans (short tables) 9,111,559 113.2 11.8
    transaction rollbacks 4,819 0.1 0.0
    transaction tables consistent rea 824 0.0 0.0
    transaction tables consistent rea 1,386,848 17.2 1.8
    user calls 159,931,913 1,986.0 206.8
    user commits 746,543 9.3 1.0
    user rollbacks 27,020 0.3 0.0
    write clones created in backgroun 7 0.0 0.0
    write clones created in foregroun 4,350 0.1 0.0
    Buffer Pool Statistics for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> Standard block size Pools D: default, K: keep, R: recycle
    -> Default Pools for other block sizes: 2k, 4k, 8k, 16k, 32k
    Free Write Buffer
    Number of Cache Buffer Physical Physical Buffer Complete Busy
    P Buffers Hit % Gets Reads Writes Waits Waits Waits
    D 774,144 95.6############ 233,869,082 10,089,734 0 0########
    K 504,000 99.9############ 3,260,227 1,070,338 0 0 65,898
    R 63,504 96.2 196,079,539 7,511,863 535 0 0 0
    Buffer wait Statistics for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> ordered by wait time desc, waits desc
    Tot Wait Avg
    Class Waits Time (s) Time (ms)
    data block 7,791,121 14,676 2
    file header block 587 101 172
    undo header 151,617 71 0
    segment header 299,312 58 0
    1st level bmb 45,235 7 0
    bitmap index block 392 1 3
    undo block 4,250 1 0
    2nd level bmb 14 0 0
    system undo header 2 0 0
    3rd level bmb 1 0 0
    Latch Activity for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    ->"Get Requests", "Pct Get Miss" and "Avg Slps/Miss" are statistics for
    willing-to-wait latch get requests
    ->"NoWait Requests", "Pct NoWait Miss" are for no-wait latch get requests
    ->"Pct Misses" for both should be very close to 0.0
    Pct Avg Wait Pct
    Get Get Slps Time NoWait NoWait
    Latch Requests Miss /Miss (s) Requests Miss
    Consistent RBA 2,686,230 0.0 0.2 0 0
    FAL request queue 86 0.0 0 0
    FAL subheap alocation 0 0 2 0.0
    FIB s.o chain latch 1,089 0.0 0 0
    FOB s.o list latch 4,589,986 0.5 0.0 2 0
    NLS data objects 1 0.0 0 0
    SQL memory manager worka 5,963 0.0 0 0
    Token Manager 0 0 2 0.0
    active checkpoint queue 719,439 0.3 0.1 0 1 0.0
    alert log latch 184 0.0 0 2 0.0
    archive control 4,365 0.0 0 0
    archive process latch 1,808 0.6 0.6 0 0
    begin backup scn array 3,387,572 0.0 0.0 0 0
    cache buffer handles 1,577,222 0.2 0.0 0 0
    cache buffers chains ############## 0.5 0.0 430 354,357,972 0.3
    cache buffers lru chain 17,153,023 0.1 0.0 1 385,505,654 0.5
    cas latch 538,804,153 0.3 0.0 7 0
    channel handle pool latc 1,776,950 0.5 0.0 0 0
    channel operations paren 2,901,371 0.3 0.0 0 0
    checkpoint queue latch 99,329,722 0.0 0.0 0 11,153,369 0.1
    child cursor hash table 3,927,427 0.0 0.0 0 0
    commit callback allocati 8,739 0.0 0 0
    dictionary lookup 7,980 0.0 0 0
    dml lock allocation 6,767,990 0.1 0.0 0 0
    dummy allocation 1,898,183 0.2 0.1 0 0
    enqueue hash chains 27,741,348 0.1 0.1 4 0
    enqueues 17,450,161 0.3 0.1 6 0
    error message lists 132,828 2.6 0.2 1 0
    event group latch 884,066 0.0 0.7 0 0
    event range base latch 1 0.0 0 0
    file number translation 34 38.2 0.9 0 0
    global tx hash mapping 577,859 0.0 0 0
    hash table column usage 4,062 0.0 0 8,757,234 0.0
    hash table modification 16 0.0 0 2 0.0
    i/o slave adaptor 0 0 2 0.0
    job workq parent latch 4 100.0 0.3 0 494 8.7
    job_queue_processes para 1,950 0.0 0 2 0.0
    ksfv messages 0 0 4 0.0
    ktm global data 8,219 0.0 0 0
    lgwr LWN SCN 2,687,862 0.0 0.0 0 0
    library cache 310,882,781 0.9 0.0 34 104,759 4.0
    library cache load lock 30,369 0.0 0.3 0 0
    library cache pin 153,821,358 0.1 0.0 2 0
    library cache pin alloca 126,316,296 0.1 0.0 4 0
    list of block allocation 2,730,808 0.3 0.0 0 0
    loader state object free 566,036 0.1 0.0 0 0
    longop free list parent 197,368 0.0 0 8,390 0.0
    message pool operations 14,424 0.0 0.0 0 0
    messages 25,931,764 0.1 0.0 1 0
    mostly latch-free SCN 40,124,948 0.3 0.0 5 0
    Latch Sleep breakdown for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    -> ordered by misses desc
    Get Spin &
    Latch Name Requests Misses Sleeps Sleeps 1->4
    cache buffers chains ############## 74,770,083 1,062,119 73803903/884
    159/71439/10
    582/0
    redo allocation 170,107,983 3,441,055 149,631 3292872/1467
    48/1426/9/0
    library cache 310,882,781 2,831,747 89,240 2754499/6780
    6/7405/2037/
    0
    shared pool 158,471,190 1,755,922 55,268 1704342/4836
    9/2826/385/0
    cas latch 538,804,153 1,553,992 6,927 1547125/6808
    /58/1/0
    row cache objects 161,142,207 1,176,998 27,658 1154070/1952
    0/2560/848/0
    process queue reference 1,893,917,184 1,119,215 106,454 78758/4351/1
    36/0/0
    Library Cache Activity for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    ->"Pct Misses" should be very low
    Get Pct Pin Pct Invali-
    Namespace Requests Miss Requests Miss Reloads dations
    BODY 3,137,721 0.0 3,137,722 0.0 0 0
    CLUSTER 6,741 0.1 4,420 0.2 0 0
    INDEX 353,708 0.8 361,065 1.2 0 0
    SQL AREA 17,052,073 0.3 54,615,678 0.9 410,682 19,628
    TABLE/PROCEDURE 3,521,884 0.2 12,922,737 0.1 619 0
    TRIGGER 1,975,977 0.0 1,975,977 0.0 1 0
    SGA Memory Summary for DB: P7IN1 Instance: P7IN1 Snaps: 582946 -583036
    SGA regions Size in Bytes
    Database Buffers 22,330,474,496
    Fixed Size 779,288
    Redo Buffers 17,051,648
    Variable Size 7,180,648,448
    sum 29,528,953,880

  • Data Handler source code need modification

    I found this in the forum and try to implement it in my page. The code is running okay but just that when I first load the page into the browser, It shows ALL the records in the database beneath the searchbox.
    All I want is first time user come to the page, it only show the search box but no result. And after user enter the data, it then performs the search and show the result beneath.
    <%@ page language="java" errorPage="errorpage.jsp" contentType="text/html;charset=ISO-8859-1" %>
    <%@ taglib uri="/webapp/DataTags.tld" prefix="jbo" %>
    <html>
    <head>
    <META NAME="GENERATOR" CONTENT="Oracle JDeveloper">
    <LINK REL=STYLESHEET TYPE="text/css" HREF="/webapp/css/bc4j.css">
    <TITLE>Browse Form</TITLE>
    </head>
    <body>
    <jbo:ApplicationModule id="am" configname="oetest.OeAppModule.OeAppModuleLocal" releasemode="Stateful" />
    <jbo:DataSource id="ds" rangesize="3" appid="am" viewobject="ProductInformationView" />
    <jbo:DataHandler appid="am" />
    <%
    /* This scriplet processes http parameters used to extract the search
    and column strings to build the criteria value string */
    /* Initialize variables to hold string if search string and column doesn't
    change between requests.*/
    String column_name=(String)session.getValue("sessionColumnName");
    String search_value=(String)session.getValue("sessionSearchValue");
    // Set search_value to empty string for first time rendering
    if (search_value == null) search_value="";
    String criteria_value;
    //get search parameter and set in session for persistence
    if (request.getParameter("search")!=null) {
    search_value=request.getParameter("search").trim();
    session.putValue("sessionSearchValue", search_value);
    //get column parameter and set in session for persistence
    if (request.getParameter("column")!=null) {
    column_name=request.getParameter("column").trim();
    session.putValue("sessionColumnName", column_name);
    //Construct the criteria value string.
    criteria_value="like '%" + search_value + "%'";
    %>
    <jbo:OnEvent name="Find" >
    <!-- Create a new ViewCiteria to filter data source. ViewCriteria is a
    container for CriteriaRow and Criteria, which filter the rowset returned
    by the the datasource (view object) -->
    <jbo:ViewCriteria id="vc" datasource="ds" action="new">
    <!-- Create CriteriaRow to restrict the query for every attribute
    Essentially, each CriteriaRow works analogous yp an 'AND' in the where
    Clause of the view object. Here we only want to restrict the
    query based on the chosen column. Using 'uppercolumns' makes the
    restricts case insensitive -->
    <jbo:CriteriaRow id="row" > <!-- uppercolumns="true" -->
    <!-- This ensures that column_name has a non empty value. -->
    <% if (column_name==null || column_name.equals("")) { %>
    <jbo:AttributeIterate id="defvc" datasource="ds" queriableonly="true" >
    <% column_name=defvc.getName(); %>
    </jbo:AttributeIterate>
    <% } // end if %>
    <!-- Create Criteria as part of a CriteriaRow to include one or more
    attributes. Essentially, each CriteriaRow works analogous to
    an 'OR' for each attribute in the where Clause of the
    view object. Specify the column_name and criteria_value in the
    dataitem and value attributes, respectively. -->
    <jbo:Criteria dataitem="<%= column_name %>" value="<%= criteria_value %>" />
    <!-- Close CriteriaRow, ViewCriteria tags, respectively -->
    </jbo:CriteriaRow>
    </jbo:ViewCriteria>
    </jbo:OnEvent>
    <h3>ProductInformationView Browse Form</h3>
    <!-- Add a form for the search input -->
    <form name="exsearch" action="ProductInformationView_Browse.jsp" method="post">
    <table>
    <tr>
    <td align="left" nowrap>
    <img border="0" src="go.gif" align="top" width="22" height="25">
    <b><font face="Arial" size="4" color="#336699"> Search </font></b>
    <!-- use a drop list of column names to select. Iterate through Attribute names -->
    <select name="column">
    <jbo:AttributeIterate id="defvc2" datasource="ds" queriableonly="true" >
    <option value="<%= defvc2.getName() %>" <%= defvc2.getName().equals(column_name)?"selected":"" %> ><%= defvc2.getName() %></option>
    </jbo:AttributeIterate>
    </select>
    <!-- Input field for search string -->
    <input name="search" value="<%= search_value %>">
    <input type="submit" name="jboEvent" value="Find">
    </td>
    </tr>
    </table>
    </form>
    <!-- The rest was generate using the BC4J JSP Wizard to build a Browse form. -->
    <table border="0">
    <tr>
    <td ALIGN="right"><jbo:DataScroller datasource="ds"/></td>
    </tr>
    <tr>
    <td><jbo:DataTable datasource="ds" /></td>
    </tr>
    </table>
    <jbo:ReleasePageResources />
    </body>
    </html>

    JosAH wrote:
    A couple of years ago Sun supplied their javax.comm package (mysteriously versioned as version 2)
    but they stopped it for the PC Windows.
    There's an alternative: rxtx which offers an identical API as Sun did and their own native implementation for serial and parallel port handling; I have used both and the rxtx version is clearly superior to Sun's old version: http://www.rxtx.org
    This question popped up recently.
    tschodt wrote in [http://forums.sun.com/thread.jspa?messageID=10863769#10863769:}
    [Java Communications API|http://java.sun.com/products/javacomm/]
    Implementations of the API are currently available for Solaris SPARC, Solaris x86, and Linux x86.follow the Download link to find
    Sun no longer offer's the Windows platform binaries of javax.comm, however javax.comm 2.0.3 can be used for the Windows platform, by using it in conjunction with the Win32 implementation layer provided by the RxTx project. To use that, download javax.comm for the 'generic' platform (which provides the front-end javax.comm API only, without platform specific back-end implementations bundled). Then acquire the Windows binary implementation rxtx-2.0.7pre1 from http://www.rxtx.org.

  • Problem with apply changes on master-detail with respect to date handling

    dear sir,
    i have one tabler where in i have created from_Date and to_date as varchar2(30) each. I stored the from date as 'dd-mon-yyyy hh24:mi' I had written before header process for calculating no of days between from_Date and to_Date using the said format. If to_Date is null then first i store the value as
    to_char(sysdate,'dd-mon-yyyy hh24:mi') and then convert it as to_Date(substr(to_Date,1,17),'dd-mon-yyyy hh24:mi'). this calculation of days procedures works fine in sql command mode.
    when i am handling through master-detail page i works fine when there is null value inf to_Date vc2 column. if i store using date picker dd-mon-yyyy hh24:mi, the data is getting stored but subsequent edit through master detail give error as ora-01722 unable to fetch row. Manually if i set the to_Date as null pages works fine. I spent lots of time and unable to understand the mistake i do.
    if anybody could help me, i will be much obliged.
    thanking you,
    yours
    dr.s.raghunatha

    I think the best would be if you create an example on apex.oracle.com and I will debug it.
    Denes Kubicek
    http://deneskubicek.blogspot.com/
    http://www.opal-consulting.de/training
    http://apex.oracle.com/pls/otn/f?p=31517:1
    ------------------------------------------------------------------------------

  • Date handling change w/upgrade

    Raj,
    Hope all is well, I upgraded from htmldb 1.38 to 1.4.0.00.21c
    Noticed dates are acting different...
    Here is just one example...Before the upgrade, I had a validation function that worked
    begin
    if :P15_SDATE < sysdate - 10950
    or :P15_EDATE < sysdate - 10950
    or :p15_sdate > :P15_edate
    then return false;
    else
    return true;
    end if;
    end;
    after the upgrade it did not work, and returned invalid month errors....
    The :p15_sdate and :p15_edate are date pickers in my native format (mm/dd/yyyy)...
    After the upgrade it only worked if I changed the code and did a
    to_date(:p15_sdate ,'MM/DD/YYYY')
    and this is fine (I know I should be comparing dates to dates and already fixed the code)...but wondering if you know about this, why the change...and am I missing something I should/could be doing to make it better.
    Thanks,
    Meyer

    though i'm pretty sure we've made no code changes to the way dates are handled in html db by default, there's a small chance that we're doing a better job of picking up the correct nls settings as determined by the Globalization options available on the application-level attributes page. to be honest, my hunch/hope is that you changed that "Application Language Derived From" attribute from the default "No NLS (Application not translated)" to "Browser (use browser language preference)". if your app's set to that "Browser (use browser language preference)", your nls_date_format will be set to the standard for your current language (in your case i'd guess that'd be en-us and dd-mon-yy, respectively). if your "Application Language Derived From" attribute is set "No NLS (Application not translated)", then we'd be using the nls_date_format of our database. so, assuming the date format of your db truly is 'MM/DD/YYYY' and your app is set up to provide no NLS support, i'm pretty sure you'd find that having a pl/sql region on a page in your app that did a...
    htp.p (sysdate);
    ...would return something like...
    10/01/2003
    ...if you then change your "Application Language Derived From" attribute to "Browser (use browser language preference)" and accessed your app from a browser using a primary language of "en-us" you'd find your pl/sql region returned a...
    01-OCT-03
    ...or at least that's my take on this stuff. i'm pretty sure i'm on the right track, though. again, after talking to our development manager about this, i'm rather sure that nothing's changed on our part (with respect to default intended functionality). though there's the off chance that we're doing a better job of paying attention to your globalization prefs., my hope is that the prefs in your app have changed since you originally coded those pages.
    hope this helps,
    raj

Maybe you are looking for

  • Imy iPhone sounds are not working. Any help please

    I have an iPhone 4s. I Powered it offhand when I opened it again all the sounds like the keyboard sounds and the sound u here when u unlock the phone. Even the music in games are not working. However, when I listen to music or watch videos on YouTube

  • Problem with importing pdf as a backgroung image

          I have a requirement for importing a pdf/doc and map xml data and create a template. Generate a tiff image as the output using the template and mapped xml data. I am looking into your Adobe LiveCycle as one option. I am able to import the pdf b

  • Resizing images

    I am using Photoshop CS (older version). I am having a problem with resizing images. Prior to sending this, everything was working okay and now nothing will resize. my file path was file menu > automate > batch > and then I would choose my source and

  • Action to print a file without opening it

    Hi. I'm using Action Script 2. I'm creating a CD ROM that launches external files such as PDF's, PowerPoint and Word Documents. My client has asked that when a user clicks on an icon to launch a file, that a dialogue appears for the user to either op

  • Purchased CS5 extended a few years back. Recently got a new computer. Can't find any download links.

    I bought CS5 extended for school a few years back. I just got a new computer and lost my installation CD a long time ago. I have my serial number, but cannot find any download that works. Is there a way I can go through Adobe to download it? Or any l