CMP 6.1 Entity bad performance.

I'am using entity 1.1 EJB on WL 6.1 and facing very bad performances:
around 150ms for an insert (i have 20 columns).
When accessing an order interface to read 2 fields in a session bean method: around
90 ms.
I'am very disapointed and confused. What should I look up for
to increase the performance ? Any important tuning or parameters ? Should I use EJB
2.0 to have significant perf ?
Thanks for any advice because we are thinking to switch all the application on stored
procedures. A solution without Entity and fewer stateless session beans.
My config:
WL: 6.1 on Sun sparc
SGBD: Sybase
Entity: WebLogic 6.0.0 EJB 1.1 RDBMS (weblogic-rdbms11-persistence-600.dtd)
Thanks

Historically its hard to get good performance & scalability out of sybase
without using stored procs. Using dynamic sql on sybase just doesnt do as
well as procs. Oracle on the other hand can get very close to stored proc
speed out of well written dynamic sql.
As far as weblogic goes, my experience is the focus of their testing for db
related stuff is Oracle, then DB2, then MSSQLServer. Sybase is usually last
on the list.
As far as the 6.1 cmp, haven't used it much, but because of these other
things I would be cautious about using it with Sybase.
Joel
"Antoine Bas" <[email protected],> wrote in message
news:3cc7cdcf$[email protected]..
>
I'am using entity 1.1 EJB on WL 6.1 and facing very bad performances:
around 150ms for an insert (i have 20 columns).
When accessing an order interface to read 2 fields in a session beanmethod: around
90 ms.
I'am very disapointed and confused. What should I look up for
to increase the performance ? Any important tuning or parameters ? ShouldI use EJB
2.0 to have significant perf ?
Thanks for any advice because we are thinking to switch all theapplication on stored
procedures. A solution without Entity and fewer stateless session beans.
My config:
WL: 6.1 on Sun sparc
SGBD: Sybase
Entity: WebLogic 6.0.0 EJB 1.1 RDBMS(weblogic-rdbms11-persistence-600.dtd)
>
Thanks

Similar Messages

  • Bad performance when open a bi publisher report in excel

    We use bi publisher(xml publisher) to create a customized report. For a small report, user like it very much. But for a bigger report, users complain about the performance when they open the file.
    I know it is not a native excel file, that may cause the bad performance. So I ask my user to save it to a new file as a native excel format. The new file still worse than a normal excel file when we open it.
    I did a test. We try to save a bi publish report to excel format, the size shrink to 4Mb. But if we "copy all" and "Paste Special" value only to a new excel file, the size is only 1Mb.
    Do I have any way to improve that, users are complaining everyday. Thanks!
    I did a test today.
    I create a test report
    Test 1: Original file from BIP in EBS is 10Mb. We save it in my local disk, when we open the file, it takes 43 sec.
    Test 2: We save the file in native excel format, the file size is 2.28Mb, it takes 7 sec. to open.
    Test 3: We copy all cell and "PasteSpecial" to a new excel file with value only. The file size is 1.66Mb, it takes only 1 sec to open.
    Edited by: Rex Lin on 2010/3/31 下午 11:26

    EBS or Standalone BIP?
    If EBS see this thread for suggestions on performance tuning and hints and tips:
    EBS BIP Performance Tuning - Definitive Guide?
    Note also that I did end up rewriting my report as PL/SQL producing a csv file and have done with several large reports in BIP on EBS.
    Cheers,
    Dave

  • Bad performance updating purchase order (ME22N)

    Hello!
    Recently, we face bad performance updating purchase orders using transaction ME22N. The problem occurs since we implemented change documents for a custom table T. T is used to store additional data to purchase order positions using BAdIs ME_PROCESS_PO_CUST and ME_GUI_PO_CUST.
    I've created a change document C_T for T using transaction SCDO. The update module of the change document is triggered in the method POST of BAdI ME_PROCESS_PO_CUST.
    Checking transaction SM13, I recognized that the update requests of ME22n have status INIT for several minutes before they are processed. I also tried to exclude the call of the update module for change document C_T (in Method POST) - the performance problem still occurs!
    The problem only occurs with transaction ME22N, thus I assume that the reason is the new change document C_T.
    Thanks for your help!
    Greetings,
    Wolfgang

    I agree with vikram, we don't have enough information, even not a small hint on usage of this field, so which answer do you expect (The quality of an answer depends ...) This analysis must be executed on your system...
    From a technical point of view, the BAPI_PO_CHANGE has EXTENSIONIN table parameter, fill it using structure BAPI_TE_MEPOITEM[X] alreading containing CI_EKPODB (*) and CI_EKPODBX (**)
    Regards,
    Raymond
    (*) I guess you have used this include
    (**) I guess you forgot this one (same field names but data element always BAPIUPDATE)

  • Reporting on master data customer and bad performances : any workaround ?

    Hello,
    I've been asked to investiguate on bad performances encountered when performing reporting
    on the specific master data zcustomer.
    Basically this master data has a quite similar design that 0customer, there are 96000 entries in the master data table.
    A simple query has been developed : the reporting is done on the master data zcustomer and its attributes : no key figure, no calculation, no restriction ...
    Nevertheless, the query can not be executed .. the query runs around 10 minute in rsrt, then the private memory is exhausted and then a short dump is generated.
    I tried to buid a very simple query on 0customer, this time, without the attributes ... and it took more than 30 sec before I get the results.
    I checked the queries statistics :
    3.x Analyzer Server 10 sec
    OLAP: Read Texts : 20 sec
    How is it that it is so long to performthe reporitng on those master data, while in the same time If i try to display the content in SAP by choosing "maintain master data", I have an immediate answer.
    I there any workaround ?
    Any help would be really appreciated.
    thank you.
    Raoul

    Hi.
    How much data have you got in the cube?
    If you make no restrictions, you are asking the system to return data for all 96.000 customers. That is one thing that might take some time.
    Also, using the attributes of this customer object, fx making selection or displaying several of them, means that the system has to run through the 96.000 records in masterdata to know what goes where in the report.
    When you display the masterdata, you are by default displaying just the 250 or so first hits, and you are not joining against any cube or sorting the result set, so that is fast.
    You should make some kind of restriction on other things than zcustomer (time, org.unit, version, etc, to limit the dataset from the cube, but also a restriction on one of the zcustomer attribs, with an index for that maybe, and performance should improve.
    br
    Jacob

  • Bad Performance in a query into table BKPF

    Hi forum i have a really problem in the second query under the table
    BKPF.. some body cans help me, please
    *THIS IS THE QUERY UNDER MSEG
      SELECT tmsegmblnr tmkpfbudat tmsegbelnr tmsegbukrs tmseg~matnr
             tmsegebelp tmsegdmbtr tmsegwaers tmsegwerks tmseg~lgort
             tmsegmenge tmsegkostl
      FROM mseg AS tmseg JOIN mkpf AS tmkpf ON tmsegmblnr = tmkpfmblnr
      INTO CORRESPONDING FIELDS OF TABLE it_docs
      WHERE
        tmseg~bukrs IN se_bukrs AND
        tmkpf~budat IN se_budat AND
        tmseg~mjahr = d_gjahr AND
        ( tmsegbwart IN se_bwart AND tmsegbwart IN (201,261) ).
      IF sy-dbcnt > 0.
    I CREATE AWKEY FOR CONSULTING BKPF
        LOOP AT it_docs.
          CONCATENATE it_docs-mblnr d_gjahr INTO it_docs-d_awkey.
          MODIFY it_docs.
        ENDLOOP.
    THIS IS THE QUERY WITH BAD BAD PERFOMANCE
    I NEED KNOW "BELNR" FOR GO TO THE BSEG TABLE
        SELECT belnr awkey
        FROM bkpf
        INTO CORRESPONDING FIELDS OF TABLE it_tmp
        FOR ALL ENTRIES IN it_docs
        WHERE
          bukrs = it_docs-bukrs AND
          awkey = it_docs-d_awkey AND
          gjahr = d_gjahr AND
          bstat = space .
    THNKS

    Hi Josue,
    The bad performance is because you're not specifying the primary keys of the table BKPF in your WHERE condition; BKPF usually is a big table.
    What you really need is to create a new index on database for table BKPF via the ABAP Dictionary on fields BUKRS, AWKEY, GJAHR & BSTAT. You'll find the performace of the program will significantly increase after the new index is activated. But I would talk to the Basis first to confirm they have no issues if you create a new index for BKPF on the database system.
    Hope this helps.
    Cheers,
    Sougata.

  • Bad performance in web intelligence reports

    Hi,
    We use Business Objects with Web Intelligence documents and Crystal Reports.
    We are supporting bad performance when we use the reports specilly when we need to change the drill options
    Can someone telling me if exists some best practices to improve performance? What features should i look to?
                      Best Regards
                            João Fernandes

    Hi,
    Thank you for your interest. I know that this a issue with many variables because that i need information about anything that could cause bad performance.
    For bad performance i mean the time that we take running and refreshing reports data.
    We have reports with many lines but the performance is bad even when a few users are in the system
                                 Best Regards
                                         João Fernandes

  • Help: Bad performance in marketing documents!

    Hello,
    When creating an AR delivery note which has about 10 lines, we have really noticed that the creation of lines becomes slower and slower. This especially happens when making tab in the system field "Quantity". In fact, before going to the next field quickly, it stays in Quantity field for about 5 seconds!
    The number of formatted searches in AR delivery note is only 5. And only one is automatic. The number of user fields is about 5.
    We have heard about the bad performance when the number of lines increases in the documents when having formatted searches, but it is odd to happen this with about 10 lines in the document.
    We are using PL16 and this issue seems to have been solved already at PL10.
    Could you throw some light on this?
    Thanks in advance,

    It is solved now.
    It had to be with the automatic formated search in 2 head fields.
    If the automatic search is removed, the performance is OK.
    Hope it helps you,

  • Bad performance on system, export/import buffer many sawps

    Hello,
    I have an ECC 6.0 system on AIX with 6 application servers. There seems to be a performance problem on the system, this issue is being noticed very well when people are trying to save a sale order for example, this operation takes about 10 minutes.
    Sometimes we get short dumps TSV_TNEW_PAGE_ALLOC_FAILED or MEMORY_NO_MORE_PAGING but not very often.
    I am not very good at studying the performance issues, but from what I could see is that there are may swaps on buffer export/import, program and generic key. Also the HitRatio is 88% at  buffer export/import, which I think is pretty low.
    I know that the maximum value accepted of swaps per day is 10000, is that right?
    Can you please advice me what needs to be done in order for these swaps to decrese and hit ratio to increase? And also what else I should do in order to analyse and root cause and the bad performance of the system?
    Many thannks,
    manoliv

    Hi,
    sappfpar determines the minimum and maximum (worst-case) swap space requirements of an R/3 application server. It also checks on shared memory requirements and that the em/initial_size_MB and abap/heap_area_total parameters are correctly set with the following procedure:
    /usr/sap/<SYSTEMNAME>/SYS/exe/run/sappfpar check pf=/usr/sap/<SYSTMENAME>/SYS/profile/<Profile name>
    At the end of the list, the program reports the minimum swap space, maximum heap space, and worst case swap space requirements:
    Additional Swap Space Requirements :
    You will probably need to increase the size of the swap space in hosts in which R/3 application servers run.
    As a rule of thumb, swap space should equal
    3 x Size of Main Storage or at least 1 GB, whichever is larger.
    SAP recommends a swap space of 2-3 GB for optimal performance.
    Determining Current Swap Space Availability: memlimits
    You can find out how much swap space is currently available in your host system with R/3’s memlimits program.
    Here’s how to run memlimits:
    From the UNIX command prompt, run the R/3 memlimits program to check on the size of the available swap space on the host system on which an R/3 application server is to run.
    The application server must be stopped, not running.
    /usr/sap/<SYSTEMNAME>/SYS/exe/run/memlimits | more
    The available swap space is reported in the output line Total available swap space: at the end of the program output. The program also indicates whether this amount of swap space will be adequate and determines the size of the data segments in the system.

  • Bad performance when deleting report column in webi(with "Design – Structure only")

    Hi all,
    One of our customer has recently upgraded from BO XI to BO4.1. In the new BO 4.1, they encountered a bad performance issue when they were deleting a column in Webi(using "Design – Structure only" mode).
    With “Design – Structure only" mode,  it took webi about 10 seconds to complete after the customer right-clicked a report column and clicked the "delete".  The customer said that they only need to wait for less than 1 second when they did the same in BO XI old version.
    The new BO version used is 4.1SP02, installed in Windows Server 2008 R2. (Server with 32 core CPU, 32G memory)
    This bad performance happened in both Webi web and Rich Client. (in Webi Rich Client, the performance is a little bit better. The 'delete column' action takes about 8 seconds to complete).
    Do anyone know how to tune this performance in webi?  Thank you.
    Besides, it seems that each time we are making change in the webi report structure in IE or Rich Client, webi need to interact with Server site to upload the changes. Is there any option to change this behavior?  Say, do not upload change to Server for when 'deleting report column', only trigger the upload after a set of actions(e.g. trigger when click the "Save" button).
    Thank you.
    Regards,
    Eton.

    Hi all,
    Could anyone help me on this?  Thanks!
    What customer concerns now is that when they did the same 'column editing' action in BO XI R2 for the same report, they did not need to wait.  And they need to wait for at least 7-8 second in the BO 4.1SP02 environment for this action to complete.(data already purged, in structure-only mode)
    One more information about the webi report being editing is: there are many sheets in the report(about 6~10 sheets in one report). Customer don't want to separate these sheet into different reports  as it will increase the effort of their end users to locate similar report sheets.
    Regards,
    Eton.

  • Temporary LOBs - bad performance when nocache is used

    Hello.
    Please, advise me what could be the reason of bad performance row 8 from the next anonymous block:
    declare
        i integer;
        c clob := 'c';
        procedure LTrimSys(InCLOB in clob ) is
            OutCLOB clob;
        begin
            DBMS_LOB.CREATETEMPORARY(OutCLOB, false, DBMS_LOB.call);
            dbms_lob.Copy(OutCLOB, InCLOB, dbms_lob.getlength(InCLOB));
            DBMS_LOB.freetemporary(OutCLOB);
        end;
    begin
        for j in 1 .. 1000 loop
            LTrimSys(c);
        end loop;
    end;
    I have two practically identical databases 10.2.0.4.0 EE 64-bit on Windows
    On first DB I have elapsed time: 4 sec, on second - 0.2 sec
    I didn't find important difference between init parameters (hidden parameters too).
    First DB has more memory (PGA) then second.
    Main time events in time of executing anonymous block on first DB are
    PL/SQL execution elapsed time
    DB CPU
    sql execute elapsed time
    DB time
    In second DB - the same but much less
    If I use caching of temporary LOBs then both DBs work fine, but I can not understand why first DB works slowly when I use nocache temporary LOBs.
    What can be the reason?

    I don't think that is the problem. See next outputs:
    select * from V$PGASTAT order by name
    NAME
    VALUE
    UNIT
    PGA memory freed back to OS
    49016834031616
    bytes
    aggregate PGA auto target
    170893312
    bytes
    aggregate PGA target parameter
    1073741824
    bytes
    bytes processed
    95760297282560
    bytes
    cache hit percentage
    93,43
    percent
    extra bytes read/written
    6724614496256
    bytes
    global memory bound
    107366400
    bytes
    max processes count
    115
    maximum PGA allocated
    2431493120
    bytes
    maximum PGA used for auto workareas
    372516864
    bytes
    maximum PGA used for manual workareas
    531456
    bytes
    over allocation count
    102639421
    process count
    57
    recompute count (total)
    117197176
    total PGA allocated
    1042407424
    bytes
    total PGA inuse
    879794176
    bytes
    total PGA used for auto workareas
    757760
    bytes
    total PGA used for manual workareas
    0
    bytes
    total freeable PGA memory
    75694080
    bytes
    select * from V$PGA_TARGET_ADVICE_HISTOGRAM where PGA_TARGET_FACTOR = 1
    PGA_TARGET_FOR_ESTIMATE
    PGA_TARGET_FACTOR
    ADVICE_STATUS
    LOW_OPTIMAL_SIZE
    HIGH_OPTIMAL_SIZE
    ESTD_OPTIMAL_EXECUTIONS
    ESTD_ONEPASS_EXECUTIONS
    ESTD_MULTIPASSES_EXECUTIONS
    ESTD_TOTAL_EXECUTIONS
    IGNORED_WORKAREAS_COUNT
    1073741824
    1
    ON
    2199023255552
    4398046510079
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    1099511627776
    2199023255551
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    549755813888
    1099511627775
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    274877906944
    549755813887
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    137438953472
    274877906943
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    68719476736
    137438953471
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    34359738368
    68719476735
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    17179869184
    34359738367
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    8589934592
    17179869183
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    4294967296
    8589934591
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    2147483648
    4294967295
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    1073741824
    2147483647
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    536870912
    1073741823
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    268435456
    536870911
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    134217728
    268435455
    0
    376
    0
    376
    0
    1073741824
    1
    ON
    67108864
    134217727
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    33554432
    67108863
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    16777216
    33554431
    1
    0
    0
    1
    0
    1073741824
    1
    ON
    8388608
    16777215
    10145
    45
    0
    10190
    0
    1073741824
    1
    ON
    4194304
    8388607
    20518
    21
    0
    20539
    0
    1073741824
    1
    ON
    2097152
    4194303
    832
    1
    0
    833
    0
    1073741824
    1
    ON
    1048576
    2097151
    42440
    0
    0
    42440
    0
    1073741824
    1
    ON
    524288
    1048575
    393113
    7
    0
    393120
    0
    1073741824
    1
    ON
    262144
    524287
    10122
    2
    0
    10124
    0
    1073741824
    1
    ON
    131072
    262143
    22712
    0
    0
    22712
    0
    1073741824
    1
    ON
    65536
    131071
    110215
    0
    0
    110215
    0
    1073741824
    1
    ON
    32768
    65535
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    16384
    32767
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    8192
    16383
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    4096
    8191
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    2048
    4095
    83409618
    0
    0
    83409618
    0
    1073741824
    1
    ON
    1024
    2047
    0
    0
    0
    0
    0
    1073741824
    1
    ON
    0
    1023
    0
    0
    0
    0
    0
    SELECT optimal_count, round(optimal_count*100/total, 2) optimal_perc,
           onepass_count, round(onepass_count*100/total, 2) onepass_perc,
           multipass_count, round(multipass_count*100/total, 2) multipass_perc
    FROM
           (SELECT decode(sum(total_executions), 0, 1, sum(total_executions)) total,
                   sum(OPTIMAL_EXECUTIONS) optimal_count,
                   sum(ONEPASS_EXECUTIONS) onepass_count,
                   sum(MULTIPASSES_EXECUTIONS) multipass_count
            FROM   v$sql_workarea_histogram);
    OPTIMAL_COUNT
    OPTIMAL_PERC
    ONEPASS_COUNT
    ONEPASS_PERC
    MULTIPASS_COUNT
    MULTIPASS_PERC
    12181507016
    100
    146042
    0
    0
    0

  • How to solve bad performance in vi server

    I'm useing vi server in a recursively programming structure in labview 7.1.
    I have very bad performance result so i made a test comparing the speed of che various vi server call and the native.
    my tests are:
    normal calling.
    vi server calling by reference node
    vi server calling by invoke node
    for each vi server calling type i tried:
    unique reference calling
    different reference calling
    different reference, preloaded,  calling
    the only 1 case i found a loosing of performace acceptable 24% respect the normal calling: vi server calling by reference node with unique reference.
    unfortunately this kind of calling is able to catch come kind of recunsion an forbidden it and this calling wait all the time the end of the execution of the vi called.
    All the oter kinds of caling have very bad performance: from 90% to 99,9% LESS then the normal calling (10 to 1000 times more in execution time).
    In addition there is a prblem in invoke colling: if you execute a invoke node with wait until done true, the mouse pointer fliks with the busy icone. that heppens in all the possible configuration.
    I would like ti know if someuone as some solution to this performance problem and if Labview 8.0 fixs this problems.
    Best regarda.
    ing. Luca Benvegnù

    Hi Luca,
    I've made some researches for you and, what can I say is that there is a certain number of requests about "vi server performances". There are some things that are in general suggested in this situation:
    1) Be "minimalist" about the use of images, graph,... And avoid Large data structures: wiring large structures into/out-of sub-VI's can take time: pass the smallest amount of data.
    2) Avoid repeated calls. Calling the sub-VI's frequently, in a loop for example, create a call over-head that could be an issue. For the same reason put the "wait until" function with a specification of time in order to run loops only as often as strictly required.
    However, about your question of fixing the problem with LabVIEW 8.0, I suggest you to download the Evaluation Version of LabVIEW 8.2 and try by yourself if performances are better.
    I hope I've been of some help.
    carlo>  

  • Sun Ray @ home  - VPN bad performance

    Hi All
    we've setup a test configuration with a Sun Ray Server (dedicated network) a CISCO PIX 501 and one DTU SR2.
    The screen refresh performance on the DTU is very bad. Other DTUs, connected directly to the SRSS are fine.
    We've played around with the bandwith limit on the SR2 (GUI firmware):
    no limit: DTU unusable
    3000: DTU slow
    1500: DTU more slower
    700: unusable
    The throughput of the PIX 501 should be around 3MPS. The PIX does nothing else than serving one DTU. All components (SR Server, PIX, DTU) are in the same room and connected directly to each other
    Is the small PIX the reason for the bad performance ?
    Should it work smoothly with only one DTU?
    Which Hardware (vpn box) is recommended to serve at least 10 SR2 DTUs ?
    Thanks a lot.
    CU
    Carsten

    You problem could be fragmentation. Have you lowered the default MTU (1500). MTU of 1356 works with my ASA 5505. Here is a nice article on the subject, http://blogs.sun.com/ThinkThin/entry/the_importance_of_mtu .

  • How to solve bad performance OSX Lion on Macbook Pro

    From beginning i install MAC OS Lion i have have a bad performance on my macbook Pro.
    Startup take about 3 minutes, after i login it takes a hour for i can use my laptop properly.
    Can some one tell my what the problem is ?
         My laptop is a MAcBook Pro 2,53 GHz Intel Core Duo with 4 GB 1067 Mhz DDR Memory
          I have a second partition with MAC OS X Leopart, Its working without any problems.
          Also i have check my Disk first, no problems found.

    Look like i have some disk issues :-(
    16/08/11 11:51:58,000 kernel: disk0s2: I/O error.
    16/08/11 11:51:58,000 kernel: hfs_clonefile: cluster_read failed - 5
    16/08/11 11:52:45,000 kernel: disk0s2: I/O error.
    16/08/11 11:52:56,441 Console: *** -[__NSArrayM objectAtIndex:]: index 992 beyond bounds for empty array

  • Master column tree - bad performance

    Hello,
    I have a table with master column that uses recursive node without Load Children event. the whole table loads at startup. I Have very bad performance : every click on the little triangle takes about 6-7 seconds, and there is even no round-trip to server because all data is already at the client!
    I need to show list of products by there family (50 families, each has 10 products). How can i imporve my performance (maybe server configuration or my code or another method)
    I work with ep 7 sp12.
    Thanks
    Roni.

    Hi Roni,
    Yes, you have to manipulate the add element on before loading the tree.  Best Practice is to use onLoadChildren event. Loading the tree at once will always have performance issues.
    Refer to these tutorial for creating tree using onLoadChildren event.
    <a href="https://www.sdn.sap.comhttp://www.sdn.sap.comhttp://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/webdynpro/wd%20java/wd%20tutorials/constructing%20a%20recursive%20and%20loadable%20web%20dynpro%20tree.pdf">Recursive Loadable Tree</a>
    <a href="https://www.sdn.sap.comhttp://www.sdn.sap.comhttp://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/webdynpro/wd%20java/wd%20tutorials/integration%20of%20a%20tree%20structure%20in%20a%20web%20dynpro%20table.pdf">Integration of Tree Structure in Table</a>
    There are no limitations on this UI control, you can stay with any number of levels.
    Regards
    Srikanth

  • Bad performance video in on vivo card?

    This opinion posted as a reply in other thread. But because I think It is proper to be an issue for discussion
    I am posting in a new thread. Let alone the difficulty in configuration, the vivo of this card seems to have a poor performance.
    I too have a GeForce 4 Ti4200-VTD8X-8894 AGP 8X 128MB.
    Vidio in working in a machine, but not in other machine.
    I also have a ati tv wonder pci, The configurations of the capture driver on both machine appeares to be the same but behave differently, which I don't know why at all.
    What I've found so far :
    1. For tv(videoin) decoding: dscaler1.7(alpha) works on both machine(but on bad one freezes often)
    2. On the bad machine using any capture driver causes pending (ie. in windows movie maker or virtual dub), drivers(1.27 or 1.22) have the same behavior.
    3. While comparing the performance of tv decoding, I found a big issue. Tvwonder was very vivid on screen and the cpu usage was only 8% while vivo was very rough(skipping, ...) and the cpu load was over 60%.
    So the conclusion?
    The tv in of this card is only for the poor. If you can bear the poor performanc, you can spend lots of time to enable it and get bad performance. I'd rather buy a h/w decoding tv card for $30 or more. With amd xp 2.1G, and Tvwonder and window media encoder, I was able to broadcast to 1.5g bandwidth of video quality while watching the movie on Window media player on the same machine. The cpu usage was around 80%. But with this card, I cannot dream of it at all.
    It is said that this card has bt848 decoder(So can be played using dscaler1.7). Then how come this card be so slow on decoding? Bad configuration? Possibly. If anyone who could watch tv(through vcr) fast enough with low cpu usage, please let me know the secret.

    I started again from scratch (3:th time), installed win98 on a clean HD and installed the vga driver from the cd in the box. Then I installed Winproducer from the cd, but the capture program I have to browse on the cd to install. The capture seems to work now, but I can not save my recording from my camcorder, some error message is poping up and the same message is when I tried to play the saved file (there is a saved file anyway).
    I have problem whith my network card also byt only in win98, complaining about ndis.vxd (can´t load ndis.vxd).
    So I have to install W2000 to get internet.
    When I installed the Winproducer on W2000 the computer rebooted all the time, is Winproducer not compatible whith W2000 ??
    I have not downloded anything yet (se abow).
    And where can I find this 28.32 or 29.42 ?
    WDM, I have installed this time.
    Thanks NE

Maybe you are looking for

  • Why no SyncML-support for Harmattan??

    I've used SyncML on my N95 with memotoo.com for years. Worked very well. I took it for granted that SyncML would be supported on the N9. But alas, it isn't. I find that really incomprehensible. Worse yet, CalDAV syncing with memotoo doesn't work - I

  • 412 E movement type - COGS posting vs. Inventory posting

    Hi All - 412/E movement type from unrestricted stock to sales order stock is sometimes triggering an accounting entry like: Dr. 133000 customer sales stock Cr. 133000 unrestricted stock but sometimes it appears that transaction key in OBYC - GBB - is

  • MDIS Question

    Could anyone clarify the following? We set the file aggregation count = 1 . This means even if we have more than one file in the ready folder still it will process the files one by one meaning first file has to complete then only the next file will p

  • Changing import setting in itunes 8 - can't find out how to do it!

    I cannot work out how to change the settings relating to the size and quality of music files ripped from CD's. The iTunes help tells me to go into the preference section and click on the advanced tab. However I Cannot see the import settings section

  • I want my users (8 company code). each company see only theirs asset class

    < MODERATOR:  Message locked.  Please read the [Rules of Engagement|https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/rulesofEngagement] before posting next time. > i have 8 company to configuration. and each company have various kind of asset