Performance Decrease with ComboBoxes

Memory consumption and bad performance decrease when closing and reopening panels containing ComboBoxes.
Context:
- Running localMode
- Jdev 9.0.3.1 DE
- Business Component Version 9.0.3.10.62
- jdk 1.3.1_02
Panel (within an JInternal frame) with some JUComboBoxes with LOV-Binding to ViewObjects, only with a few entries (2-20)
We observed, that the count of JUIteratorBindings encreases with every new Panel (re)open.
After closing the Internal Frame (including the Panel) the IteratorBindings of all ComboBoxes keep alive and a set of new ones are added to the PanelBinding when opening the (same) Panel once more.
Is it a bug?
Have one to remove the IteratorBindings explicitly?
Is there any configuration possibility, that the IteratorBindings die with the Panel?
Has anyone the same problem?
Roswitha Stepputtis

Yes, we experience similar problems. My suggestion would be to simply hide (setVisible(false)) your internal frames and reuse them when you open them again.
So, instead of dynamically re-creating your JInternalFrame instances, you keep a reference to them once they are created and reuse them each time they should be made visible. I think this will make your application more efficient as a whole.

Similar Messages

  • SQL Server 2005 performance decreases with DB size while SQL Server 2012 is fine

    Hi,
    We have a C# windows service running that polls some files and inserts/updates some fields in database.
    The service was tested on a local dev machine with SQL Server 2012 running and performance was quite decent with any number of records. Later on the service was moved to a test stage environment where SQL Server 2005 is installed. At that point database
    was still empty and service was running just fine but later on, after some 500k records were written, performance problems came to light. After some more tests we've founds out that, basically, database operation performance in SQL Server 2005 decreases with
    a direct correlation with the database size. Here are some testing results:
    Run#
    1
    2
    3
    4
    5
    DB size (records)
    520k
    620k
    720k
    820k
    920k
    SQL Server 2005
    TotalRunTime
    25:25.1
    32:25.4
    38:27.3
    42:50.5
    43:51.8
    Get1
    00:18.3
    00:18.9
    00:20.1
    00:20.1
    00:19.3
    Get2
    01:13.4
    01:17.9
    01:21.0
    01:21.2
    01:17.5
    Get3
    01:19.5
    01:24.6
    01:28.4
    01:29.3
    01:24.8
    Count1
    00:19.9
    00:18.7
    00:17.9
    00:18.7
    00:19.1
    Count2
    00:44.5
    00:45.7
    00:45.9
    00:47.0
    00:46.0
    Count3
    00:21.7
    00:21.7
    00:21.7
    00:22.3
    00:22.3
    Count4
    00:23.6
    00:23.9
    00:23.9
    00:24.9
    00:24.5
    Process1
    03:10.6
    03:15.4
    03:14.7
    03:21.5
    03:19.6
    Process2
    17:08.7
    23:35.7
    28:53.8
    32:58.3
    34:46.9
    Count5
    00:02.3
    00:02.3
    00:02.3
    00:02.3
    00:02.1
    Count6
    00:01.6
    00:01.6
    00:01.6
    00:01.7
    00:01.7
    Count7
    00:01.9
    00:01.9
    00:01.7
    00:02.0
    00:02.0
    Process3
    00:02.0
    00:01.8
    00:01.8
    00:01.8
    00:01.8
    SQL Server 2012
    TotalRunTime
    12:51.6
    13:38.7
    13:20.4
    13:38.0
    12:38.8
    Get1
    00:21.6
    00:21.7
    00:20.7
    00:22.7
    00:21.4
    Get2
    01:38.3
    01:37.2
    01:31.6
    01:39.2
    01:37.3
    Get3
    01:41.7
    01:42.1
    01:35.9
    01:44.5
    01:41.7
    Count1
    00:20.3
    00:19.9
    00:19.9
    00:21.5
    00:17.3
    Count2
    01:04.5
    01:04.8
    01:05.3
    01:10.0
    01:01.0
    Count3
    00:24.5
    00:24.1
    00:23.7
    00:26.0
    00:21.7
    Count4
    00:26.3
    00:24.6
    00:25.1
    00:27.5
    00:23.7
    Process1
    03:52.3
    03:57.7
    03:59.4
    04:21.2
    03:41.4
    Process2
    03:05.4
    03:06.2
    02:53.2
    03:10.3
    03:06.5
    Count5
    00:02.8
    00:02.7
    00:02.6
    00:02.8
    00:02.7
    Count6
    00:02.3
    00:03.0
    00:02.8
    00:03.4
    00:02.4
    Count7
    00:02.5
    00:02.9
    00:02.8
    00:03.4
    00:02.5
    Process3
    00:21.7
    00:21.0
    00:20.4
    00:22.8
    00:21.5
    One more thing is that it's not Process2 table that constantly grows in size but is Process1 table, that gets almost 100k records each run.
    After that SQL Server 2005 has also been installed on a dev machine just to test things and we got exactly the same results. Both SQL Server 2005 and 2012 instances are installed using default settings with no changes at all. The same goes for databases
    created for the service.
    So the question is - why are there such huge differences between performance of SQL Server 2005 and 2012? Maybe there are some settings that are set by default in SQL Server 2012 database that need to be set manually in 2005?
    What else can I try to test? The main problem is that production SQL Server will be updated god-knows-when and we can't just wait for that.
    Any suggestions/advices are more than welcome.

    ...One more thing is that it's not Process2 table that constantly grows in size but is
    Process1 table, that gets almost 100k records each run....
    Hi,
    It is not clear to me what is that you are doing, but now we have a better understanding on ONE of your tables an it is obviously you will get worse result as the data become bigger. Actually your table look like a automatic build table by ORM like Entity
    Framework, and it's DDL probably do not much your needs. For example if your select query is using a filter on the other column that [setID] then you have no index and the server probably have to scan the entire table in order to find the records that you
    need.
    Forum is a suitable place to seek advice about a specific system (as I mentioned before we are not familiar with your system), and it is more suitable for general questions. For example the fact that you have no index except the index on the column [setID]
    can indicate a problem. Ultimately to optimize the system will need to investigate it more thoroughly (as it is no longer appropriate forum ... but we're not there yet). Another point is that now we can see that you are using [timestamp] column, an this
    implies that your are using this column as a filter for selecting the data. If so, then maybe a better DDL will be to use clustered index on this column and if needed a nonclustered index on the [setID] if it is needed at all...
    what is obviously is that next is to check if this DDL fit
    your specific needs (as i mentioned before).
    Next step is to understand what action do you do with this table. (1) what is your query which become slowly in a bigger data set. (2) Are you using ORM (object relational mapping, like Entity Framework
    code first), and if so then which one.
    [Personal Site] [Blog] [Facebook]

  • JDev10g: General performance decrease with JClients like Oracle BC Browser

    Hi,
    I'm observing a significant performance decrease of JClient applications (homegrown and Oracle Business Components Browser)in JDev10g compared against JDev9i (9033)?
    Can anyone confirm my observation?
    Just create a default Dept-Emp business components package and start the Business components browser. Scroll to some data (it takes some time...) and resize the window (takes even more time)!
    My guess: JDev9i-style data binding for JClients is simulated in JDev10g with ADF-Data-Binding?? My own application is a simple "recompile"-migration of a JDev9i project with JDev10g. Does this apply to Oracle Business Components Browser too?
    My own application runs still faster with JDev9i!
    What is the problem? Any hints are welcome!
    Thanks,
    Markus

    Excellent Blog. Thank You
    Small clarification on Step **6) Oracle Home Directory, ...a) Resize the Root Partition**
    Ubuntu 11.10 has Gparted available as a Ubuntu software download, DONT use that while trying the above step, instead download the ISO file from http://sourceforge.net/projects/gparted/files/gparted-live-stable/ gparted-live-0.12.0-5.iso (124.6 MB)
    Burn that ISO file on a Blank DVD, reboot the Ubuntu , during startup select Boot from DVD Option if not already selected. this will take to Boot Menu Options of Gparted Live then select the first menu option, and this allows to do further action such as Re-sizing .
    and once you have chosen and executed step a) . do NOT run step b) also that is "Setup External Storage"
    I hope this minor clarification can avoid some confusion
    Regards
    Madhusudhan Rao
    Edited by: MadhusudhanRao on Mar 24, 2012 11:30 PM

  • Performance decrease with xorg 1.6 and intel

    i have a box with
    gigabyte ga-8ig1000mk mainboard
    intel 865G graphics
    celeron 2.6 cpu
    512 M ram
    arch linux
    x does not start with xorg 1.5/intel or xorg 1.6/intel (video memory related error message)
    kde 4 performance with xorg 1.6/intel-legacy is noticably lower than with xorg 1.4/intel
    glxgears in its default window size gives 460 fps with xorg 1.4/intel and 300 fps with xorg 1.6/intel-legacy
    currently the only way to have xorg 1.4 is
    IgnorePkg = xf86-video-intel
    IgnorePkg = intel-dri
    IgnorePkg = glproto
    IgnorePkg = xcb-proto
    IgnorePkg = libxcb
    IgnorePkg = libx11
    IgnorePkg = dri2proto
    IgnorePkg = libdrm
    IgnorePkg = mesa
    IgnorePkg = libgl
    IgnorePkg = libxdamage
    IgnorePkg = damageproto
    IgnorePkg = xcb-util
    IgnorePkg = gtk2
    IgnorePkg = pango
    IgnorePkg = libthai
    IgnorePkg = libxinerama
    IgnorePkg = libxcursor
    IgnorePkg = libxklavier
    IgnorePkg = xkeyboard-config
    IgnorePkg = libxi
    IgnorePkg = libxext
    IgnorePkg = libx11
    IgnorePkg = libxcb
    IgnorePkg = xcb-proto
    IgnorePkg = xextproto
    IgnorePkg = inputproto
    IgnorePkg = cairo
    IgnorePkg = libpng
    IgnorePkg = pixman
    IgnorePkg = xf86-input-keyboard
    IgnorePkg = xf86-input-mouse
    IgnorePkg = xf86-video-vesa
    IgnorePkg = libfontenc
    IgnorePkg = freetype2
    IgnorePkg = fontsproto
    IgnorePkg = libxfont
    IgnorePkg = xproto
    IgnorePkg = xorg-font-utils
    IgnorePkg = xorg-fonts-encodings
    IgnorePkg = xorg-fonts-alias
    IgnorePkg = expat
    IgnorePkg = fontconfig
    IgnorePkg = xorg-fonts-100dpi
    IgnorePkg = xorg-fonts-75dpi
    IgnorePkg = libxau
    IgnorePkg = libxdmcp
    IgnorePkg = xcb-proto
    IgnorePkg = libxcb
    IgnorePkg = kbproto
    IgnorePkg = inputproto
    IgnorePkg = bigreqsproto
    IgnorePkg = libx11
    IgnorePkg = xextproto
    IgnorePkg = libxext
    IgnorePkg = libice
    IgnorePkg = libsm
    IgnorePkg = libxt
    IgnorePkg = libxmu
    IgnorePkg = libxpm
    IgnorePkg = printproto
    IgnorePkg = libxp
    IgnorePkg = libxaw
    IgnorePkg = xorg-res-utils
    IgnorePkg = libxkbfile
    IgnorePkg = libxkbui
    IgnorePkg = xf86miscproto
    IgnorePkg = libxxf86misc
    IgnorePkg = xf86vidmodeproto
    IgnorePkg = libxxf86vm
    IgnorePkg = xcursor-themes
    IgnorePkg = xorg-xkb-utils
    IgnorePkg = xkeyboard-config
    IgnorePkg = fontcacheproto
    IgnorePkg = libxfontcache
    IgnorePkg = libxi
    IgnorePkg = liblbxutil
    IgnorePkg = renderproto
    IgnorePkg = libxrender
    IgnorePkg = randrproto
    IgnorePkg = libxrandr
    IgnorePkg = trapproto
    IgnorePkg = libxtrap
    IgnorePkg = mcpp
    IgnorePkg = xtrans
    IgnorePkg = xorg-server-utils
    IgnorePkg = xf86dgaproto
    IgnorePkg = libxxf86dga
    IgnorePkg = recordproto
    IgnorePkg = libxtst
    IgnorePkg = xineramaproto
    IgnorePkg = libxinerama
    IgnorePkg = videoproto
    IgnorePkg = libxv
    IgnorePkg = dmxproto
    IgnorePkg = libdmx
    IgnorePkg = libxft
    IgnorePkg = xorg-utils
    IgnorePkg = libdrm
    IgnorePkg = xorg-fonts-misc
    IgnorePkg = pixman
    IgnorePkg = xbitmaps
    IgnorePkg = xorg-server
    IgnorePkg = xorg-twm
    IgnorePkg = xorg-xauth
    IgnorePkg = xorg-xinit
    IgnorePkg = fixesproto
    IgnorePkg = libxfixes
    IgnorePkg = libxcursor
    IgnorePkg = libpng
    IgnorePkg = xorg-apps
    IgnorePkg = xterm
    i want a better solution
    Last edited by ash (2009-04-19 22:48:49)

    you do not need nearly as many packages on hold. Haven't logged into my 32-bit arch install for a month now, but I've maintained a xorg 1.4 install with this line just fine:
    IgnorePkg = xf86-video-intel xf86-input-evdev xorg-server xorg-apps xorg-xkb-utils xorg-server-utils xorg-utils intel-dri synaptics libgl libdrm xorg-res-utils xf86driproto xf86-input-keyboard xkeyboard-config
    That said, I expext to upgrade next time I use it (probably to play around with some fast boot). I would look into troubleshooting with the latest xorg and intel driver as it really should work, if I were you. I presume you have already tried starting x without a xorg.conf as a baseline test with those versions?

  • Performance decrease with custom cell renderer

    Hi,
    Finally, I have my cell renderer working fine (it was simpler than I thought). Anyway, I'm suspecting something is not working quite well yet. When I try to reorder my colums (by dragging them) they move like slow motion. If I don't register my custom cell renderer, I can move my columns fine. Next is the code of my custom renderer.
    class PriorityRenderer extends DefaultTableCellRenderer {   
      public Component getTableCellRendererComponent(
      JTable table, Object value,     boolean isSelected,
      boolean hasFocus, int row, int column) {              
        if(value instanceof coloredValue) {
        coloredValue cv = (coloredValue)value;
        if(isSelected) this.setBackground(table.getSelectionBackground());
        else this.setBackground(cv.getbkColor());
        this.setForeground(cv.gettxColor());
        this.setText(value.toString());
        this.setOpaque(true);
        return this;
    }All the cells of my JTable are "coloredValue"s. This is a little class with a String, two colors and some getter methods.
    Can anyone giveme a hint ?
    Thanks!!

    OK! Now the performance problem is gone!! (I don't know why, I didn't touch a thing)
    Thanks anyway

  • Performance problems with combobox and datasource

    We have a perfomance problem, when we are connecting a datatable object or something like this to a datasource property of a combobox. Below you find the source code. The SQL-Statement reads about 40000 rows and the result (all 40000) should be listed in the combobox. There is duration about 30 second before this process has finished. Any suggestions?
    Dim ds As New DataSet
    strSQL = "Select * from am.city"
    conn = New Oracle.DataAccess.Client.OracleConnection(Configuration.ConfigurationSettings.AppSettings("conORA"))
    comm = New Oracle.DataAccess.Client.OracleCommand(strSQL)
    da = New Oracle.DataAccess.Client.OracleDataAdapter(strSQL, conn)
    conn.Open()
    da.Fill(ds)
    conn.Close()
    Dim dt As New DataTable
    dt = ds.Tables(0)
    ComboBox1.DataSource = dt
    ComboBox1.ValueMember = dv.Table.Columns("id").ColumnName
    ComboBox1.DisplayMember = dv.Table.Columns("city").ColumnName

    But how long does it take to fill the DataTable?
    I can fill a 40000 row datatable in under 4 seconds.
    DataBinding a combo box to that many rows is pretty expensive, and not normally recommended.
    David
    Dim strConnection As String = "Data Source=oracle;User ID=scott;Password=tiger;"
    Dim conn As OracleConnection = New OracleConnection(strConnection)
    conn.Open()
    Dim cmd As New OracleCommand("select * from (select * from all_objects union all select * from all_objects) where rownum <= 40000", conn)
    Dim ds As New DataSet()
    Dim da As New OracleDataAdapter(cmd)
    Dim begin As Date = Now
    da.Fill(ds)
    Console.WriteLine(ds.Tables(0).Rows.Count & " rows loaded in " & Now.Subtract(begin).TotalSeconds & " seconds")
    outputs
    40000 rows loaded in 3.734375 seconds

  • Performance decrease with flash animation/games

    When I try to view embedded flash using firefox or safari, it seems to hog a ton of memory. It starts out at a reasonable ammount (usually ~300MB of Physical plus Virtual) but eventually it gets up to 1.39 GB (out of my 2.00 GB available). The whole system slows considerably and often crashes.
    I have noticed that the same flash animations work fine on my windows based PC with only 1 GB of physical memory. I have a 256 MB graphics card along with 2 GB of memory and a fast processor. Is there a way I can pipeline the system more to use the graphics chip rather than the memory or any software that might prevent this issue?
    Thank You

    Make sure you have the latest versions of Adobe Flash Player Software.
    Try running repair permissions from your HD and afterwards run +repair disk+ _from your software install DVD._

  • Performance Decreased with group by

    Hi,
    I am tuning a query in oracle. when I run the query without group by clause, the query runs in 5-6 sec, but after including the group by clause. The same query takes 14mins to run. If anybody has any idea of what is the impact of group by clause on performance of a SQL query, please help me . The table has 1million rows of data.

    Definitely group by clause will take more time.. Do you really need group by clause in the query ?
    Please check
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::p11_question_id:561712900346367451
    REgards
    Rajesh
    Edited by: Rajesh on Jun 10, 2010 12:20 PM

  • IPhone 3G Performance Worse with 4.2.1

    Has anyone had their performance decrease with the 4.2.1 update? My phone is responding much more poorly. I can't open maps or any other application that uses location services over the cellular network. Totally bombs. Works ok with wifi, load is still slow but once loaded works ok. Also calendar and camera take even longer to load than they did before. Phone hangs a lot more also and many more calls failing.
    Anyone else? I want to chuck my phone out the window. 6 months of fail.

    I'm new to this but if you're using a PC (can't speak on Mac) try going to your Control Panel, click on the Add or Remove Programs icon, when the list populates scroll down to the Apple Mobile Device Support icon, click on Add/Remove (on Windows 7 its a right click on the line anywhere) a menu drops down or a window appears depending on your system, and it offers you the choice to Unintall, Change, or Repair. Choose the Repair option and it should update your system. Perhaps that's your problem. There are other options in that list to chose such as a link to click on which takes you to the webpage where help is available. I've discovered that cleaning out my Add/Remove program list to save storage space is good for my computer and also made me realize that we, the computer using public, have real power in that we can thin out that list considerably by just seeing which programs we use frequently or rarely and delete accordingly, scares the heck out of the big boys who design browsers and other useless programs that clog up our system. I just deleted Google Chrome who were so sorry to see me leave and was promptly sent a survey so that they could understand my reasons for leaving. I felt on par with them, for once and let them know that Internet Explorer 9 ripped them off. Good luck.

  • [URGENT] Performance problem with BC4J and partioned data

    Hi all,
    I have a big performance probelm with BC4J and partitioned data. As as partitioned table shouldn't have a primary key like a sequence (or something else) my partitioned table doesn't have any primary key.
    When I debug my BC4J application I can see a message showing me "ignoring row with no primary key" from EntityCache. It takes a long time to retrieve my data even if I use the partition keys. A quick & dirty forms application was multiple times faster!
    Is this a bug in BC4J, or is BC4J not suitable for partitioned data? Can anyone give me a hint what to do, do make the BC4J application fast even with partitioned data? In a non-partitioned environment the application works quite well. So it seams that it must be an "error" somewhere in this part.
    Thanks,
    Axel

    Here's a SQL statement that creates the table.
    CREATE TABLE SEARCH
    (SEAR_PARTKEY_DAY              NUMBER(4)        NOT NULL
    ,SEAR_PARTKEY_EMP            VARCHAR2(2)      NOT NULL
    ,SEAR_ID                     NUMBER(20)       NOT NULL
    ,SEAR_ENTRY_DATE             TIMESTAMP        NOT NULL
    ,SEAR_LAST_MODIFIED            TIMESTAMP             NOT NULL
    ,SEAR_STATUS                 VARCHAR2(100)    DEFAULT '0'
    ,SEAR_ITC_DATE               TIMESTAMP        NOT NULL
    ,SEAR_MESSAGE_CLASS          VARCHAR2(15)     NOT NULL
    ,SEAR_CHIPHERING_TYPE        VARCHAR2(256)   
    ,SEAR_GMAT                   VARCHAR2(1)      DEFAULT 'U'
    ,SEAR_NATIONALITY            VARCHAR2(3)      DEFAULT 'XXX'
    ,SEAR_MESSAGE_ID             VARCHAR2(32)     NOT NULL
    ,SEAR_COMMENT                VARCHAR2(256)    NOT NULL
    ,SEAR_NUMBER_OF              NUMBER(3)        NOT NULL
    ,SEAR_INTERCEPTION_SYSTEM    VARCHAR2(40)    
    ,SEAR_COMM_PRIOD_H           NUMBER(5)        DEFAULT -1
    ,SEAR_PRIOD_R                  NUMBER(5)        DEFAULT -1
    ,SEAR_INMARSAT_CES           VARCHAR2(40)    
    ,SEAR_BEAM                   VARCHAR2(10)    
    ,SEAR_DIALED_NUMBER          VARCHAR2(70)    
    ,SEAR_TRANSMIT_NUMBER        VARCHAR2(70)    
    ,SEAR_CALLED_NUMBER          VARCHAR2(40)    
    ,SEAR_CALLER_NUMBER          VARCHAR2(40)    
    ,SEAR_MATERIAL_TYPE          VARCHAR2(3)      NOT NULL
    ,SEAR_SOURCE                 VARCHAR2(10)    
    ,SEAR_MAPPING                VARCHAR2(100)    DEFAULT '__REST'
    ,SEAR_DETAIL_MAPPING         VARCHAR2(100)
    ,SEAR_PRIORITY               NUMBER(3)        DEFAULT 255
    ,SEAR_LANGUAGE               VARCHAR2(5)      DEFAULT 'XXX'
    ,SEAR_TRANSMISSION_TYPE      VARCHAR2(40)    
    ,SEAR_INMARSAT_STD           VARCHAR2(1)     
    ,SEAR_FILE_NAME              VARCHAR2(100)    NOT NULL
    PARTITION BY RANGE (SEAR_PARTKEY_DAY, SEAR_PARTKEY_EMP)
      PARTITION SEARCH_MAX VALUES LESS THAN (MAXVALUE, MAXVALUE) MIRA4_SEARCH_EVEN
    );of course SEAR_ID is filled by a sequence but the field is not the primary key as it would decrease the performance of partitioned data.
    We moved to native JDBC with our application and the performance is like we never expected to be!

  • Severe Samba performance decrease after u6 - u7 upgrade

    I just upgraded my Solaris 10 x86 system from u6 (kernel 138889-08) to u7 using LiveUpgrade. I have samba sharing some ZFS filesystems, and the first thing I noticed after the upgrade was a severe performance decrease in reading files via smb.
    For example, using smbget from a MacOS client to retrieve a particular test file, pre-upgrade I was getting about 30Mb/sec, but after the upgrade to u7 I'm only getting 100Kb/sec. If I boot back to my alternate boot environment that still has u6, performance returns to normal. This happened when u7 was first released, so I just ignored it and went back to u6, and now I tried u7 again yesterday and applied all of the current security and most recommended patches (including the latest samba patch) and am still experiencing the same behavior.
    I don't currently have any socket options or other performance tuning in my smb.conf, it is about as simple as possible with just a couple share definitions.
    Any ideas why the upgrade would cause such problems for samba? The only troubleshooting I can think to check is that the network card (e1000g) isn't getting errors all of a sudden after the upgrade, but netstat -i shows everything is working as expected; 0 errors/collisions/etc.
    Any suggestions would be appreciated!
    Thanks,
    Matthew

    Hi Experts,
    After i tried a few times, i can successfully start prepare with upgrade asistant monitor - Administrator >> Start Prepare,
    Thanks

  • Performance decrease due to build options

    I am trying to use the performance monitor remote tool to mainly view the process processor time for multiple processes on my WEC7 device with a native application code, however, the OS build with the required build options shows a significant decrease in
    performance to the point where it eventually crashes. Is there a work around for this performance decrease due to build options?
    I asked in the native application development section but received 1 response which was only a question to
    my question.
    Thanks,

    Hi Jaydeep,
    You create one J2ee library project. In this project include all required external jars. Build it and deploy on the server.
    Now, these jars are available on server. So, just make a reference to this library project in your webdynpro application.
    To accomplish this, goto webdynpro application's property. These select webdynpro references-> Library references.
    Here add the refernce as following:
    <Application provider>/<Library project name>
    By default application provider name is sap.com.
    Regards,
    Bhavik

  • Performance Issue with SXMB_MONI

    Hi All,
    I have a typical performance issue with SXMB_MONI, when I trigger this T.code it is taking around 20-24 hrs to execute.
    Here I have found some tables which actually stores these processed xml messages,
    SXMSPFADDRESS
    SXMSPFRAWH
    RSXMB_REMOTE_SERVICE
    SXMSPFAGG
    SXMSCONFVL
    SXMSPMAST
    SXMSPEMAS, SXMSPERROR, SXMSPMAST & SXMSPVERS.
    SXMSPMAST, SXMSPMAST2, SXMSCLUR, SXMSCLUR2,
    SXMSCLUP, SMXSLUP2, SXMSPFRAWH,
    Here I want to increase the performance of sxmb_moni, firstly I want to know from which tables does the sxmb_moni fetches data and more over is it a single table or multiple table.
    And please suggest any technique which can decrease the latency time in executing sxmb_moni.
    Regards,
    Vijay N

    Hi,
    Periodically you need to archive the XI messages, that allows you to maintain sufficient performance level.
    Create archive jobs in SXMB_ADM to archive data which is 15 days old from XI related growing tables like SXMSCLUR, SXMSPEMAS, SXMSPHIST, SXMSPMAST, SXMSPVERS, SXMSPFRAW ,SWWWIHEAD. This archive job created archive files at the OS Level
    For XI tables refer
    /people/gourav.khare2/blog/2007/12/12/interesting-abap-tables-in-xi-150-part-i
    SXMSPMAST, SXMSCLUP, SXMSPCLUR
    the last two are cluster tables
    and you won't get XML messages directly from them
    have a look inside them
    The classes that reads this information in SXMB_MONI are abap classes,
    (can be seen at SE24) it is quite difficult to use them,
    you might debug SXMB_MONI or use SE30 and see all the classes
    that have been used.
    You can use value mapping if you are not looking at picking up values from application system.It is just like SM30 transcation.You can get the info under SAP XI->Design and Configuration->Configuration->Value Mapping.
    Also see the
    these tables,
    /SAPDMC/LSOMAP Field Mapping
    /SAPTRX/SCAOTMAP
    /SAPTRX/SCCNDMAP /SAPTRX/SCEVTMAP
    /SAPTRX/SCFUNMAP /SAPTRX/SCSOMAP
    /people/udo.martens/blog/2006/02/16/own-logging-of-xi-messages
    message-mappings: stored in which database-table?
    sxmb_moni, table sxmspmast, Messages with ICON_LED_RED, report RSXMB_SELECT
    http://help.sap.com/saphelp_nw04s/helpdata/en/44/a1b46c4c686341e10000000a114a6b/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/ef/45393c3eb3036be10000000a11402f/frameset.htm
    Thanks
    Swarup

  • Is there a recommended limit on the number of custom sections and the cells per table so that there are no performance issues with the UI?

    Is there a recommended limit on the number of custom sections and the cells per table so that there are no performance issues with the UI?

    Thanks Kelly,
    The answers would be the following:
    1200 cells per custom section (NEW COUNT), and up to 30 custom sections per spec.
    Assuming all will be populated, and this would apply to all final material specs in the system which could be ~25% of all material specs.
    The cells will be numeric, free text, drop downs, and some calculated numeric.
    Are we reaching the limits for UI performance?
    Thanks

  • Performance issues with pipelined table functions

    I am testing pipelined table functions to be able to re-use the <font face="courier">base_query</font> function. Contrary to my understanding, the <font face="courier">with_pipeline</font> procedure runs 6 time slower than the legacy <font face="courier">no_pipeline</font> procedure. Am I missing something? The <font face="courier">processor</font> function is from [url http://www.oracle-developer.net/display.php?id=429]improving performance with pipelined table functions .
    Edit: The underlying query returns 500,000 rows in about 3 minutes. So there are are no performance issues with the query itself.
    Many thanks in advance.
    CREATE OR REPLACE PACKAGE pipeline_example
    IS
       TYPE resultset_typ IS REF CURSOR;
       TYPE row_typ IS RECORD (colC VARCHAR2(200), colD VARCHAR2(200), colE VARCHAR2(200));
       TYPE table_typ IS TABLE OF row_typ;
       FUNCTION base_query (argA IN VARCHAR2, argB IN VARCHAR2)
          RETURN resultset_typ;
       c_default_limit   CONSTANT PLS_INTEGER := 100;  
       FUNCTION processor (
          p_source_data   IN resultset_typ,
          p_limit_size    IN PLS_INTEGER DEFAULT c_default_limit)
          RETURN table_typ
          PIPELINED
          PARALLEL_ENABLE(PARTITION p_source_data BY ANY);
       PROCEDURE with_pipeline (argA          IN     VARCHAR2,
                                argB          IN     VARCHAR2,
                                o_resultset      OUT resultset_typ);
       PROCEDURE no_pipeline (argA          IN     VARCHAR2,
                              argB          IN     VARCHAR2,
                              o_resultset      OUT resultset_typ);
    END pipeline_example;
    CREATE OR REPLACE PACKAGE BODY pipeline_example
    IS
       FUNCTION base_query (argA IN VARCHAR2, argB IN VARCHAR2)
          RETURN resultset_typ
       IS
          o_resultset   resultset_typ;
       BEGIN
          OPEN o_resultset FOR
             SELECT colC, colD, colE
               FROM some_table
              WHERE colA = ArgA AND colB = argB;
          RETURN o_resultset;
       END base_query;
       FUNCTION processor (
          p_source_data   IN resultset_typ,
          p_limit_size    IN PLS_INTEGER DEFAULT c_default_limit)
          RETURN table_typ
          PIPELINED
          PARALLEL_ENABLE(PARTITION p_source_data BY ANY)
       IS
          aa_source_data   table_typ;-- := table_typ ();
       BEGIN
          LOOP
             FETCH p_source_data
             BULK COLLECT INTO aa_source_data
             LIMIT p_limit_size;
             EXIT WHEN aa_source_data.COUNT = 0;
             /* Process the batch of (p_limit_size) records... */
             FOR i IN 1 .. aa_source_data.COUNT
             LOOP
                PIPE ROW (aa_source_data (i));
             END LOOP;
          END LOOP;
          CLOSE p_source_data;
          RETURN;
       END processor;
       PROCEDURE with_pipeline (argA          IN     VARCHAR2,
                                argB          IN     VARCHAR2,
                                o_resultset      OUT resultset_typ)
       IS
       BEGIN
          OPEN o_resultset FOR
               SELECT /*+ PARALLEL(t, 5) */ colC,
                      SUM (CASE WHEN colD > colE AND colE != '0' THEN colD / ColE END)de,
                      SUM (CASE WHEN colE > colD AND colD != '0' THEN colE / ColD END)ed,
                      SUM (CASE WHEN colD = colE AND colD != '0' THEN '1' END) de_one,
                      SUM (CASE WHEN colD = '0' OR colE = '0' THEN '0' END) de_zero
                 FROM TABLE (processor (base_query (argA, argB),100)) t
             GROUP BY colC
             ORDER BY colC
       END with_pipeline;
       PROCEDURE no_pipeline (argA          IN     VARCHAR2,
                              argB          IN     VARCHAR2,
                              o_resultset      OUT resultset_typ)
       IS
       BEGIN
          OPEN o_resultset FOR
               SELECT colC,
                      SUM (CASE WHEN colD > colE AND colE  != '0' THEN colD / ColE END)de,
                      SUM (CASE WHEN colE > colD AND colD  != '0' THEN colE / ColD END)ed,
                      SUM (CASE WHEN colD = colE AND colD  != '0' THEN 1 END) de_one,
                      SUM (CASE WHEN colD = '0' OR colE = '0' THEN '0' END) de_zero
                 FROM (SELECT colC, colD, colE
                         FROM some_table
                        WHERE colA = ArgA AND colB = argB)
             GROUP BY colC
             ORDER BY colC;
       END no_pipeline;
    END pipeline_example;
    ALTER PACKAGE pipeline_example COMPILE;Edited by: Earthlink on Nov 14, 2010 9:47 AM
    Edited by: Earthlink on Nov 14, 2010 11:31 AM
    Edited by: Earthlink on Nov 14, 2010 11:32 AM
    Edited by: Earthlink on Nov 20, 2010 12:04 PM
    Edited by: Earthlink on Nov 20, 2010 12:54 PM

    Earthlink wrote:
    Contrary to my understanding, the <font face="courier">with_pipeline</font> procedure runs 6 time slower than the legacy <font face="courier">no_pipeline</font> procedure. Am I missing something? Well, we're missing a lot here.
    Like:
    - a database version
    - how did you test
    - what data do you have, how is it distributed, indexed
    and so on.
    If you want to find out what's going on then use a TRACE with wait events.
    All nessecary steps are explained in these threads:
    HOW TO: Post a SQL statement tuning request - template posting
    http://oracle-randolf.blogspot.com/2009/02/basic-sql-statement-performance.html
    Another nice one is RUNSTATS:
    http://asktom.oracle.com/pls/asktom/ASKTOM.download_file?p_file=6551378329289980701

Maybe you are looking for