Table per locale - i18n application development approach

I learn that I shall use table per locale approach for a multi-language application development. And I am in a situation that no all entries will have multi-language versions. I can image some possible issues on data retrieval and update. Can anyone share experience in this regard?
Thanks very much in advance.

I would share my experice here.
For the internationalization of our product, we do not have a table per locale, rather we have a row per locale in the appropriate table. We have localeCode, messageCode and messageDescription columns in that table and the application picks up the right row depending on user locale.
By default we have English data in the database. For inserting french data we run a database script which inserts same rows as English rows appended by some french characters. This is for internal testing purpose only.
In addition, we also run another translation script for creating locale specific resource bundle files and tiles files.
So, labels will come from tiles files and data/messages would come from resource bundle and database. These scripts will not be delivered to the customer. They have their own translators and its their job to provide translations at all necessary places.
After running this script we have experienced that the server takes more time to startup.
Rahul

Similar Messages

  • I18n application development

    Dear Experts,
    I would like to know what is the recommended way to create an i18n compatible composite application using NW CE 7.1. The following is my current finding.
    In SDN, I have found the following article called u201CDeveloping an International Web Dynpro Applicationu201D.
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/70361a88-d6ce-2a10-b1bc-c357097a7d26
    The main problem of this approach is that I have to put my validation logic in the Web DynPro and which means in my UI layer. For performance boost, I believe it is nice to validate the data in the UI layer. However, I believe this validation must be finally implemented in the business logic layer and that is the Application Service. In that way, my Application Services can validate all the inputs no matter what UI I am currently using. However, by implementing the validation in Application Services, I got no clue on how to write such an i18n application. Can anyone show me a more realistic sample? Thanks in advance.
    Regards,
    Alex

    Hi Ayyapparaj,
    Is there anyway that I can made it more smoothly with the NWDS? Those xlf files are generated by NWDS and I don't think it's wise to made too much manual modification to the generated file structure. Is there a way to generate the xlf files in your recommended way? Thanks!
    Regards,
    Alex

  • What i use to develop complex tree and table in my web application?

    hi guys
    simply most of my work in my web application will be depend on:
    1-tree
    2-table
    i need this tree and table deal with database table (or even tables) must be able to do every thing like add node/row/column, delete, update, drag-n-drop, various object nodes/cells like check-box, list, images...etc
    i used JTree and JTable then stoped because i faced problems in dealing with applet/database, and i really didnt see this is perfect way to do that even i agree that JTree and JTable are excellent objects. i also start using ajax tree but i didnt see it robust to have all this complicated work with database.
    by the way i need also master-detail UI table, but its easy to get this if i have good and flexable table object.
    my rest pages developed by jsp, servlet, javascript, html...etc
    so any suggestion to how i build this tree and table? i also heared about JSF, is it serve me?
    thanks

    so if the only way to use JTree and JTable,There are other ways, but they're not Java and might be more complicated.
    are there
    any links you have discussing dealing database with
    applet by keep applet away from direct connection to
    my database?I'm not aware of links, but it's common sense. You don't want to expose the DB server to the internet, and instead contact a server for authentication that then forwards the request to the DB.

  • Iphone application development, how to save data locally?

    I am now developing a small application, need to be a small part of the data stored locally. Whenever a user when the next time you open this application, can call this data, the present flash cs5 iphone application development, can achieve this functionality?

    answer -> http://forums.adobe.com/thread/514065?tstart=0

  • Magic Values - A Question of Development Approach

    Hello folks,
    I have a question for you PL/SQ developers out there. This isn't a specific problem or query I'm raising here, more a question of general approach. I'm probably not using the correct terms here so forgive me. Also, I've already posted this in the ApEx forum, however there's a degree of overlap into pure PL/SQL so I thought you were all bound to have experience of something along the same lines.
    Anyhoo...
    How to deal with Magic values - i.e. values which hold no intrinsic value in and of themselves, other than for state, process or conditional logic control. I use them quite a lot in my PL/SQL code (as I'm sure most developers do in one context or another).
    From a Data architecture perspective, I'll generally have some sort of table for storing the 'facts': names, addresses etc, etc. Any application-specific magic values ('status', 'type') will be held as a foreign key in this table, which will reference a form of lookup table.
    Example:
    EMOTION
    ID   Description
    ==   ===========
    1    HAPPY
    2    SAD
    3    NEUTRAL
    PERSON
    NAME ... EMOTIONAL_STATE
    ====       ===============
    BILL       1
    JERRY    1
    BRIAN      3
    DONNA    2So far, so banal...
    Now, say I have a process that needs to reference someone's emotional state for some sort of conditional logic:
    declare
       n_estate number;
    begin
       select emotional_state into n_estate
         from Person
        where name = 'BILL';
       case when v_state = 1 then
          -- do something
       case when v_state = 2 then 
          -- do something else
       else
          -- otherwise something else again
    end case;
    end;straight away your bad code radar should be going crazy: you're coding literals in there! So, the old java programmer in me wants to store these as constants - I'll generally square them away inside a package somewhere, like so:
    create or replace package PKG_CONSTANTS as
       ES_HAPPY constant number:= 1;
       ES_SAD constant number := 2;
       ES_NEUTRAL constant number := 3;
    end PKG_CONSTANTS;Thus the code becomes
    Case when v_state = PKG_CONSTANTS.ES_HAPPY then ...Herein lies the crux of the issue. I'm effectively defining the same value twice: once in the lookup table (for data integrity) and once in the package. If new values are defined (say "Existential Ennui") or existing values are changed, I need to make sure the two are aligned, which hinders maintainability.
    I thought about initialising the values as sort of pseudo-constants in the package initialise code but then you end up replacing one literal with another; you end up with code like:
    create or replace package PKG_CONSTANTS as
       ES_HAPPY number;
       ES_SAD constant number;
       ES_NEUTRAL constant number;
    end PKG_CONSTANTS;
    create or replace package body PKG_CONSTANTS as 
       rf_curs sys_refcursor;
    begin
       for rf_curs in
          select ID
                 ,description
            from EMOTIONAL_STATE;
       loop
          case description
          when 'HAPPY' then
             ES_HAPPY := ID;
          when 'SAD' then
             ES_SAD := ID;
          when 'NEUTRAL' then
             ES_NEUTRAL := ID;
          else
             null;
          end case;
       end loop;
    end PKG_CONSTANTS;I also thought about using dynamic PL/SQL to re-write and recompile the constants package in the event of a value being changed in the lookup table...seems like quite a lot of work, given that the magic value is pretty much meaningless outside of the scope of the application.
    So... how to deal with this? What approach to you take? Does data integrity over-ride application programming style?
    Any contributions would be welcome!

    Hello,
    I had a look through the article (8 year's worth of thread? Sheesh that's dedication!) and yet it doesn't quite express exactly what I'm meaning. The argument there appears to be between dynamic SQL with bind variables versus static SQL. I'm not talking dynamically building queries or the use of bind variablers per se - its more related to how one makes use of magic values within the context of conditional logic and application code.
    The example I chose happened to use a case statement, which maybe blurs the line with the syntax of pure SQL query and perhaps why you thought I was going down the dynamic SQL route, but I could just have easily replaced them with a series of 'if elsif else end' type expressions.
    From an application developer point of view, the mantra of 'abstraction through constants' is the norm - referencing literals in expressions is generally frowned upon, with the possible exception of special numbers such as 1 or 0 (for incrementing counters, referring to the start of arrays etc, etc). One only has to look at the work of Feuerstein to see this - time and again in his books, the concept of delegating constant values (and subtypes) to well-defined areas (the "Single Source of Truth") rears it's head.
    Now in the Oracle world, data architecture generally has primacy, which in this case manifests itself as the use of foreign keys in data tables referencing the equivalent lookups (dimensional modelling, star diagrams and the rest) - thus even special, application-specific values, i.e. with no intrinsic value in the real world, end up in your ERD. There appears to be a bit difference of opinon, depending on the background of the developer.
    Hence my question - how do you, as developers, deal with these sorts of situations?

  • Devleoper Studio and Workplace for Java application development

    1. Can someone please help me know what is netweaver developer studio and developer workplace ?
    2. I came across EP master guide that says :
    You install an AS Java centrally and the developer workplace on each developer host.
    This option is recommended for large development projects with a need for local and central testing. Developers can test their developments locally using the developer workplace. For integration tests, they deploy their changes on the central AS Java.Ûê.® |
    You install an AS Java centrally and Developer Studio instances on each developer host.
    This option requires fewer hardware resources per developer host. In this landscape scenario you can set up an AS Java in development mode centrally and connect to it from the other hosts in the landscape using the Developer Studio.
    How does Java application development take place ? Does this say that we need Developer Studio instances installed on every developer's host (suppose I have 3 developers) or do I need one server setup for Dev Studio and developers using some client program to connect to it ?
    I suppose developers create programs on their local pc's, port them to Dev studio for testing and transfer them to EP Dev server ? Is that correct.. ? Thereby, we need CTS+ to promote developments to Qlty and Prod ?

    Hi,
    Please check-out the following links to understand NWDS, Developer studio and the installation procedure:
    SAP NetWeaver Developer Studio 7.1 Overview
    http://help.sap.com/saphelp_47x200/helpdata/en/cb/f4bc3d42f46c33e10000000a11405a/content.htm
    /thread/705153 [original link is broken]
    http://www.sapgeek.net/2010/05/how-to-install-nwds/
    Hope these help.
    Regards,
    Varun

  • Application Developed in Apex 3.2 is not working Apex 4.1

    Hi,
    We have few application developed using oracle apex 3.2 and oracle 10g enterprise edition as the database also have a Apache tomcat server configured in the system. These application are running fine have no issue in production instance.
    I recently installed Oracle Apex 4.1 and Oracle 10g Express Edition in my local machine to test these application in oracle 4.1. I took the same database dump and installed in the local machine no issues with the database side. I installed the application that are developed in 3.2 and installed in new local instance but this is not working.
    It show >>Internet Explorer cannot display the webpage
    Please suggest me what might be the issue how to trouble shoot this issue.
    Thanks
    Sudhir

    I don't understand reading so many tract file starting with
    xe_s00??Most of the trace file show about {Embedded PL/SQL Gateway:} it shows is metioned below
    Embedded PL/SQL Gateway: (wppr.c,495) Pl/sql block parsed...
    Embedded PL/SQL Gateway: (wpdenv.c,1528) CGI Environment has 33 vars. Max name len 128, Max Value Len 512
    Embedded PL/SQL Gateway: PLSQL_GATEWAY(14)=(6)WebDb
    Embedded PL/SQL Gateway: GATEWAY_IVERSION(17)=(2)2
    Embedded PL/SQL Gateway: SERVER_SOFTWARE(16)=(42)Oracle >>Embedded PL/SQL Gateway/10.2.0.1.0
    Embedded PL/SQL Gateway: GATEWAY_INTERFACE(18)=(8)CGI/1.1
    Embedded PL/SQL Gateway: SERVER_PORT(12)=(5)8080
    Embedded PL/SQL Gateway: SERVER_NAME(12)=(16)XDB HTTP Server
    Embedded PL/SQL Gateway: REQUEST_METHOD(15)=(4)GET
    Embedded PL/SQL Gateway: QUERY_STRING(13)=(36)p=4000:1500:3165980474173701::NO:::
    Embedded PL/SQL Gateway: PATH_INFO(10)=(3)/f
    Embedded PL/SQL Gateway: SCRIPT_NAME(12)=(6)/apex
    Embedded PL/SQL Gateway: REMOTE_HOST(12)=(1)
    Embedded PL/SQL Gateway: REMOTE_ADDR(12)=(10)127.0.0.1
    Embedded PL/SQL Gateway: SERVER_PROTOCOL(16)=(9)HTTP/1.1
    Embedded PL/SQL Gateway: REQUEST_PROTOCOL(17)=(4)tcp
    Embedded PL/SQL Gateway: REMOTE_USER(12)=(10)ANONYMOUS
    Embedded PL/SQL Gateway: ORACLE_SSO_USER(16)=(1)
    Embedded PL/SQL Gateway: HTTP_CONTENT_LENGTH(20)=(2)0
    Embedded PL/SQL Gateway: HTTP_CONTENT_TYPE(18)=(1)
    Embedded PL/SQL Gateway: HTTP_USER_AGENT(16)=(71)Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; WOW64; Trident/5.0)
    Embedded PL/SQL Gateway: HTTP_HOST(10)=(15)127.0.0.1:8080
    Embedded PL/SQL Gateway: HTTP_ACCEPT(12)=(10)text/html
    Embedded PL/SQL Gateway: HTTP_ACCEPT_ENCODING(21)=(5)gzip
    Embedded PL/SQL Gateway: HTTP_ACCEPT_LANGUAGE(21)=(6)en-IN
    Embedded PL/SQL Gateway: HTTP_COOKIE(12)=(344)ORA_WWV_REMEMBER_UN=SUDHIR:DEVELOPMENT; PUBLIC_ORA_WWV_USER_63113903525053=-1; ORA_WWV_USER_63113903525053=2AD259DDAF6770BE05FEF29509B68F5F; LOGIN_USERNAME_COOKIE=sudhir; WWV_CUSTOM-F_1638610445865577_111=CC022FAB479054CF32361C22B789D7EE; ORA_WWV_ATTRIBUTE_PAGE=4495%2C%23ALL; WWV_CUSTOM-F_1638610445865577_107=19291B8FC60BA1A47916DE46EE85F01C
    Embedded PL/SQL Gateway: HTTP_REFERER(13)=(127)http://127.0.0.1:8080/apex/f?p=4000:1:3165980474173701::NO:RP:FB_FLOW_ID,F4000_P1_FLOW,P0_FLOWPAGE,RECENT_PAGES:109,109,109;|,
    Embedded PL/SQL Gateway: WEB_AUTHENT_PREFIX(19)=(1)
    Embedded PL/SQL Gateway: DAD_NAME(9)=(5)apex
    Embedded PL/SQL Gateway: DOC_ACCESS_PATH(16)=(5)docs
    Embedded PL/SQL Gateway: DOCUMENT_TABLE(15)=(23)wwv_flow_file_objects$
    Embedded PL/SQL Gateway: PATH_ALIAS(11)=(1)
    Embedded PL/SQL Gateway: REQUEST_CHARSET(16)=(9)AL32UTF8
    Embedded PL/SQL Gateway: REQUEST_IANA_CHARSET(21)=(6)UTF-8
    Embedded PL/SQL Gateway: SCRIPT_PREFIX(14)=(1)
    *** ACTION NAME:() 2012-06-27 16:52:32.213
    *** MODULE NAME:() 2012-06-27 16:52:32.213
    *** CLIENT ID:() 2012-06-27 16:52:32.212
    Embedded PL/SQL Gateway: (wpd.c,1960) Begin header parsing...
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (47 bytes): X-ORACLE-IGNORE: IGNORE
    Embedded PL/SQL Gateway: (wpd.c,2027) X-ORACLE-IGNORE parsed
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (47 bytes): X-ORACLE-IGNORE: IGNORE
    Embedded PL/SQL Gateway: (wpd.c,2027) X-ORACLE-IGNORE parsed
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (47 bytes): X-ORACLE-IGNORE: IGNORE
    Embedded PL/SQL Gateway: (wpd.c,2027) X-ORACLE-IGNORE parsed
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (47 bytes): X-ORACLE-IGNORE: IGNORE
    Embedded PL/SQL Gateway: (wpd.c,2027) X-ORACLE-IGNORE parsed
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (77 bytes): Content-type: text/html; charset=utf-8
    Embedded PL/SQL Gateway: (wpd.c,2108) Parsed header - Content-Type:text/html; charset=utf-8
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (43 bytes): X-Frame-Options: DENY
    Embedded PL/SQL Gateway: (wpd.c,2168) Parsed header - X-Frame-Options:DENY
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (43 bytes): Content-length: 40155
    Embedded PL/SQL Gateway: (wpd.c,2082) Parsed header - Content-Length:40155
    Embedded PL/SQL Gateway: (wpd.c,2009) Got a line (1 bytes):
    Embedded PL/SQL Gateway: (wpd.c,2016) End of headers detected
    Embedded PL/SQL Gateway: (wpd.c,1882) Parsed begin
    owa.get_page(:data__,:ndata__);
    end;
    Embedded PL/SQL Gateway: (wpd.c,1898) Bound drec variablesEmbedded PL/SQL Gateway: (wpcs.c, 76) Executed 'begin dbms_session.reset_package; end;' (rc=0)
    Embedded PL/SQL Gateway: (wpd.c,1818) Going to close cursor
    Embedded PL/SQL Gateway: (wpx.c,660) Normal completion
    Embedded PL/SQL Gateway: (wpx.c,688) Shutdown has been called
    Embedded PL/SQL Gateway: (wpx.c,700) Going to logoff
    *** SESSION ID:(527.737) 2012-06-27 16:52:37.533
    Embedded PL/SQL Gateway: (wpdenv.c,679) script_name='/apex' path_info='/f'script_prefix='' dad_name='apex'
    Embedded PL/SQL Gateway: (wpdenv.c,792) User-Agent is Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; WOW64; Trident/5.0)
    Embedded PL/SQL Gateway: (wpdenv.c,1412) dadname = 'apex', path_info = 'f'
    Embedded PL/SQL Gateway: (wpdenv.c,1451) Service will NOT use dynamic auth
    Embedded PL/SQL Gateway: (wpx.c,394) Initialized successfully 0
    Embedded PL/SQL Gateway: (wpx.c,316) SetRemoteUser : Remote User set to ANONYMOUS for this request.
    Embedded PL/SQL Gateway: (wpx.c,480) Auth info from .APP file is being used
    Embedded PL/SQL Gateway: (wpd.c,1729) Attempting to logon with '(unknown)'
    Embedded PL/SQL Gateway: (wpu.c,1499) DBCharSet=>AMERICAN_AMERICA.AL32UTF8 OWAVersion=10.1.2.0.4 (101204) OWAMatch=>1 (rc=0)
    Embedded PL/SQL Gateway: (wpd.c,1763) Logged in as (unknown)
    Embedded PL/SQL Gateway: (wpx.c,593) Going to select...
    Embedded PL/SQL Gateway: (wpx.c,647) Have been asked to execute a request
    Embedded PL/SQL Gateway: (wppa.c,335) Building Arglist based on Parsed Content from WRB
    Embedded PL/SQL Gateway: (wppa.c,1015) Enter ParseUrlData
    Embedded PL/SQL Gateway: GET
    Embedded PL/SQL Gateway: (wppa.c,1064) Getting Values from QUERY_STRING
    Embedded PL/SQL Gateway: [headers begin]
    Embedded PL/SQL Gateway: HTTP_COOKIE(11): (343)ORA_WWV_REMEMBER_UN=SUDHIR:DEVELOPMENT; PUBLIC_ORA_WWV_USER_63113903525053=-1; ORA_WWV_USER_63113903525053=2AD259DDAF6770BE05FEF29509B68F5F; LOGIN_USERNAME_COOKIE=sudhir; WWV_CUSTOM-F_1638610445865577_111=CC022FAB479054CF32361C22B789D7EE; ORA_WWV_ATTRIBUTE_PAGE=4495%2C%23ALL; WWV_CUSTOM-F_1638610445865577_107=19291B8FC60BA1A47916DE46EE85F01C
    Embedded PL/SQL Gateway: [headers end]
    Embedded PL/SQL Gateway: p=109:1:3165980474173701:::::
    Embedded PL/SQL Gateway: Identifier legal p(notconverted)
    Embedded PL/SQL Gateway: (wppa.c,1514) indx = 1, entryCnt = 1
    Embedded PL/SQL Gateway: (wppa.c,1834) Listing distinct actual names:
    Embedded PL/SQL Gateway: (wppa.c,1836) p
    Embedded PL/SQL Gateway: (wppa.c,1838) Listing actuals of array with large entries:
    Embedded PL/SQL Gateway: (wppa.c,1843) Listing distinct actual names and values:
    Embedded PL/SQL Gateway: (wppa.c,1875) p, type = 0, value (28) = 109:1:3165980474173701:::::
    Embedded PL/SQL Gateway: (wppa.c,430) Arglist built, 1 unique entries
    Embedded PL/SQL Gateway: (wpx.c,654) Going to wpprodb_OciDoBlock...
    Embedded PL/SQL Gateway: Identifier legal f(notconverted)
    Embedded PL/SQL Gateway: (wppr.c,435) start working with f
    Embedded PL/SQL Gateway: (wppr.c,1017) lcase(procname): f
    Embedded PL/SQL Gateway: (wppr.c,1296) The CALL block: len=1067, bind_count=9What is the next step to make the 3.2 developed application work.
    Thanks
    Sudhir
    Edited by: Sudhir_Meru on Jun 27, 2012 5:09 AM
    Edited by: Sudhir_Meru on Jun 27, 2012 5:11 AM

  • JSP application developed for PC does not work on IPAQ

    Hi all,
    I have developed a database/web application using JSP, Apache Tomcat, MS Access. It works great on Windows PC. Now I want to make sure that it works on PDAs also.
    I tried using an IPAQ which runs Win CE 3.0 but I get some errors while interacting with the database.
    One of the webpages in the application has a form through which the user can submit a text message and upload a file. Following are the problems that I am facing:
    1. The 'UPLOAD' button through which the user can browse his local hard drive and select a file is invisible if I am browsing with an IPAQ. I am using multipart formdata (MultipartRequest.java provided by Oreilly) for submitting a message and uploading the file. What could be the problem?
    2. If I forget about the attachment file and just try to submit a text message (which works on PCs) , I get the error message: "IOEXception: submitted data is not multipart formdata".
    I am very new to PDA application development, and any inputs will be very valuable.
    Thanks in advance.
    m_asu.

    I was wondering if anybody can give me some pointers as to how can I make an application developed for PCs work on IPaqs.
    Thanks
    m_asu

  • ABAP Code for Backup the entire table data in the application server

    Hello Friends,
    I have to create the table data Backup and Store the entire table data in the application server and also be able to restore the data back if needed.
    this should be dynamic program for any table based on the table name given on the application server.. I have developed a program for this but its having problems with the Quantity, amount. Its not writing it correctly at the application level.
    ANy Suggestions on this.
    Below is the program for this.
    Thanks,
    Ster.
    * Report  YWMM_TABLE_DUMP                                             *
    REPORT ywmm_table_dump .
    TABLES :
            dd03l.
    * Type spool declaration
    TYPE-POOLS:
            abap, slis.
    DATA: i_table_data1  TYPE REF TO data.
    DATA : it_dd03l LIKE dd03l OCCURS 0 WITH HEADER LINE.
    *DATA : gt_fieldcat TYPE lvc_s_fcat.
    DATA : i_fcat      TYPE STANDARD TABLE OF lvc_s_fcat,
           l_dr_line         TYPE   REF TO data,
           l_v_as4vers       TYPE as4vers.
    FIELD-SYMBOLS: <f_table_data1>     TYPE STANDARD TABLE,
                   <f_wa_table_data1>  TYPE ANY.
    SELECTION-SCREEN: BEGIN OF BLOCK bl1 WITH FRAME TITLE text-001.
    PARAMETERS: rb_copy RADIOBUTTON GROUP map DEFAULT 'X',
                rb_rest RADIOBUTTON GROUP map.
    SELECTION-SCREEN: END   OF BLOCK bl1.
    SELECTION-SCREEN: BEGIN OF BLOCK bl2 WITH FRAME TITLE text-002.
    PARAMETERS: p_table  TYPE tabname OBLIGATORY,
                p_plfld TYPE dd03l-fieldname.
    SELECTION-SCREEN SKIP 1.
    PARAMETERS: p_bkfile TYPE localfile OBLIGATORY.
    SELECTION-SCREEN: END   OF BLOCK bl2.
    PERFORM get_data.
    IF rb_copy = 'X'.
      PERFORM backup.
    ELSEIF rb_rest = 'X'.
      PERFORM database_update.
    ENDIF.
    *&      Form  get_data
    FORM get_data.
      CLEAR   i_fcat.
      REFRESH i_fcat.
      CALL FUNCTION 'LVC_FIELDCATALOG_MERGE'
           EXPORTING
                i_structure_name = p_table  " Table Name
           CHANGING
                ct_fieldcat      = i_fcat
           EXCEPTIONS
                OTHERS           = 1.
      CALL METHOD cl_alv_table_create=>create_dynamic_table
      EXPORTING
        it_fieldcatalog = i_fcat
      IMPORTING
        ep_table        = i_table_data1.
      IF sy-subrc = 0.
        ASSIGN i_table_data1->* TO <f_table_data1>.
      ELSE.
        WRITE: 'Error creating internal table'.
      ENDIF.
      IF rb_copy = 'X'.
        SELECT  * FROM (p_table) INTO CORRESPONDING FIELDS OF
                  TABLE <f_table_data1> UP TO 20 ROWS.
      ELSEIF rb_rest = 'X'.
        CREATE DATA l_dr_line LIKE LINE OF <f_table_data1>.
        ASSIGN l_dr_line->* TO <f_wa_table_data1>.
    *Get Data from Application Server
    * Opening the dataset P_BKFILE given in the selection screen
        TRANSLATE p_bkfile TO LOWER CASE.
        OPEN DATASET p_bkfile FOR INPUT IN TEXT MODE." ENCODING DEFAULT.
        IF sy-subrc NE 0.
    *    MESSAGE:
        ELSE.
          DO.
    * Reading the file from application server
            READ DATASET p_bkfile INTO <f_wa_table_data1>.
            IF sy-subrc = 0.
              APPEND <f_wa_table_data1> TO <f_table_data1>.
            ELSE.
              EXIT.
            ENDIF.
          ENDDO.
    * Closing the dataset
          CLOSE DATASET p_bkfile.
        ENDIF.
      ENDIF.
    ENDFORM.                    " get_data
    *&      Form  backup
    *       text
    *  -->  p1        text
    *  <--  p2        text
    FORM backup.
      TRANSLATE p_bkfile TO LOWER CASE.
      OPEN DATASET p_bkfile FOR OUTPUT IN TEXT MODE.
      IF sy-subrc NE 0.
        WRITE: text-017.
        STOP.
      ELSE.
        LOOP AT <f_table_data1> ASSIGNING <f_wa_table_data1>.
          TRANSFER <f_wa_table_data1> TO p_bkfile.
        ENDLOOP.
      ENDIF.
      CLOSE DATASET p_bkfile.
    ENDFORM.                    " backup
    *&      Form  database_update
    FORM database_update.
      DATA : i_mara_u TYPE STANDARD TABLE OF mara WITH HEADER LINE,
             i_ekpo_u TYPE STANDARD TABLE OF ekpo WITH HEADER LINE,
             i_eban_u TYPE STANDARD TABLE OF eban WITH HEADER LINE,
             i_resb_u TYPE STANDARD TABLE OF resb WITH HEADER LINE,
             i_plpo_u TYPE STANDARD TABLE OF plpo WITH HEADER LINE,
             i_stpo_u TYPE STANDARD TABLE OF stpo WITH HEADER LINE,
             i_vbap_u TYPE STANDARD TABLE OF vbap WITH HEADER LINE,
             i_vbrp_u TYPE STANDARD TABLE OF vbrp WITH HEADER LINE,
             i_lips_u TYPE STANDARD TABLE OF lips WITH HEADER LINE,
             i_afvc_u TYPE STANDARD TABLE OF afvc WITH HEADER LINE,
             i_asmd_u TYPE STANDARD TABLE OF asmd WITH HEADER LINE,
    *           i_cooi_u TYPE STANDARD TABLE OF cooi WITH HEADER LINE,
             i_qmel_u TYPE STANDARD TABLE OF qmel WITH HEADER LINE,
             i_cooi_u TYPE STANDARD TABLE OF cooi WITH HEADER LINE,
             i_esll_u TYPE STANDARD TABLE OF esll WITH HEADER LINE,
             i_t165_u  TYPE STANDARD TABLE OF t165 WITH HEADER LINE,
             i_t165e_u TYPE STANDARD TABLE OF t165e WITH HEADER LINE,
             i_twpko_u TYPE STANDARD TABLE OF twpko WITH HEADER LINE,
             i_tpext_u TYPE STANDARD TABLE OF tpext WITH HEADER LINE,
             i_ce4mxpa_u TYPE STANDARD TABLE OF ce4mxpa WITH HEADER LINE,
             i_ce4mxpa_acct_u TYPE STANDARD TABLE OF ce4mxpa_acct WITH
                                                             HEADER LINE,
             i_zaim_u  TYPE STANDARD TABLE OF zaim WITH HEADER LINE,
             i_s012_d TYPE STANDARD TABLE OF s012 WITH HEADER LINE,
             i_s012_i TYPE STANDARD TABLE OF s012 WITH HEADER LINE,
             i_dummy  TYPE STANDARD TABLE OF mara.
      CASE p_table.
        WHEN 'MARA'.
    *     Non-Key
          PERFORM move_to_table USING   <f_table_data1>
                                CHANGING i_mara_u[]
                                         i_mara_u.
          PERFORM update_table USING i_mara_u[].
      ENDCASE.
    ENDFORM.                    " database_update
    *&      Form  move_to_mara
    FORM move_to_table USING    p_tab_from TYPE STANDARD TABLE
                       CHANGING p_tab_to   TYPE STANDARD TABLE
                                p_w_table.
      DATA:  l_wa_fcat TYPE lvc_s_fcat.
      FIELD-SYMBOLS: <f_field_from> TYPE ANY,
                     <f_field_to>   TYPE ANY.
      LOOP AT p_tab_from ASSIGNING <f_wa_table_data1>.
        LOOP AT i_fcat INTO l_wa_fcat.
          ASSIGN COMPONENT l_wa_fcat-fieldname
         OF STRUCTURE <f_wa_table_data1> TO <f_field_from>.
          ASSIGN COMPONENT l_wa_fcat-fieldname
         OF STRUCTURE p_w_table TO <f_field_to>.
          <f_field_to> = <f_field_from>.
        ENDLOOP.
        APPEND p_w_table TO p_tab_to.
      ENDLOOP.
    ENDFORM.                    " move_to_mara
    *&      Form  update_table
    FORM update_table  USING p_table_update TYPE STANDARD TABLE.
      SELECT SINGLE *
        FROM dd03l
       WHERE fieldname = p_plfld
         AND tabname   = p_table
         AND keyflag   <> 'X'
         AND as4local = 'A'
         AND   as4vers = l_v_as4vers
         AND   ( comptype = 'E' OR comptype = space ).
      IF sy-subrc = 0.
    *   Do update
        IF NOT p_table_update IS INITIAL.
          UPDATE (p_table) FROM TABLE p_table_update.
          IF sy-subrc = 0.
            COMMIT WORK.
          ELSE.
            ROLLBACK WORK.
            WRITE: text-003.
            STOP.
          ENDIF.
        ENDIF.
      ELSE.
    *delete and insert.
        IF NOT p_table_update IS INITIAL.
    *      DELETE (p_table).
          IF sy-subrc = 0.
            INSERT (p_table) FROM TABLE p_table_update.
            IF sy-subrc = 0.
              COMMIT WORK.
            ELSE.
              ROLLBACK WORK.
              WRITE: text-018.
              STOP.
            ENDIF.
          ELSE.
            ROLLBACK WORK.
            WRITE: text-018.
            STOP.
          ENDIF.
        ENDIF.
      ENDIF.
    ENDFORM.                    " update_table
    Edited by: Julius Bussche on Jul 18, 2008 1:43 PM
    Please use a meaningfull subject title!

    ARS,
    I am struggling a bit to get this.
    there is a syntax error,
    Field "FIELDS_INT-TYPE" is unknown. It is neither in one of thespecified tables nor defined by a "DATA" statement.     
    Again you have asked to move to a diffrent table. What is that table and how to build it.
        LOOP AT <f_table_data1> ASSIGNING <f_wa_table_data1>.
          LOOP AT i_fcat INTO l_fcat.
            IF l_fcat-inttype EQ 'P'.
              ASSIGN COMPONENT l_fcat-fieldname
                  OF STRUCTURE <f_wa_table_data1> TO <f_field>
                  TYPE     fields_int-type
                  DECIMALS fields_int-decimals.
            ELSE.
              ASSIGN COMPONENT l_fcat-fieldname
                  OF STRUCTURE <f_wa_table_data1> TO <f_field>
                  TYPE     fields_int-type.
            ENDIF.
            " Move <f_field> to a new table and use this table for download
          ENDLOOP.
          TRANSFER <f_wa_table_data1> TO p_bkfile.
        ENDLOOP.
    Ster

  • Firefox is creating endless copies of C:\Documents and Settings\Catain Kirk\Local Settings\Application Data\Mozilla\Firefox\Profiles\afbrfhnj.default\Cache ... how do I stop it

    I just installed 9.0.1 and it continually creates copies of C:\Documents and Settings\Catain Kirk\Local Settings\Application Data\Mozilla\Firefox\Profiles\afbrfhnj.default\Cache ... how do I stop it !!!
    I'm running Windows XP Version 5.1 SP3.

    Hi,
    That is the disk cache of Firefox, where contents of loaded pages are kept for faster retrieval and it also helps to conserve bandwidth. To disable this: '''Tools''' ('''Alt''' + '''T''') > '''Options''' > '''Advanced''' > '''Network''', check (tick) '''Override automatic cache management ''' and put the value at zero.
    [https://support.mozilla.org/en-US/kb/Options%20window%20-%20Advanced%20panel?as=u#advanced_network Cache]
    Useful links:
    [https://support.mozilla.com/en-US/kb/Options%20window All about Tools > Options]
    [http://kb.mozillazine.org/About:config Going beyond Tools > Options - about:config]
    [http://kb.mozillazine.org/About:config_entries about:config Entries]
    [https://support.mozilla.com/en-US/kb/Page%20Info%20window Page Info] Tools (Alt + T) > Page Info, Right-click > View Page Info
    [https://support.mozilla.com/en-US/kb/Keyboard%20shortcuts Keyboard Shortcuts]
    [https://support.mozilla.com/en-US/kb/Viewing%20video%20in%20Firefox%20without%20a%20plugin Viewing Video without Plugins]
    [http://kb.mozillazine.org/Profile_folder_-_Firefox Firefox Profile Folder & Files]
    [https://developer.mozilla.org/en/Command_Line_Options#Browser Firefox Commands]
    [https://support.mozilla.com/en-US/kb/Basic%20Troubleshooting Basic Troubleshooting]
    [https://support.mozilla.com/en-US/kb/common-questions-after-upgrading-firefox-36 After Upgrading]
    [https://support.mozilla.com/en-US/kb/Safe%20Mode Safe Mode]
    [http://kb.mozillazine.org/Problematic_extensions Problematic Extensions]
    [https://support.mozilla.com/en-US/kb/Troubleshooting%20extensions%20and%20themes Troubleshooting Extensions and Themes]
    [https://support.mozilla.com/en-US/kb/Troubleshooting%20plugins Troubleshooting Plugins]
    [http://kb.mozillazine.org/Testing_plugins Testing Plugins]

  • Entity framework code first inheritance table per type problem

    hello
    i am using ODP.NET version 11.2.0.3.20. i am trying to get entity framework code first to work. one problem i have encountered is with inheritance. i have a "table per type" inheritance scenario. i have 3 tables involved. the base table is called "S_PERIOD" which maps to the base "Period" class. i have a derived table called "S_SEASON_QUARTER" that maps to the derived "Quarter" class. And I have another derived table called "S_SEASON_PRICE_PERIOD" that maps to the derived "PriceBreak" class. for testing purposes i am trying to load all records. the problem is that every class instance ends up being of type "Quarter", which is incorrect. for whatever reason the provider thinks that every "S_PERIOD" record has a corresponding "S_SEASON_QUARTER" record. i took a look at the generated SQL and posted it down below. looking at this SQL it's clear to me why every object ends up being of type "Quarter". It looks like the provider is checking the "C2" field to determine the concrete type to instantiate. the problem however is that the "C2" field value provided by each derived table is ALWAYS equal to "1", which is a problem since this will result in the "case" statement always hitting the second "when" condition. does anyone have any idea how to fix this? thank you
    SELECT
    CASE WHEN (( NOT (("Project1"."C2" = 1) AND ("Project1"."C2" IS NOT NULL))) AND ( NOT (("Project2"."C2" = 1) AND ("Project2"."C2" IS NOT NULL)))) THEN '0X' WHEN (("Project1"."C2" = 1) AND ("Project1"."C2" IS NOT NULL)) THEN '0X0X' ELSE '0X1X' END AS "C1",
    CAST( "Extent1"."PERIOD_ID" AS number(10,0)) AS "C2",
    "Extent1"."START_DATE" AS "START_DATE",
    "Extent1"."END_DATE" AS "END_DATE",
    "Extent1"."NAME" AS "NAME",
    "Extent1"."TYPE_CODE" AS "TYPE_CODE",
    CAST( "Extent1"."CREATE_USER_ID" AS number(10,0)) AS "C3",
    "Extent1"."CREATE_DATE" AS "CREATE_DATE",
    CAST( "Extent1"."MODIFY_USER_ID" AS number(10,0)) AS "C4",
    "Extent1"."MODIFY_DATE" AS "MODIFY_DATE",
    CASE WHEN (( NOT (("Project1"."C2" = 1) AND ("Project1"."C2" IS NOT NULL))) AND ( NOT (("Project2"."C2" = 1) AND ("Project2"."C2" IS NOT NULL)))) THEN NULL WHEN (("Project1"."C2" = 1) AND ("Project1"."C2" IS NOT NULL)) THEN "Project1"."QUARTER_NAME" END AS "C5"
    FROM "DBO_SPACE_DEV"."S_PERIOD" "Extent1"
    LEFT OUTER JOIN (SELECT
         "Extent2"."QUARTER_NAME" AS "QUARTER_NAME",
         CAST( "Extent2"."QUARTER_ID" AS number(10,0)) AS "C1",
         1 AS "C2"
         FROM "DBO_SPACE_DEV"."S_SEASON_QUARTER" "Extent2" ) "Project1" ON ( CAST( "Extent1"."PERIOD_ID" AS number(10,0))) = "Project1"."C1"
    LEFT OUTER JOIN (SELECT
         CAST( "Extent3"."PRICE_PERIOD_ID" AS number(10,0)) AS "C1",
         1 AS "C2"
         FROM "DBO_SPACE_DEV"."S_SEASON_PRICE_PERIOD" "Extent3" ) "Project2" ON ( CAST( "Extent1"."PERIOD_ID" AS number(10,0))) = "Project2"."C1"
    Edited by: 997830 on Apr 3, 2013 8:40 AM

    An update:
    I tried again following this example to the letter:
    Using NuGet to Install and Configure Oracle Data Provider for .NET
    This time I used a console application as described in the example. Yes, I rebuilt the project after the NuGet install.
    I made the appropriate mods to App.config. I get the same error message as with the MVC example above.
    Does the ODP.Net driver really work with EF? If so, can anyone provide me with a working sample project?

  • How to improve speed of queries that use ORM one table per concrete class

    Hi,
    Many tools that make ORM (Object Relational Mapping) like Castor, Hibernate, Toplink, JPOX, etc.., have the one table per concrete class feature that maps objects to follow structure:
    CREATE TABLE ABSTRACTPRODUCT (
        ID VARCHAR(8) NOT NULL,
        DESCRIPTION VARCHAR(60) NOT NULL,
        PRIMARY KEY(ID)
    CREATE TABLE PRODUCT (
        ID VARCHAR(8) NOT NULL REFERENCES ABSTRACTPRODUCT(ID),
        CODE VARCHAR(10) NOT NULL,
        PRICE DECIMAL(12,2),
        PRIMARY KEY(ID)
    CREATE UNIQUE INDEX iProduct ON Product(code)
    CREATE TABLE BOOK (
        ID VARCHAR(8) NOT NULL REFERENCES PRODUCT(ID),
        AUTHOR VARCHAR(60) NOT NULL,
        PRIMARY KEY (ID)
    CREATE TABLE COMPACTDISK (
        ID VARCHAR(8) NOT NULL REFERENCES PRODUCT(ID),
        ARTIST VARCHAR(60) NOT NULL,
        PRIMARY KEY(ID)
    there is a way to improve queries like
    SELECT
        pd.code CODE,   
        abpd.description DESCRIPTION,
        DECODE(bk.id,NULL,cd.artist,bk.author) PERSON
    FROM
        ABSTRACTPRODUCT abpd,
        PRODUCT pd,
        BOOK bk,
        COMPACTDISK cd
    WHERE
        pd.id = abpd.id AND
        bk.id(+) = abpd.id AND
        cd.id(+) = abpd.id AND
        pd.code like '101%'
    or like this:
    SELECT
        pd.code CODE,   
        abpd.description DESCRIPTION,
        DECODE(bk.id,NULL,cd.artist,bk.author) PERSON
    FROM
        ABSTRACTPRODUCT abpd,
        PRODUCT pd,
        BOOK bk,
        COMPACTDISK cd
    WHERE
        pd.id = abpd.id AND
        bk.id(+) = abpd.id AND
        cd.id(+) = abpd.id AND
        abpd.description like '%STARS%' AND
        pd.price BETWEEN 1 AND 10
    think in a table with many rows, then exists something inside MaxDB to improve this type of queries? like some anotations on SQL? or declare tables that extends another by PK? on other databases i managed this using Materialized Views, but i think that this can be faster just using PK, i'm wrong? the better is to consolidate all tables in one table? what is the impact on database size with this consolidation?
    note: with consolidation i will miss NOT NULL constraint at database side.
    thanks for any insight.
    Clóvis

    Hi Lars,
    i dont understand because the optimizer get that Index for TM at execution plan, and because dont use the join via KEY column, note the WHERE clause is "TM.OID = MF.MY_TIPO_MOVIMENTO" by the key column, and the optimizer uses an INDEX that the indexed column is ID_SYS, that isnt and cant be a primary key, because its not UNIQUE, follow the index columns:
    indexes of TipoMovimento
    INDEXNAME     COLUMNNAME          SORT     COLUMNNO     DATATYPE     LEN     INDEX_USED     FILESTATE     DISABLED
    ITIPOMOVIMENTO     TIPO               ASC     1          VARCHAR          2     220546          OK          NO
    ITIPOMOVIMENTO     ID_SYS               ASC     2          CHAR          6     220546          OK          NO
    ITIPOMOVIMENTO     MY_CONTA_DEBITO          ASC     3          CHAR          8     220546          OK          NO
    ITIPOMOVIMENTO     MY_CONTA_CREDITO     ASC     4          CHAR          8     220546          OK          NO
    ITIPOMOVIMENTO1     ID_SYS               ASC     1          CHAR          6     567358          OK          NO
    ITIPOMOVIMENTO2     DESCRICAO          ASC     1          VARCHAR          60     94692          OK          NO
    after i create the index iTituloCobrancaX7 on TituloCobranca(OID,DATA_VENCIMENTO) in a backup instance and get surprised with the follow explain:
    OWNER     TABLENAME     COLUMN_OR_INDEX          STRATEGY                    PAGECOUNT     
         TC          ITITULOCOBRANCA1     RANGE CONDITION FOR INDEX          5368     
                   DATA_VENCIMENTO               (USED INDEX COLUMN)          
         MF          OID               JOIN VIA KEY COLUMN               9427     
         TM          OID               JOIN VIA KEY COLUMN               22     
                                  TABLE HASHED          
         PS          OID               JOIN VIA KEY COLUMN               1350     
         BOL          OID               JOIN VIA KEY COLUMN               497     
                                       NO TEMPORARY RESULTS CREATED          
         JDBC_CURSOR_19                    RESULT IS COPIED   , COSTVALUE IS     988
    note that now the optimizer gets the index ITITULOCOBRANCA1 as i expected, if i drop the new index iTituloCobrancaX7 the optimizer still getting this execution plan, with this the query executes at 110 ms, with that great news i do same thing in the production system, but the execution plan dont changes, and i still getting a long execution time this time at 413516 ms. maybe the problem is how optimizer measure my tables.
    i checked in DBAnalyser that the problem is catalog cache hit rate (we discussed this at [catalog cache hit rate, how to increase?|;
    ) and the low selectivity of this SQL command, then its because of this that to achieve a better selectivity i must have an index with, MF.MY_SACADO, MF.TIPO and TC.DATA_VENCIMENTO, as explained in previous posts, since this type of index inside MaxDB isnt possible, i have no choice to speed this type of query without changing tables structure.
    MaxDB developers can develop this type of index? or a feature like this dont have any plans to be made?
    if no, i must create another schema, to consolidate tables to speed queries on my system, but with this consolidation i will get more overhead, i must solve the less selectivity because i think if the data on tables increase, the query becomes impossible, i see that CREATE INDEX supports FUNCTION, maybe a   FUNCTION that join data of two tables can solve this?
    about instance configuration it is:
    Machine:
    Version:       '64BIT Kernel'
    Version:       'X64/LIX86 7.6.03   Build 007-123-157-515'
    Version:       'FAST'
    Machine:       'x86_64'
    Processors:    2 ( logical: 8, cores: 8 )
    data volumes:
    ID     MODE     CONFIGUREDSIZE     USABLESIZE     USEDSIZE     USEDSIZEPERCENTAGE     DROPVOLUME     TOTALCLUSTERAREASIZE     RESERVEDCLUSTERAREASIZE     USEDCLUSTERAREASIZE     PATH     
    1     NORMAL     4194304          4194288          379464          9               NO          0               0               0               /db/SPDT/data/data01.dat     
    2     NORMAL     4194304          4194288          380432          9               NO          0               0               0               /db/SPDT/data/data02.dat     
    3     NORMAL     4194304          4194288          379184          9               NO          0               0               0               /db/SPDT/data/data03.dat     
    4     NORMAL     4194304          4194288          379624          9               NO          0               0               0               /db/SPDT/data/data04.dat     
    5     NORMAL     4194304          4194288          380024          9               NO          0               0               0               /db/SPDT/data/data05.dat
    log volumes:
    ID     CONFIGUREDSIZE     USABLESIZE     PATH               MIRRORPATH
    1     51200          51176          /db/SPDT/log/log01.dat     ?
    parameters:
    KERNELVERSION                         KERNEL    7.6.03   BUILD 007-123-157-515
    INSTANCE_TYPE                         OLTP
    MCOD                                  NO
    _SERVERDB_FOR_SAP                     YES
    _UNICODE                              NO
    DEFAULT_CODE                          ASCII
    DATE_TIME_FORMAT                      ISO
    CONTROLUSERID                         DBM
    CONTROLPASSWORD                       
    MAXLOGVOLUMES                         2
    MAXDATAVOLUMES                        11
    LOG_VOLUME_NAME_001                   /db/SPDT/log/log01.dat
    LOG_VOLUME_TYPE_001                   F
    LOG_VOLUME_SIZE_001                   6400
    DATA_VOLUME_NAME_0005                 /db/SPDT/data/data05.dat
    DATA_VOLUME_NAME_0004                 /db/SPDT/data/data04.dat
    DATA_VOLUME_NAME_0003                 /db/SPDT/data/data03.dat
    DATA_VOLUME_NAME_0002                 /db/SPDT/data/data02.dat
    DATA_VOLUME_NAME_0001                 /db/SPDT/data/data01.dat
    DATA_VOLUME_TYPE_0005                 F
    DATA_VOLUME_TYPE_0004                 F
    DATA_VOLUME_TYPE_0003                 F
    DATA_VOLUME_TYPE_0002                 F
    DATA_VOLUME_TYPE_0001                 F
    DATA_VOLUME_SIZE_0005                 524288
    DATA_VOLUME_SIZE_0004                 524288
    DATA_VOLUME_SIZE_0003                 524288
    DATA_VOLUME_SIZE_0002                 524288
    DATA_VOLUME_SIZE_0001                 524288
    DATA_VOLUME_MODE_0005                 NORMAL
    DATA_VOLUME_MODE_0004                 NORMAL
    DATA_VOLUME_MODE_0003                 NORMAL
    DATA_VOLUME_MODE_0002                 NORMAL
    DATA_VOLUME_MODE_0001                 NORMAL
    DATA_VOLUME_GROUPS                    1
    LOG_BACKUP_TO_PIPE                    NO
    MAXBACKUPDEVS                         2
    LOG_MIRRORED                          NO
    MAXVOLUMES                            14
    LOG_IO_BLOCK_COUNT                    8
    DATA_IO_BLOCK_COUNT                   64
    BACKUP_BLOCK_CNT                      64
    _DELAY_LOGWRITER                      0
    LOG_IO_QUEUE                          50
    _RESTART_TIME                         600
    MAXCPU                                8
    MAX_LOG_QUEUE_COUNT                   0
    USED_MAX_LOG_QUEUE_COUNT              8
    LOG_QUEUE_COUNT                       1
    MAXUSERTASKS                          500
    _TRANS_RGNS                           8
    _TAB_RGNS                             8
    _OMS_REGIONS                          0
    _OMS_RGNS                             7
    OMS_HEAP_LIMIT                        0
    OMS_HEAP_COUNT                        8
    OMS_HEAP_BLOCKSIZE                    10000
    OMS_HEAP_THRESHOLD                    100
    OMS_VERS_THRESHOLD                    2097152
    HEAP_CHECK_LEVEL                      0
    _ROW_RGNS                             8
    RESERVEDSERVERTASKS                   16
    MINSERVERTASKS                        28
    MAXSERVERTASKS                        28
    _MAXGARBAGE_COLL                      1
    _MAXTRANS                             4008
    MAXLOCKS                              120080
    _LOCK_SUPPLY_BLOCK                    100
    DEADLOCK_DETECTION                    4
    SESSION_TIMEOUT                       180
    OMS_STREAM_TIMEOUT                    30
    REQUEST_TIMEOUT                       5000
    _IOPROCS_PER_DEV                      2
    _IOPROCS_FOR_PRIO                     0
    _IOPROCS_FOR_READER                   0
    _USE_IOPROCS_ONLY                     NO
    _IOPROCS_SWITCH                       2
    LRU_FOR_SCAN                          NO
    _PAGE_SIZE                            8192
    _PACKET_SIZE                          131072
    _MINREPLY_SIZE                        4096
    _MBLOCK_DATA_SIZE                     32768
    _MBLOCK_QUAL_SIZE                     32768
    _MBLOCK_STACK_SIZE                    32768
    _MBLOCK_STRAT_SIZE                    16384
    _WORKSTACK_SIZE                       8192
    _WORKDATA_SIZE                        8192
    _CAT_CACHE_MINSIZE                    262144
    CAT_CACHE_SUPPLY                      131072
    INIT_ALLOCATORSIZE                    262144
    ALLOW_MULTIPLE_SERVERTASK_UKTS        NO
    _TASKCLUSTER_01                       tw;al;ut;2000*sv,100*bup;10*ev,10*gc;
    _TASKCLUSTER_02                       ti,100*dw;63*us;
    _TASKCLUSTER_03                       equalize
    _DYN_TASK_STACK                       NO
    _MP_RGN_QUEUE                         YES
    _MP_RGN_DIRTY_READ                    DEFAULT
    _MP_RGN_BUSY_WAIT                     DEFAULT
    _MP_DISP_LOOPS                        2
    _MP_DISP_PRIO                         DEFAULT
    MP_RGN_LOOP                           -1
    _MP_RGN_PRIO                          DEFAULT
    MAXRGN_REQUEST                        -1
    _PRIO_BASE_U2U                        100
    _PRIO_BASE_IOC                        80
    _PRIO_BASE_RAV                        80
    _PRIO_BASE_REX                        40
    _PRIO_BASE_COM                        10
    _PRIO_FACTOR                          80
    _DELAY_COMMIT                         NO
    _MAXTASK_STACK                        512
    MAX_SERVERTASK_STACK                  500
    MAX_SPECIALTASK_STACK                 500
    _DW_IO_AREA_SIZE                      50
    _DW_IO_AREA_FLUSH                     50
    FBM_VOLUME_COMPRESSION                50
    FBM_VOLUME_BALANCE                    10
    _FBM_LOW_IO_RATE                      10
    CACHE_SIZE                            262144
    _DW_LRU_TAIL_FLUSH                    25
    XP_DATA_CACHE_RGNS                    0
    _DATA_CACHE_RGNS                      64
    XP_CONVERTER_REGIONS                  0
    CONVERTER_REGIONS                     8
    XP_MAXPAGER                           0
    MAXPAGER                              64
    SEQUENCE_CACHE                        1
    _IDXFILE_LIST_SIZE                    2048
    VOLUMENO_BIT_COUNT                    8
    OPTIM_MAX_MERGE                       500
    OPTIM_INV_ONLY                        YES
    OPTIM_CACHE                           NO
    OPTIM_JOIN_FETCH                      0
    JOIN_SEARCH_LEVEL                     0
    JOIN_MAXTAB_LEVEL4                    16
    JOIN_MAXTAB_LEVEL9                    5
    _READAHEAD_BLOBS                      32
    CLUSTER_WRITE_THRESHOLD               80
    CLUSTERED_LOBS                        NO
    RUNDIRECTORY                          /var/opt/sdb/data/wrk/SPDT
    OPMSG1                                /dev/console
    OPMSG2                                /dev/null
    _KERNELDIAGFILE                       knldiag
    KERNELDIAGSIZE                        800
    _EVENTFILE                            knldiag.evt
    _EVENTSIZE                            0
    _MAXEVENTTASKS                        2
    _MAXEVENTS                            100
    _KERNELTRACEFILE                      knltrace
    TRACE_PAGES_TI                        2
    TRACE_PAGES_GC                        20
    TRACE_PAGES_LW                        5
    TRACE_PAGES_PG                        3
    TRACE_PAGES_US                        10
    TRACE_PAGES_UT                        5
    TRACE_PAGES_SV                        5
    TRACE_PAGES_EV                        2
    TRACE_PAGES_BUP                       0
    KERNELTRACESIZE                       5369
    EXTERNAL_DUMP_REQUEST                 NO
    _AK_DUMP_ALLOWED                      YES
    _KERNELDUMPFILE                       knldump
    _RTEDUMPFILE                          rtedump
    _UTILITY_PROTFILE                     dbm.utl
    UTILITY_PROTSIZE                      100
    _BACKUP_HISTFILE                      dbm.knl
    _BACKUP_MED_DEF                       dbm.mdf
    _MAX_MESSAGE_FILES                    0
    _SHMKERNEL                            44601
    __PARAM_CHANGED___                    0
    __PARAM_VERIFIED__                    2008-05-03 23:12:55
    DIAG_HISTORY_NUM                      2
    DIAG_HISTORY_PATH                     /var/opt/sdb/data/wrk/SPDT/DIAGHISTORY
    _DIAG_SEM                             1
    SHOW_MAX_STACK_USE                    NO
    SHOW_MAX_KB_STACK_USE                 NO
    LOG_SEGMENT_SIZE                      2133
    _COMMENT                              
    SUPPRESS_CORE                         YES
    FORMATTING_MODE                       PARALLEL
    FORMAT_DATAVOLUME                     YES
    OFFICIAL_NODE                         
    UKT_CPU_RELATIONSHIP                  NONE
    HIRES_TIMER_TYPE                      CPU
    LOAD_BALANCING_CHK                    30
    LOAD_BALANCING_DIF                    10
    LOAD_BALANCING_EQ                     5
    HS_STORAGE_DLL                        libhsscopy
    HS_SYNC_INTERVAL                      50
    USE_OPEN_DIRECT                       YES
    USE_OPEN_DIRECT_FOR_BACKUP            NO
    SYMBOL_DEMANGLING                     NO
    EXPAND_COM_TRACE                      NO
    JOIN_TABLEBUFFER                      128
    SET_VOLUME_LOCK                       YES
    SHAREDSQL                             YES
    SHAREDSQL_CLEANUPTHRESHOLD            25
    SHAREDSQL_COMMANDCACHESIZE            262144
    MEMORY_ALLOCATION_LIMIT               0
    USE_SYSTEM_PAGE_CACHE                 YES
    USE_COROUTINES                        YES
    FORBID_LOAD_BALANCING                 YES
    MIN_RETENTION_TIME                    60
    MAX_RETENTION_TIME                    480
    MAX_SINGLE_HASHTABLE_SIZE             512
    MAX_HASHTABLE_MEMORY                  5120
    ENABLE_CHECK_INSTANCE                 YES
    RTE_TEST_REGIONS                      0
    HASHED_RESULTSET                      YES
    HASHED_RESULTSET_CACHESIZE            262144
    CHECK_HASHED_RESULTSET                0
    AUTO_RECREATE_BAD_INDEXES             NO
    AUTHENTICATION_ALLOW                  
    AUTHENTICATION_DENY                   
    TRACE_AK                              NO
    TRACE_DEFAULT                         NO
    TRACE_DELETE                          NO
    TRACE_INDEX                           NO
    TRACE_INSERT                          NO
    TRACE_LOCK                            NO
    TRACE_LONG                            NO
    TRACE_OBJECT                          NO
    TRACE_OBJECT_ADD                      NO
    TRACE_OBJECT_ALTER                    NO
    TRACE_OBJECT_FREE                     NO
    TRACE_OBJECT_GET                      NO
    TRACE_OPTIMIZE                        NO
    TRACE_ORDER                           NO
    TRACE_ORDER_STANDARD                  NO
    TRACE_PAGES                           NO
    TRACE_PRIMARY_TREE                    NO
    TRACE_SELECT                          NO
    TRACE_TIME                            NO
    TRACE_UPDATE                          NO
    TRACE_STOP_ERRORCODE                  0
    TRACE_ALLOCATOR                       0
    TRACE_CATALOG                         0
    TRACE_CLIENTKERNELCOM                 0
    TRACE_COMMON                          0
    TRACE_COMMUNICATION                   0
    TRACE_CONVERTER                       0
    TRACE_DATACHAIN                       0
    TRACE_DATACACHE                       0
    TRACE_DATAPAM                         0
    TRACE_DATATREE                        0
    TRACE_DATAINDEX                       0
    TRACE_DBPROC                          0
    TRACE_FBM                             0
    TRACE_FILEDIR                         0
    TRACE_FRAMECTRL                       0
    TRACE_IOMAN                           0
    TRACE_IPC                             0
    TRACE_JOIN                            0
    TRACE_KSQL                            0
    TRACE_LOGACTION                       0
    TRACE_LOGHISTORY                      0
    TRACE_LOGPAGE                         0
    TRACE_LOGTRANS                        0
    TRACE_LOGVOLUME                       0
    TRACE_MEMORY                          0
    TRACE_MESSAGES                        0
    TRACE_OBJECTCONTAINER                 0
    TRACE_OMS_CONTAINERDIR                0
    TRACE_OMS_CONTEXT                     0
    TRACE_OMS_ERROR                       0
    TRACE_OMS_FLUSHCACHE                  0
    TRACE_OMS_INTERFACE                   0
    TRACE_OMS_KEY                         0
    TRACE_OMS_KEYRANGE                    0
    TRACE_OMS_LOCK                        0
    TRACE_OMS_MEMORY                      0
    TRACE_OMS_NEWOBJ                      0
    TRACE_OMS_SESSION                     0
    TRACE_OMS_STREAM                      0
    TRACE_OMS_VAROBJECT                   0
    TRACE_OMS_VERSION                     0
    TRACE_PAGER                           0
    TRACE_RUNTIME                         0
    TRACE_SHAREDSQL                       0
    TRACE_SQLMANAGER                      0
    TRACE_SRVTASKS                        0
    TRACE_SYNCHRONISATION                 0
    TRACE_SYSVIEW                         0
    TRACE_TABLE                           0
    TRACE_VOLUME                          0
    CHECK_BACKUP                          NO
    CHECK_DATACACHE                       NO
    CHECK_KB_REGIONS                      NO
    CHECK_LOCK                            NO
    CHECK_LOCK_SUPPLY                     NO
    CHECK_REGIONS                         NO
    CHECK_TASK_SPECIFIC_CATALOGCACHE      NO
    CHECK_TRANSLIST                       NO
    CHECK_TREE                            NO
    CHECK_TREE_LOCKS                      NO
    CHECK_COMMON                          0
    CHECK_CONVERTER                       0
    CHECK_DATAPAGELOG                     0
    CHECK_DATAINDEX                       0
    CHECK_FBM                             0
    CHECK_IOMAN                           0
    CHECK_LOGHISTORY                      0
    CHECK_LOGPAGE                         0
    CHECK_LOGTRANS                        0
    CHECK_LOGVOLUME                       0
    CHECK_SRVTASKS                        0
    OPTIMIZE_AGGREGATION                  YES
    OPTIMIZE_FETCH_REVERSE                YES
    OPTIMIZE_STAR_JOIN                    YES
    OPTIMIZE_JOIN_ONEPHASE                YES
    OPTIMIZE_JOIN_OUTER                   YES
    OPTIMIZE_MIN_MAX                      YES
    OPTIMIZE_FIRST_ROWS                   YES
    OPTIMIZE_OPERATOR_JOIN                YES
    OPTIMIZE_JOIN_HASHTABLE               YES
    OPTIMIZE_JOIN_HASH_MINIMAL_RATIO      1
    OPTIMIZE_OPERATOR_JOIN_COSTFUNC       YES
    OPTIMIZE_JOIN_PARALLEL_MINSIZE        1000000
    OPTIMIZE_JOIN_PARALLEL_SERVERS        0
    OPTIMIZE_JOIN_OPERATOR_SORT           YES
    OPTIMIZE_QUAL_ON_INDEX                YES
    DDLTRIGGER                            YES
    SUBTREE_LOCKS                         NO
    MONITOR_READ                          2147483647
    MONITOR_TIME                          2147483647
    MONITOR_SELECTIVITY                   0
    MONITOR_ROWNO                         0
    CALLSTACKLEVEL                        0
    OMS_RUN_IN_UDE_SERVER                 NO
    OPTIMIZE_QUERYREWRITE                 OPERATOR
    TRACE_QUERYREWRITE                    0
    CHECK_QUERYREWRITE                    0
    PROTECT_DATACACHE_MEMORY              NO
    LOCAL_REDO_LOG_BUFFER_SIZE            0
    FILEDIR_SPINLOCKPOOL_SIZE             10
    TRANS_HISTORY_SIZE                    0
    TRANS_THRESHOLD_VALUE                 60
    ENABLE_SYSTEM_TRIGGERS                YES
    DBFILLINGABOVELIMIT                   70L80M85M90H95H96H97H98H99H
    DBFILLINGBELOWLIMIT                   70L80L85L90L95L
    LOGABOVELIMIT                         50L75L90M95M96H97H98H99H
    AUTOSAVE                              1
    BACKUPRESULT                          1
    CHECKDATA                             1
    EVENT                                 1
    ADMIN                                 1
    ONLINE                                1
    UPDSTATWANTED                         1
    OUTOFSESSIONS                         3
    ERROR                                 3
    SYSTEMERROR                           3
    DATABASEFULL                          1
    LOGFULL                               1
    LOGSEGMENTFULL                        1
    STANDBY                               1
    USESELECTFETCH                        YES
    USEVARIABLEINPUT                      NO
    UPDATESTAT_PARALLEL_SERVERS           0
    UPDATESTAT_SAMPLE_ALGO                1
    SIMULATE_VECTORIO                     IF_OPEN_DIRECT_OR_RAW_DEVICE
    COLUMNCOMPRESSION                     YES
    TIME_MEASUREMENT                      NO
    CHECK_TABLE_WIDTH                     NO
    MAX_MESSAGE_LIST_LENGTH               100
    SYMBOL_RESOLUTION                     YES
    PREALLOCATE_IOWORKER                  NO
    CACHE_IN_SHARED_MEMORY                NO
    INDEX_LEAF_CACHING                    2
    NO_SYNC_TO_DISK_WANTED                NO
    SPINLOCK_LOOP_COUNT                   30000
    SPINLOCK_BACKOFF_BASE                 1
    SPINLOCK_BACKOFF_FACTOR               2
    SPINLOCK_BACKOFF_MAXIMUM              64
    ROW_LOCKS_PER_TRANSACTION             50
    USEUNICODECOLUMNCOMPRESSION           NO
    about send you the data from tables, i dont have permission to do that, since all data is in a production system, the customer dont give me the rights to send any information. sorry about that.
    best regards
    Clóvis

  • Can we Customize Edit User-Defined Table Screen in Fusion Application

    Hi All,
    We have a requirement to add a ADFDI functionality to the 'Edit USer-Defined Table' screen by adding a 'Mass Upload' Button.
    Normally for Some of the screens (Fusion Compensation) i am seeing 'Personalize -> Edit Current Page' Option on Menu items on top. But on 'Edit User-Defined Table' i am not seeing this option in 'Personalize' Menu.
    So the questions is is this screen Customizable in Fusion Application? If So What will be EAR file which we need to use for Customization.
    The screen where we need to add ADFDI Functionality is under 'oracle/apps/setup' path.
    Can anyone point me to the Exact EAR on this and also let me know if we can customize it or not
    Thanks
    Kaja

    Generally the code structure can be observed run-time [url https://blogs.oracle.com/fadevrel/entry/finding_code_artifacts_for_customization]using the composer, however in your case the content comes through a portlet in which case the method does not work as you cannot drill down to the specific region details in the composer. Based on a search through the code I think the regions you are interested are:
    <li>hcmPayroll/userTables/ui/public_html/oracle/apps/hcm/userTables/ui/page/UserTableSearch.jsff --> This is the details page with header "Manage User-Defined Tables"
    <li>hcmPayroll/userTables/ui/public_html/oracle/apps/hcm/userTables/ui/page/ManageUserStructure.jsff --> This is the details page with header "User-Defined Table: <TABLE_NAME>"
    These would presumably be in EarHcmPayroll. I'll see if I can find out how the portlet content can be observed from the RT application..
    Jani Rautiainen
    Fusion Applications Developer Relations
    https://blogs.oracle.com/fadevrel/

  • Application Development using Jdeveloper using Oracle Lite

    Hi all,
    I need to develop an application that works with the oracle database lite technology.
    Ritght now I am working on some 10 tables .The Scenario is like the application developed is deployed on the middle tier and the user downloads it and goes offline. And he uses the application offline with the same tables present (that are used in developing the application) in another database present on the user machine . I have gone through the ADF BC tutorials but I couldn’t figure out whether the business components (view objects, entity objects) that are created by using tables (connecting to a particular schema in the database) can work if the schema changes(coz I tried doing that and faced errors that connection to database couldn’t be established ). I think this might be because the schema has changed.
    Does this Business components created work if the database schema changes?
    Is it possible to develop an application using java swing that works on the oracle database lite? I need a standalone Windows 32 application to be developed.
    Thanks in advance

    Let's assume you have two database schema USER1 and USER2
    And let's assume that USER2 grants the appropriate privileges so that USER1 can see his tables.
    If you connect as USER1 in JDeveloper and create an entity object based on a table USER2.SOMETABLE, then the SomeTable entity object will remember the USER2 schema name as part of the table name.
    If instead if USER1 creates an entity object against a table in his own USER1 schema, then the entity object will only record the "TABLENAME" in the entity metadata, without a schema qualifier. In this case, if you connect to a different schema at runtime which still has the expected TABLENAME table in it, then it will work fine.
    Generally you would want to use an entity based on a synonym instead of an entity based directly on a SCHEMANAME.TABLENAME so that it wouldn't matter if the schema changed.
    Yes, you can build Swing, JSF, or JSP applications against Oracle Lite using ADF BC.

  • CAF-Core application development in CE 7.1

    Hi Experts,
    I am new to CAF. I need to help how to created CAF-Core Applications.
    1. How Link RFC's to external services.
    2. How to created BO.
    3. How to expose BO to Web services.
    4. How to return more than one table values to external service.
    5. How to created WebDynpro  application from CAF Services.
    Please guide me how to learn CAF- Core application developement.
    Regards,
    Satya.

    Hi Satya,
    Following are some pointers for you to start:
    CAF in CE:
    http://help.sap.com/saphelp_nwce10/helpdata/en/43/0034e14db01a85e10000000a422035/frameset.htm
    Link RFC's to External Services:
    http://help.sap.com/saphelp_nwce10/helpdata/en/af/7ad9beab5e824cb2403f6821cde7dd/frameset.htm
    BO to Web Service:
    http://help.sap.com/saphelp_nwce10/helpdata/en/44/24128b347c1c9fe10000000a11466f/frameset.htm
    Regards,
    Alka.

Maybe you are looking for

  • Populating the data grid

    OK I think that is resolved but I cannot pull the Data. The Grid does not populate with data......Just keeps trying to pull it. THE CFC FUNCTION: <cffunction name="comments" access="remote" returntype="struct">   <cfargument name="page" required="tru

  • How can I copy my folders or files in my Ipad?

    I'm a biginner. How can I copy on my iPad some windows files or folder?

  • Down Payment with a normal account G/L

    Hi guys, i 'd like to know if is possible use a down payment with a no Special G/l account in SAP. Exaclty: Edit a document FI like a Down Payment,  For example: Down Payment Vendor to Vendor "X" and insert the number of PO in the field BSEG_EBELN. T

  • Spry: Conditionally Fields

    Hi, I am using DW CS3 to create spry fields. I have a situation in which I want one field to become required based upon the value of another field. In my page I added the following: var chkOtherValid = function(value, options){ var sel = document.get

  • Interactive report - strip HTML

    Hi, Where is the option to strip HTML for an Interactive report? Is it possible to set this option for each column? N.B. This setting can be found under Layout and Pagination for classic report. Thanks, Louis-Guillaume Homepage: http://www.insum.ca B