Container and pluggable databases

Hi,
Does anyone know if it is possible to mix enterprise edition and standard edition PDBs in a cdb?
Regards
Alan

Oracle 12c standard edition comes with below details:
Support for one pluggable database for entry-level cloud computing and consolidation
Cross-platform recovery
Built-in Oracle Real Application Clusters support for higher levels of system uptime
Simplified installation and configuration
Suitable for all types of data and all of your applications
Upwardly compatible with Oracle Database 12c Enterprise Edition, protecting your initial investment
check with the licencing part!!!
Thanks,
http://gssdba.wordpress.com

Similar Messages

  • How to use expdp and impdp in oracle 12c on a pluggable database

    Hi,
    I have a pluggable database in a container DB. I have to use expdp to export a table and later import the table into another pluggable database.
    How can i tell expdp to which pluggable database I want to connect?
    Thanks,
    Sarayu

    Then
    1. Add newpdb to  tnsnames.ora ?
    2. connect to pluggable database create a directory.
    sqlplus sys/<password>@newpdb as sysdba
    create or replace directory exp_dir as '/u01/app';
    grant read, write on directory exp_dir to scott;
    3. export emp table of scott user.  -- with sys user
    expdp directory=exp_dir dumpfile=TDUMP.DMP TABLES='SCOTT'.'EMP'
    username: sys@newpdb as sysdba
    password: <password>
    Regards
    Mahir M. Quluzade
    Paste here result.

  • Can I use flashback database to flashback a pluggable database?

    Hi All,
    I created a container database and then created a pluggable database.
    In the pluggable 12.1 datbase, I loaded the user data.
    I performed the below
    sqlplus / as sysdba;
    create restore point CLEAN_DB guarantee flashback database;
    And I performed the transactions on the pdb.
    Now my intention is to restore to the restore point CLEAN_DB as created from the above step.
    If I perform the below, will it restore my pdb to the initial state?
    flashback database to restore point CLEAN_DB;
    Regards,
    Kamal.

    Hi,
    You cannot use FLASHBACK DATABASE for pluggable database.
    Flashback query/versions queryis UNDO based and work in PDB’s
    Flashback transaction query/flashback transactionis UNDO and REDO based and work in PDB’s
    Flashback data archiveUNDO gernerated archives
    Flashback table
    before drop –> Rename segment name of recylebin, which is based on UNDO, work in PDB’s
    flashback table to –> is UNDO based, work in PDB’s
    Flashback databasedoesn’t work in PDB’s:
    RMAN> flashback pluggable database prmdb01 to time "to_date('23:00 20-11-2013','hh24:mi dd-mm-yyyy')";
    Starting flashback at 25-NOV-13
    using channel ORA_DISK_1
    using channel ORA_DISK_2
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of flashback command at 11/20/2013 23:00:00
    RMAN-05108: Command is not supported for pluggable database
    Regards
    Mahir M. Quluzade
    p.s. Command exists but not supporting, may be next release come support.

  • Dropping public synonyms in a pluggable database

    I have an Oracle 12c container database with several pluggable databases in it.  In one of the pluggable databases, I de-installed OLAP by running the following scripts:
    SQL> @%oracle_home%/olap/admin/catnoamd.sql
    SQL> @%oracle_home%/olap/admin/olapidrp.plb
    SQL> @%oracle_home%/olap/admin/catnoxoq.sql
    SQL> @%oracle_home%/olap/admin/catnoaps.sql
    After running the scripts, I then generated the SQL statements to remove invalid OLAP synonyms, by executing the following:
    select 'drop public synonym ' || object_name || ';'
    from dba_objects
    where status <> 'VALID'
    and owner = 'PUBLIC'
    and object_type = 'SYNONYM'
    and object_name like '%OLAP%';
    When I execute the resulting SQL statements in the pluggable database, I get the following error for all of them:
    ERROR at line 1:
    ORA-65040: operation not allowed from within a pluggable database
    If I switch back to the CDB$ROOT container and execute the same statements, I get the following error:
    ERROR at line 1:
    ORA-01432: public synonym to be dropped does not exist
    How can I remove these invalid public synonyms from my pluggable database?

    OK - let's go back to the basics
    For 12c you should NOT be manually removing/uninstalling/disabling Oracle options.
    There is a CHOPT utitlity specifically designed to enable or disable certain options.
    http://docs.oracle.com/cd/E11882_01/install.112/e48740/post_inst_task.htm#CHDFEIJF
    5.2.7 Enabling and Disabling Database Options
    When you install Oracle Database, some options are enabled and others are disabled. To enable or disable a particular database feature for an Oracle home, shut down the database and use the chopt tool. See Example 5-1.
    The chopt tool is a command-line utility that is located in the ORACLE_HOME/bin directory. The syntax for chopt is:
    That doc section shows that one of the 'db_option' values available is 'olap' to disable. NOTE: the uttility DISABLES the option. It is still not clear to me why you would attempt to actually deinstall it.
    If your concern is licensing then all you need to do is disable it. It is very possible that you have corrupted that PDB by attempting a manual uninstall. Only Oracle support could tell you for sure.
    If it is a new PDB and licensing is your issue you may be far better off starting over and using CHOPT to disable the options you don't want.
    1. why are you 'unistalling' olap in the pdb? What PROBLEM are you trying to solve?
    2. why are you manually dropping synonyms?
    3. how did you decide to execute those particular scripts? Is there an Oracle doc you are following that has instructions for deinstalling olap? Does that doc give you the order to execute the scripts in?
    If you review the catnoamd.sql script you will see that just about ALL that it does is drop public synonyms.
    Most of those synonyms that it drops have 'olap' in the name.
    Which suggests that if Oracle wanted those synonyms YOU are trying to drop to be dropped you have to wonder why it didn't drop them itself in that script.
    Was that script the FIRST script you executed?
    1. Modify your 'drop' script to simply create the DDL to drop the synonyms.
    2. Then examine the data dictionary to see what objects those synonyms refer to.
    3. Then check if those objects still exist or not
    4. Also check those scripts to see if they did anything (e.g. drop) those objects.
    Post the results of the above.

  • Permissions required to create a pluggable database

    Hi
    using 12.1.0.1.5 on oracle linux 6.3
    I am trying to setup a new user with permissions to create pluggable databases within my CDB, i have created the user with the following
    create user c##admin identified by oracle12c container=all;
    grant connect, resource, dba to c##admin container=all;
    grant create pluggable database to c##admin container=all;
    however, when i connect as the user to the CDB and try to clone a PDB
    sqlplus c##admin/oracle12c
    create pluggable database pdb99 from pdb1 file_name_convert=('/vol1/u01/app/oracle/oradata/cdb1/pdb1', '/vol1/u01/app/oracle/oradata/cdb1/pdb99');
    i get the following permissions error
    SQL> create pluggable database pdb99 from pdb1 file_name_convert=('/vol1/u01/app/oracle/oradata/cdb1/pdb1', '/vol1/u01/app/oracle/oradata/cdb1/pdb99');
    create pluggable database pdb99 from pdb1 file_name_convert=('/vol1/u01/app/oracle/oradata/cdb1/pdb1', '/vol1/u01/app/oracle/oradata/cdb1/pdb99')
    ERROR at line 1:
    ORA-01031: insufficient privileges
    i have manually created the folder into which the new PDB datafiles will be created (/vol1/u01/app/oracle/oradata/cdb1/pdb99),
    any ideas what i have done wrong?
    thanks
    Dave

    i get the following permissions error
    SQL> create pluggable database pdb99 from pdb1 file_name_convert=('/vol1/u01/app/oracle/oradata/cdb1/pdb1', '/vol1/u01/app/oracle/oradata/cdb1/pdb99');
    create pluggable database pdb99 from pdb1 file_name_convert=('/vol1/u01/app/oracle/oradata/cdb1/pdb1', '/vol1/u01/app/oracle/oradata/cdb1/pdb99')
    ERROR at line 1:
    ORA-01031: insufficient privileges
    i have manually created the folder into which the new PDB datafiles will be created (/vol1/u01/app/oracle/oradata/cdb1/pdb99),
    any ideas what i have done wrong?
    The user executing the statement has to have privileges to create files in that directory.
    See the doc:
    https://docs.oracle.com/database/121/ADMIN/cdb_plug.htm
    CREATE_FILE_DEST Clause
    If the PDB will use Oracle Managed Files, then the CREATE_FILE_DEST clause of the CREATE PLUGGABLE DATABASE statement specifies the default file system directory or Oracle ASM disk group for the PDB's files.
    If a file system directory is specified as the default location in this clause, then the directory must exist. Also, the user who runs the CREATE PLUGGABLE DATABASE statement must have the appropriate privileges to create files in the specified directory.

  • HELP! FOR LOOP TO SCROLL THROUGH TABLE AND CREATE DATABASE LINK

    Hi,
    Here's the scenario, not much of a PL programmer, just basic SQL so really need some help people!
    I have 2 tables. 1 contains list of DB's and the other contains rules to follow.
    I need to create a loop that goes through the table containing the DB's and on each row a DB link is created (Only 1 link allowed!)
    Once created, the schema currently logged in with also has an account on the linked DB in order to run scripts- The scripts are stored centrally hence the requirement for the link to the target DB.
    There are numerous scripts that need to be executed and can all be called from 1 script, once executed the loop exists and the database link needs to be dropped.
    Once dropped, the first loop continues, creating a DB link for the next DB listed in the table (and all the scripts are fired again)
    This continues against all the DB's listed in the table.

    Hi BlueShadow,
    Thanks again for the response, you've hit the nail on the head. SQL scripts on a unix server, a loop goes through a table 1 at a time. Each row gets a link created and then all the scripts stored on the server are executed against the db linked to. So I'm assuming this is a loop within a loop.
    1 loop to go through the table to create the link and then another loop within once connected to execute all the scripts against the connected DB. Once the scripts are run, the loop exits and moves onto the next server and so on until all the servers have the scripts are run.
    It's PL/SQL scripts we're after and not shell scripts as this would free us from the OS constraints.
    We have to drop the links due to security. Any idea on o

  • Database schema SCM does not contain the associated database objects

    I am getting the following error when i am trying to migrate the form to apex using application migration.
    "*Database schema SCM does not contain the associated database objects for the project, aafs.*
    *Ensure the database schema associated with the project contains the database objects associated with the uploaded Forms Module .XML file(s).* ".
    Actully i am having one schema which i named as SCM, and i have defined one table TT.
    I created one form test.fmb in which i used TT table.its compiled successfully.
    Then i generated the xml file using frmf2xml from fmb file. After that, I created the project in appication migration wizard in SCM schema.
    Project creattion is working fine.but when i m trying to create application,it is showing me above error.
    can any one help in solving this problem.

    Hi Hilary,
    Thanks for your response/feedback.
    1. The schema associated with the project does not contain the necessary objects Can you please verify that the schema associated with your Forms conversion project does in fact contain the objects associated with the uploaded files. Could you also verify that the object names referenced in the error message do not exist within the schema associated with your workspace. Ensure that the schema associated with the project contains the necessary database objects before proceeding to the generation phase of the conversion process.
    Ans:
    Yes it does contain the objects (See results from SQL query Commands below):
    SELECT MWRA_CONTRACT_NO, OLD_CONTRACT_NO FROM PROJECTS@CONTRACT_LX19.MWRA.NET
    ORDER BY MWRA_CONTRACT_NO
    000000569 551TA
    000000570 553TA
    000000575 560TA
    000000576 561TA
    000107888 502TA
    000108498 500TA
    000108502 503TA
    2. The block being converted contains buttons, which may have been incorrectly identified as database columns, and included in the original or enhanced query associated with your block This is a known issue ,bug 9827853, and a fix will be available in our upcoming 4.0.1 patch release. Some possible solutions to this issue are:
    -> delete the buttons before generating the XML
    -> delete the button tags from the XML
    -> add "DatabaseItem=No" for the button in the XML file before importing it in Apex.The button is excluded when creating the Application.
    Ans
    yes it does contain push buttons to transfer to another forms and these are defined as Non data base items. Parial XML code provided below:
    - <Item Name="REPORTS" FontSize="900" DirtyInfo="true" Height="188" XPosition="4409" FontName="Fixedsys" ForegroundColor="black" DatabaseItem="false" Width="948" CompressionQuality="None" YPosition="3709" FontSpacing="Normal" Label="REPORTS" BackColor="canvas" FillPattern="transparent" ShowHorizontalScrollbar="false" FontWeight="Medium" ShowVerticalScrollbar="false" FontStyle="Plain" ItemType="Push Button">
    <Trigger Name="WHEN-BUTTON-PRESSED" TriggerText="GO_BLOCK('REPORT_1');" />
    </Item>
    - <Item Name="TRACKHDR" FontSize="900" DirtyInfo="true" Height="188" XPosition="3409" FontName="Fixedsys" ForegroundColor="black" DatabaseItem="false" Width="948" CompressionQuality="None" YPosition="3709" FontSpacing="Normal" Label="TRACK" BackColor="canvas" FillPattern="transparent" ShowHorizontalScrollbar="false" FontWeight="Medium" ShowVerticalScrollbar="false" FontStyle="Plain" ItemType="Push Button">
    <Trigger Name="WHEN-BUTTON-PRESSED" TriggerText="GO_BLOCK('TRACKHDRS');" />
    </Item>
    - <Item Name="SUBAWRD" FontSize="900" DirtyInfo="true" Height="188" XPosition="2429" FontName="Fixedsys" ForegroundColor="black" DatabaseItem="false" Width="948" CompressionQuality="None" YPosition="3719" FontSpacing="Normal" Label="SUBAWARDS" BackColor="canvas" FillPattern="transparent" ShowHorizontalScrollbar="false" FontWeight="Medium" ShowVerticalScrollbar="false" FontStyle="Plain" ItemType="Push Button">
    3. If you are still experiencing issues, then please create a testcase on apex.oracle.com and update this thread with the workspace details so I can take a look.
    Test case details are given below. It was created per ORACLE for open Service Request Number 3-1938902931 on ORACLE Metalink.
    Workspace: contract4
    username: [email protected] (my email)
    Password: contract4
    Comments:
    For my migration/testing purpose a dabatase link and synonyms have been setup by our ORACLE DBA. Could this be causing this problem?
    Do we know when the fix 4.0.1 patch release will be available?
    Thanks for your help.
    Indra

  • Co-location of persistent chat and monitoring databases on Standard Edition server

    Hello, I've seen all kinds of different answers to this question around the web with conflicting answers.
    We have a basic IM/Presence only install of Lync 2013 and now want to deploy Monitoring and Persistent Chat roles.
    As per this blog, it states you can co-locate all of these on one FE server.
    http://windowspbx.blogspot.ca/2012/07/aaa-donotpost-install-lync-standard.html
    Which does match what this TechNet article states:
    http://technet.microsoft.com/en-us/library/gg398131.aspx
    However, in the Blog, it says you need to install Full SQL Server 2008 in order to collocate, yet this goes against what TechNet says is possible, in that you can have all databases in one instance:
    Each SQL instance can contain only a single back-end database (for an Enterprise Edition Front End pool), single Monitoring database, single Archiving database, single persistent chat database, and single persistent chat compliance database.
    I guess my question is, can I simply install Monitoring and Persistent Chat using the same default SQL Server 2012 instance that was installed along with the Front End Standard Edition server (RTC instance)? Or should I create a new instance? Or do I need
    to install a new SQL Server 2012 somewhere on a different server altogether?
    Thanks!

    Hi,
    Lync standard edition server has SQL express database by default to host RTC instance. Monitoring server required SQL reporting services, SQL express does not have reporting components part of it. Hence, you required full SQL version for monitoring
    and archiving deployment.
    Technically, you can install full SQL server on the standard edition server itself. But , it is not recommended in production environment. It's good idea to install SQL server on a dedicated server for hosting monitoring and archiving database.
    Thanks
    Saleesh
    If answer is helpful, please hit the green arrow on the left, or mark as answer. Blog : http://blogs.technet.com/b/saleesh_nv/

  • Production and Development Database Connections

    Within JDeveloper I have created a database connection to a DEVELOPMENT database and am exposing PL/SQL packages in the development database as web services.
    I can develop and test successfully the exposed packages as web services, and can deploy and run them to a standalone OC4J container on my local machine.
    I then handover the generated EAR file to the production team who deploy the EAR as a web service to the Oc4j container in a remote 9i Application Server.
    This runs successfully from 9iAS.
    So far so good.
    If I have the same PL/SQL package in a live database and development database, but only have access to the development database, how can I expose the package in the live database?
    Can I have a database connection in JDeveloper with a connection name pointing to the development database connection, and have a connection defined in 9iAS with the same name pointing to the live database?
    Any information will be useful.
    Thanks,
    Claire.

    Per the J2EE "roles" definition this is a deployment role todo.
    I.E. the theory is; take a 3rd party ear, depoyee with your appserver vendor's
    deployment tool, edit the data-sources.xml imbeded in the ear on the fly, as
    deployment happens.
    Bake to reality-
    OC4J 903 is intended to be used with OEM for the J2EE spec'ed role of deployer. I'm
    headed this way, but not started the 903/OEM todo.
    Today, OC4J 902 and 902 Jdev, I deploy to a local instance of OC4J. I take that
    ear to production. I've written procedural instructions on poping open the ear and
    editing the data-sources.xml for the DB connection URL and possible login/passwords.
    BTW all your AppModule configurations will have to use data-sources DB connections else
    you'll have DB connection info littered in your bc4j.xcfg's for each AM deeper in your war's.
    I'd like to hear how this is supposed to work from some other folks? :)
    Good luck,
    curt

  • How to conncet mysql database and oracle database 11g2 on linux 5.4

    Hi,
    Can anyone help me to connect mysql server database to Oracle
    database on Linux OS?
    I have done all things, but I am getting a problem in installing dg4odbc
    driver on Oracle database. In a simple word, I don't know how to create
    listener between mysql database and Oracle database.
    If anyone knows that please help me in configuring dg4odbc for Oracle. OS
    -Linux 5.4 and refer any doc or link.
    oracle 11g2 11.2.0.1
    i follow the metalink id - 466228.1 ( but cant successed).
    Thanks and regards,
    pritesh ranjan

    hi pl z clear me it is required to install manully the dg4odbc driver for oracle to connect MySQL server database. because in my $ORACLE_HOME/hs/admin/initdg4odbc.ora is present automatically.
    details of the initdg4odbc.ora
    # This is a sample agent init file that contains the HS parameters that are
    # needed for the Database Gateway for ODBC
    # HS init parameters
    HS_FDS_CONNECT_INFO = mysql_odbc
    HS_FDS_TRACE_LEVEL = off
    HS_FDS_SHAREABLE_NAME = /usr/lib/libmyodbc5w.so
    # ODBC specific environment variables
    set ODBCINI=/etc/odbc.ini
    set LD_LIBRARY_PATH=/usr/lib
    # Environment variables required for the non-Oracle system
    #set HOME=/home/oracle
    ~
    ~
    ~
    "initdg4odbc.ora" 20L, 437C
    details of the listener.ora file
    DG4ODBC =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = localhost)(PORT = 1521))
    (CONNECT_DATA=(SID=dg4odbc))
    (HS=OK)
    SID_LIST_ORCL =
    (SID_LIST =
    (SID_DESC =
    (GLOBAL_DBNAME = orcl)
    (ORACLE_HOME = /u01/app/oracle/product/11.2.0/dbhome_1)
    (SID_NAME = ORCL)
    ADR_BASE_DG4ODBC = /u01/app/oracle
    ORCL =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = pritesh11gprim.oracle.com)(PORT = 1521))
    ADR_BASE_ORCL = /u01/app/oracle
    "listener.ora" [readonly] 37L, 779C 35,0-1 85%
    details of tnsname.ora file
    # tnsnames.ora Network Configuration File: /u01/app/oracle/product/11.2.0/dbhome_1/network/admin/tnsnames.ora
    # Generated by Oracle configuration tools.
    ORCL =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.1.132)(PORT = 1521))
    (CONNECT_DATA =
    (SERVICE_NAME = ORCL)
    DG4ODBC =
    (DESCRIPTION=
    (ADDRESS=(PROTOCOL=tcp)(HOST=localhost)(PORT=1521))
    (CONNECT_DATA=(SID=DG4ODBC))
    (HS=OK)
    MYSQL =
    (DESCRIPTION=
    (ADDRESS=(PROTOCOL=tcp)(HOST=localhost)(PORT=1521))
    (CONNECT_DATA=(SID=MYSQL))
    (HS=OK)
    when i m going to start the listener it could not be started..
    details of the mysql server database
    [root@pritesh11gprim ~]# mysql -u root -p
    Enter password:
    Welcome to the MySQL monitor. Commands end with ; or \g.
    Your MySQL connection id is 2
    Server version: 5.6.10 MySQL Community Server (GPL)
    Copyright (c) 2000, 2013, Oracle and/or its affiliates. All rights reserved.
    Oracle is a registered trademark of Oracle Corporation and/or its
    affiliates. Other names may be trademarks of their respective
    owners.
    Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
    mysql>
    mysql> show databases;
    | Database |
    | information_schema |
    | demo |
    | mysql |
    | performance_schema |
    | pritesh |
    | test |
    6 rows in set (0.01 sec)
    if any where i m wrong then plz suggest me......
    thanks & regards
    pritesh ranjan

  • Transactions involving multiple caches and a database

    Hi all,
    I'm curious if the following is possible with the transaction support in Coherence.
    I have the need to write data to two caches and a database from within my Tomcat container, and the whole operation must be atomic.
    Example:
    Write to database (this is NOT via a CacheStore)
    Write to cache 1 (uses write-through to database)
    Write to cache 2 (uses write-through to database)
    If any operation fails, the whole transaction needs to rollback. This feels like an XA transaction, but it looks like it won't work like I expect because the cache must be the last resource, but I have two caches. The ordering of operations is not important.
    Thanks,
    Rob

    MagnusE wrote:
    When using write-through there is (as far as I know) no way to get a fully transactional behaviour (assuming you have more than one cache node) since each node is responsible for persisting its own data items (they will each use a separate connection to the database).
    If you on the other hand uses a "cache beside" pattern this can be made transactional using XA. As long as both caches belong to the same cache service they count as a single "last resource"....
    /MagnusYou also need to use the same transaction isolation and concurrency setting for it being the same last resource. Practically you should only have a single CacheAdapter instance enrolled in the same transaction.
    Best regards,
    Robert

  • Do I need to cleanup runtime binding container and iterators?

    If I create a runtime binding container and iterator in a JSF session bean as follows:
    FacesContext facesContext = FacesContext.getCurrentInstance();
    ValueBinding valueBinding = facesContext.getApplication().createValueBinding("#{data}");
    BindingContext bindingContext = (BindingContext)valueBinding.getValue(facesContext);
    // To create JUIteratorBinding, we need the following objects
    DCDataControl dataControl = bindingContext.findDataControl("MyDataControlName");
    ApplicationModule appModule = (ApplicationModule)dataControl.getDataProvider();
    DCBindingContainer bindingContainer = new DCBindingContainer();
    // Create dynamic iterator.  Need to set bindingContainer otherwise executeQuery throws exception
    JUIteratorBinding iterator = new JUIteratorBinding(bindingContext, dataControlName, "MyViewObjectName", "MyIteratorName", -1);
    iterator.setBindingContainer(bindingContainer);....do I have to later clean down the iterator and binding container from the data control to release the memory (and what calls do I need to make), or once the objects fall out of scope in my code are they marked for GC?
    CM.
    PS. JDev 10.1.3.1, ADF BC and JSF

    Hi Steve, thanks for the follow up.
    I'm coding a dynamic menu <af:page> control that maps to the MenuTreeModelAdapter (session bean) as detailed in the JDev 4GL/Forms guide.
    I've made a change that instead of loading the menu items from the hardcoded faces-config.xml file, they're actually sourced from the database via the code above, which is placed in the MenuTreeModelAdapter constructor. As the <af:page> tag exists on a number of pages, and thus the MenuTreeModelAdapter may be instantiated for the first time from a number of different locations, I can't hard-code the binding container or iterator into any single pageDef file. Thus I dynamically construct it in the constructor when the session bean is created for each user. I require a session bean as the menu items are different for each user.
    However I'm concerned once I've created the binding container and iterator, that even though the session bean may become invalidated, the 2 binding layer constructs may remain, consuming resources.
    Thanks for your assistance.
    CM.

  • Transaction Help ? JMS and db2 database without appserver

    We have a Queue and db2 database.
    need to read the message from queue and write it to the database in a standalone application without appserver. Now question is how to have a single transaction for both the jobs?
    Please advice!!!

    If you want to receive a JMS message and write to a database in the same transaction, then it sounds as if you are asking for the ability to enlist both transactional resources within the same XA transaction and then perform a two-phase commit of the two resources. If you use an application server then this is easy to configure.
    Even if you don't want to use an application server (or some other container offering similar services), you will still need a transaction manager, but will need to enlist the transactional resources yourself. Possible but not easy. If an application server is too heavyweight, then look for lighter-weight alternatives.
    Nigel

  • V1 and V2 database updates???

    Hi...
    can anyone pl give me idea abt V1 and V2 database updates???
    and diffence in b/w them...

    Hi,
    Refer this:
    V1 and V2 Update Modules
    An update is divided into different modules (see also Update Request). Each module corresponds to an update function module.
    There are two types of module.
    The SAP System makes a distinction between primary, time-critical (V1) and secondary, non-time-critical (V2) update modules. The system also supports collective runs for function modules that are used on a regular basis.
    This distinction allows the system to process critical database changes before less critical changes.
    V1 modules describe critical or primary changes; these affect objects that have a controlling function in the SAP System, for example order creation or changes to material stock.
    V2 modules describe less critical secondary changes. These are pure statistical updates, for example, such as result calculations.
    The V1 modules are processed consecutively in a single update work process on the same application server. This means that they belong to the same database LUW and can be reversed. Furthermore, V1 updates are carried out under the SAP locks of the transaction that creates the update (see  The SAP Lock Concept). This ensures that the data remains consistent; simultaneous changes to the objects to be updated are not possible.
    All V2 updates are carried out in a separate LUW and not under the locks of the transaction that creates them. If your SAP System contains a work process for V2 updates, these are only carried out in this work process. If this is not the case, the V2 components are processed by a V1 update process.
    All V1 modules of an update must be processed before the V2 modules.
    Let us assume that a transaction makes planning changes to a material and balance sheet, and updates two sets of statistics.
    Each of these changes is represented by means of an update module (call update function module) in the update request - the two planning changes by a V1 update module (time critical), and the statistical changes by a V2 update module (less critical). (The V1 modules have priority, although the V2 modules are usually also processed straight away).
    Jogdand M B

  • Problems logging into pluggable database 'admin' user on 12c

    Hi,
    I've created a multi-tenant 12c database. During the creation of the pluggable database the administrator account 'admin' was created.
    The database is on a Windows server running under a system account, our clients are all on Windows 7 and Active Directory.
    *I* can successfully log into the admin user e.g. conn admin/my_admin_password@//my_server/my_pdb.my_server as sysdba from my PC, but another user on the *same* PC gets an invalid username/password response. Both of us can successfully access other local users on the same pluggable database. Both of us can access the PDB as the 'sys' user.
    Any ideas as to what might be going on?
    Does the 'admin' user login make use of NTS privileges in some way (we are in different AD groups)?

    Good Afternoon MelCZier,
    The resource below will provide some steps to try if you are having trouble signing in to Messages on your iMac.  Do you have two-step verification enabled?  If so, that may be why you are being asked for an additional password.  
    Get help signing in to FaceTime, Game Center, Messages - Apple Support
    https://support.apple.com/en-us/HT204408
    Hope that helps ...
    - Judy

Maybe you are looking for

  • Flex - scroll wheel ok using IE, but not on FF or Chrome?

    Switching between browsers during debugging and noticed the scroll  wheel does not work in Firefox or Chrome, but is fine in Internet  Explorer. Anyone else notice this? Suppose I could use a scroll listener  and manually do it, but would much rather

  • How to Currency Sign

    How can I change the currency sign in report like from dollar to pound sign. Any help and advice will be highly appreciated.

  • KB21N - cross-compnay posting - BSET

    dear all, I have following Problem and I hope, someone can help: We use Transaction KB21N and execute: 1. an direct activity allocation between two cost center in one compnay and different profit center 2. an direct activity allocation between two co

  • Can I make a 'Contact Form' in Fireworks CS4?

    I would like to inbed a simple 'Contact Form' into one of the webpages that I am creating in Fireworks CS4 so that people could type a message, and then it would go straight to my email. Any help that you could provide will be appreciated! Thank You!

  • SQL 2008 Overflowed an int but the value is..integer...

    Hi guys, having create table #rete (mpo int) insert into rete (mpo) values ('2380001112857') The conversion of the varchar value '2380001112857' overflowed an int column Why? I'm inserting two million rows but I'm struck because this stupid issue. I