IOT Distirbution and Replication

I am currently taking Oracle's 9i DBA training on CD. I was surprised when the course told me that unlike regular tables, Index Organized Tables do NOT support distribution and replication. This surprised me enough to come here to confirm this fact. A quick google search led me to the following Oracle document (http://otn.oracle.com/products/oracle9i/datasheets/iots/iot_ds.html) which contradicts Oracle's courseware by saying it DOES support distribution and replication.
Can anyone confirm this one way or another? And if it does support distribution and replication, how accurate is the rest of Oracle's 9i courseware?
Thank you,
Scott Davies

Check below.
[http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:595657388029]
[http://www.oracle.com/technology/products/oracle9i/datasheets/iots/iot_ds.html]

Similar Messages

  • Database Mirroring and Replication in SQL Server 2008 R2

    I have configured the mirroring and replication between 4 servers (A,B,C,D).i.e, Mirroring between A to B and C to D, Replication between A and C. The configuration was a success and I am able test the replication(B to C) when I have failed over the mirroring
    dbs(A to B). The replication works fine after the mirroring fail over but I am not able to check its status in the Replication monitor. When I am having any insert in a table which is replicated in B, it is reflected back to C..it means the replication is
    working fine.
    Any thought on how I can view the status of replication from mirrored server. I tried adding the publisher in the monitor, but no luck. If I check the snapshot agent status, it says could not retrieve the info, same with the log reader agent status.
    Any suggestions on this please.
    Thanks, Siri

    For example in your case...
    Server A is principal and Server B is Mirror with either Manual or Automatic Failover.
    Server A is replicated to Server B ( publisher & B is subscriber )
    In Server A Database named Test_Mirror_Replication is configured for both mirroring and Replication.
    Now you have failed over your Database 'Test_Mirror_Replication' from Server A to Server B.
    After the Failover Server A will act as Mirror for DB 'Test_Mirror_Replication' & Server B will act as Principal for DB 'Test_Mirror_Replication'
    Hope my understanding is correct ?
    If yes then have you tried monitoring the replication after registering in other instance with current principal database sql instance name ? I mean current Publisher database name sql instance ? not your old sql instance name which was before the mirroring
    role change or failover...
    Hope you are trying with mirror database server name ?
    Raju Rasagounder MSSQL DBA

  • Only Remote-Desktop and Replication with 0.0 Reduction

    Hi guys,
    I have a solution with 16 WAAS 4.4.3 and I got good numbers in all default applications except Remote-Desktop and Replication
    ( All Traffic 53% of Reduction).
    I can understand not otimize Remote-Desktop Applications cause the customer is using Metaframe with crypto and already compacted and
    using too RDP with crypto/compacted.
    But in the Replication I would expect some gain in Active Directory and MS-FRS Replication.
    I got 0.0, Actually, the optmized traffic is biggest that the Original Traffic.
    I have to modify something in the Microsoft servers to be able to got reduction in Replication Application ??
    I'm using all the Default Classifiers in Replication Application
    Double-Take                            LZ+TFO+DRE       1100, 1105
    EMC-Celerra-Replicator            LZ+TFO+DRE       8888
    MS-AD-Replication1                 LZ+TFO+DRE       UUID:e3514235-4b06-11d1-ab04-00c04fc2dcd2
    ms-content-repl-srv                   TFO                     507, 560
    MS-FRS1                                LZ+TFO+DRE       UUID:f5cc59b4-4264-101a-8c59-08002b2f8426
    netapp-snapmirror                    LZ+TFO+DRE       10565-10569
    pcsync-http                             LZ+TFO+DRE       8444
    pcsync-https                           TFO                      8443
    rrac                                        TFO                      5678
    Rsync                                     LZ+TFO+DRE       873
    Thanks a Lot
    My Best Regards,
    Andre Lomonaco

    As you said, you need to only send the changes over the line.
    Java can not tell you what has changed, so you need to either write some JNI, or do a check on the current image, with the last image. (this would eat quite a bit of CPU time on the server).
    It might be worth looking at the source for VNC, and other simler products.
    The simpliest solution would be to split the image into blocks of x by y pixels, then do a pixel-by-pixel check on that block, and only send the block if it has changed.
    Next thing to think about is scrapping RMI, and using raw sockets. RMI has a fair sized overhead.
    Zipping a JPEG is unlikely to have any effect. A JPEG is already compressed.
    You might want to look at a lossless compression format, like PNG, as the JPEG will have articles. This might be acceptable. If so, changing the compression ratio of the JPEG might be an option.
    Finally, the bottle neck might be the call to Robot.createScreenCapture. This is not the fastest method.

  • How to automate checking status of Replication agent and Replication queue in CQ5

    As part of monitoring the status of replication agents and replication queue, is it possible to automate the checking process?
    We can check them from GUI, but from a monitoring perspective we are loking for a mechanism where the status can be checked continuously and we are alerted whenever there is some issue.
    1. Check status of replication agent (enabled/disabled)
    2. Check status of replication queue (if the queue is blocked)
    Thanks in advance
    nkrao

    you can pull the queue from the page the same way the UI gets the list
    in the table:
    curl -u admin:admin
    "http://localhost:4502/etc/replication/agents.author/publish/jcr:content.queue.json?agent=p ublish"
    and then use a shell library like ticktick.sh to parse the json
    response. A sample of how to get the bundles loaded and installed from
    CQ using the ticktick library can be found here
    https://github.com/kristopolous/TickTick
    http://ruben42.wordpress.com/2012/04/24/using-ticktick-sh-to-parse-json-responses-from-ado be-cq5-5/

  • Best Practice Adding New Target to Namespace and Replication

    Hi,
    whats the best way to add a new target to Namespace and replication. Goal is to replace a old file Server at the end.
    I did the following:
    - copied the share with robocopy incl timestamps of files and folders
    - created share
    - added the new share as a new target as well as meshd member of the replication connection
    - disabled the new member in the Namespace, so no one can Access it until dfsr is fully done and initialized
    After the the new dfsr Connection was replicated through AD to all 4 Members (3 different site, 1 same site) the
    following happend:
    dfsr begin and almost every file was in a conflicted and copied over the the Conflict Folder. Almost all timestamps
    of the Folders were changed to the current date, but the timestamps of the files not.
    Thousands of eventlogs: 4412
    The DFS Replication service detected that a file was changed on multiple servers. A conflict resolution algorithm was used to determine the winning file. The losing file was moved to the Conflict and Deleted folder.
    Any idea why? Later on i disabled the Connections to the remote Fileservers, but that did not stop it.
    My idea was to pre-seed the files with robocopy. So what would be the best way to prevent that for the next share ? Is it a better way to just add the target to a bi-directional Connection to the local Fileserver without adding to DFS-N and without copying
    the files before ? Is it better to let DFSR do the hole Initial sync incl Files ?
    At the end i have no loss of date but to check almost every file for conflict took Ages to finish.
    Thanks a lot,
    Marco

    Hi,
    The steps you performed are correct - compare with waiting for DFS initial replication, a manually pre-staging is recommended.
    When doing the Robocopy step, wether all attributes are copied such as NTFS permissions?
    After robocopy, you can add that folder as a folder target of DFS replication group - you can add it to DFS namespace after replication finished.
    And if it is a Windows 2012 R2, you can prestaging DFS database for a better result.
    https://social.technet.microsoft.com/Forums/windows/en-US/a06c9d25-ed04-44e9-a1f7-e1506e645d53/forum-faq-how-to-prestaging-dfsr-database-on-windows-server-2012-r2?forum=winserverfiles
    Please remember to mark the replies as answers if they help and un-mark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • How to use hotback to sync master and replication?

    Hi,
    If replication is down for some time, the data difference between master and replication will be large.
    My question is: is it possible we just hotbackup the data and log of master, and then let replication recover from the backup?
    I did 3 test:
    (1) Test condition: master application keep writing, db_hotbackup is not run. Result: the log of master application is good.
    (2) Test condition: master application make no writing, run db_hotbackup. Result: the log of master application is good.
    (3) Test condition: master application keep writing, run db_hotbackup. Result: the log of master application will be corrupted. It reports 'checksum mismatch' error when I try to recover the data.
    I am wondering why (3) could happen?
    Thank you very much.
    -yuan

    At this time there is no connection between iPhoto for the Mac and for IOS. Hopefully that will come in the future.
    LN

  • Database Mirroring and replication

    Hi,
    can mirroring and replication is possible on same instance of sql
    of sql server 2008. if yes then please provide the  information.
    thanks in advance.

    You may need to use traceflag 1448 to get this to work.
    looking for a book on SQL Server 2008 Administration?
    http://www.amazon.com/Microsoft-Server-2008-Management-Administration/dp/067233044X looking for a book on SQL Server 2008 Full-Text Search?
    http://www.amazon.com/Pro-Full-Text-Search-Server-2008/dp/1430215941

  • Password Managerment and Replication

    Hi All,
    I came across this problem of Replication and Password Policy. I have a Multi master with 3 dedicated consumers set up with replication. Account lock out policy is set up to lock the account in 3 attempts. Once the account is locked out, I reset it in the master (clear the account unlocktime, reset the failure counter and reset the retrucountresettime. But this is not getting replicated to the consumer, I read in the Admin's Manual certain attributes like account lockout counters are not replicated !!! If so since it's a dedicated consumer, how the values can be reset , to allow the user to logon ? Your help is very much appreciated,
    Solaris, iDS5.1 SP1
    Thanks
    chandra mouli

    Some clever programmer at Sun will have to figure out how to integrate the chaining database, entry distribution, and replication to chain bind requests to the master.

  • Issue with CDC and Replication enabled

    Hello,
    We have this strange issue with CDC and replication. Let me explain
    1. We have a database on write server and we replicate some tables to the read server. There are 15 tables that we replication and 8 of them have computed columns that are persisted.
    2. We also have CDC enabled on the same database where we have transactional replication enabled. I know that both CDC and replication uses replication log reader. Some how, all the time we see the log_reuse_wait says replication
    3. If I add around 100-200 MB into these tables, with these persisted columns, it will be around 500 MB of data. But the replication is queuing up 10-15 GB of data.
    4. I checked CDC tables, and the updates are in cdc tables. Also, I don't see CDC capture job. Is this because there is already replication enabled?
    What might be the issue that's causing the log to hold for a very long extended periods of time? We don't see any issue with log reader and CDC.

    2. Log_reuse_wait will show replication status both for CDC and replication.
    4. Yes as you are using transaction replication, Log Reader Agent is created for the database and the capture job won’t exist.
    When the Log Reader Agent is used for both change data capture and transactional replication, replicated changes are first written to the distribution database.
    Then, captured changes are written to the change tables. Both operations are committed together. If there is any latency in writing to the distribution database, there will be a corresponding latency before changes appear in the change tables.
    https://msdn.microsoft.com/en-us/library/cc645938.aspx?f=255&MSPPError=-2147217396
    As you said CDC updates are in cdc table I don’t see any issue.
    You could run DBCC OPENTRAN to see the old active transaction? It will give you more info.

  • Afaria Database on SQL Anywhere 12 and Replication

    Hi,
    I have Afaria Farm and I Use SQL Anywhere 12 as database server.
    I want to replicate this database to another SQL Anywhere 12 Server.
    Which one is good way to replication
    And can I have some documentation that tells this step by step?
    I am not a DB admin.

    Hello Tevfik,
    I think we need to understand your requirements a little more: why are you interested in SQL Anywhere replication and what are your plans for it?
    If you're looking for an always-available SQL Anywhere database for Afaria, there is high-availability (database mirroring). This will allow you to maintain two database server partners (with a third arbiter), where the primary has read/write capabilities and the mirror partner is read-only. There is a tutorial here in the documentation.
    If you're looking for a reporting server to do read-only queries for reporting, there is read-only scale out (copy nodes). The tutorial can be found here.
    Are you trying to keep a backup copy of your database via replication? If so, you may want to consider live backups (see the wiki), or just regular backups (also see the wiki). Regular backups should still be used in conjunction with any of the high-availability scenarios described above.
    If you're really looking for data movement, particularly to multiple database nodes, there is synchronization to other enterprise databases (including SQL Anywhere, but this also includes HANA, ASE, IQ, etc.) with MobiLink and replication to other SQL Anywhere databases, via SQL Remote. MobiLink uses a session-based HTTP or TCP/IP connection as a transport, while SQL Remote can use offline FILE, FTP, SMTP, and online HTTP as a transport.
    You may read elsewhere that historically SQL Anywhere can also be used with SQL Remote for Adaptive Server Enterprise (ASE) or Replication Server for replication, but both of these methods have been deprecated and removed in current versions and are not supported by development.
    Regards,
    Jeff Albion
    SAP Active Global Support

  • Unable to auto-start cache and replication agents

    Hi,
    Everytime we start server, we need to manually start cache and replication agents!
    Is there any way to configure them starting automatically???

    Yes:
    ttAdmin -repPolicy always -cachePolicy always DSN
    Chris

  • Session migration and replication

    Hi All,
    I am having a hard time in configuring my application for HTTP session migration. Our weblogic server consists of two managed servers running in same cluster. Each server has an Ehcache that stores some information of user with key as session ID and value as info object. In case, if a server needs restart, we would want to take this updated info Object from cache residing on server being restarted to another managed server within the same cluster.
    I browsed through many documentations online. Most of them explained about session replication but not migration. so I followed replication (I don't want a real time sync up of HTTP session. I want it to migrate if something goes wrong with one of the managed server).
    However, I could not achieve this task after following the steps to configure this feature. I would appreciate a lot if someone can help me figuring out the issue here.
    Here is what I did.
    1) Weblogic.xml
         <session-descriptor>
    <persistent-store-type>replicated_if_clustered</persistent-store-type>
    </session-descriptor>
    2) An implementation class of HttpSessionActivationListener, HttpSessionListener, HttSessionAttributeListener:
    public void sessionDidActivate(HttpSessionEvent sessionEvent) {
    // NEVER GETS CALLED
    over here i would check if session has any attribute with name 'CACHE_ELEMENT'. If yes, then it is a migration case for current managed server
              Log.info(UserCacheMigrationListener.class, "inside sessionDidActivate");//
    public void sessionWillPassivate(HttpSessionEvent sessionEvent) {
    // NEVER GETS CALLED
    Over here I would set attribute 'CACHE_ELEMENT' so that it is available for target managed server when it's sessionDidActivate is called
              Log.info(UserCacheMigrationListener.class, "inside sessionWillPassivate");
    public void sessionCreated(HttpSessionEvent sessionEvent) {
    THIS GETS CALLED! and I set the following attribute
    sessionEvent.getHttpSession().setAttribute(UserCacheMigrationListener.class.getName(), this);
    public void sessionDestroyed(HttpSessionEvent arg0) {
    THIS GETS CALLED!
    public void valueBound(HttpSessionBindingEvent arg0) {
    THIS GETS CALLED! WHICH MEANS THAT I WAS ABLE TO setAttribute this class's instance in sessionCreatedMethod
    In the above code, the setAttribute method used inside the sessionCreated(..) method successfully sets the attribute to this session. This is apparent because valueBound(..) method is called when session is created. But why does it not call the sessionWillActivate method???
    3.) An entry in web.xml for this listener.
         <listener>
              <listener-class>com.xyz.UserCacheMigrationListener</listener-class>
         </listener>
    4) from weblogic config.xml. I am copying all meaning stuffs from config.xml to describe as much as I can.
    <server>
    <name>AdminServer</name>
    <ssl>
    <enabled>false</enabled>
    </ssl>
    <listen-address>localhost</listen-address>
    <network-access-point>
    <name>AdminChannel</name>
    <protocol>t3</protocol>
    <listen-address>localhost</listen-address>
    <http-enabled-for-this-protocol>true</http-enabled-for-this-protocol>
    <tunneling-enabled>false</tunneling-enabled>
    <outbound-enabled>false</outbound-enabled>
    <enabled>true</enabled>
    <two-way-ssl-enabled>false</two-way-ssl-enabled>
    <client-certificate-enforced>false</client-certificate-enforced>
    </network-access-point>
    <data-source>
    <rmi-jdbc-security xsi:nil="true"></rmi-jdbc-security>
    </data-source>
    </server>
    <server>
    <name>Node1</name>
    <ssl>
    <enabled>false</enabled>
    </ssl>
    <machine>DevMachine</machine>
    <listen-port>7002</listen-port>
    <cluster>DevCluster</cluster>
    <replication-group>devGroup1</replication-group>
    <preferred-secondary-group>devGroup2</preferred-secondary-group>
    <web-server>
    <keep-alive-secs>500</keep-alive-secs>
    <post-timeout-secs>120</post-timeout-secs>
    </web-server>
    <listen-address>localhost</listen-address>
    <jta-migratable-target>
    <user-preferred-server>Node1</user-preferred-server>
    <cluster>DevCluster</cluster>
    </jta-migratable-target>
    <data-source>
    <rmi-jdbc-security xsi:nil="true"></rmi-jdbc-security>
    </data-source>
    </server>
    <server>
    <name>Node2</name>
    <ssl>
    <enabled>false</enabled>
    </ssl>
    <machine>DevMachine</machine>
    <listen-port>7003</listen-port>
    <cluster>DevCluster</cluster>
    <replication-group>devGroup2</replication-group>
    <preferred-secondary-group>devGroup1</preferred-secondary-group>
    <listen-address>localhost</listen-address>
    <network-access-point>
    <name>Node2Channel</name>
    <protocol>t3</protocol>
    <listen-address>localhost</listen-address>
    <http-enabled-for-this-protocol>true</http-enabled-for-this-protocol>
    <tunneling-enabled>true</tunneling-enabled>
    <outbound-enabled>false</outbound-enabled>
    <enabled>true</enabled>
    <two-way-ssl-enabled>false</two-way-ssl-enabled>
    <client-certificate-enforced>false</client-certificate-enforced>
    </network-access-point>
    <jta-migratable-target>
    <user-preferred-server>Node2</user-preferred-server>
    <cluster>DevCluster</cluster>
    </jta-migratable-target>
    <data-source>
    <rmi-jdbc-security xsi:nil="true"></rmi-jdbc-security>
    </data-source>
    </server>
    <cluster>
    <name>DevCluster</name>
    <cluster-messaging-mode>unicast</cluster-messaging-mode>
    </cluster>
    <machine>
    <name>DevMachine</name>
    <node-manager>
    <nm-type>Plain</nm-type>
    </node-manager>
    </machine>
    <migratable-target>
    <name>Node1 (migratable)</name>
    <notes>This is a system generated default migratable target for a server. Do not delete manually.</notes>
    <user-preferred-server>Node1</user-preferred-server>
    <cluster>DevCluster</cluster>
    </migratable-target>
    <migratable-target>
    <name>Node2 (migratable)</name>
    <notes>This is a system generated default migratable target for a server. Do not delete manually.</notes>
    <user-preferred-server>Node2</user-preferred-server>
    <cluster>DevCluster</cluster>
    </migratable-target>
    ------------------------------------------------------------------------------------------------------------------

    Hi,
    So you want to migrate your server to server.
    Here are the following links which help you.
    http://docs.oracle.com/cd/E15051_01/wls/docs103/cluster/migration.html
    http://www.oracle.com/technetwork/middleware/weblogic/messaging/wlasm-1853193.pdf
    let me know the status if you need any further help on this issue.
    Regards,
    Kal

  • Server 2012 R2 Domain Controllers stop accepting log in and replication failure.

    We have recently completed a Domain upgrade to 2012 R2 AD DS from native 2003 AD. The issue we are experiencing is that for seemingly no reason the 2012 hosts will randomly stop accepting (RDP) logins and we see replication failures. We also cannot
    otherwise remotely manage the host (remote service management etc...). The only way I've been able to recover from this condition is to restart the affected host, after which (RDP) logins are again accepted and successful replication starts again.
    Logs aren't showing any issues prior to the issue happening and it is not specific to one host. I have been digging around to find a solution to this abnormal behavior but as yet haven't come across anything specific to what I am seeing.
    Domain level is 2008 R2 and our few remaining 2008 R2 hosts never experience this. Only seems to be the 2012 hosts and randomly among all the DCs. Not one specific 2012 host is affected.
    We have 20 DCs total.
    Are there any known issues? Any advice on what I may be able to look at?
    We are otherwise healthy aside from this intermittent replication/login issue.

    This sounds more like a network type issue than AD type problem.
    Are any firewalls in the way? 
    Do the servers respond to a ping when they are not responding to RDP.
    What are you seeing in event logs prior to the change. 
    Are the servers physical or virtual? Are you able to get onto the console - either through your HyperVisor or through ILO type management?
    Thanks
    Regards,
    Denis Cooper
    MCITP EA - MCT
    Help keep the forums tidy, if this has helped please mark it as an answer
    Blog: http://www.windows-support.co.uk 
    Twitter:   LinkedIn:

  • Document/customer pricing and replication of pricing data from r/3 to crm

    Hi,
    can any one tel me how can we replicate the r/3 pricing condition, pricing procedure to crm.
    once it is replicated  to crm what customization is require in crm.
    tel me abt customer and document pricing procedure, how it will be effected in the transactions.
    plz any one help me to know this scenarios.
    thanks
    kumar

    Hello Sathish,
    Start from here: [Data Transfer from SAP ECC (Download)|http://help.sap.com/saphelp_crm60/helpdata/en/51/62431b7b3511d3ad650000e8a5bd28/content.htm] The help document gives you the steps required. You can also see the [Master Guide |https://service.sap.com/~sapdownload/011000358700001398312008E/MG_CRM2007V2.pdf]or [Customizing Replication Guide|http://help.sap.com/bp_crmv250/CRM_DE/BBLibrary/Documentation/B09_BB_ConfigGuide_EN_US.doc] for similar configuration information.
    Essentially after you download the pricing related customizing data from ECC into CRM, you have to download the conditions. the standard conditions get downloaded without any issues since the download adapters are defined already. However, for downloading custom conditions in ECC, you have to do further configuration. You have to define the adapter objects and do the initial download of those conditions. Such an information is available in the [Master data and Transaction data Replication Guide|http://help.sap.com/bp_crmv250/CRM_DE/BBLibrary/Documentation/C03_BB_ConfigGuide_EN_US.doc].
    Easwar Ram
    http://www.parxlns.com

  • Oracle Text and Replication

    We need to use an Oracle Text index on an application table in a read-only materialised-view replicated environment. The Oracle Text index needs to exist at both the master site and the materialised-view site.
    We see two possible alternatives:
    (1) Replicate read-only copies of the Oracle Text Index tables and indexes to the materialised view site, in addition to replication of the application table.
    or
    (2) Replicate only the master application table and create local instances of the Oracle Text index at both the master and materialised view sites.
    Which method is best? Any experience, advice or references that you can provide would be much appreciated.
    Many thanks for your help,
    Peter

    See [url http://download-uk.oracle.com/docs/cd/B19306_01/server.102/b14226/repoverview.htm#i15730]Introduction to Advanced Replication, [url http://download-uk.oracle.com/docs/cd/B19306_01/server.102/b14229/strms_over.htm#i1006084] Introduction to Streams and [url http://download-uk.oracle.com/docs/cd/B19306_01/server.102/b14228/gen_rep.htm#i1007573]Understanding Streams Replication

Maybe you are looking for