Question about advance compression in Oracle 11gR2

Hi,
I am on Oracle 11gR2 on Solaris 10. I want to run the oracle advance compression advisor for my database and get compression ratios for the tables, how can I do it? I am looking for a sample command to run this advisor package (dbms_compression) from sqlplus.
Can someone please suggest.
Thanks,
Nirav

Thanks SriniChavali and Stefan. I can't make more answer "Helpful' so I couldn't do that for your answers! Here is my point. In Jonathan's blog i see this remarks:
"Sadly it seems that “compress for OLTP” (formerly “compress for all operations”) doesn’t compress for all operations, it compresses only for inserts, and the benefits it has over basic compression are that (a) it leaves 10% of the block free for updates, and (b) it doesn’t require direct path inserts to trigger compression. Given the limitations on how it works you may find that the problems it brings might make it something you want to avoid.... this is at this link: http://allthingsoracle.com/compression-in-oracle-part-3-oltp-compression/
To date I’ve only heard complaints about OLTP compression (there’s an element of self-selection there as no-one ever calls me to look at their system because it’s running so well and has no problems). A common thread in the complaints I have heard, though, is about the significant amount of row migration (once it has been noticed), the extra CPU, and ”buffer busy waits”.
Compression for OLTP is (according to the manuals) supposed to be able to compress during updates – but it doesn’t (at least, as far as I can tell); this means that you can easily end up suffering a large number of row migrations on updates, which can result in extra random I/Os, buffer busy waits, and increased CPU and latch activity.
If you can work out a good strategy for using OLTP compression, though, think carefully about making a choice between freelist management and ASSM – there seem to be some undesirable side effects that appear when you mix OLTP compression with ASSM."
Note that I am not that technical and I hope I have not mis-quoted. I am trying to figure out if this is a good option or not and seeing some findings that it may not be so good and seeing at some other places notes that it is indeed very good.
Best regards

Similar Messages

  • Basic vs Advanced Compression in Oracle 11g

    Hi,
    We are going to install Oracle 11gR2 in one new database server. Since the database will be used for Data Warehouse purposes and our company has rejected to pay licenses for "Oracle Advanced Compression", I wanted to know which options of "Basic Compression" are suitable for us, if we want to compress the biggest tables in our environment.
    Thanks in advance for your feedback.
    Regards,
    Rubén

    Hi;
    OAC mean you need to pay extra money to oracle. We are using basic compression in our env. Please check this search. Check first 4 links which is explain already how&what basic compress
    PS: This is installation related forum site. For your similar issue please use Oracle Discussion Forums » Oracle Database » General Questions forum part
    Regard
    Helios

  • Question about table compression

    I administer several 10g databases, and have been asked to compress tables in several tablespaces. I am unfamiliar with using compression. I have read some Oracle documentation on the ALTER TABLE... MOVE COMPRESSION command, but still have several questions. Can anyone recommend a good white paper, or other documentation on how best to compress tables with existing data? Thank you in advance for your guidance!

    Basically, I have questions about how to select which tables will benefit from compression. I also read somewhere that, once a table has been compressed, new updates to the compressed data will not be compressed and so I will need to re-compress the tables periodically.
    I also needed to know if the ALTER TABLE statement will compress the existing data or just future inserted data. Through further research I found that by using the MOVE clause, it will compress the existing data. I will also need to rebuild the indexes once the ALTER TABLE statement is complete.
    Thank you for your help and encouragement!

  • Question regarding shared_pool sizes in Oracle 11gR2

    Hi! I administer an Oracle 11gR2 database that runs on IBM hardware, OS AIX 6.1. The LPAR has 44 Gb of available memory and 8 dual-core processors assigned.
    Normally, everything runs fine and we have no problems. Today the database started rejecting connections suddenly. The issue lasted for about 14 minutes and then cleared itself. When checking the alert log, it indicated an out of memory problem with the shared_pool, ORA-4031 error.
    I am utilizing the automatic memory management feature of 11g, and memory_max_target is set to 14 Gb. I am going to increase memory_max_target to 18 Gb as part of my solution, but will that also then increase the size of the shared_pool in the SGA? Is it best to let the database manage the size of the shared_pool, or can (and should) I set it to a minimum size?
    Any help, or links to documentation or MOS notes, is greatly appreciated. Thank you!
    Mark

    Before suggesting you change your memory managment in use I would want more information such as how long have your been running the system this way?  How long had the system been running since its last startup?
    If you have access to the AWR then what does the AWR report for the problem period show?  Any unusual activity?  Were any new features or major processing added to the application recently.
    When you look at the memory management views how frequently does Oracle move granules from PGA to SGA, from the buffer cache to the shared pool and so on.
    The following support note may help with trying to investigate what issues you face in relation to ORA-04031 providing the issue re-occurs.
    Oracle document# 1088239.1 Master Note for Diagnosing ORA-4031
    HTH -- Mark D Powell --

  • Questions about deploying compressed WAR

    I've been having issues when trying to deploy a compressed
    Flex application into my Weblogic Server. I read in the
    Installation Guide (
    http://www.adobe.com/support/documentation/en/flex/2/install.html)
    that you need to some "hacking" to make it work but i have a big
    question:
    The guide says:
    Locate the servlet definition for MessageBrokerServlet, and
    update the flex.write.path init-param element to specify an
    existing, writeable, directory path:
    <init-param>
    <param-name>flex.write.path</param-name>
    <param-value>C:/bea/flex</param-value>
    </init-param>
    I have no clue about this, i mean what is supposed to be in
    the path C:/bea/flex??? Can anyone explain what the
    "flex.write.path" means or referrers to?
    Diego

    I tried setting the write path to any folder like the example
    and i got this during the server startup:
    [Flex] [WARN] The watch-file,
    /WEB-INF/flex/services-config.xml, could not be resolved to a path
    and will be ignored.
    [Flex] [WARN] The watch-file, /WEB-INF/flex/proxy-config.xml,
    could not be resolved to a path and will be ignored.
    [Flex] [WARN] The watch-file,
    /WEB-INF/flex/remoting-config.xml, could not be resolved to a path
    and will be ignored.
    [Flex] [WARN] The watch-file,
    /WEB-INF/flex/messaging-config.xml, could not be resolved to a path
    and will be ignored.
    [Flex] [WARN] The watch-file,
    /WEB-INF/flex/data-management-config.xml, could not be resolved to
    a path and will be ignored.
    [Flex] [WARN] The touch-file, /WEB-INF/web.xml, could not be
    resolved to a path and will be ignored.
    And then when i tried to run the sample application CRM
    (which is included in the samples war when you install flex 2) i
    get this error:
    java.sql.SQLException: The url cannot be null
    so what's wrong?

  • Question about "RE-COMPRESSION" of mpeg 4 files

    First off this is a great forum. It has answered many questions. This is my first post.
    This is my workflow.
    1. I have a massive library of Betacam tapes that I am archiving ( over 2000) average length 45 minutes)
    2. I capture them from a betacam deck, thru a canopus adcvc 110, at SD, No problems.
    3. I wind up with media files that are like 15 gigs. I have over 40 terabytes of external storage, I use firewire 800 (so far, so good)
    4. I then take the media files, drag them to the timeline, edit etc, add graphics from motion etc
    5. I then simply play the timeline via firewire to a dv, video deck, and send the tape to the tv station, they closed caption it and it hits the air
    (so life is great, this works perfect)
    Here comes the question
    1. After having done that, I use the "send to compressor" function, to generate an mpeg 4 file, which instead of 12 gigs is like less than one, it looks great.
    2. This file is then uploaded to an ftp site, given a description, and then put on the website for viewing or purchase.
    3, since I have like 2000 tapes, do the math, if each one takes 12 gigs....lots of drives.
    4. Most tapes are NOT edited for television broadcast, I just find a good in and out point for a fade, compress it, shoot it to ftp and im done.
    5. Recently I have begun deleting the 12 gig files, sine I have the messages in mpeg 4...to avoud buying several petabytes of storge.
    SOOOO
    After all that, here is my question.
    If I want to edit a program that I have in the archive, that the decision has been made to broadcsat for some reason, If I drag the MPEG$ file onto the timeline, edit it from a 50 minute down to a 28:30, drop in green screen cut aways, grapics, lower thirds animated transitions between clips, titles, etc...
    Render it.
    THEN play it directly from the timeline via firewire to the mini dv deck...
    How much quality will I lose on the base video, which is mpeg 4, as compared to the 12 gig media file?
    Does it "re- compress" in some way?
    should I just suck it up and pay for storage for the 12 gig files that probably will never use for broadcast?
    or can I use an MPEG 4 file as my "main video" and not come out looking like an amature moron?
    Does any of that make sense? or is it TMI from an idiot?
    I come from the analoge days, D-3's and D-5's, grass valley 200 switcher, pinnacle dve's, 18 Ikegamis shooting NFL football. Betacams, D-1's, sattelite uplinks. all in good old NTSC 525...
    So bear with an old man.
    Any assistance would be great.
    Can I edit using an MPEG4 file and not get burned?
    Be well all.
    Michael D, Fodor.
    heritagechristiancenter.com
    San diego.
    cya stay out of the heat...

    You probably won't have a huge loss in quality-- as you say yourself, the MPEG4 looks pretty good for it's file size. I won't get into a debate about whether or not it's the best option for archival, that's another thread.
    But as long as you're editing in FCP (7), you'll want to convert it to something else before importing. ProRes is a pretty standard choice. Otherwise it'll be nothing but pain. You can still archive in MPEG-4, but for those that you pull from the archive to re-broadcast, recompress them to ProRes.

  • Questions About Chapter 2 in Oracle DB 10g: SQL Fundamentals II

    Hello,
    first of all i'm glad to be a member of your forum. I have joined a beginner Oracle Course: Intro to SQL. I'm facing some problems understanding some concepts in Chapter 2 of Oracle Database 10g: SQL Fundamentals II text book. I got about 15 questions. However, i will only ask two questions at first. Since i'm a newbie, please answer it in a simplistic form. Excuse me if you see grammatical mistakes.
    Dropping a column can take a while if the column has a large number of values. In this case it may be better to set it to be unused and drop it when the number of users on the system are fewer to avoid extended locks.
    Questions:
    "when the number of users on the system are fewer to avoid extended locks."
    1. Can you explain this to me please?! fewer than before? fewer than? What if users kept increasing during the years! then this "fewer" may not happen until the company collapse!
    2. Why do we need to use unused columns? When should we use unused columns?

    Great! .... I got more questions, i just do not want to open a new same thread. Thus, i will just post the questions in here and i hope i will get help from experts...Please bare with me guys...The questions are numbered, unnumbered parts are information that helps you understand my question.
    Note: just answer what you are willing to, one question or whatever you want. I'm not expecting to get all the answers from one member :)
    Thanks for understanding
    Page 2-7:
    Certain columns can never be dropped such as columns that form part of the partitioning
    key for a partitioned table or columns that form part of the primary key of an index- organized table.
    Questions:
    "columns that form part of the partitioning key for a partitioned table"
    *1. Do they mean one table can be split into two different storage? What is the thing that*
    link these both tables to make Oracle Server realize these two tables are actually one  table? Is is tablespace_name?
    "columns that form part of the primary key of an index-organized table."
    *2. Can you clarify the above sentence please*
    *3. If i have set of columns that has large amount of data, i rather set them unused then*
    drop them because the response time is going to be faster! I do not get it, can you
    explain please? What i know is drop drops the column and release the disk space whilst
    unused column make the columns useless and does not release disk space yet until we drop them, so
    drop column does it in one step unlike taking the unused column process. In brief, i would like to know
    why dropping unused columns that has large set of data is faster then dropping the column
    directly...
    Page 2-12
    4. ALTER TABLE emp2 ADD CONSTRAINT emp_dt_fk
    FOREIGN KEY (Department_id)
    REFERENCES departments ON DELETE CASCADE);
    The above query is written in text book. I think it should be written as
    ALTER TABLE emp2 ADD CONSTRAINT emp_dt_fk
    FOREIGN KEY (Department_id)
    REFERENCES departments(dept_id) ON DELETE CASCADE;
    Am i correct?
    *5. Can you tell me what deferring constraints is in one sentence please? Why do we need it? When do we need it in real-life?*
    *7. You can defer checking constraints for validity until the end of the transaction. A*
    constraint is deferred if the system checks that it is satisfied only on commit. If a
    deferred constraint is violated, then commit causes the transaction to roll back.
    I do not understand the above paragraph, please explain. What i know is "end of
    transaction" ends with ; or commit
    Page 2-18
    create table test1 (
    pk NUMBER PRIMARY KEY,
    fk NUMBER,
    col1 NUMBER,
    col2 NUMBER,
    CONSTRAINT fk_constraint FOREIGN KEY (fk) REFERENCES test1,
    CONSTRAINT ck1 CHECK (pk > 0 and col1 > 0),
    CONSTRAINT ck2 CHECK (col2 > 0) );
    -- "CONSTRAINT fk_constraint FOREIGN KEY (fk) REFERENCES test1"
    *8. This is wrong isn't it? It references to test1 but no column specified.*
    An error is returned for the following statements:
    ALTER TABLE test1 DROP (pk); -- pk is a parent key.
    *9. We can not drop it because we did not mention ON DELETE CASCADE. Am i right?*
    ALTER TABLE test1 DROP (col1) -- col1 is referenced by multicolumn constraint ck1.
    *10. I do not get it, can you explain please. col1 is not referenced, i see CHECK constraint is applied*
    but no references made. Secondly, is ck1 considered multicolumn because it check two columns?
    Or multicolumn here represents something else?
    ALTER TABLE emp2
    DROP COLUMN employee_id CASCADE CONSTRAINTS;
    *11. This drop employee_id column and all its child. Correct?*
    ALTER TABLE test1
    DROP (pk, fk, col1) CASCADE CONSTRAINTS;
    *12. This drops three columns and all its child if there are any. Correct?*
    *13. Then what's the difference between ON DELETE CASCADE and CASCADE CONSTRAINTS?*
    For example, What if employee_id in emp2 table definition does not have ON DELETE CASCADE,
    will CASCADE CONSTRAINTS work? Please explain...
    Page 2-22
    When you are expecting a large data load and want to speed the operation. You may want
    to disable the constraints while performing the load and then enable them, in which case
    having a unique index on the primary key will still cause the data to be verified during
    the load. So you can first create a nonunique index on the column designated as PRIMARY
    KEY, and then create the PRIMARY KEY column and specify that it should use the existing
    index.
    Example:
    1. create the table
    create table new_emp
    (employee_id number(6),
    first_name varchar2(10)
    2. create the index
    create index emp_id_idx2 on new_emp(employee_id);
    "You may want to disable the constraints while performing the load and then enable them"
    so i suggest to load all data i want into new_emp.
    3. create the primary key
    alter table new_emp ADD primary key (employee_id) USING index emp_id_idx2;
    What i understand is the following:
    If we want to load large data into the new_emp, its better to create the table without any
    constraints - in our case the constraint is primary key. After that, we create nonunique
    index points to employee_id and then load data into new_emp. Finally, specify employee_id
    as primary key using the nonunique index.
    *14. Is my explanation correct?*
    "in which case having a unique index on the primary key will still cause the data to be
    verified during the load."
    *15. Data to be verified against what? Is it to be verified whether its NULL or NOT NULL? I*
    know primary key does not take NULL and every value must be unique.
    After loading all data we want, what if i did
    "alter table new_emp ADD primary key (employee_id);"
    *16. Will i face any problems or inefficient process?*
    I do not think we need step two, we could do the following:
    1. create the table
    create table new_emp
    (employee_id number(6),
    first_name varchar2(10)
    "You may want to disable the constraints while performing the load and then enable them"
    so i suggest to load all data i want itno new_emp.
    2. create the primary key
    alter table new_emp ADD primary key (employee_id);
    *17. The above steps are as efficient as the three steps i mentioned above. The only difference*
    is we let index be created implicitly. Right? If no, why?
    Page 2-23
    CREATE INDEX upper_dept_name_idx ON dept2(UPPER(department_name));
    The following statement may use the index, but without the WHERE clause the
    Oracle server may perform a full table scan:
    select *
    from employees
    where UPPER(last_name) IS NOT NULL
    ORDER BY UPPER (last_name);
    "but without the WHERE clause the Oracle server may perform a full table scan"
    *18. The above query let oracle server perform full table scan anyway! Right? It has to go*
    through every field and check is it not null or not. I know we are using function-based
    index but there are alot of not null last_name! so oracle server must scan one by one. If
    we only had one not null field, then i would say Oracle server can point to that field
    immediately by the aid of function-based index we created above. Can you clarify please...
    Another related topic statement that i do not get it yet:
    "The oracle server treats indexes with columns marked DESC as function-based indexes."
    *19. The bove statements is so general. What if we have a column ordered by DESC order and we*
    did not create any function-based indexes, will statement be true?!
    Lets go back the above query:
    ORDER BY UPPER (last_name);
    *20. Its not DESC. To me, the above query does not flow with this statement "The oracle server treats*
    *indexes with columns marked DESC as function-based indexes."?*
    Page 2-27
    Regarding FLASHBACK TABLE, you can invoke a flashback table operation on one or more
    tables, even on tables in different schema. You specify the point in time to which you
    want to revert by providing a valid timestamp. By default, database triggers are disabled
    for all tables involved. You can override this default behavior by specifying the ENABLE
    TRIGGERS clause.
    "By default, database triggers are disabled for all tables involved. You can override this
    default behavior by specifying the ENABLE TRIGGERS clause."
    *21. What are database triggers?*
    *22. About External Tables. What are external tables? When is it used in real-life? Why do*
    we want External Tables?
    Page 2-30
    Oracle server provides two major access drivers for external tables. They are
    ORACLE_LOADER access driver and ORACLE_DATAPUMP access driver. ORACLE_DATAPUMP used to
    both import and export data using a platform-independent format.
    "platform-independent format."
    *23. What is the format? Is it .dat?*
    Page 2-35
    CREATE TABLE oldemp ( fname char(25), lname char(25) )
    ORGANIZATION EXTERNAL
    (TYPE ORACLE_LOADER
    DEFAULT DIRECTORY emp_dir
    ACCESS PARAMETERS
    (RECORDS DELIMINATED BT NEWLINE
    NOBADFILE
    NOLOGFILE
    FIELDS TERMINATED BY ',' (fname POSITION (1:20) CHAR, lname POSITION (22:41) CHAR)
    LOCATION ('emp.dat') )
    PARALLEL 5
    REJECT LIMIT 200;
    *24. Can you please explain the below part:*
    ACCESS PARAMETERS
    (RECORDS DELIMINATED BT NEWLINE
    NOBADFILE
    NOLOGFILE
    FIELDS TERMINATED BY ',' (fname POSITION (1:20) CHAR, lname POSITION (22:41) CHAR)
    *25. Can you please explain what is PARALLEL 5? and Why do we need it?*
    Again, any help is appreciated...
    Edited by: user11164565 on Jul 21, 2009 4:41 AM

  • Question About Advanced Replication

    I red advanced replication concept on your blogs and i done sucessfully in my test database.
    you choosed to configured only one user and taken two tables.its ok. i want replicate all the users include all the tables to replication database just once time..i don't want to take one table and do it and then take another table and do it not like this.. i want configure all the tables to replication database, is there any procedure? may i want to know about this?
    Thanks & Regards
    S.Mohamed Azar
    Edited by: azarmohds on Jul 12, 2009 6:51 AM

    Request: Please remove my name from subject, it is public forum anyone can be participate.
    Please email "[email protected]" for personal.
    Check : http://download.oracle.com/docs/cd/B19306_01/server.102/b14226/repoverview.htm#REPLN001
    Thanks
    Regards,
    Taj
    Edited by: Mohammed Taj on Jul 12, 2009 5:51 PM
    Edited by: Mohammed Taj on Jul 12, 2009 6:02 PM
    Edited by: Mohammed Taj on Jul 12, 2009 6:08 PM

  • Question about BC4J data tags, Oracle sessions and Locking!

    Hi ,
    I have seen numerous examples of JSPs using data tags and in all the examples the data tag for the application module has the "username" and "password" harcoded in it.
    My questions are:
    1) For a stateful application should we be including the username and password in every JSP page. I personally believe that we should not.
    2) If we have a username and password in every JSP page will it not start a new ORACLE user session and if so will it not cause locking problems?
    3) If we don't hard code the username and password in every JSP page, will it reuse the same ORACLE session ?
    4) How do we avoid locking problems when we use data tags?
    5)I can understand the inclusion of username and password in every JSP page if it is a stateless application but again Is there a way we can avoid hardcoding the username and password in every single page?
    I would appreciate if some one can let me know if any of my assumptions are incorrect.
    JDeveloper Team/Juan any advice?

    The username and password are optional. They can be provided via the connections.properties file. The multple entries for username and password don't mean that separate connection are made. The first time the ApplicationModule tag is encountered, your application instance is created. If you are running in reserved mode (look at your releasePageResources tag) the application instance is kept until your Http session times out. If you are running in Stateful or Stateless mode, you application instance is returned to the application pool and retrieved the next time you need an instance. Please refer to the application pool documentation and to the source in oracle\jbo\common\ampool provided in jbohtmlsrc.zip.

  • Questions about JAVA/JBOSS on Oracle VM 2.2.1 - Performance

    I recently upgraded to VM 2.2.1 on an older 1.86 quad-processor (intel) system (DELL 2950). We have 3 VM's setup. One as an application server running JBOSS 5.1 and Jira and SubVersion. The second is a 10G Oracle Database server and the third is a 11G Database Server. This is all for development. THe Dell has 32GB of memory, 1 quad processor and 2 NIC (TOE) cards. I'm using SATA RAID 5.
    The developers are noting that the startup of JBOSS is taking 6-9 minutes which is substantially longer and they say the application is now slower. I've looked at everything and don't see any one blocking resource. I've increased the number of VCPU's to 4 for the Application Server and raised it's priority in OVM. There was no impact to the speed of the application or the boot process.
    JBoss seems to initiate something every few seconds that raises the CPU usage to 70-99% (out of 400%). On Windows systems that seems to occur as well.
    I'm curious if anyone else has a similar environment and what types of performance you are seeing with JBOSS in and Oracle VM environment on boots.
    Thanks.

    Maybe the following can useful:
    http://www.mastertheboss.com/en/jboss-howto/42-jboss-config/83-jboss-cpu-monitor.html
    What was the environment when you did not experience this problem? Can you provide more info?
    Network changes? NIC/Switch settings mismatch? Bad cabling? DNS problems?
    Does any of the following apply?
    http://stackoverflow.com/questions/1927071/improving-jboss-5-slow-startup
    Edited by: Markus Waldorf on Aug 25, 2010 2:19 PM

  • Quick question about advanced pulldown

    Hi,
    So, I am trying to make a MiniDV backup of a 23.98 sequence using FCP 4.5 and a Sony DCR-VX2000 camera. The footage in the sequence is non-time coded 23.98 footage from a DVD. To test if my backups are actually backups that I could later edit in 23.98, I have printed to tape, recaptured and removed the pulldown with either Cinema Tools or using Remove Advanced Pulldown, and placed the recaptured footage in a new 23.98 sequence. However, neither method has worked-- the wrong frames are removed or the captured footage just looked screwed up.
    I tried two methods: printed to MiniDV with a 2:3:2:3 pulldown in the Playback settings, recaptured into FCP using normal DV capture, then used Cinema Tools'"batch reverse telecine" to remove the pulldown, then placed the footage in a new 23.98 sequence When that didn't work, I tried a 2:3:3:2 pulldown print to tape and captured using Advanced Pulldown Removal. That looked even worse.
    Any suggestions? Thanks in advance.
    Powerbook G4 Mac OS X (10.3.8)

    Well, this is what I did:
    I created 4 tests clips, each printed to tape with a different playback setting (2323, 2332, 2224, full quality).
    I then recaptured them all using both "Advanced Pulldown Removal" and no pulldown removal with Cinema Tools Batch Reverse Telecine used to remove pulldown on the clips.
    Results: The proper frames seemed to be removed when 1) I printed to tape with a 2:2:2:4 pulldown and Reverse Telecined through Cinema Tools and 2) when I printed to tape with Full Quality and used Advanced Pulldown Removal. Nothing else worked.
    HOWEVER, with both "succesful" methods, I found that when I played the clips in a 23.98 sequence, they had noise in he picture -- thin black horizontal lines on areas of quick movement. However, when clicked through frame by frame, this noise does not appear.
    So, Apple users of ther world, who can explain a) why full quality works with Cinema Tools and b) what to do about the noise in the playback picture (but again not seen frame by frame), and whether it is part of the actual clip or not?
    Powerbook G4 Mac OS X (10.3.8)

  • Question about SAP installation with Oracle 10g

    I have successfully installed Oracle 10g on Windows server 2008 R2 64bit. I can connect to the database using mysql or through command prompt and run SQL commands.
    I started installing the SAP EHP1 for SAP NetWeaver 7.0 ("51038620_13      BS 7 SR1 Installation Master Windows Server on x64 64bit") The installation goes very smooth. I choose the SAPSID and the DBSID sid, but then it gives an error stating to install oracle and continue or stop the installation & install Oracle. I have tried on continuing the installation since Oracle is already installed but I get the same message again and again.
    Does it have to do with ORACLE 10g Sid. My Oracle SID is set to "ORCL" should it be the same as DBSID (during SAP Installation). Or the Oracle needs to be installed after SAP installation gets to the point for Oracle Installation. Please assist what needs to be changed. I can reinstall Oracle if needed. Any tips or comments would be nice.

    > Does it have to do with ORACLE 10g Sid. My Oracle SID is set to "ORCL" should it be the same as DBSID (during SAP Installation). Or the Oracle needs to be installed after SAP installation gets to the point for Oracle Installation.
    You can't "change". You installed the Oracle software using "setup.exe" - which is wrong. You also created an example database - which is also not correct.
    The Oracle software is installed, according to the installation guide, as
    - start "sapserver.cmd" on the Oracle DVD (the SAP media)
    - install the software according to the installation guide
    - install the latest patchset
    - install the latest PSU patch
    --> and then start sapinst.
    Markus

  • Oracle spatial data compression (using advance compression).

    What are the bast practice for oracle spatial to compress data using advance compression).
    ver. 11.2.0.3

    Details about Advanced Compression can be found in:
    Oracle E-Business Suite Release 12.1 with Oracle Database 11g Advanced Compression (Doc ID 1110648.1)
    Is Advanced Compression Supported In The E-business Suite ? (Doc ID 1368152.1)
    https://blogs.oracle.com/stevenChan/entry/using_advanced_compression_with_e-business_suite
    Thanks,
    Hussein

  • About pessimistick lock in Oracle

    Hello,
    I have some questions about pessimistick lock in Oracle:
    I know if i do a SELECT..FOR UPDATE statement, those records are locked, and just me can opperate on them.. but me - WHO? How i'm identified in Oracle? If i do a SELECT..FOR UPDATE select in an A procedure, then i call the update procedure B, how can i know that both calls belongs to the same user? If someone tries to update the same records, how the server knows if that person is allowed to make those updates or not? because everyone uses the same connection string.. isn't it?
    Probably it works if i make a SELECT..FOR UPDATE and i make the effective update in that procedure, but this means that the Business Logic is in that procedure..
    Please correct me if i'm wrong somewhere and clarify me.
    Thanks!

    By default the other session has to wait (no error message):
    From SQL Reference SELECT section: http://download.oracle.com/docs/cd/B28359_01/server.111/b28286/statements_10002.htm#SQLRF01702
    >
    NOWAIT | WAIT
    The NOWAIT and WAIT clauses let you tell the database how to proceed if the SELECT statement attempts to lock a row that is locked by another user.
    Specify NOWAIT to return control to you immediately if a lock exists.
    Specify WAIT to instruct the database to wait integer seconds for the row to become available and then return control to you.
    If you specify neither WAIT nor NOWAIT, then the database waits until the row is available and then returns the results of the SELECT statement.
    >
    Edited by: P. Forstmann on 7 janv. 2010 09:00
    Edited by: P. Forstmann on 7 janv. 2010 09:01

  • ####a question about jdbc driver

    A question about jdbc driver: for Oracle JDBC drivers which will not allow O/R
    tools to access columns over 4 kb. This affects some of the functionality of my
    application.A workaround is to use the my JDBC driver that acts as a wrapper to
    ensure that columns over 4 kb read and persist properly. my JDBC driver determines
    which actual driver to load based on the connection URL. If you pass in a URL
    that contains the string jdbc:oracle:thin, it will know to load oracle.jdbc.driver.OracleDriver.
    If the URL contains the string jdbc:db2, it will know to load COM.ibm.db2.jdbc.net.DB2Driver,
    and so on.
    now i have set CLASSPATH for my jdbc wapper,when I Creating and Configuring a
    JDBC Connection Pool,i have to input my driver name such com.lawrence.driver.....then
    i did,all
    is well,then i restart my WLS,when load config.xml,some errors such as "No registered
    driver accept URL:jdbc:oracle:thin:@localhost:1521:lawrence",if i set driver name
    is oracle.jdbc.driver.OracleDriver,all is good,no any errors.
    I find some questions like:
    1.for WLS8.1 ,it is only find some registed driver?
    2.why it cannot load my driver wapper when it is loading jdbc Connection Pool,but
    when the WLS start is finished, i Creating and Configuring a JDBC Connection Pool
    then it is right?

    my JDBC driver like this:
    public class myDriver implements Driver {
         public static String DB2_DRIVER = "COM.ibm.db2.jdbc.net.DB2Driver";
         public static String DB2_URL = "jdbc:db2:";
         public static String HYPERSONIC_DRIVER = "org.hsqldb.jdbcDriver";
         public static String HYPERSONIC_URL = "jdbc:hsqldb:";
         public static String MYSQL_DRIVER = "com.mysql.jdbc.Driver";
         public static String MYSQL_URL = "jdbc:mysql:";
         public static String ORACLE_DRIVER = "oracle.jdbc.driver.OracleDriver";
         public static String ORACLE_URL = "jdbc:oracle:";
         public static String POSTGRESQL_DRIVER = "org.postgresql.Driver";
         public static String POSTGRESQL_URL = "jdbc:postgresql:";
         public static String SQLSERVER_DRIVER =
              "com.microsoft.jdbc.sqlserver.SQLServerDriver";
         public static String SQLSERVER_URL = "jdbc:microsoft:";
         public boolean acceptsURL(String url) throws SQLException {
              return true;
         public synchronized Connection connect(String url, Properties props)
              throws SQLException {
              if (_driver == null) {
                   Class driverClass = null;
                   try {
                        if (url.startsWith(DB2_URL)) {
                             driverClass = Class.forName(DB2_DRIVER);
                        else if (url.startsWith(HYPERSONIC_URL)) {
                             driverClass = Class.forName(HYPERSONIC_DRIVER);
                        else if (url.startsWith(MYSQL_URL)) {
                             driverClass = Class.forName(MYSQL_DRIVER);
                        else if (url.startsWith(ORACLE_URL)) {
                             driverClass = Class.forName(ORACLE_DRIVER);
                        else if (url.startsWith(POSTGRESQL_URL)) {
                             driverClass = Class.forName(POSTGRESQL_DRIVER);
                        else if (url.startsWith(SQLSERVER_URL)) {
                             driverClass = Class.forName(SQLSERVER_DRIVER);
                        _driver = (Driver)driverClass.newInstance();
                   catch (Exception e) {
                        throw new SQLException(e.getMessage());
              Connection con = _driver.connect(url, props);
              return new myConnection(con);
         public int getMajorVersion() {
              return _driver.getMajorVersion();
         public int getMinorVersion() {
              return _driver.getMinorVersion();
         public DriverPropertyInfo[] getPropertyInfo(String url, Properties props)
              throws SQLException {
              return _driver.getPropertyInfo(url, props);
         public boolean jdbcCompliant() {
              return _driver.jdbcCompliant();
         private Driver _driver;

Maybe you are looking for

  • Macbook Core 2 Duo Keyboard - Any Improvement On Previous Generation?

    It's a few hundred kilometers to the nearest Apple Store, and local stores which sell Apple products don't have very many models on display or change them often. I have tried the keyboards on the previous models (pre-core-2) and found that the keyboa

  • BDC for Inspection plan

    Dear QM Gurus, I have recorded BDC for QP01. At the time of recording system was not asking for "Inspection Method". Also we have not created inspection method for any material. But after completion of recording, abaper has done his job to run that B

  • Why is my navbar not loading in edge animate 2014?

    Just upgraded Edge animate to the new Adobe Edge animate 2014 but now my navigation won't load in preview browser.

  • What i have to do when my iphone has turned off and i does not turn on

    my iphone has turned off by itself and and does not turn on ??? what i should to do ??

  • My media has been deleted

    My media card fell out of my phone and when i put it back in all my music, videos, songs, pictures etc had been deleted and a message came up on the screen saying 'A Media Card has been inserted that is not formatted. Do you wish to format the Media