Problem with Exporting a Table with BLOB data.

Hello All,
I get error messages while exporting a table from Oracle8i ( 8.1.7.4.1), containing BLOB data:
ORA-01555: snapshot too old: rollback segment number with name "" too small
ORA-22924: snapshot too old
Metalink says there is a bug in release 8.1.7.4.1 and suggests to create a new tablespace (with manual space management) ant move BLOB object to this TB.
Can i solve this problem without creating a new tablespace ?
Maybe somebody's experienced in such problem.
thank you in advance.

Maybe you could solve Your problem without creating new tablespace.
Consider changing the value of MAX_ROLLBACK_SEGMENTS to max value if currently is not at that value and creating new bigger rollback segments.

Similar Messages

  • How to export a table with BLOBs

    Hi,
    when I have to export a table with its own data, I generally export it by using the export on UNIX.
    I've never tried to export a table with a BLOB. Is it possible in the same manner or what else?
    Thanks!

    Hi mark;
    Please see below which could be helpful for your issue:
    Export and Import of Table with LOB Columns (like CLOB and BLOB) has Slow Performance [ID 281461.1]
    Master Note for Data Pump [ID 1264715.1]
    Regard
    Helios

  • Export - only tables with data

    Hai Everybody!
    I have a database in Oracle, which is having hundreds of tables. When i export the database, all the tables will be exported into the .dmp file.
    i want to export only the tables, which are having atleast one row to the .dmp file.
    Can anyone help me by providing the solution.
    Thanks
    JD

    JayaDev(JD) wrote:
    version 10gIn 11g, there is feature called "deferred segment creation". It means, if you create a new table and do not insert any data into it, Oracle will not create segment for that table, hence it will not be displayed in dba_segments.
    That is why, if you try export that table with exp utility, it will not be exported.

  • Export/import table with XMLTYPE data_type and fine_grained policy

    Hi friends!
    I'm trying to export a table with XMLTYPE and faine-grained policy.
    Source: HP-UX - Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    Target: Linux 2.6.18-238.el5 - Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    I do it thru exp/imp old utilities as mention in metalink ID 1318012.1.
    One of the things that surprised me after exp/imp is that the number of objects increased:
    Objects at Source:
    exp \"/ as sysdba\" owner=xml log=xml file=xml.dmp
    Type
    Number of Objects
    TYPE
    431
    TRIGGER
    6
    TABLE
    17
    PROCEDURE
    2
    LOB
    120
    INDEX
    17
    FUNCTION
    1
    On Target:
    imp \"/ as sysdba\" file=xml.dmp fromuser=xml touser=xml log=xml.log
    Type
    Number of Objects
    TYPE
    431
    TABLE
    32
    PROCEDURE
    2
    LOB
    429
    INDEX
    478
    FUNCTION
    1
    Why is this happening? Is it normal?
    Other problem that I found is triggers are not imported, why!?!?
    Thank you very much for your help!
    José

    Hi!
    The new squema was created empty. If I run the following query:
    select '10g', TABLE_NAME,COUNT(*)
    from  DBA_LOBS@DSN_HP
    where OWNER='XML'
    group by TABLE_NAME
    union
    select '11g', table_name,count(*)
    from  DBA_LOBS
    where OWNER='XML'
    group by TABLE_NAME
    order by 2,1
    As a result:
    Version
    Table_name
    Total
    10g
    ACTION_TABLE
    1
    11g
    ACTION_TABLE
    1
    10g
    Document1767_TAB
    14
    11g
    Document1767_TAB
    13
    10g
    Document1852_TAB
    14
    11g
    Document1852_TAB
    13
    10g
    Document1941_TAB
    16
    11g
    Document1941_TAB
    15
    10g
    Document2016_TAB
    14
    11g
    Document2016_TAB
    13
    10g
    Document2087_TAB
    13
    11g
    Document2087_TAB
    12
    10g
    IBT_XML_RECIBIDOS
    1
    11g
    IBT_XML_RECIBIDOS
    1
    10g
    LINEITEM_TABLE
    2
    11g
    LINEITEM_TABLE
    2
    10g
    PURCHASEORDER
    7
    11g
    PURCHASEORDER
    7
    10g
    PurchaseOrder1145_TAB
    9
    11g
    PurchaseOrder1145_TAB
    7
    10g
    RICARDO
    13
    10g
    RICARDO2
    1
    11g
    RICARDO2
    1
    10g
    RITNTFER
    1
    11g
    RITNTFER
    1
    10g
    RITNTFRE_08
    13
    11g
    SYS_NT3+LEU6vbfGLgQ18DLgrURw==
    69
    11g
    SYS_NT3+LEU6vffGLgQ18DLgrURw==
    76
    11g
    SYS_NT3+LEU6vjfGLgQ18DLgrURw==
    63
    11g
    SYS_NT3+LEU6vpfGLgQ18DLgrURw==
    1
    11g
    SYS_NT3+LEU6vqfGLgQ18DLgrURw==
    2
    11g
    SYS_NT3+LEU6vTfGLgQ18DLgrURw==
    65
    11g
    SYS_NT3+LEU6vXfGLgQ18DLgrURw==
    66
    10g
    TESTCLOB
    1
    11g
    TESTCLOB
    1
    There are many new tables created...I suppose because the differences between versions...
    Any ideas?
    Thanks a lot!
    José

  • How Can I Export A Table With Its Entries ( values )

    Hi guys,
    How can i export a table with the entries that the table is populated with ??!!
    Best Regards,
    Fateh

    To export a table and its data you can do it a number of ways.. Do you have access to SQL Developer? If not you can do it through the APEX SQL Workshop..
    Under SQL Workshop, Utilities, Data Unload, To text (or XML, whichever you are more comfortable with..).. Select your schema, click next button, select the table, click next, select the columns you want exported (select ALL ITEMS in the select List), click next, on optionally enclosed by I usually enter a " (Sometimes I have long strings in columns), click the include column names checkbox to get the column names at the top of the document, click unload data button..
    To get the ddl (code to build table) Sql Workshop, utilities, generate ddl, create script, select schema, select next, click table checkbox, select next, click the checkbox of table you want to get ddl for...
    Thank you,
    Tony Miller
    Webster, TX
    There are two kinds of pedestrians -- the quick and the dead.
    If this question is answered, please mark the thread as closed and assign points where earned..

  • ORA-00600 problem when create XMLType table with registerd schema

    Hi,
    I am using Oracle9i Enterprise Edition Release 9.2.0.4.0 on RedHat Linux 7.2
    I found a problem when I create table with registered schema with follow content:
         <xs:element name="body">
              <xs:complexType>
                   <xs:sequence>
                   </xs:sequence>
                   <xs:attribute name="id" type="xs:ID"/>
                   <xs:attribute name="class" type="xs:NMTOKENS"/>
                   <xs:attribute name="style" type="xs:string"/>
              </xs:complexType>
         </xs:element>
         <xs:element name="body.content">
              <xs:complexType>
                   <xs:choice minOccurs="0" maxOccurs="unbounded">
                        <xs:element ref="p"/>
                        <xs:element ref="hl2"/>
                        <xs:element ref="nitf-table"/>
                        <xs:element ref="ol"/>
                   </xs:choice>
                   <xs:attribute name="id" type="xs:ID"/>
              </xs:complexType>
         </xs:element>
    Does Oracle not support element reference to other element with dot?
    For instance, body -> body.content
    Thanks for your attention.

    Sorry, amendment on the schema
         <xs:element name="body">
              <xs:complexType>
                   <xs:sequence>
                        <xs:element ref="body.head" minOccurs="0"/>
                        <xs:element ref="body.content" minOccurs="0" maxOccurs="unbounded"/>
                        <xs:element ref="body.end" minOccurs="0"/>
                   </xs:sequence>
                   <xs:attribute name="id" type="xs:ID"/>
                   <xs:attribute name="class" type="xs:NMTOKENS"/>
                   <xs:attribute name="style" type="xs:string"/>
              </xs:complexType>
         </xs:element>

  • Data pump export single tables with specific criteria with the API

    Hello, I'm trying to export some table data from a schema using dbms_datapump API, but it gives me problems and it takes too much time to elaborate and I don't understand why! I want to export data only in a dmp file, that I will use later to import in an other schema. I'm using Oracle 10g R2 .
    I want to export data from TABLE1 and TABLE2, and ONLY the first 10 rows (this is for test now). I used the data_filter to write the subquery to filter the rows, and NAME_LIST to filter table names. I've set INCLUDE_METADATA to 0, to not export metadata. But it takes 10 minutes to run, and the output log says that there was an error (after 10 minutes??!):
    content of : table_dump.log
    Starting "MYSCHEMANAME"."SYS_EXPORT_TABLE_02": 
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    ORA-39166: Object IN ('TABLE1' was not found.
    ORA-39166: Object 'TABLE2') was not found.
    ORA-31655: no data or metadata objects selected for job
    Job "MYSCHEMANAME"."SYS_EXPORT_TABLE_02" completed with 3 error(s) at 15:58:47
    This is the code I use:
    DECLARE
      handle NUMBER;
      status VARCHAR2(20);
    BEGIN
      handle := DBMS_DATAPUMP.OPEN ('EXPORT', 'TABLE');
      dbms_datapump.add_file(handle => handle,filename => 'table_dump.log',directory => 'DATAPUMP_DIR',filetype => DBMS_DATAPUMP.KU$_FILE_TYPE_LOG_FILE);
      dbms_datapump.add_file(handle => handle,filename => 'table_dump.dmp',directory => 'DATAPUMP_DIR',filetype => DBMS_DATAPUMP.KU$_FILE_TYPE_DUMP_FILE);
      dbms_datapump.metadata_filter(handle, 'SCHEMA_EXPR', 'IN (''MYSCHEMANAME'')');
      dbms_datapump.metadata_filter (handle, 'NAME_LIST', 'IN (''TABLE1'',''TABLE2'')');
      dbms_datapump.data_filter(handle, 'SUBQUERY', 'WHERE rownum <= 10', 'TABLE1', 'MYSCHEMANAME');
      dbms_datapump.data_filter(handle, 'SUBQUERY', 'WHERE rownum <= 10', 'TABLE2', 'MYSCHEMANAME');
      dbms_datapump.set_parameter(HANDLE => handle,NAME => 'INCLUDE_METADATA', VALUE => 0) ;
      dbms_datapump.START_JOB(handle);
      dbms_datapump.WAIT_FOR_JOB(handle, status);
    END;
    /Edited by: user10396517 on 27-feb-2012 9.17 - added the code formatting

    Welcome to the forums. When pasting code use the {  code  } tags for better readability. See the FAQ for other details. And always feel free to include all the ddl/dml for your test cases so we don't have to do much more than run your code to reproduce it ourselves.
    I've had very little luck with NAME_LISTs. Though I know you can do similar with NAME_EXPR:
    SQL> create table table1 as select * from dba_tables where rownum <= 20;
    Table created.
    SQL> create table table2 as select * from dba_tables where rownum <= 20;
    Table created.
    -- note 20 rows each.
    SQL> DECLARE
    handle NUMBER;
    status VARCHAR2(20);
      2    3    4  BEGIN
      5  handle := DBMS_DATAPUMP.OPEN ('EXPORT', 'TABLE');
      6  dbms_datapump.add_file(handle => handle,filename => 'table_dump.log',directory => 'DATA_PUMP_DIR',filetype => DBMS_DATAPUMP.KU$_FILE_TYPE_LOG_FILE);
      7  dbms_datapump.add_file(handle => handle,filename => 'table_dump.dmp',directory => 'DATA_PUMP_DIR',filetype => DBMS_DATAPUMP.KU$_FILE_TYPE_DUMP_FILE);
      8  dbms_datapump.metadata_filter(handle, 'SCHEMA_EXPR', 'IN (''ANDY'')');
      9  dbms_datapump.metadata_filter (handle, name=> 'NAME_EXPR', value=> 'IN (''TABLE1'',''TABLE2'')');
    10  dbms_datapump.data_filter(handle, 'SUBQUERY', 'WHERE rownum <= 10', 'TABLE1', 'ANDY');
    11  dbms_datapump.data_filter(handle, 'SUBQUERY', 'WHERE rownum <= 10', 'TABLE2', 'ANDY');
    12  dbms_datapump.set_parameter(HANDLE => handle,NAME => 'INCLUDE_METADATA', VALUE => 0) ;
    dbms_datapump.START_JOB(handle);
    13   14  dbms_datapump.WAIT_FOR_JOB(handle, status);
    15  END;
    16  /
    PL/SQL procedure successfully completed.
    SQL> !cat /u01/app/oracle/admin/test1/dpdump/table_dump.log
    Starting "SYS"."SYS_EXPORT_TABLE_18":
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 128 KB
    . . exported "ANDY"."TABLE1"                             29.32 KB      10 rows
    . . exported "ANDY"."TABLE2"                             29.32 KB      10 rows
    Master table "SYS"."SYS_EXPORT_TABLE_18" successfully loaded/unloaded
    Dump file set for SYS.SYS_EXPORT_TABLE_18 is:
      /u01/app/oracle/admin/test1/dpdump/table_dump.dmp
    Job "SYS"."SYS_EXPORT_TABLE_18" successfully completed at 17:01:41 As for your time issues, nothing is preventing you from seeing what your datapump session is doing. If it is waiting on something, doing a full scan of something, etc.
    Good luck.

  • Exporting a table with Long datatype col. name

    Hello,
    I need help exporting a table, One of the column is long datatype.
    How can I do this without using Export Util?
    Is it possible.
    (If someone has a solution then PLEASE Email
    me)
    URGENT.
    Thanks.
    Pankaj Patel.

    Just wanted to find out, if you already solved this problem. I have a similar issue with long column. I am trying to sql dump a table with long column that will be imported into another database(probably using sql*load), but the spooled file puts the data from the long column in separate lines 80 char long and not on a single line. I have set the long to 64000 and linesize to 32000. Wrap is on too.
    null

  • [Solved]Problem in showing ADF tables with 50,000 Records and more in TP4

    Hi,
    When I tried to view readonly ADF table in TP4, it loaded perfectly, but when i tried to run the vertical scrollbar to bottom record, it showed Fetching Data .... for few minutes and finally it showed a message. java.lang.OutOfMemoryError: Java heap space.
    When I tried to view normal ADF table in TP4, it loaded perfectly, but when i selected one row a popup Error message came and reported all the columns in the table with java.lang.NullPointerException.
    Similar functionality works beautifully in JDeveloper 10g version. Seems to be a serious problem in handling cached rows.
    Regards,
    Santhosh
    I am also getting a warning in OC4J console as given below:
    WARNING: Application loader current-workspace-app.web.SRApplication-ViewController-webapp:0.0.0 may not use /D:/Appl/jdevstudio1111/j2ee/home/lib/oc4j-internal.jar (from <code-source> in META-INF/boot.xml in D:\Appl\jdevstudio1111\j2ee\home\oc4j.jar)
    Message was edited by:
    Santhosh1234
    Message was edited by:
    Santhosh1234

    Hi!
    You have two ways:
    1. build your own table CollectionModel that will use TopLink paged data source access (see Pedja's thread here Paging of large data sets - SOLUTION
    2. use new Paged EJB finders that started to appear with TP4 when you use wizard to generate SessionBeans (named something like "query<Entity>FindAllByRange") . I posted already question about this new paged finders here How to use new "query<Entity>FindAllByRange" EJB finders? but no answers from devs yet :)
    Personally, I opt for no. 2 (paged finders) as it looks more platform-supported that writing custom CollectionModels for each finder... But until devs enlighten us with details how the paged finders are to be used (the af:table is not [yet?] supportive for drag-and-drop of paged finders), we are stuck with 1. option.
    Regards,
    Pavle

  • How to export a table with a mixed case name using EXP untility

    i'm trying to export a table which has a name with mixed upper and lower case. The command i'm trying is
    exp USER/password TABLES=("MyTableName") FILE=ExportCNCS1.dat
    and does not work.
    Error:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Produc
    tion
    With the Partitioning, OLAP and Data Mining options
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    About to export specified tables via Conventional Path ...
    EXP-00011: USER.MYTABLENAME does not exist
    Export terminated successfully with warnings.
    I presume its saying the table doesnt exist because its looking for MYTABLENAME rather than MyTableName and doing this because its ignoring the double quotes. Any ideas?

    Or, let the export prompt you for the table name:
    C:\Temp>exp
    Export: Release 9.2.0.7.0 - Production on Mon Aug 14 11:33:28 2006
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Username: scott/tiger
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.7.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.7.0 - Production
    Enter array fetch buffer size: 4096 >
    Export file: EXPDAT.DMP >
    (2)U(sers), or (3)T(ables): (2)U > T
    Export table data (yes/no): yes >
    Compress extents (yes/no): yes > no
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    About to export specified tables via Conventional Path ...
    Table(T) or Partition(T:P) to be exported: (RETURN to quit) > MyTableName
    EXP-00011: SCOTT.MYTABLENAME does not exist
    Table(T) or Partition(T:P) to be exported: (RETURN to quit) > "MyTableName"
    . . exporting table                    MyTableName         14 rows exported
    Table(T) or Partition(T:P) to be exported: (RETURN to quit) >
    Export terminated successfully with warnings.
    C:\Temp>

  • JSF: Problems adding rows to table with custom method

    Since this being my first post, I find it only appropriate to thank the development team for such a huge addition to the application development world. You guys rock. You have cured my Java identity crisis. That being said...
    I have been stumped for days. I'm making a simple web cart. I am trying to get two string values (name and price) passed through a custom exposed application module method ran on a backing bean. What I want the exposed VO module to do is, upon receiving the two strings, create a new row on the empty data controlled table I have built to store them. Basically I just need to add rows to a table with my custom managed bean. Most recent version.
    I have written a test method to make sure my backing bean and exposed app module impl were working. This test works:
    //WORKS
    //CartInfo.java   
        public String cb1_action() {
            AppModuleImpl am = new AppModuleImpl();
            am.testMethod();
            return null;       
    //AppModuleImp.java
        public void testMethod() {
            System.out.println("WORKS");      
        }I found some code on how to add rows and unfortunately it isn't working out for some reason. Here is what is producing the error "Caused By: java.lang.NullPointerException" :
    //DOESN'T WORK IN CUSTOM METHOD BUT WORKS WHEN DRAGGED ON TO THE PAGE AS A COMMAND BUTTON
    //CartInfo.java
        public String cb1_action() {
            AppModuleImpl am = new AppModuleImpl();
            am.testMethod("test", "test");
            return null;
    //AppModuleImp.java
           public void testMethod(String pName,String pPrice) {
            ViewObjectImpl vo = this.getCartVO1();
            Row r = vo.createRow();
            r.setAttribute("NAME", pName);
            r.setAttribute("PRICE", pPrice);
            vo.insertRow(r);
    //The VO attributes for the DC are:
    // Name : Type : Alias Name : Entity Usage : Info
    NAME : String : NAME : (blank) : Transient
    PRICE : String : PRICE : (blank) : TransientSo what I have so far is a table linked to my database. A user can select and click on a product on the row's individual Add button and using a modified selection listener, it returns the values to the backing_bean.
    Then what I want it to do is have those variables passed to the other table. I just can't put it together.
    Please if there is anyone that could send me in the right direction I would greatly appreciate it!
    Thank you,
    Jack

    if i understand you, you want to call application module method when user clicks on add button on the UI and your table is from VO
    so what i can come up for now is:
    1) create variable binding from the iterator for both name and price.
    2) on your UI action button :
    <af:commandButton actionListener="yourbean.addToCart">
    <f:attribute name="pName" value ="binding.<created name bind variable>"
    <f:attribute name="pPrice" value = "binding.<created price bind variable>"
    </af:commandButton>
    3) custom Managed Bean:
    //yourBean.java
    public void addToCart(ActionEvent e){
    // get binding here
    // use executeWitParams method to send parameter to the function "testMethod" and execute
    //AppModuleImp.java
    public void testMethod(String pName,String pPrice) {
    CardVOImpl vo = this.getCartVO1();
    CardVORowImpl r = (CardVORowImpl)vo.createRow();
    r.setAttribute("NAME", pName);
    r.setAttribute("PRICE", pPrice);
    vo.insertRow(r);
    Hope this helps...
    Let me know if not..
    Thanks
    Edited by: MavenDev on Oct 30, 2011 8:08 PM

  • Problem with version enabling tables with ric.

    Hello,
    i have the a problem when i want to version enable tables having ref. int. constraits having delete rule cascade or set null.
    Is it possible that i can't version enable the tables because of these constraints? How i could solve this problem if i want to keep the delete rule?
    thanks,
    Orsi Gyulai

    Hi,
    We are internally creating a <table_name>_g procedure that transfers privileges to the necessary users.  If you have a table with that name, it would explain the error.
    When using ignore_last_error, it will skip the statement that is selected from the all_wm_vt_errors view. Sometimes, the statement can be safely skipped, while in other cases it cannot be. This procedure will always eventually complete when it is repeatedly called with ignore_last_error set to true. However in doing so, some required objects or privileges may not exist or be in an invalid state.
    In your case, you most likely had to skip the 3 or so statements that dealt with the <table_name>_g procedure.  Typically, these statements should not be skipped, but you may or may not see a problem with it due to a number of factors.
    The best course of action may be to drop the trigger in a beginDDL/commitDDL session, and then recreate it in a separate session. Of course, only do this after renaming the <table_name>_g table that you have.  Unfortunately, there is currently no way of getting around this naming convention.
    Thanks,
    Ben

  • Quality problem when export to DV with QT

    Hello,
    With Imovie 11, Mountain Lion, Imac 24"
    When I export my movie with HD 1080 quality no problem,
    When I export my movie with QT to DV, I have huge pixelisation on vertical line on the movie. When people move for example, their legs and arms got zebra movement.
    That didn't happen before with 4/3 screen, looks like It's coming with HD 16/9 record.
    Any Idea ?
    Bruno

    killvanrea wrote:p.P.:I have many customization on it like - intel HD graphic driver, sound manager, compiz, eclipse  and so on... so I'll be happy if I can upgrade and dont losing all of them
    Running `pacman -R` without the -n switch will leave all of your configuration files alone and just remove the package files. In this case it might be helpful to remove anything you can without removing packages from the "base" group, to minimize conflicts. After successfully upgrading the remaining packages, run `pacman -S <packages>` to (re)install the newest version of the packages you removed. At least it has worked for me to avoid package conflicts in the past; your mileage may vary. See removing packages and pacman tips.
    If all else fails, in theory you could run something like `pacman -r /mnt -R $(pacman -r /mnt -Q | cut -d ' ' -f 1)` from a live CD to remove all packages, then use pacstrap to reinstall them. Use this as a last resort and make a backup first!

  • How does Sharepoint 2013 Server farm using BLOB data, Hyper-V with virtual machines storing the BLOB data?

    Hello Community
        A Sharepoint 2013 Server farm created on a physical server
    stores it's documents in SQL Server according to how SQL Server is setup.
        If a Sharepoint 2013 Server farm is using Hyper-V and Virtual machines
    how is BLOB data stored ?  
        Thank you
        Shabeaut

    Are you looking to use Remote Blob Storage? or RBS of SQL with SP?
    It is going to be same way as it is configured for physical server.
    Thanks, Ashish | Please mark a post helpful/answer if it is helpful or answer your query.

  • Problem in using CREATE TABLE with Execute Immediate

    I'm trying to create a table using Native Dynamic SQL. the code of the pl/sql block is
    BEGIN
    EXECUTE IMMEDIATE 'create table demo (ddate date)';
    END;
    The problem is that the above block is executed successfully as an anonymous PL/SQL block. The same block when written in a procedure it gives an error
    "ORA-01031 Insufficient privelages"
    at the time of execution. The procedure is complied successfully.
    null

    Your user needs direct system privs to create tables. You are receiving your privs properly by the role RESOURCE. Connect as system and grant CREATE TABLE directly to your user - that should do it.
    Regards
    Peter Larsen

Maybe you are looking for

  • Windows 7  home premium 64bit installation problem

    HI, all, I installed oracle 11g r2 enterprise version database on windows 7 home premium laptop, everything looks fine after installation, but whenever I restart the computer or relogin to the computer, the screen is only back. Could anyone tell me w

  • Desktop Image wont display in Windows 8.1 Enterprise

    Hi So everything worked fine on my laptop yesterday ( Toshiba)My laptop is running Windows 8.1 Enterprise.  It is activated. When I started my laptop, only the initial startup image showed.  When I go into Personalize, it shows only solid colours, no

  • Video Playback Android (Kindle Fire)

    I have been trying for several days to get video playback working on my Kindle Fire.  This is my first attempt at playing video on Android at all. Upon buildling my .apk, my video file resides at assets/cinematics/intro.mp4.  I would like to play thi

  • F-03 gl taging

    hi, I am using t-code f-03 for gl  clearing but i have problem that i have number of  entries  with different profit center  i want to tag that entry by profitcenter wies how can i can do that? or tell if any other method to tag entry by profitcenter

  • Tty1-6 freeze in the middle of the boot process, tty7 and X works fine

    I'm running Arch+KDE (and some alternative like xmonad as wm). I have the following strange problem: Booting works fine, KDM starts, I can login, everythings works. However, although I can change to tty1-6 via ctrl+alt+f1...6 they are frozen in the m