Log in to database

After long long installation, When I try to log in i can't remember my user name nor my password
Can somebody help? where is this information can be obtain?

Hello,
What's your os? What's your oracle version?
On windows
C:> sqlplus "/as sysdba"
SQL>  select * from dba_users; -- This will display list of users
Also you can try
C:> sqlplus scott/tiger
C:> sqlplus system/manager
On Unix/linux
as an oracle user
$> sqlplus '/as sysdba'
SQL>select * from dba_users;
And if you don't remember the password you can change the password of non-admin users other than SYS/SYSTEM
SQL> ALTER USER YOUUSERNAME IDENTIFIED BY "mynewpassword";Regards
Edited by: OrionNet on Jan 13, 2009 10:31 PM
Edited by: OrionNet on Jan 13, 2009 10:33 PM

Similar Messages

  • SQL Server wants to open properties of wrong database and cannot restore logs of different databases at the same time

    We have implemented our own way of logshipping by backing up the transaction logs of all our non-system databases every hour. Then we transfer the backups (before SQL Server 2008 R2 with additional zip and unzip steps) to the second environment where we
    restore them only during the night on databases in standby-state (Like this, our developers can access the databases without disturbing the productive system and we can restore the t-logs during night, so not shutting out the developers during the day).
    Now since we replaced our hardware in the productive environment, we replaced the hardware in the second environment with the servers from previous productive environment. Same SQL Server 2008 R2, same Service Pack (SP2), same installation procedure. But
    since then the SQL Server in the second environment behaves different:
    - When the logs of some database "A" are restored, I cannot access the properties of any database, i.e. database "B", I only get the error message "Cannot show requested dialog. (SqlMgmt) Database 'A' cannot be opened. It is in the
    middle of a restore. (Microsoft SQL Server, Error: 927)"
    - Our logshipping runs multiple instance at the same time, so restoring multiple databases' logs. But since the change, two of the instances fail restoring of datatbases, i.e. "B" and "C" with the error message "Database 'A' cannot
    be opened. It is in the middle of a restore.". The jobs guarantee to restore only logs of the database they need to restore, they have strictly separated sets of target databases and only try to restore the databases of their sets.
    Before the new installations, we did not have these strange behaviours. But as said, it is the same SQL Server version, the same service pack, and we followed the same installation document but maybe with the small changes in settings. The user the jobs
    are running under is a domain admin, the login they use has dbcreator, processadmin, db_owner and db_backupoperator roles. My login on the SQL Server has sysadmin rights and as described similar problems.
    Is this problem known or does somebody have an idea where to go looking for?
    Thanks, Chris

    Hi Christar,
    According to your description, after you configuring Log shipping between two database, you remove the hardware from production environment to dev environment, in this hardware, there are saved the transaction logs of your database.
     When restoring  your database, there will occurs error 927, right?
    SQL Server Log shipping allows you to automatically send transaction log backups from a
    primary database on a primary server instance to one or more
    secondary databases on separate secondary server instances. It consists of three operations:
    1.Back up the transaction log at the primary server instance.
    2.Copy the transaction log file to the secondary server instance.
    3.Restore the log backup on the secondary server instance.
    Since you want to move your hardware avoiding copy job, we need to verify if .mdf and .ldf file are not saved in your hardware, or else,
     if you move your hardware, you could not open your database in your Primary Server. I recommend you can refer to the following steps.
    Full back up your primary database and restore it into the secondary database. It would have been set up as WITH STANDBY ,the databases would read StandBy/ Read Only.
    Perform transaction log backups via a SQL Server Agent job running on the primary server instance, save them in your hardware.
    In night, you move hardware to your dev environment, you can manually restore these transaction log backups or set up a SQL Server Agent job to restoring them from your hardware.
    There is detail about how to configure log shipping in SQL Server 2008R2 step by step, you can review it.
    http://mostafaelmasry.wordpress.com/2013/01/02/log-shipping-in-sql-server-2008r2-step-by-step/
    Regards,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

  • How to use the mirrored and log shipped secondary database for update or insert operations

    Hi,
    I am doing a DR Test where I need to test the mirrored and log shipped secondary database but without stopping the mirroring or log shipping procedures. Is there a way to get the data out of mirrored and log shipped database to another database for update
    or insert operations?
    Database snapshot can be used only for mirrored database but updates cannot be done. Also the secondary database of log shipping cannot used for database snapshot. Any ideas of how this can be implemented?
    Thanks,
    Preetha

    Hmm in this case I think you need Merge Replication otherwise it breaks down the purpose of DR...again in that case.. 
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • How to delete Transaction Logs in SQL database

    Hi,
    Can any one explain me the process how to delete the transcation logs in SQL database.
    Thanks
    Sunil

    Sunil,
    Yes you can take online backup in MS SQL server.
    The transaction log files contain information about all changes made to the database. The log files are necessary components of the database and may never be deleted. Why you want to delete it?
    I am taking any backup, do i need to turn off the SAP server that is running at the moment or can i take it online
    There are three main types of SQL Server Backup: Full Database Backup, Differential Database Backup and Transaction Log Backup. All these backups can be made when the database is online and do not require you to stop the SAP system.
    Check below link for details
    http://help.sap.com/erp2005_ehp_04/helpdata/EN/89/68807c8c984855a08b60f14b742ced/frameset.htm
    Thanks
    Sushil

  • When is anything written to standby redo logs on standby database?

    I am on Oracle 10.2.0.4 on HP UNIX. I have read Oracle 10.2 concepts guide on technet.oracle.com, have read may article on metalink and internet, yet I am unable to verify when anything is written to standby redo logs on stand by database.
    I have a simple database reconfiguration: a primary database and one standby database.
    I created primary database and set up log_archive_dest_2 to use LGWR SYNC AFFIRM
    I have created standby redo logs on primary.
    alter database add standby logfile GROUP
    I create standby control file on primary.
    I copied all the primary information to create standby database. I have put standby database in managed recovery.
    I did archive log switches, I created a table and inserted information in table.
    I never saw standby redo logs updated on standby database by looking at timestamp of standby redo log files.
    I then setup database in maximum availability mode by running following on primary:
    Alter database set standby database to maximize availability
    When I do insert into my tables, I do see standby redo log files on primary database being updated, but I have never seen standby redo logs on standby database updated. Why?
    I am still at loss when actually standby redo logs are updated on standby database.
    When I read Oracle 9i database documentation on data guard, it says that you do not need standby redo logs on primary instead you need them on standby. Only reason, you need them on primary is from primary changes role to standby database, so standby redo logs on standby database should be updated instead of standby redo logs on primary.

    What is the PROTECTION_MODE ,PROTECTION_LEVEL values of your database.
    As per metalink:--
    Create standby redo log files, if necessary:
    Standby redo logs are necessary for the higher protection levels such as
    Guaranteed, Instant, and Rapid. In these protection modes LGWR from the
    Primary host writes transactions directly to the standby redo logs.
    This enables no data loss solutions and reduces the amount of data loss
    in the event of failure. Standby redo logs are not necessary if you are using
    the delayed protection mode.
    If you configure standby redo on the standby then you should also configure
    standby redo logs on the primary database. Even though the standby redo logs
    are not used when the database is running in the primary role, configuring
    the standby redo logs on the primary database is recommended in preparation
    for an eventual switchover operation.
    Standby redo logs must be archived before the data can be applied to the
    standby database. The standby archival operation occurs automatically, even if
    the standby database is not in ARCHIVELOG mode. However, the archiver process
    must be started on the standby database. Note that the use of the archiver
    process (ARCn) is a requirement for selection of a standby redo log.
    METALINK ID:- Doc ID: Note:219344.1
    Edited by: Anand... on Sep 15, 2008 2:15 AM

  • When creating a tablespace why should we enable LOGGING when a database is already on ARCHIVE LOG mode

    Question :
    When creating a tablespace why should we enable LOGGING when a database is already on ARCHIVE LOG mode ?
    Example:
    Create Tablespace
    CREATE SMALLFILE TABLESPACE "TEST_DATA"
    LOGGING
    DATAFILE '+DG_TEST_DATA_01(DATAFILE)' SIZE 10G
    AUTOEXTEND ON NEXT  500K MAXSIZE 31000M
    EXTENT MANAGEMENT LOCAL
    SEGMENT SPACE MANAGEMENT AUTO;
    LOGGING: Generate redo logs for creation of tables, indexes and  partitions, and for subsequent inserts. Recoverable
    Are they not logged and not recoverable if we do not enable LOGGING? What is that ARCHIVELOG mode does?

    What is that ARCHIVELOG Mode Does?
    Whenever your database is in archive log mode , Oracle will backup the redo log files in the form of Archives so that we can recover the database to the consistent state in case of any failure.
    Archive logging is essential for production databases where the loss of a transaction might be fatal.
    Why Logging?
    Logging is safest method to ensure that all the changes made at the tablespace will be captured and available for recovery in the redo logs.
    It is just the level at which we defines:
    Force Logging at DB level
    Logging at Tablespace Level
    Logging at schema Level
    Before the existence of FORCE LOGGING, Oracle provided logging and nologging options. These two options have higher precedence at the schema object level than the tablespace level; therefore, it was possible to override the logging settings at the tablespace level with nologging setting at schema object level.

  • Report Viewer unable to log into the database

    I'm fairly novice to ZFD Inventory and have a reporting issue when attempting to run from ConsoleOne (1.3.6f). Receive a Report Viewer error that states "unable to log into the database. Ensure the ODBC driver...".
    TID 10094443 describes my symptom exactly but apparently doesn't apply. It applies to ZFD6.5 and we're running 7. Does the Oracle fix it refers to even apply to ZFD7 Inventory? If not, what does address my issue?
    Is it SQL now?
    Thanks for any help.

    Originally Posted by youngj
    I'm fairly novice to ZFD Inventory and have a reporting issue when attempting to run from ConsoleOne (1.3.6f). Receive a Report Viewer error that states "unable to log into the database. Ensure the ODBC driver...".
    TID 10094443 describes my symptom exactly but apparently doesn't apply. It applies to ZFD6.5 and we're running 7. Does the Oracle fix it refers to even apply to ZFD7 Inventory? If not, what does address my issue?
    Is it SQL now?
    Thanks for any help.
    If you are using the Sybase database with ZDM7, take a look at the file \database drivers\odbcreadme.txt file on the Novell ZENworks 7 Companion 2 CD (downloadable from novell.com if you don't already have it). Page 347 of the ZDM 7 Install Doc also mentions this procedure:
    http://www.novell.com/documentation/...dm7install.pdf

  • Configurin​g Database for logging to citadel Database using shared variable engine

    Hello All,
    I have two systems with me here, one with LabVIEW 8.5 and one with LabVIEW 8.6, I'm using Shared Variables in my code and I am Logging to citadel Database. In a PC with LabVIEW 8.5, I am able to Log Data to citadel Database, but with the same code I am not able to Log Data from the PC with LabVIEW 8.6. Both the PC's have Database Installed and the connection with the Database Exists when I test connection with database through control panel. I would Like to know any configurations (in LabVIEW or in code or in Database or in PC) that have to be done for Logging to Database because the PC with LabVIEW 8.6 was added recently and the code was upgraded to 8.6.

    It was due to a dll in LabVIEW, a dll named nitaglv.dll
    Now I have a issue with Data Logging from the EXE, once the shared variables are deployed programatically from the EXE the Data Logging stops, 
    can I get any input on this issue..???

  • Total no. of users logged into the database

    Can we find out total number of users logged into the database? I do not have DBA privileges.
    If it is not possible by a straight SQL then do I have any other options?
    Thanks.

    You can find the list of all users using the query "Select * from V$Sesssion_connect_info". This requires access to system tables. But this query gives the OS user information and not the database user information.
    <BLOCKQUOTE><font size="1" face="Verdana, Arial, Helvetica">quote:</font><HR>Originally posted by Krishnan:
    Can we find out total number of users logged into the database? I do not have DBA privileges.
    If it is not possible by a straight SQL then do I have any other options?
    Thanks.<HR></BLOCKQUOTE>
    null

  • SQL0964C  The transaction log for the database is full

    Hi,
    i am planning to do QAS refresh from PRD system using Client export\import method. i have done export in PRD and the same has been moved to QAS then started imported.
    DB Size:160gb
    DB:DB2 9.7
    os: windows 2008.
    I facing SQL0964C  The transaction log for the database is full issue while client import and regarding this i have raised incident to SAP then they replied to increase some parameter like(LOGPRIMARY,LOGSECOND,LOGFILSIZ) temporarily and revert them back after the import. Based on that i have increased from as below mentioned calculation.
    the filesystem size of /db2/<SID>/log_dir should be greater than LOGFILSIZ*4*(Sum of LOGPRIMARY+LOGSECONDARY) KB
    From:
    Log file size (4KB)                         (LOGFILSIZ) = 60000
    Number of primary log files                (LOGPRIMARY) = 50
    Number of secondary log files               (LOGSECOND) = 100
    Total drive space required: 33GB
    To:
    Log file size (4KB)                         (LOGFILSIZ) = 70000
    Number of primary log files                (LOGPRIMARY) = 60
    Number of secondary log files               (LOGSECOND) = 120
    Total drive space required: 47GB
    But still facing the same issue. Please help me to resolve the ASAP.
    Last error TP log details:
    3 ETW674Xstart import of "R3TRTABUFAGLFLEX08" ...
    4 ETW000   1 entry from FAGLFLEX08 (210) deleted.
    4 ETW000   1 entry for FAGLFLEX08 inserted (210*).
    4 ETW675 end import of "R3TRTABUFAGLFLEX08".
    3 ETW674Xstart import of "R3TRTABUFAGLFLEXA" ...
    4 ETW000  [     dev trc,00000]  Fri Jun 27 02:20:21 2014                                             -774509399  65811.628079
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] CON = 0 (BEGIN)                    85  65811.628164
    4 ETW000  [     dev trc,00000]  &+     DbSlModifyDB6( SQLExecute ): [IBM][CLI Driver][DB2/NT64] SQL0964C  The transaction log for the database is full. 
    4 ETW000                                                                                                  83  65811.628247
    4 ETW000  [     dev trc,00000]  &+     SQLSTATE=57011 row=1                                                                                             
    4 ETW000                                                                                                  51  65811.628298
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  67  65811.628365
    4 ETW000  [     dev trc,00000]  &+     DELETE FROM "FAGLFLEXA" WHERE "RCLNT" = ?                                                                        
    4 ETW000                                                                                                  62  65811.628427
    4 ETW000  [     dev trc,00000]  &+       cursor type=NO_HOLD, isolation=UR, cc_release=YES, optlevel=5, degree=1, op_type=8, reopt=0                    
    4 ETW000                                                                                                  58  65811.628485
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  53  65811.628538
    4 ETW000  [     dev trc,00000]  &+     Input SQLDA:                                                                                                     
    4 ETW000                                                                                                  52  65811.628590
    4 ETW000  [     dev trc,00000]  &+                        1 CT=WCHAR           T=VARCHAR         L=6     P=9     S=0   
    4 ETW000                                                                                                  49  65811.628639
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  50  65811.628689
    4 ETW000  [     dev trc,00000]  &+     Input data:                                                                                                      
    4 ETW000                                                                                                  49  65811.628738
    4 ETW000  [     dev trc,00000]  &+     row 1:             1 WCHAR           I=6       "210"               34  65811.628772
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  51  65811.628823
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  50  65811.628873
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] (END)                              27  65811.628900
    4 ETW000  [    dbtran  ,00000]  ***LOG BY4=>sql error -964   performing DEL on table FAGLFLEXA                    
    4 ETW000                                                                                                3428  65811.632328
    4 ETW000  [    dbtran  ,00000]  ***LOG BY0=>SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1
    4 ETW000                                                                                                  46  65811.632374
    4 ETW000  [     dev trc,00000]  dbtran ERROR LOG (hdl_dbsl_error): DbSl 'DEL'                             59  65811.632433
    4 ETW000                         RSLT: {dbsl=99, tran=1}
    4 ETW000                         FHDR: {tab='FAGLFLEXA', fcode=194, mode=2, bpb=0, dbcnt=0, crsr=0,
    4 ETW000                                hold=0, keep=0, xfer=0, pkg=0, upto=0, init:b=0,
    4 ETW000                                init:p=0000000000000000, init:#=0, wa:p=0X00000000020290C0, wa:#=10000}
    4 ETW000  [     dev trc,00000]  dbtran ERROR LOG (hdl_dbsl_error): DbSl 'DEL'                            126  65811.632559
    4 ETW000                         STMT: {stmt:#=0, bndfld:#=1, prop=0, distinct=0,
    4 ETW000                                fld:#=0, alias:p=0000000000000000, fupd:#=0, tab:#=1, where:#=2,
    4 ETW000                                groupby:#=0, having:#=0, order:#=0, primary=0, hint:#=0}
    4 ETW000                         CRSR: {tab='', id=0, hold=0, prop=0, max.in@0=1, fae:blk=0,
    4 ETW000                                con:id=0, con:vndr=7, val=2,
    4 ETW000                                key:#=3, xfer=0, xin:#=0, row:#=0, upto=0, wa:p=0X00000001421A3000}
    2EETW125 SQL error "-964" during "-964" access: "SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1"
    4 ETW690 COMMIT "14208" "-1"
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] CON = 0 (BEGIN)                 16208  65811.648767
    4 ETW000  [     dev trc,00000]  &+     DbSlModifyDB6( SQLExecute ): [IBM][CLI Driver][DB2/NT64] SQL0964C  The transaction log for the database is full. 
    4 ETW000                                                                                                  75  65811.648842
    4 ETW000  [     dev trc,00000]  &+     SQLSTATE=57011 row=1                                                                                             
    4 ETW000                                                                                                  52  65811.648894
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  51  65811.648945
    4 ETW000  [     dev trc,00000]  &+     INSERT INTO DDLOG (SYSTEMID, TIMESTAMP, NBLENGTH, NOTEBOOK) VALUES (  ? , CHAR(   CURRENT TIMESTAMP - CURRENT TIME
    4 ETW000                                                                                                  50  65811.648995
    4 ETW000  [     dev trc,00000]  &+     ZONE   ), ?, ? )                                                                                                 
    4 ETW000                                                                                                  49  65811.649044
    4 ETW000  [     dev trc,00000]  &+       cursor type=NO_HOLD, isolation=UR, cc_release=YES, optlevel=5, degree=1, op_type=15, reopt=0                   
    4 ETW000                                                                                                  55  65811.649099
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  49  65811.649148
    4 ETW000  [     dev trc,00000]  &+     Input SQLDA:                                                                                                     
    4 ETW000                                                                                                  50  65811.649198
    4 ETW000  [     dev trc,00000]  &+                        1 CT=WCHAR           T=VARCHAR         L=44    P=66    S=0   
    4 ETW000                                                                                                  47  65811.649245
    4 ETW000  [     dev trc,00000]  &+                        2 CT=SHORT           T=SMALLINT        L=2     P=2     S=0   
    4 ETW000                                                                                                  48  65811.649293
    4 ETW000  [     dev trc,00000]  &+                        3 CT=BINARY          T=VARBINARY       L=32000 P=32000 S=0   
    4 ETW000                                                                                                  47  65811.649340
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  50  65811.649390
    4 ETW000  [     dev trc,00000]  &+     Input data:                                                                                                      
    4 ETW000                                                                                                  49  65811.649439
    4 ETW000  [     dev trc,00000]  &+     row 1:             1 WCHAR           I=14      "R3trans"           32  65811.649471
    4 ETW000  [     dev trc,00000]  &+                        2 SHORT           I=2        12744              32  65811.649503
    4 ETW000  [     dev trc,00000]  &+                        3 BINARY          I=12744   00600306003200300030003900300033003300310031003300320036003400390000...
    4 ETW000                                                                                                  64  65811.649567
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  52  65811.649619
    4 ETW000  [     dev trc,00000]  &+                                                                                                                      
    4 ETW000                                                                                                  51  65811.649670
    4 ETW000  [     dev trc,00000]  *** ERROR in DB6Execute[dbdb6.c, 4980] (END)                              28  65811.649698
    4 ETW000  [    dbsyntsp,00000]  ***LOG BY4=>sql error -964   performing SEL on table DDLOG                36  65811.649734
    4 ETW000  [    dbsyntsp,00000]  ***LOG BY0=>SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1
    4 ETW000                                                                                                  46  65811.649780
    4 ETW000  [    dbsync  ,00000]  ***LOG BZY=>unexpected return code 2          calling ins_ddlog           37  65811.649817
    4 ETW000  [     dev trc,00000]  db_syflush (TRUE) failed                                                  26  65811.649843
    4 ETW000  [     dev trc,00000]  db_con_commit received error 1024 in before-commit action, returning 8
    4 ETW000                                                                                                  57  65811.649900
    4 ETW000  [    dbeh.c  ,00000]  *** ERROR => missing return code handler                               1974  65811.651874
    4 ETW000                         caller does not handle code 1024 from dblink#5[321]
    4 ETW000                         ==> calling sap_dext to abort transaction
    2EETW000 sap_dext called with msgnr "900":
    2EETW125 SQL error "-964" during "-964" access: "SQL0964C  The transaction log for the database is full.  SQLSTATE=57011 row=1"
    1 ETP154 MAIN IMPORT
    1 ETP110 end date and time   : "20140627022021"
    1 ETP111 exit code           : "12"
    1 ETP199 ######################################
    Regards,
    Rajesh

    Hi Babu,
    I believe you should have taken a restart of your system if log primary are changed.  If so, then increase log primary to 120 and secondary to 80 provide size and space are enough.
    Note 1293475 - DB6: Transaction Log Full
    Note 1308895 - DB6: File System for Transaction Log is Full
    Note: 495297 - DB6: Monitoring transaction log    
    Regards,
    Divyanshu

  • Log suspended for database sybsystemdb

    This morning, I got error message:
    1 task(s) are sleeping waiting for space to become available in the log segment for database sybsystemdb.
    then I extend the space for sybsystempdb(data and log).  the process for logsuspended automatically gone.  Then, I got more error message:
    Error: 21, Severity: 21, State: 1
    00:00000:00037:2014/07/08 09:43:24.80 server  WARNING - Fatal Error 806 occurred at Jul  8 2014  9:43AM.  Please note the error and time, and contact a user with System Administrator (SA) authorization.
    00:00000:00037:2014/07/08 09:43:24.81 server  Error: 806, Severity: 21, State: 1
    00:00000:00037:2014/07/08 09:43:24.81 server  Could not find virtual page for logical page 1919221760  in database 'mydb'.
    00:00000:00037:2014/07/08 09:43:24.81 server  Error while undoing log row in database 'mydb'.  Rid pageid = 0x0; row num = 0x0.
    00:00000:00037:2014/07/08 09:43:24.81 server  Error: 6103, Severity: 17, State: 1
    00:00000:00037:2014/07/08 09:43:24.81 server  Unable to do cleanup for the killed process; received Msg 3300.
    00:00000:00037:2014/07/08 09:43:24.83 server  WARNING: spid 37 with suid 37 and curdbid 5 has an active transaction in dbid 5 with xactid (0, 0).
    then check doc for error 806, 6103. Looks like something possibly wrong. System looks like fine at this time. Want to know how to do more check for this case to make sure ase(12.5) it's okay.

    For the issue of sybsystemdb/log filling up ... enable the auto-truncation of the log (sp_dboption sybsystemdb,'trunc','true').
    NOTE: sybsystemdb is typically used to manage distributed transactions and can include keeping track of some proxy calls; for example the sybsystemdb log can fill up due to a high volume of proxy calls; unless you're working with distributed transactions the recovery of the sybsystemdb database usually isn't a concern, ie, just enable 'trunc log on chkpt' to keep the log trimmed
    For the 806/6103 errors ... without knowing more details about what processing was hung, which processes may have been killed, etc ... I'd bounce the dataserver and run 'dbcc checkalloc()' against mydb to see if this is a permanent or transient error.

  • Error upon logging into the database

    We've been having this intermittent occurrence, where the following error pops up upon logging into a database instance (SQL Developer is already open)
    java.lang.NullPointerException
         at oracle.dbtools.raptor.utils.TNSHelper$2.compare(TNSHelper.java:349)
         at java.util.Arrays.mergeSort(Arrays.java:1270)
         at java.util.Arrays.mergeSort(Arrays.java:1281)
         at java.util.Arrays.mergeSort(Arrays.java:1281)
         at java.util.Arrays.mergeSort(Arrays.java:1281)
         at java.util.Arrays.mergeSort(Arrays.java:1282)
         at java.util.Arrays.sort(Arrays.java:1210)
         at java.util.Collections.sort(Collections.java:159)
         at oracle.dbtools.raptor.utils.TNSHelper.getTNSEntries(TNSHelper.java:355)
         at oracle.dbtools.raptor.utils.TNSHelper.getEntry(TNSHelper.java:216)
         at oracle.dbtools.raptor.utils.Connections.populateConnectionInfo(Connections.java:645)
         at oracle.dbtools.raptor.standalone.connection.RaptorConnectionCreator.getConnection(RaptorConnectionCreator.java:62)
         at oracle.dbtools.raptor.dialogs.conn.ConnectionPrompt.promptForPassword(ConnectionPrompt.java:67)
         at oracle.jdeveloper.db.adapter.DatabaseProvider.getConnection(DatabaseProvider.java:321)
         at oracle.jdeveloper.db.adapter.DatabaseProvider.getConnection(DatabaseProvider.java:254)
         at oracle.jdevimpl.db.adapter.CADatabaseFactory.createConnectionImpl(CADatabaseFactory.java:60)
         at oracle.javatools.db.DatabaseFactory.createConnection(DatabaseFactory.java:534)
         at oracle.javatools.db.DatabaseFactory.createDatabase(DatabaseFactory.java:208)
         at oracle.jdeveloper.db.DatabaseConnections.getDatabase(DatabaseConnections.java:607)
         at oracle.jdeveloper.db.DatabaseConnections.getDatabase(DatabaseConnections.java:562)
         at oracle.dbtools.raptor.utils.Connections$ConnectionInfo.getDatabase(Connections.java:145)
         at oracle.dbtools.raptor.utils.Connections.getConnection(Connections.java:874)
         at oracle.dbtools.raptor.utils.Connections.getConnection(Connections.java:855)
         at oracle.dbtools.raptor.navigator.DatabaseConnection.openConnectionImpl(DatabaseConnection.java:117)
         at oracle.dbtools.raptor.navigator.AbstractConnectionNode.getConnection(AbstractConnectionNode.java:30)
         at oracle.dbtools.raptor.navigator.ConnectionFilter.getFactory(ConnectionFilter.java:92)
         at oracle.dbtools.raptor.navigator.ConnectionFilter$1.doWork(ConnectionFilter.java:117)
         at oracle.dbtools.raptor.navigator.ConnectionFilter$1.doWork(ConnectionFilter.java:102)
         at oracle.dbtools.raptor.backgroundTask.RaptorTask.call(RaptorTask.java:193)
         at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
         at java.util.concurrent.FutureTask.run(FutureTask.java:138)
         at oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask.run(RaptorTaskManager.java:492)
         at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
         at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
         at java.util.concurrent.FutureTask.run(FutureTask.java:138)
         at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
         at java.lang.Thread.run(Thread.java:619)What's strange is if that person closes SQL Developer and refrains from using it for about 20 minutes, everything works fine the next time they re-open the tool & try to log in. The issue seems to affect the Developer team in a round-robin fashion too, never hitting the same person 2 times in a row.
    We're using SQL Developer in a Citrix environment, and no one else has the problem simultaneously. Does anyone see something useful in the error message above?
    Thanks
    --=Chuck
    SQL Developer Version 2.1.0.63

    A good number of folks are having problems with this same error message at login, and nobody seems to have found a fix.
    I believe I have stumbled upon, if not a solution, at least part of the answer. If you're logged into a 10.4 workstation, open a Finder window and click the "Network" icon. The first time you do this, you'll only see "Library" and "Servers" icons, but if you let it sit for a moment, other folders will magically appear. One of these is "My Network" which is the critical part.
    In Workgroup Manager, edit one of your users' home directory configuration. There's a URL piece, usually "afp://server.domain.com/Users" (which is what shows up in the window when you click on the "Home" tab for that user), and then there's a "Home" path that is usually something like "/Network/Servers/servername/Users/username". Trouble is, you've noticed that 10.4 doesn't put servers into the "Servers" path, but rather into the "My Network" path.
    So, if you change that "Home" path to read "/Network/My Network/servername/Users/username", it should work--with one caveat. Apparently 10.4 doesn't create the "My Network" path on a given client until you actually browse the network on that client. If you haven't logged into a 10.4 workstation and browsed the network, the "My Network" path won't be there, and you'll still get the error. Login as a local user, browse the network (so that the "My Network" folder shows up), logout, and then login as an Open Directory user, and it should work.
    I'm betting that 10.3, 10.4, and 10.5 all handle this issue differently, so if you're in a mixed-version situation, you may have to experiment to find a workable fix.
    Derndest thing I ever saw.

  • The log file for database 'tempdb' is full

    Hi Experts,
    in my sync RFC to JDBC scenario I get the following error message which says:
    11.02.2009 11:23:20 Error Unable to execute statement for table or stored procedure.
    'Trns' (Structure 'XYZ')
    due to java.sql.SQLException:
    [Microsoft][SQLServer 2000 Driver for JDBC][SQLServer]
    The log file for database 'tempdb' is full. Back up the transaction
    log for the database to free up some log space.
    11.02.2009 11:23:20 Error JDBC message processing failed;
    reason Error processing request in sax parser: Error when executing
    statement for table/stored proc. 'Trns' (structure 'XYZ'):
    java.sql.SQLException: [Microsoft][SQLServer 2000 Driver for JDBC][SQLServer]
    The log file for database 'tempdb' is full. Back up the transaction log
    for the database to free up some log space.
    see here:
    http://img23.imageshack.us/img23/6541/unbenannthd9.jpg
    does anyone know how to solve this problem?
    thanks
    chris

    Christian,
    See this page for more information on the error you're getting: http://sqlserver2000.databases.aspfaq.com/why-is-tempdb-full-and-how-can-i-prevent-this-from-happening.html.
    Kind regards,
    Koen

  • Configuring OID and logging into a database with a OID username\password

    I've installed the Oracle database software (v9.2.0.2.1) and created a database called OIDDB. I then installed the "Management and Integration" portion of the Oracle database software which includes "Oracle Management Server" and "Oracle Internet Directory". All of this installation is on the same database box.
    On another server, my other database (called: BAHDB) resides that I OID enabled. I've already used the "Database Configuration Assistant" to register it with the directory service.
    What else needs to be done at this point to log into the database without a Oracle username and password? What high level main steps am I missing? Do I need to do anything with Oracle wallets?

    There's now a difference when you use domains in your network configuration:
    tnsping server.at.your.domain should get resolved through tnsnames.
    tnsping server shoudl get resolved through the LDAP server.
    The illegal userid/password error makes me suspect you try to logon with the fully qualified server name; LDAP should respond with something like "no nickname found".
    Also - the guest user is only used as a "stub" - you should be able to login with a name, defined in OID - not defined in the database!
    Things to check:
    - rdbms_server_dn - DBCA should take care of it during registration of the database in OID, but it does not always do that. Setting it requires a restart. Remeember, it is case sensitive! Format is cn=SID,cn=OracleContext,dc=<your realm here>
    - Wallet. Was it created, what is in it?
    - User mapping (Use the Enterprise Security Manager - part of the client install); open Realms, open OracleDefaultDomain within Enterprise Domains, third tab ("Database Schema Mapping"), select "Add".
    Navigate to cn=Users,dc=<your realm>. Make sure Subtree level is selected, and use your globally identified user (you used guest) for schema; click OK.
    Click Apply
    You should now be able to logon as any user, defined in the OID under cn=Users,dc=<your realm> - orcladmin should do.
    when logged on, show user will respond with "guest"; use select sys_context('userenv','external_name') from dual; to find out the real user.
    The way you now set it up probably will fail to login with 'user lacks create session privilege' - for test purposes, grant create session to guest - in a 'normal' eenviroment, you would use enterprise roles, and assign these to the OID defined users, not directly to the "stub user" (guest in your example)

  • UUT Result log in SQL database?

    Hello All
    I am facing a requirement where i want to log results of sequence execution[only the complete seq result and NOT step results] to SQL database BUT as and when the UUT_ID varies the entry should be made into a new column with the same name as UUT_ID. As of now whenever i execute a sequence with different UUT_ID, a new row gets added as a result.
    Please reply back with your suggestions asap, as this very urgent for me :-)
    Regards
    Nitin Goel

    But in my opinion, that's just a way to display the results - not a way to store the data. I also have thousands of test cases and tens of thousands of different UUT each month. Each UUT has a separate record (a unique row) and yet if I design a proper query and report format, I can show the results exactly like you have. The design of a relational database should not be the same as a report design, imho.
    A database is designed to have unlimited rows (dependent only on storage space). I'm not sure a db can have unlimited columns. That would be something you could check with your dba. Dynamically adding a new column and doing an update to an existing row would probaby be much slower than just doing an insert.

  • How to get log information from database server?

    Oracle 9.2 in Unix
    we got blocking error in the production database last midnight for 40 min and nothing was being processed during that time which delayed our production night batch process. Then it was gone after 40min. I like to know what was happening during that time in the database. My question is where I can find error in the database server? In bdump? Appreciate any ideas.
    Thanks
    S.

    If you have jobs regularly collect statspack report, then it will be easier to troubleshoot. You can just pull the report from the time period.
    A couple of things you can check if there's any storage issue during the time, like some backup or other jobs running taken all OS storage, so Oracle hung because no space to archive logs. Usually this will log errors in alert.log but if bdump destination also full, the error will not be logged.
    Check OS activity during the time frame, what's CPU and IO activity. See if you have sar installed and configured
    check OS message file.

Maybe you are looking for