Truncating a large database failed

if (0 != dbp->truncate(dbp, NULL, &countp, 0)){
out_string(c, "ERROR");
}else{
out_string(c, "OK");
If only a few records in database, it works. But when millions of records, it fails.
Is there a need to set following three parameters?
DB_ENV->set_lk_max_lockers     Set maximum number of lockers
DB_ENV->set_lk_max_locks     Set maximum number of locks
DB_ENV->set_lk_max_objects     Set maximum number of lock objects
Regards
Steve Chu

Hi Steve,
It's highly likely that you haven't configured the locking subsystem properly. Locking is involved indeed when truncating a database that is part of an environment with the locking subsystem initialized. For each database page a lock is needed; so if you know the size of your database and the page size you can estimate how many locks you will need (you can also use the "db_stat" utility to find out this information). The default values for the maximum number of locks, lockers and lock objects is 1000, which may be insufficient in the context of your large database.
Information on how to configure the locking subsystem is here:
http://you've-db/db/ref/lock/max.html
Also, I would suggest testing the value returned by DB->truncate() and printing out the eventual error using the error reporting means you'e configured.
Regards,
Andrei

Similar Messages

  • Import to new Server with Standard Database Failing at 5%

    Since the new database edition (basic, standard, premium) preview has gone live I have been trying to import an automatically exported bacpac onto a new server and into a Standard S1 database using the portal UI. So far every attempt has failed
    at roughly 5% which it never appears to move past. It is roughly 26GB, the bacpac was created with the automatic export on a business edition database and I have selected Standard, S1 and 30GB for the new database. I have tried 3 times now with
    no success and have no idea what to try next.
    Any ideas on how to get this database into the new edition?

    Hello,
    I had also the same problem with much smaller databases. The only solution for the moment is to use Premium (maybe Premium 2 or 3) to do the import the database.
    The reason is that when I am trying to import the database, I see on the dashboard  monitor that I use 100% of available Write_log metric for the edition, and in some way the server is "slowing" my import requests to keep me in my edition limits. Of
    course with this policy is almost impossible to import relatively large databases to these editions. I hope Microsoft engineers find a solution to this problem, or remove the limit for this metric.
    Regards,
    Dimitris

  • Proc_getObject with nolock causes "read operation on a large object failed while sending data to the client"

    SharePoint 2013 code for the SharePoint Config database stored procedure dbo.proc_getObject has been changed from SharePoint 2010 with the difference of select with nolock.
    I am seeing numerous Error 7886, Severity-20 errors nightly:
    "A read operation on a large object failed while sending data to the client. A common cause for this is if the application is running in READ UNCOMMITTED isolation level. This connection will be terminated."
    Here I have commented out "with (nolock)" to 'fix' the error and to align the store proc with SharePoint 2010 code : 
    ALTER PROCEDURE [dbo].[proc_getObject]
    @Id uniqueidentifier,
    @RequestGuid uniqueidentifier = NULL OUTPUT
    AS
    SET NOCOUNT ON
    SELECT
    Id,
    ParentId,
    ClassId,
    Name,
    Status,
    Version,
    Properties
    FROM
    Objects --with (nolock)
    WHERE
    Id = @Id
    RETURN 0
    How is this stored procedure called and for what use?
    Why did Microsoft make this change to SharePoint 2013 - some performance gain through dirty reads at the cost of stability? Or what could be a reason for this to happen that can be fixed?

    GetObject retrieves any farm object that is selected (by ID). Changing this is obviously not supported, and NoLock was used to (help) prevent locks when performing a select query.
    I've never seen this issue, and would guess it is unique to your environment given this is such a core piece of the SharePoint infrastructure.
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • Problem with  large databases.

    Lightroom doesn't seem to like large databases.
    I am playing catch-up using Lightroom to enter keywords to all my past photos. I have about 150K photos spread over four drives.
    Even placing a separate database on each hard drive is causing problems.
    The program crashes when importing large numbers of photos from several folders. (I do not ask it to render previews.) If I relaunch the program, and try the import again, Lightroom adds about 500 more photos and then crashes, or freezes again.
    I may have to go back and import them one folder at a time, or use iView instead.
    This is a deal-breaker for me.
    I also note that it takes several minutes after opening a databese before the HD activity light stops flashing.
    I am using XP on a dual core machine with, 3Gigs of RAM
    Anyone else finding this?
    What is you work-around?

    Christopher,
    True, but given the number of posts where users have had similar problems ingesting images into LR--where LR runs without crashes and further trouble once the images are in--the probative evidence points to some LR problem ingesting large numbers.
    I may also be that users are attempting to use LR for editing during the ingestion of large numbers--I found that I simply could not do that without a crash occuring. When I limited it to 2k at a time--leaving my hands off the keyboard-- while the import occured, everything went without a hitch.
    However, as previously pointed out, it shouldn't require that--none of my other DAMs using SQLite do that, and I can multitask while they are ingesting.
    But, you are right--multiple single causes--and complexly interrated multiple causes--could account for it on a given configuration.

  • Error message: "Querying or saving changes to the database failed."

    Hi all,
    I can sense I'm becoming a regular on this particular forum - the problem of spending 8 hours a day trying to get this software working!
    _*The problem*_
    When i try and setup a new device, or modify an existing one, I get an error message when i get to the Transcode Settings page. The error message is "Querying or saving changes to the database failed." After this I get a blank screen and can no longer see my settings.
    No amount of restarting has fixed this. And I cannot see a way to create a device through the client software and add transcode settings.
    Help?
    Thanks in advance,
    Ben

    Ah very good!
    Ok - it seems then that this works absolutely fine as a work around. I can create a device and then assign the device to transcode settings through the client adminsitration panel (which is indeed what I meant).
    So, there is a work around. Far less pressing now, but I would like to know why I'm getting the error message all the same - I'm now concerned about the integrity of the database...
    Thanks Tony!
    B

  • How can we suggest a new DBA OCE certification for very large databases?

    How can we suggest a new DBA OCE certification for very large databases?
    What web site, or what phone number can we call to suggest creating a VLDB OCE certification.
    The largest databases that I have ever worked with barely over 1 Trillion Bytes.
    Some people told me that the results of being a DBA totally change when you have a VERY LARGE DATABASE.
    I could guess that maybe some of the following topics of how to configure might be on it,
    * Partitioning
    * parallel
    * bigger block size - DSS vs OLTP
    * etc
    Where could I send in a recommendation?
    Thanks Roger

    I wish there were some details about the OCE data warehousing.
    Look at the topics for 1Z0-515. Assume that the 'lightweight' topics will go (like Best Practices) and that there will be more technical topics added.
    Oracle Database 11g Data Warehousing Essentials | Oracle Certification Exam
    Overview of Data Warehousing
      Describe the benefits of a data warehouse
      Describe the technical characteristics of a data warehouse
      Describe the Oracle Database structures used primarily by a data warehouse
      Explain the use of materialized views
      Implement Database Resource Manager to control resource usage
      Identify and explain the benefits provided by standard Oracle Database 11g enhancements for a data warehouse
    Parallelism
      Explain how the Oracle optimizer determines the degree of parallelism
      Configure parallelism
      Explain how parallelism and partitioning work together
    Partitioning
      Describe types of partitioning
      Describe the benefits of partitioning
      Implement partition-wise joins
    Result Cache
      Describe how the SQL Result Cache operates
      Identify the scenarios which benefit the most from Result Set Caching
    OLAP
      Explain how Oracle OLAP delivers high performance
      Describe how applications can access data stored in Oracle OLAP cubes
    Advanced Compression
      Explain the benefits provided by Advanced Compression
      Explain how Advanced Compression operates
      Describe how Advanced Compression interacts with other Oracle options and utilities
    Data integration
      Explain Oracle's overall approach to data integration
      Describe the benefits provided by ODI
      Differentiate the components of ODI
      Create integration data flows with ODI
      Ensure data quality with OWB
      Explain the concept and use of real-time data integration
      Describe the architecture of Oracle's data integration solutions
    Data mining and analysis
      Describe the components of Oracle's Data Mining option
      Describe the analytical functions provided by Oracle Data Mining
      Identify use cases that can benefit from Oracle Data Mining
      Identify which Oracle products use Oracle Data Mining
    Sizing
      Properly size all resources to be used in a data warehouse configuration
    Exadata
      Describe the architecture of the Sun Oracle Database Machine
      Describe configuration options for an Exadata Storage Server
      Explain the advantages provided by the Exadata Storage Server
    Best practices for performance
      Employ best practices to load incremental data into a data warehouse
      Employ best practices for using Oracle features to implement high performance data warehouses

  • Move large database to other server using RMAN in less downtime

    Hi,
    We have large database around 20TB. We want to migrate (move) the database from one server to other server. We do not want to use standby option.
    1)     How can we move database using RMAN in less downtime
    2)     Other than RMAN is there any option is available to move the database to new server
    For option 1 (restore using RMAN),
    Whether below options are valid?
    If this option is valid, how to implement this?
    1)     How can we move database using RMAN in less downtime
    a)     Take the full backup from source (source db is up)
    b)     Restore the full backup in target (source db is up)
    c)     Take the incremental backup from source (source db is up)
    d)     Restore incremental backup in target (source db is up)
    e)     Do steps c and d, before taking downtime (source db is up)
    f)     Shutdown and mount the source db, and take the incremental backup (source db is down)
    g)     Restore last incremental backup and start the target database (target is up and application is accessing this new db
    database version: 10.2.0.4
    OS: SUN solaris 10
    Edited by: Rajak on Jan 18, 2012 4:56 AM

    Simple:
    I do this all the time to relocate file system files... But the principle is the same. You can do this in iterations so you do not need to do it all at once:
    Starting 8AM move less-used files and more active files in the afternoon using the following backup method.
    SCRIPT-1
    RMAN> BACKUP AS COPY
    DATAFILE 4 ####"/some/orcl/datafile/usersdbf"
    FORMAT "+USERDATA";
    Do as many files as you think you can handle during your downtime window.
    During your downtime window: stop all applications so there is no contention in the database
    SCRIPT-2
    ALTER DATABASE DATAFILE 4 offline;
    SWITCH DATAFILE 4 TO COPY;
    RECOVER DATAFILE 4;
    ALTER DATABASE DATAFILE 4 online;
    I then execute the delete of the original file at somepoint later - after we make sure everything has recovered and successfully brought back online.
    SCRIPT-3
    DELETE DATAFILECOPY "/some/orcl/datafile/usersdbf"
    For datafiles/tablespaces that are really busy, I typically copy them later in the afternoon as there are fewer archivelogs that it has to go through in order to make them consistent. The ones in the morning have more to go through, but less likelihood of there being anything to do.
    Using this method, we have moved upwards 600G at a time and the actual downtime to do the switchover is < 2hrs. YMMV. As I said, this can be done is stages to minimize overall downtime.
    If you need some documentation support see:
    http://docs.oracle.com/cd/E11882_01/server.112/e18951/asm_rman.htm#CHDBDJJG
    And before you do ANYTHING... TEST TEST TEST TEST TEST. Create a dummy tablespace on QFS and use this procedure to move it to ASM to ensure you understand how it works.
    Good luck! (hint: scripts to generate these scripts can be your friend.)

  • Connect to database failed,rc=-10709 Connection failed (RTE:Database name)

    Hi gurues, I'm facing this issue with an IDES on MaxDB with Windows.
    System was up and running until someone tries to make an upgrade of the kernel. Since then we can not start our system. we tried to rollback kernel release to the previous one but it didn't work.
    When we start our system from MMC console we can see all process in green but work processes are in RUN state and after a few minutes they goes down with the following error
    trc file: "dev_w0", trc level: 1, release: "701"
    ACTIVE TRACE LEVEL           1
    ACTIVE TRACE COMPONENTS      all, MJ

    B Mon May 16 23:57:31 2011
    B  create_con (con_name=R/3)
    B  Loading DB library '
    sirillium1\sapmnt\SED\SYS\exe\uc\NTAMD64\dbsdbslib.dll' ...
    B  Library '
    sirillium1\sapmnt\SED\SYS\exe\uc\NTAMD64\dbsdbslib.dll' loaded
    B  Version of '
    sirillium1\sapmnt\SED\SYS\exe\uc\NTAMD64\dbsdbslib.dll' is "700.08", patchlevel (0.24)
    B  New connection 0 created
    M sysno      00
    M sid        SED
    M systemid   562 (PC with Windows NT)
    M relno      7010
    M patchlevel 0
    M patchno    32
    M intno      20020600
    M make:      multithreaded, Unicode, 64 bit, optimized
    M pid        212
    M
    M  kernel runs with dp version 241000(ext=110000) (@(#) DPLIB-INT-VERSION-241000-UC)
    M  length of sys_adm_ext is 576 bytes
    M  ***LOG Q0Q=> tskh_init, WPStart (Workproc 0 212) [dpxxdisp.c   1348]
    I  MtxInit: 30000 0 0
    M  DpSysAdmExtCreate: ABAP is active
    M  DpSysAdmExtCreate: VMC (JAVA VM in WP) is not active
    M  DpShMCreate: sizeof(wp_adm)          22784     (1752)
    M  DpShMCreate: sizeof(tm_adm)          5912704     (29416)
    M  DpShMCreate: sizeof(wp_ca_adm)          24064     (80)
    M  DpShMCreate: sizeof(appc_ca_adm)     8000     (80)
    M  DpCommTableSize: max/headSize/ftSize/tableSize=500/16/552064/552080
    M  DpShMCreate: sizeof(comm_adm)          552080     (1088)
    M  DpSlockTableSize: max/headSize/ftSize/fiSize/tableSize=0/0/0/0/0
    M  DpShMCreate: sizeof(slock_adm)          0     (104)
    M  DpFileTableSize: max/headSize/ftSize/tableSize=0/0/0/0
    M  DpShMCreate: sizeof(file_adm)          0     (72)
    M  DpShMCreate: sizeof(vmc_adm)          0     (1864)
    M  DpShMCreate: sizeof(wall_adm)          (41664/36752/64/192)
    M  DpShMCreate: sizeof(gw_adm)     48
    M  DpShMCreate: SHM_DP_ADM_KEY          (addr: 000000000B2C0050, size: 6607136)
    M  DpShMCreate: allocated sys_adm at 000000000B2C0050
    M  DpShMCreate: allocated wp_adm at 000000000B2C2270
    M  DpShMCreate: allocated tm_adm_list at 000000000B2C7B70
    M  DpShMCreate: allocated tm_adm at 000000000B2C7BD0
    M  DpShMCreate: allocated wp_ca_adm at 000000000B86B450
    M  DpShMCreate: allocated appc_ca_adm at 000000000B871250
    M  DpShMCreate: allocated comm_adm at 000000000B873190
    M  DpShMCreate: system runs without slock table
    M  DpShMCreate: system runs without file table
    M  DpShMCreate: allocated vmc_adm_list at 000000000B8F9E20
    M  DpShMCreate: allocated gw_adm at 000000000B8F9EA0
    M  DpShMCreate: system runs without vmc_adm
    M  DpShMCreate: allocated ca_info at 000000000B8F9ED0
    M  DpShMCreate: allocated wall_adm at 000000000B8F9EE0

    M Mon May 16 23:57:32 2011
    M  rdisp/queue_size_check_value :  -> off
    M  ThTaskStatus: rdisp/reset_online_during_debug 0
    X  EmInit: MmSetImplementation( 2 ).
    X  MM global diagnostic options set: 0
    X  <ES> client 0 initializing ....
    X  Using implementation view
    X  <EsNT> Using memory model view.
    M  <EsNT> Memory Reset disabled as NT default
    X  ES initialized.
    M  ThInit: running on host sirillium1

    M Mon May 16 23:57:33 2011
    M  calling db_connect ...

    C  DBSDBSLIB : version 700.08, patch 0.024 (Make PL 0.32)
    C  MAXDB shared library (dbsdbslib) patchlevels (last 10)
    C    (0.024) Default value for max. input variables is 2000 (note 655018)
    C    (0.024) Profile parameter to define max. input variables (note 655018)
    C    (0.024) Switch SQLMODE after CREATE INDEX SERIAL (note 1267841)
    C    (0.024) Input parameters for SQL statements increased (note 655018)
    C    (0.018) Create index serial for MaxDB 7.6 (note 1267841)
    C    (0.018) More trace in case of packed to string conversion error (note 1262799)
    C    (0.016) R3trans export aborts with signal 6 (note 1262245)
    C    (0.009) IA64 alignment errors (note 1245982)
    C    (0.007) Support DB-Type 'SAP DB' by UPDSTAT (note 1225668)


    C  Loading SQLDBC client runtime ...
    C  SQLDBC SDK Version : SQLDBC.H  7.6.0    BUILD 007-123-091-175
    C  SQLDBC Library Version : libSQLDBC 7.6.5    BUILD 015-123-202-938
    C  SQLDBC client runtime is MaxDB 7.6.5.015 CL 202938
    C  SQLDBC supports new DECIMAL interface : 0
    C  SQLDBC supports VARIABLE INPUT data   : 1
    C  SQLDBC supports keepAlive indicator   : 0
    C  INFO : SQLOPT= -I 0 -t 0 -S SAPR3
    C  Try to connect (DEFAULT) on connection 0 ...
    C  *** ERROR => Connect to database failed, rc = -10709 (Connection failed (RTE:Database name is missing))
    [dbsdbsql.cpp 137]
    B  ***LOG BV3=> severe db error -10709    ; work process is stopped [dbsh#2 @ 1203] [dbsh    1203 ]
    B  ***LOG BY2=> sql error -10709 performing CON [dblink#3 @ 431] [dblink  0431 ]
    B  ***LOG BY0=> Connection failed (RTE:Database name is missing) [dblink#3 @ 431] [dblink  0431 ]
    M  ***LOG R19=> ThInit, db_connect ( DB-Connect 000256) [thxxhead.c   1449]
    M  in_ThErrHandle: 1
    M  *** ERROR => ThInit: db_connect (step 1, th_errno 13, action 3, level 1) [thxxhead.c   10563]

    M  Info for wp 0

    M    pid = 212
    M    severity = 0
    M    status = 0
    M    stat = WP_RUN
    M    waiting_for = NO_WAITING
    M    reqtype = DP_RQ_DIAWP
    M    act_reqtype = NO_REQTYPE
    M    rq_info = 0
    M    tid = -1
    M    mode = 255
    M    len = -1
    M    rq_id = 65535
    M    rq_source =
    M    last_tid = 0
    M    last_mode = 0
    M    semaphore = 0
    M    act_cs_count = 0
    M    csTrack = 0
    M    csTrackRwExcl = 0
    M    csTrackRwShrd = 0
    M    mode_cleaned_counter = 0
    M    control_flag = 0
    M    int_checked_resource(RFC) = 0
    M    ext_checked_resource(RFC) = 0
    M    int_checked_resource(HTTP) = 0
    M    ext_checked_resource(HTTP) = 0
    M    report = >                                        <
    M    action = 0
    M    tab_name = >                              <
    M    attachedVm = no VM

    M  *****************************************************************************
    M  *
    M  *  LOCATION    SAP-Server sirillium1_SED_00 on host sirillium1 (wp 0)
    M  *  ERROR       ThInit: db_connect
    M  *
    M  *  TIME        Mon May 16 23:57:33 2011
    M  *  RELEASE     701
    M  *  COMPONENT   Taskhandler
    M  *  VERSION     1
    M  *  RC          13
    M  *  MODULE      thxxhead.c
    M  *  LINE        10783
    M  *  COUNTER     1
    M  *
    M  *****************************************************************************

    M  PfStatDisconnect: disconnect statistics
    M  Entering TH_CALLHOOKS
    M  ThCallHooks: call hook >BtcCallLgCl< for event BEFORE_DUMP
    M  ThCallHooks: call hook >ThrSaveSPAFields< for event BEFORE_DUMP
    M  *** ERROR => ThrSaveSPAFields: no valid thr_wpadm [thxxrun1.c   723]
    M  *** ERROR => ThCallHooks: event handler ThrSaveSPAFields for event BEFORE_DUMP failed [thxxtool3.c  261]
    M  Entering ThSetStatError
    M  ThIErrHandle: do not call ThrCoreInfo (no_core_info=0, in_dynp_env=0)
    M  Entering ThReadDetachMode
    M  call ThrShutDown (1)...
    M  ***LOG Q02=> wp_halt, WPStop (Workproc 0 212) [dpnttool.c   334]
    We check several entries on this SDN site, and look for a lot of notes on the marketplace, and they always are related to default user in xuser files, but our xuser list output is ok. We have a DEFAULT connection to SAPSID user, a DBM connection to CONTROL user and a USERDBA connection.
    XUSER Entry  1
    Key         :DEFAULT          
    Username    :SAPSID                                                         
    UsernameUCS2:S.A.P.S.E.D. . . . . . . . . . . . . . . . . . . . . . . . . . .
    Password    :?????????
    PasswordUCS2:?????????
    Dbname      :SID              
    Nodename    :server1                                                     
    Sqlmode     :SAPR3  
    Cachelimit  :-1
    Timeout    
    Isolation  
    Charset     :<unspecified>    
    XUSER Entry  2
    Key         :c                
    Username    :CONTROL                                                        
    UsernameUCS2:C.O.N.T.R.O.L. . . . . . . . . . . . . . . . . . . . . . . . . .
    Password    :?????????
    PasswordUCS2:?????????
    Dbname      :SID              
    Nodename    :server1                                                     
    Sqlmode     :INTERNAL
    Cachelimit  :-1
    Timeout     :-1
    Isolation   :-1
    Charset     :<unspecified>    
    XUSER Entry  3
    Key         :c_J2EE           
    Username    :CONTROL                                                        
    UsernameUCS2:C.O.N.T.R.O.L. . . . . . . . . . . . . . . . . . . . . . . . . .
    Password    :?????????
    PasswordUCS2:?????????
    Dbname      :SID              
    Nodename    :server1                                                     
    Sqlmode     :SAPR3  
    Cachelimit  :-1
    Timeout    
    Isolation  
    Charset     :<unspecified>    
    XUSER Entry  4
    Key         :w                
    Username    :SUPERDBA                                                       
    UsernameUCS2:S.U.P.E.R.D.B.A. . . . . . . . . . . . . . . . . . . . . . . . .
    Password    :?????????
    PasswordUCS2:?????????
    Dbname      :SID              
    Nodename    :server1                                                     
    Sqlmode     :INTERNAL
    Cachelimit  :-1
    Timeout     :-1
    Isolation   :-1
    Charset     :<unspecified>    
    Any idea? we are out of ideas at this moment.
    Regards
    Edited by: Gustavo Goicochea on May 17, 2011 6:37 PM

    Natalia, this system was not resored nor copied, it was up and running before kernel update. We did some of the tests you said and someothers, here are some outputs.
    We update kernel to release 142
    R3trans -d and -x exit with (all with user SIDADM)
    SVERS table does exist and belongs to SAPSID
    dbmcli on SID>sql_execute select * from domain.tables where tablename='SVERS'
    OK
    END
    'SAPSID';'SAPSID';'SVERS';'';'TABLE';'TABLE';'20100406';'00054637';'20100406';'00054637';'20100406';'00054638';'NO';(null);20000;(null);x'0000000000004842';'NO';'NO';'YES';'NO';(null)
    Why it still says that SVERS table does not exist and SQL error 942?
    C  Now I'm connected to SAP DB
    C  00: server1-SID, since=20110518015619, ABAP= <unknown> (0)
    B  Connection 0 opened (DBSL handle 0)
    C  *** ERROR =>   prepare() of C_0000, rc=1, rcSQL=-942 (POS(28) Unknown table name:SVERS) [
    dbsdbsql.cpp 1604]
    C  *** ERROR => SQL PREPARE on connection 0, rc=-942 (POS(28) Unknown table name:SVERS) [dbslsdb.cpp  7253]
    C sc_p=000000000CFAECC0,no=0,idc_p=0000000000000000,con=0,act=0,slen=34,smax=256,#vars=0,stmt=0000000002EA0AF0,table=SVERS                        
    C  SELECT VERSION INTO ? FROM SVERS  ;
    B  ***LOG BZA=> table SVERS      does not exist on database            [dblink#4 @ 1301] [dblink  1301 ]
    M  ***LOG R19=> ThInit, db_connect ( DB-Connect 004096) [thxxhead.c   1515]
    M  in_ThErrHandle: 1
    M  *** ERROR => ThInit: db_connect (step 1, th_errno 13, action 3, level 1) [thxxhead.c   10824]

    Users seem to be ok
    dbmcli on SID>sql_execute SELECT username, usermode FROM users
    OK
    END
    'CONTROL';'ADMIN'
    'SUPERDBA';'SYSDBA'
    'SAPSID';'DBA'
    'SAPR3';'DBA'
    User SAPR3 does not have anything on database
    dbmcli on SID>sql_execute select * from domain.tables where owner='SAPR3'
    ERR
    -24988,ERR_SQL: SQL error
    100,Row not found
    Any idea?
    regards

  • Error in /tmp/applianceagent/scripts/Database.sh: Start database  failed with exit code 2

    Hello
    I have set up the Demo appliance landscape for the openSAP course "Introduction to SAP Fiori UX", on AWS.
    The frontend machine is OK, but the configuration and startup of the server machine has failed for some reason. When logging in with ssh to the server, this message is presented:
    The Appliance Agent detected an error and terminated.
    When asked, please provide the file /tmp/applianceagent_20141002-145658.tar.bz2.
    Note that a restrictive iptables based firewall is still active.
    I have attached the file (had to add .txt to the name to circumvent file type restrictions).
    I noticed that the appliancedeploy.log ended with these lines:
    Start database  Execute /usr/sap/hostctrl/exe/sapcontrol -nr 02 -function StartWait 3600 20
    02.10.2014 13:56:58
    Start
    OK
    02.10.2014 14:56:58
    StartWait
    FAIL: Timeout
    error in /tmp/applianceagent/scripts/Database.sh: Start database  failed with exit code 2
    Notice the 1 hour jump in time, and then the timeout error message.
    Do any of you have any idea what could be causing this, and how to fix it?
    Please let me know if you need any further information.
    Best regards,
    Bo

    Hi Hannes.
    Trying to create and activate a new instance (version 5 of the Fiori appliance), I first got this message in the CAL instance overview list:
    Instance Operation Failed
    Operation Start failed on step Start start VAs with message: Start failed for VM SAP Server due to Amazon Web Services issue : AMAZON : The snapshot 'snap-c779ff64' does not exist.
    Then I terminated the instance and tried over again. And this time the deployment worked - I have a running server, which responds to the SAP GUI, and shows the Fiori web page.
    Thanks a lot for your quick help.
    Best regards,
    Bo

  • Constant failures in setting up scalable wordpress. Update ClearDB database failed.

    I've been having Constant failures in setting up scalable wordpress with no luck on finding a fix. 
    Update ClearDB database failed is the most common issue, with Update deployment being the second issue. I though maybe it had to do with Azure changes over the last few months (beta portal), but have yet to successfully deploy a site. 
    Any suggestions / fixes would be greatly appreciated. 
    Also, i've tried utilizing the support link but it seems to hang perpetually (regardless of browser) when checking my subscription. 

    Hi,
    Two points need confirmed:
    1.Have you already purse the ClearDB database?
    2.What's error did you encounter when you scalable WP site?
    Regards,
    Will 
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Connection test to database failed: could not connect to server

    I restarted one of my HPOV NNMi 9.23 servers this morning and NNMi is no longer loading.  I am seeing the following errors in the logs indicated:
    /var/opt/OV/log/nnm/public/nmsdbmgr.log
    04/06/2015 09:28:05 AM Connection test to database failed: could not connect to server: Connection refused
    Is the server running on host "127.0.0.1" and accepting
    TCP/IP connections on port 5432?
    /var/opt/OV/log/nnm/public/postgres.log
    2015-04-06 09:25:55.312 EDT: :8580:0LOG: could not translate host name "localhost", service "5432" to address: Name or service not known
    2015-04-06 09:25:55.312 EDT: :8580:0WARNING: could not create listen socket for "localhost"
    2015-04-06 09:25:55.312 EDT: :8580:0FATAL: could not create any TCP/IP sockets
    I checked /etc/hosts and found the following entry:
    #; Local
    127.0.0.1 localhost
    From the command line I can ping localhost with no issue.  Not sure why NNMi is not starting at this point. 
    This question was solved.
    View Solution.

    I also attempted to do a reset using nnmresetebdb.ovpl.  I recieved the following:
    Attempting to reset the embedded database...
    Starting nmsdbmgr process for database reset...
    nmsdbmgr process did not restart successfully - please try reset again.
    When I run ovstart -c I receive the following:
    Name PID State Last Message(s)
    OVsPMD 29831 RUNNING -
    pmd 29832 RUNNING Initialization complete.
    nmsdbmgr - NOT_RUNNING - Exit(1)
    ovjboss - UNSTARTABLE -
    nnmaction - UNSTARTABLE -
    ovspmd: Attempt to start HP OpenView services is complete.
    I definitely need some help with this please!  
    Thanks!

  • Attached database failed for sever "blah". (Microsoft.SqlServer.Smo)

    ===================================
    Attach database failed for Server 'DENR-HP'.  (Microsoft.SqlServer.Smo)
    For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&ProdVer=10.50.1600.1+((KJ_RTM).100402-1539+)&EvtSrc=Microsoft.SqlServer.Management.Smo.ExceptionTemplates.FailedOperationExceptionText&EvtID=Attach+database+Server&LinkId=20476
    Program Location:
       at Microsoft.SqlServer.Management.Smo.Server.AttachDatabase(String name, StringCollection files)
       at Microsoft.SqlServer.Management.SqlManagerUI.AttachDatabaseData.PrimaryFile.Attach()
       at Microsoft.SqlServer.Management.SqlManagerUI.AttachDatabase.SendDataToServer()
    ===================================
    An exception occurred while executing a Transact-SQL statement or batch. (Microsoft.SqlServer.ConnectionInfo)
    Program Location:
       at Microsoft.SqlServer.Management.Common.ServerConnection.ExecuteNonQuery(String sqlCommand, ExecutionTypes executionType)
       at Microsoft.SqlServer.Management.Common.ServerConnection.ExecuteNonQuery(StringCollection sqlCommands, ExecutionTypes executionType)
       at Microsoft.SqlServer.Management.Smo.ExecutionManager.ExecuteNonQuery(StringCollection queries)
       at Microsoft.SqlServer.Management.Smo.Server.AttachDatabaseWorker(String name, StringCollection files, String owner, AttachOptions attachOptions)
       at Microsoft.SqlServer.Management.Smo.Server.AttachDatabase(String name, StringCollection files)
    ===================================
    The database 'PlanningDB' cannot be opened because it is version 655. This server supports version 611 and earlier. A downgrade path is not supported.
    Could not open new database 'PlanningDB'. CREATE DATABASE is aborted. (.Net SqlClient Data Provider)
    For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&EvtSrc=MSSQLServer&EvtID=948&LinkId=20476
    Server Name: (local)
    Error Number: 948
    Severity: 20
    State: 1
    Line Number: 1
    Program Location:
       at Microsoft.SqlServer.Management.Common.ConnectionManager.ExecuteTSql(ExecuteTSqlAction action, Object execObject, DataSet fillDataSet, Boolean catchException)
       at Microsoft.SqlServer.Management.Common.ConnectionManager.ExecuteTSql(ExecuteTSqlAction action, Object execObject, DataSet fillDataSet, Boolean catchException)
       at Microsoft.SqlServer.Management.Common.ServerConnection.ExecuteNonQuery(String sqlCommand, ExecutionTypes executionType)

    Hi,
    As Olaf’s post, you cannot restore or attach a database which is created from a higher version of SQL Server to a lower version of SQL Server.
    However, there are some ways which can help you to move a database to a lower version of SQL Server. I recommend you to use Generate Scripts Wizard of SQL Server Management Studio (SSMS), you can script all the database objects in SQL Server 2008, then run
    the SQL scripts in SQL Server 2005, for more details, you can follow the steps in this blog:
    How to migrate a SQL Server database to a lower version. Also you can try to use the Import and Export Wizard to move database from SQL Server 2008 to SQL Server 2005,for more details, please review this article:
    How to: Run the SQL Server Import and Export Wizard.
    Thanks
    Lydia Zhang

  • (2008 R2) AD RMS Installation Failure: ALTER DATABASE failed because a lock could not be placed on database

    First-time install on a SBS 2011 server, name replaced as <SERVER> for this post. I had created a separate IIS site rather than using the Default Web Site and I am guessing this might be a permission issue on the folder...
    Installation failed with Event ID 110 and a repeating error that states, "<Error>: Attempt to configure Active Directory Rights Management Server failed. The provisioning process failed to create the Configuration database due to the following
    error from Sql Server: System.Data.SqlClient.SqlException: ALTER DATABASE failed because a lock could not be placed on database 'DRMS_Config_adrms_<SERVER>_local_443'. Try again later. ALTER DATABASE statement failed. ALTER DATABASE failed because a
    lock could not be placed on database ." Configuration as follows:
    <Informational>: This server might need to be restarted after the installation completes.
    Active Directory Rights Management Services
    Cluster Type
    Root cluster
    Trust Hierarchy
    Production
    Configuration Database Server
    Windows Internal Database
    Service Account
    <SERVER>\adrms
    Cluster Key Storage
    AD RMS centrally managed key storage
    Cluster Web Site
    Active Directory RMS
    Cluster Internal Address
    https://adrms.<SERVER>.local:443/
    SSL Certificate
    Create a self-signed certificate
    Licensor Certificate Name
    <SERVER>
    Register SCP
    Register now

    Hi Ahmad, I have the same problem, I have a server windows server 2008 R2 standard. While
    doing an Enterprise to Standard downgrade, I have removed enterprise edition of SQL Server 2008, rebooted and while installing standard edition, I get the below error saying "You SQL server 2008 installation completed with failures. Some or all identity
    references could not be translated"
    Bit of background. This server has another express instance of SQL Server 2008. I have
    used local system account, network service as well as our general SQL service account, but I get the error towards the end of the installation and it fails saying “Some or all identity references could not be translated.”
    I am installing a named instance with the same name as before but fails. I have tried
    creating a new test instance that fails too. Any ideas? I have uninstalled SQL enterprise edition from
    Add and Remore Programs/Programs and Features. I've also run the Windows Installer CleanUp Utility but no joy.
    Cheers

  • 9.2 create Database fails in sql.bsq

    I have installed Oracle 9.2.0.1.0. It appears to have installed correctly.
    I am attempting to create a new database. The script sql.bsq in %ORACLE_HOME%\rdbms\admin appears to be trying to create a cluster around line 1329. I don't use clusters (per se).
    I am rcving an the following error a trace file.
    *** SESSION ID:(10.1) 2003-09-08 16:59:56.000
    ORA-00604: error occurred at recursive SQL level 1
    ORA-01569: data file too small for system dictionary tables
    Offending statement at line 1332
    create cluster c_cobj# (obj# number)
    pctfree 0 pctused 50
    /* space for: update cdef$ set condition = 'col IS NOT NULL' at // */
    size 300
    storage (initial 50K) /* avoid space management during IOR I */
    ORA-01501: CREATE DATABASE failed
    ORA-01519: error while processing file '%ORACLE_HOME%\RDBMS\ADMIN\SQL.BSQ' near line 1332
    ORA-00604: error occurred at recursive SQL level 1
    ORA-01569: data file too small for system dictionary tables
    Can someone point me to an answer for this?
    I have 2 95MB files being created in the system tablespace. and I can provide you either my create database script and/or my init.ora file, if they will assist.
    Thnak you

    Thank you.
    I have changed the create database statement several times to try different possiblities.
    I have put maxsize unlimited. That failed.
    I have put maxsize 200M (with a second file of maxsize 200M) that failed.
    I have added the temp storage syntax. That failed.
    All of these failed in sql.bsq and all used the ORA 1569 error about a data file being too small.
    WHAT DATA FILE? The system tablespace data file(s) have never been increased (by Oracle) from the original size value I provided (whether that size was 500K, or 95M). So what data file is Oracle complaining about?
    In addition, I have seen the following message in the alert_project.log:
    ========================================
    Assigning activation ID 2846189359 (0xa9a5672f)
    Thread 1 opened at log sequence 1
    Current log# 1 seq# 1 mem# 0: D:\DATA\ORACLE\ORADATA\BURKE\LOGS\LOGFILE1.LOG
    Current log# 1 seq# 1 mem# 1: D:\ORACLE\ORADATA\BURKE\LOGS\LOGFILE1.LOG
    Successful open of redo thread 1.
    Tue Sep 09 11:42:39 2003
    SMON: enabling cache recovery
    Tue Sep 09 11:42:39 2003
    create tablespace SYSTEM datafile 'd:\data\Oracle\oradata\burke\system01.dbf' size 500K REUSE
              autoextend on next 500K maxsize 200M,
    'd:\data\Oracle\oradata\burke\system02.dbf' size 500k REUSE
              autoextend on next 500K maxsize 200M
    default storage (initial 10K next 10K) EXTENT MANAGEMENT DICTIONARY online
    Tue Sep 09 11:42:39 2003
    Completed: create tablespace SYSTEM datafile 'd:\data\Oracle
    Tue Sep 09 11:42:39 2003
    create rollback segment SYSTEM tablespace SYSTEM
    storage (initial 50K next 50K)
    Completed: create rollback segment SYSTEM tablespace SYSTEM
    Tue Sep 09 11:42:40 2003
    Errors in file d:\data\oracle\admin\burke\udump\project_ora_1324.trc:
    ORA-00604: error occurred at recursive SQL level 1
    ORA-01569: data file too small for system dictionary tables
    ========================================
    The line that says
    default storage (initial 10K next 10K) EXTENT MANAGEMENT DICTIONARY online
    is not in my create database statement. Can it be a portion of my problem?
    I don't see it in the 9.2 SQL manual for create database.
    I appreciate the assistance that has been provided, but I still need some more.
    Thanks you all again.

  • Best approach to archival of large databases?

    I have a large database (~300 gig) and have a data/document retention requirement that requires me to take a backup of the database once every six months to be retained for 5 years. Other backups only have to be retained as long as operationally necessary, but twice a year, I need these "reference" backups to be available, should we need to restore the data for some reason - usually historical research for data that extends beyond what's currently in the database.
    What is the best approach for making these backups? My initial response would be to do a full export of the database, as this frees me from any dependencies on software versions, etc. However, an export takes a VERY long time. I can manage it by doing concurrent multiple exports by tablespace - this is able to be completed in < 1 day. Or I can back up the software directory + the database files in a cold backup.
    Or is RMAN well-suited for this? So far, I've only used RMAN for my operational-type backups - for short-term data recovery needs.
    What are other people doing?

    Thanks for your input. How would I do this? My largest table is in monthly partitions each in its own tablespace. Would the process have to be something like: alter table exchange partition-to-be-rolled-off with non-partitioned-table; then export that tablespace?

Maybe you are looking for