Sql server 2005 log file shrinking issue.

hi,
one of my production server database  log file size is increasing continuously,we can take log backup but no use,
we can change the database recovery model from  full to simple we shrink the log file, log file shrink successfully but space is not released,and we are getting the alerts continuously your physical disk is 95% like that.
here what we can do how reduced the log file size and how to control the log file size.
please give me the answer  ASAP.
Thanks,

Hello,
In SQL Server, log truncation occurs automatically as follows:
• Under the simple recovery model, after a checkpoint.
• Under the full recovery model or bulk-logged recovery model, after a log backup, if a checkpoint has occurred since the previous backup.
As Dan posted earlier, when we run the above code, we will get an integer value. The log_reuse_wait_desc value tells us why the transaction log space is not reused. You can refer to the article below to check the root cause. 
http://msdn.microsoft.com/en-us/library/ms345414.aspx
Regards,
Alisa Tang
Alisa Tang
TechNet Community Support

Similar Messages

  • Migrating From Sql Server 2005 to Oracle (Uniqueidentifier Issue)

    If i want to migrate from SQL SERVER 2005 to Oracle, How Oracle will deal with my primary key columns in almost all tables having datatype of UNIQUEIDENTIFIER ???
    is there any concept of uniqueidentifier in Oracle same as in SQL SERVER ????
    Regards

    When you migrate your SP's from SQL Server to Oracle the parser basically does 'as good a job as it can' to migrate them.
    Luckily, using SQL Developer you are able to look at the source SP and converted SP in windows next to each other and see what oracle has done to them.
    You can also paste your SQL Server code into a window and ask SQL Developer to translate it to oracle for you. Use the 'Translation Scratch Editor' on the migration menu if this is something that interests you. I was told that the parser used by the Translation Scratch Editor is different to the parser used when migrating the whole database.

  • Help! SQL server database log file increasing enormously

    I have 5 SSIS jobs running in sql server job agent and some of them are pulling transactional data into our database over the interval of 4 hours frequently. The problem is log file of our database is growing rapidly which means in a day, it eats up 160GB of
    disk space. Since our requirement dont need In-point recovery, so I set the recovery model to SIMPLE, eventhough I set it to SIMPLE, the log
    data consumes more than 160GB in a day. Because of disk full, the scheduled jobs getting failed often.Temporarily I am doing DETACH approach
    to cleanup the log.
    FYI: All the SSIS packages in the job is using Transaction on
    some tasks. for eg. Sequence Cointainer
    I want a permanent solution to keep log file in a particular memory limit and as I said earlier I dont want my log data for future In-Point recovery, so no need to take log backup at all.
    And one more problem is that in our database,the transactional table has 10 million records in it and some master tables have over 1000 records on them but our mdf file
    size is about 50 GB now.I dont believe that this 10 million records should make it to 50GB memory consumption.Whats the problem here?
    Help me on these issues. Thanks in advance.

    And one more problem is that in our database,the transactional table has 10 million records in it and some master tables have over 1000 records on them but our mdf file
    size is about 50 GB now.I dont believe that this 10 million records should make it to 50GB memory consumption.Whats the problem here?
    Help me on these issues.
    For SSIS part of question it would be better if you ask in SSIS forum although noting is going to change about logging behavior. You can increase some space on log file and also should batch your transactions as already suggested
    Regarding memory question about SQL Server, once it utilizes memory is not going to release unless there is windows OS faces memory pressure and SQLOS asks SQL Server to trim down its memory consumption. So again if you have set max server memory to some
    where near 50 SQL Server will utilize that much memory eventually. So what you are seeing is totally normal. Remember its costtly task for SQL Server to release and take memory so it avoids it by caching as much as possible plus it caches more so as to avoid
    physical reads which are costly
    When log file is getting full what does below transaction return
    select log_reuse_wait_desc from sys.databases where name='db_name'
    Can you manually introduce chekpoint in ETL query. Try this it might help you
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • SQL Server 2008 .MDF File attaching issue

    Hy all guys here is big Happy news about the SQL Server 2008 database.mdf file attaching issue on Windows 8.1 and SQL Server 2008
    I found the solution like this
    first you go to the directory of your computer and go to that folder where your database and other files are like
    folder having database and files then > Right click and > security > then give full rights from which user you LOGIN and then apply > ok after that go to SQL Server 2008 > Right Click > run as administrator
    Hurray your problem will be resolve I Resolved my issue too  thanks to providing every solution
    if you guys find solution please give me best regards thanks  

    Hi Farhan-Islam,
    Glad to hear that your issue had been solved by yourself. Thank you for your sharing which will help other forum members who have the similar issue.
    Regards,
    Lydia Zhang

  • SQL Server 2005 High Availability and Disaster Recovery options

    Hi, We are are working on a High Availability & Disaster Recovery Planning solution for an application database which is on SQL Server 2005. What different options have we got to implement this for SQL Server 2005 and after we have everything setup how
    do we test the failover is working?
    Thanks in advance.........
    Ione

    DR : Disaster recovery is the best option for the business to minimize their data loss and downtime. The SQL server has a number of native options. But, everything is depends upon your recovery time objective RTO and recovery point objective RPO.
    1. Data center disaster
    Geo Clustering
    2. Server(Host)/Drive (Except shared drive) disaster
    Clustering
    3. Database/Drive disaster     
    Database mirroring
    Log shipping
    Replication
    Log shipping
    Log shipping is the process of automating the full database backup and transaction log on a production server and then automatically restores them on to the secondary (standby) server.
    Log shipping will work either Full or Bulk logged recovery model.
    You can also configure log shipping in the single SQL instance.
    The Stand by database can be either restoring or read only (standby).
    The manual fail over is required to bring the database online.
    Some data can be lost (15 minutes).
    Peer-to-Peer Transactional Replication
    Peer-to-peer transactional replication is designed for applications that might read or might modify the data in any database that participates in replication. Additionally, if any servers that host the databases are unavailable, you can modify the application
    to route traffic to the remaining servers. The remaining servers contain same copies of the data.
    Clustering
    Clustering is a combination of one or more servers it will automatically allow one physical server to take over the tasks of another physical server that has failed. Its not a real disaster recovery solution because if the shared drive unavailable we cannot
    bring the database to online.
    Clustering is best option it provides a minimum downtime (like 5 minutes) and data loss in case any data center (Geo) or server failure.
    Clustering needs extra hardware/server and it’s more expensive.
    Database mirroring
    Database mirroring introduced in 2005 onwards. Database Mirroring maintain an exact copy of a database on a different server. It has automatic fail over option and mainly helps to increase the database availability too.
    Database mirroring only works FULL recovery model.
    This needs two instances.
    Mirror database always in restoring state.
    http://msdn.microsoft.com/en-us/library/ms151196%28v=sql.90%29.aspx
    http://blogs.technet.com/b/wbaer/archive/2008/04/19/high-availability-and-disaster-recovery-with-microsoft-sql-server-2005-database-mirroring-and-microsoft-sql-server-2005-log-shipping-for-microsoft-sharepoint-products-and-technologies.aspx
    http://www.slideshare.net/rajib_kundu/disaster-recovery-in-sql-server
    HADR Considerations
    Need to Understand the business motivations and regulatory requirements that are driving the customer's HA/DR requirements. Understand how your customer categorizes the workload from an HA/DR perspective. There is likely to be an alignment between the needs
    and categorization.
    Check for both the recovery time objective (RTO) and the recovery point objective (RPO) for different workload categories, for both a failure within a data center (local high availability) and a total data center failure (disaster recovery). While RPO and
    RTO vary for different workloads because of business, cost, or technological considerations, customers may prefer a single technical solution for ease in operations. However, a single technical solution may require trade-offs that need to be discussed with
    customers so that their expectations are set appropriately.
    Check and understand if there is an organizational preference for a particular HA/DR technology. Customers may have a preference because of previous experiences, established operational procedures, or simply the desire for uniformity across databases from
    different vendors. Understand the motives behind a preference: A customers' preference for HA/DR may not be because of the functions and features of the HA/DR technology. For example, a customer may decide to adopt a third-party solution for DR to maintain
    a single operational procedure. For this reason, using HA/DR technology provided by a SAN vendor (such as EMC SRDF) is a popular approach.
    To design and adopt an HA/DR solution it is also important to understand the implications of applying maintenance to both hardware and software (including Windows security patching). Database mirroring is often adopted to minimize the service disruption
    to achieve this objective.
    HADR Options :
    Failover clustering for HA and database mirroring for DR.
    Synchronous database mirroring for HA/DR and log shipping for additional DR.
    Geo-cluster for HA/DR and log shipping for additional DR.
    Failover clustering for HA and storage area network (SAN)-based replication for DR.
    Peer-to-peer replication for HA and DR (and reporting).
    Backup & Restore ( DR)
    keep your server DB backups in network location ( DR)
    Always keep your sql server 2005 upto date, in case if you are not getting any official support from MS then you have to take care of any critical issues and more..
    Raju Rasagounder Sr MSSQL DBA

  • I plan to upgrade the SSRS 2005 SP2 enterprise edition in our internal web server to 2005 SP4, the Reporting server database is hosted in another sql server in sql server 2005 SP4. Do I need to do anything on the reporting server database side?

    My question is what the steps do I need to take to upgrade SSRS from 2005 SP2 to SP4.  The web server that host the SSRS is in 2005 SP2, and the OS is in window 2003. 
    Our SSRS report server and report server database are in different servers.  The SSRS in the web server is in 2005 SP2 enterprise edition, the report server database is in sql server 2005 SP4 enterprise edition.
    To upgrade the SSRS in web server from 2005 sp2 to sp4, do I need to backup/restore the encryption key?  Nothing will be changed in the report server database.  We will still pointing to the same database in the current server, all
    I wanted to do is performing a inplace upgrade of SSRS from 2005 SP2 to SP4.  
    Any response will be greate appreciated.  Thank you!
    Li-hui Chen

    Hi Lihui Chen,
    According to your description, you want to install the Services Pack 4 for SQL Server. Right?
    In SQL Server, Services Packs are used for fixing issues of current version product. It's not an Upgrade, you don't have to backup/restore your encryption key. You just need to download the Service Pack 4 on:
    Microsoft SQL Server 2005 Service Pack 4 RTM  . Please make sure you have administrative rights on the computer to install SQL Server 2005 SP4. For more information, see links below:
    How to obtain the latest service pack for SQL Server 2005
    List of the issues that are fixed in SQL Server 2005 Service Pack 4
    SQL Server 2005 SP4, KBA 2463332, Installation Issues
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

  • Log File Issue In SQL server 2005 standard Edition

    We have database of size 375GB .The data file has 80 GB free space within .When trying to rebuild the index we had 450 GB free space on the disk where Log file is residing.The rebuild index activity failed due to space issue.added more space and got the
    job done successfully
    The Log file has grow up to 611GB to complete the rebuild index.
    version :SQL server 2005 Standard Edition .Is ther a way to estimate the space required for rebuild index in this version.
    I am aware we notmaly allocate 1.5 times of data file.But in this case It was totaly wrong.
    Any suggestion with examples would be appreciated.
    Raghu

    OK, there's a few things here.
    Can you outline for everybody the recovery model you are using, the frequency with which you take full, differential and transaction log backups.
    Are you selectively rebuilding your indexes or are you rebuilding everything?
    How often are you doing this? Do you need to?
    There are some great resources on automated index maintenance, check out
    this post by Kendra Little.
    Depending on your recovery point objectives I would expect a production database to be in the full recovery mode and as part of this you need to be taking regular log backups otherwise your log file will just continue to grow. By taking a log backup it will
    clear out information from inactive VLF's and therefore allow SQL Server to write back to those VLF's rather than having to grow the log file. This is a simplified version of events, there are caveats.
    A VLF will be marked as active if it still has an open transaction in it or there is a HA option that still requires that data to be available as that data has not been copied to another node yet.
    Most customers that I see take transaction log backups every 15 - 30 minutes, but this really does depend upon how much data your company can afford to lose. That's another discussion for another day.
    Make sure that you take a transaction log backup prior to your job that does your index rebuilds (hopefully a smart job not a sledge hammer job).
    As mentioned previously swapping to bulk logged can help to reduce the size of the amount of information logged during index rebuilds. If you do this make sure to swap back into the full recovery model straight after and perform a full backup. There are
    problems with the ability to do point in time restores whilst in the bulk logged recovery model, so you need to reduce the amount of time you use it.
    Really you also need to look at how your indexes are created does the design of them lead to them being fragmented on a regular basis? Are they being used? Are there better indexes out there that can help performance?
    Hopefully that should put you on the right track.
    If you find this helpful, please mark the post as helpful,
    If you think this solves the problem, please propose or mark it an an answer.
    Please provide details on your SQL Server environment such as version and edition, also DDL statements for tables when posting T-SQL issues
    Richard Douglas
    My Blog: Http://SQL.RichardDouglas.co.uk
    Twitter: @SQLRich

  • Issue with migration DB from SQL Server 2005 to Oracle 11.2 using SQL Developer Migration workbench

    Hi,
    We face an issue while migrating an SQL Server 2005 DB to Oracle 11.2.  It fails during the process.  I hope someone on the forum has seen this before and can give us some advice.
    I use the latest version of SQL Developer with JRE included (3.2.20.09).  The JDBC driver to connect to SQL Server 2005 is 1.2.0
    Here are the steps we take in the Migration Workbench wizard:
    I made an online extract of the SQL Server database.
    Using the workbench in SQL Developer I created a migration repository in schema PDM_MIGRATION.
    Next I start the migration, it captures the tables fine, but immediately after the start of the conversion I get a failed message without further explanations.
    This is the content of the error xml:
    <?xml version="1.0" encoding="windows-1252" standalone="no"?>
    <log>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>SEVERE</level>
      <class>oracle.dbtools.migration.workbench.core.logging.MigrationLogUtil</class>
      <message>Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@4c12ab</param>
      <exception>
        <message>oracle.dbtools.migration.convert.ConvertException: Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>1078</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>316</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>1002</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>303</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>205</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>159</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTask</class>
          <line>193</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask</class>
          <line>515</line>
        </frame>
        <frame>
          <class>java.util.concurrent.Executors$RunnableAdapter</class>
          <line>441</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>886</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>908</line>
        </frame>
        <frame>
          <class>java.lang.Thread</class>
          <line>662</line>
        </frame>
      </exception>
    </record>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>SEVERE</level>
      <class>oracle.dbtools.migration.workbench.core.logging.MigrationLogUtil</class>
      <message>Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
      <param>oracle.dbtools.migration.convert.ConverterWorker.copyModel(ConverterWorker.java:1078)</param>
      <param>oracle.dbtools.migration.convert.ConverterWorker.runConvert(ConverterWorker.java:316)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doConvert(FullMigrateTask.java:1002)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doMaskBasedActions(FullMigrateTask.java:303)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doWork(FullMigrateTask.java:205)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doWork(FullMigrateTask.java:159)</param>
      <param>oracle.dbtools.raptor.backgroundTask.RaptorTask.call(RaptorTask.java:193)</param>
      <param>java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)</param>
      <param>java.util.concurrent.FutureTask.run(FutureTask.java:138)</param>
      <param>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask.run(RaptorTaskManager.java:515)</param>
      <param>java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)</param>
      <param>java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)</param>
      <param>java.util.concurrent.FutureTask.run(FutureTask.java:138)</param>
      <param>java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)</param>
      <param>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)</param>
      <param>java.lang.Thread.run(Thread.java:662)</param>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@5dc1bc</param>
      <exception>
        <message>oracle.dbtools.migration.convert.ConvertException: Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>1078</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>316</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>1002</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>303</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>205</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>159</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTask</class>
          <line>193</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask</class>
          <line>515</line>
        </frame>
        <frame>
          <class>java.util.concurrent.Executors$RunnableAdapter</class>
          <line>441</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>886</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>908</line>
        </frame>
        <frame>
          <class>java.lang.Thread</class>
          <line>662</line>
        </frame>
      </exception>
    </record>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>WARNING</level>
      <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
      <message>Building converted model: FAILED : Database Migration : FAILED</message>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@15a3779</param>
    </record>
    Does anybody know what this error means and what steps we should take to continue the migration?
    I see the PDM_MIGRATION.MIGR_FILTER is a type.
    Many thanks in advance,
    Kris

    Hi Wolfgang,
    Thanks for your reply.
    This is how the type MIGR_FILTER looks like:
    create or replace
    TYPE MIGR_FILTER IS OBJECT (
      FILTER_TYPE INTEGER, -- Filter Types are 0-> ALL, 1->NAMELIST, 2->WHERE CLAUSE, 3->OBJECTID LIST
      OBJTYPE VARCHAR2(40),
      OBJECTIDS OBJECTIDLIST,
      NAMES NAMELIST,
      WHERECLAUSE VARCHAR2(1000));
    I think the repository user has the correct privileges.  This is the overview of privileges it has:
    SQL> select * from dba_sys_privs where GRANTEE in ('PDM_MIGRATION') order by GRANTEE;
    GRANTEE PRIVILEGE ADM
    PDM_MIGRATION ALTER SESSION NO
    PDM_MIGRATION CREATE CLUSTER NO
    PDM_MIGRATION CREATE DATABASE LINK NO
    PDM_MIGRATION CREATE PROCEDURE NO
    PDM_MIGRATION CREATE SEQUENCE NO
    PDM_MIGRATION CREATE SESSION NO
    PDM_MIGRATION CREATE SYNONYM NO
    PDM_MIGRATION CREATE TABLE NO
    PDM_MIGRATION CREATE TRIGGER NO
    PDM_MIGRATION CREATE VIEW NO
    PDM_MIGRATION UNLIMITED TABLESPACE NO
    SQL> select * from dba_role_privs where GRANTEE in ('PDM_MIGRATION') order by GRANTEE;
    GRANTEE GRANTED_ROLE ADM DEF
    PDM_MIGRATION CONNECT NO  YES
    PDM_MIGRATION RESOURCE NO  YES
    Best regards,
    Kris

  • Oracle provider Issue in SQL Server 2005

    Hi all,
    On the dev server, I'm using SSAS 2005 (32-bytes) + Oracle provider for OLE DB to access my Oracle Database and it works perfectly.
    On the production server, SSAS 2005 is running on a 64-bytes platform.
    The Oracle client 10.2.0.1.0 (64 bits) has been installed containing the Oracle provider for OLE DB but unfortunately the Oracle provider for OLE DB does not appear in the list in BIDS when trying to create a date source.
    Could someone help ?
    Regards,
    Juan

    After much grief I figured this out and this works for me going from x64 sql 05 on windows server 2003 to x86 oracle 10g on another windows server 2003 box...
    Software Prerequisites
    •     x64 Oracle Client:
    http://www.oracle.com/technology/software/products/database/oracle10g/htdocs/10201winx64soft.html
    •     x64 Oracle Data Access Components (ODAC) http://www.oracle.com/technology/software/tech/windows/odpnet/64-bit/index.html
    •     x86 Oracle Client:
    http://www.oracle.com/technology/software/products/database/oracle10g/htdocs/10201winsoft.html
    •     SQL Server 2005 SP2 (x64):
    http://www.microsoft.com/downloads/details.aspx?FamilyID=D07219B2-1E23-49C8-8F0C-63FA18F26D3A&displaylang=en
    Patch Level Required (As of Thursday, July 31, 2008):
    •     Microsoft SQL Server Management Studio: 9.00.3042.00
    •     Microsoft Analysis Services Client Tools: 2005.090.3042.00
    •     Microsoft Data Access Components (MDAC): 2000.086.3959.00 (srv03_sp2_rtm.070216-1710)
    •     Microsoft MSXML: 2.6 3.0 4.0 6.0
    •     Microsoft Internet Explorer: 7.0.5730.13
    •     Microsoft .NET Framework: 2.0.50727.42
    •     Operating System: 5.2.3790
    •     Oracle Data Access Components (ODAC): 10.2.0.3.00 (64-bit Oracle10g Release 2 ODAC for Windows x64)
    1. install the x64 and x86 oracle clients on the mssql box
    2. install the odac on the mssql
    3. Verify that the registry settings at
    HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\MSDTC\MTxOCI and
    HKEY_LOCAL_MACHINE\SOFTWARE\Wow6432Node\Microsoft\MSDTC\MTxOCI
    are correct:
    a.     OracleOciLib = oci.dll
    b.     OracleSqlLib = orasql10.dll
    c.     OracleXaLib = oraclient10.dll
    4. Verify that your tnsnames.ora and sqlnet.ora files have the correct info to connect to the source Oracle 10g (x86) 32 bit source. Note if there is a DNS or other name resolution problem in the environment you may need to change your host entry in your tnsnames file to be the actual I.P. address of the source Oracle box, until the network issue is resolved by the applicable environments network engineer. Additionally, note that if your client tnsnames.ora file is not configured to match the network topology in use, you will have issues… loopback configurations particularly cause grief.
    5. Update Machine.config Add the following binding policy:
    <runtime>
    <assemblyBinding xmlns="urn:schemas-microsoft-com:asm.v1">
    <dependentAssembly>
    <assemblyIdentity name="Oracle.DataAccess" publicKeyToken="89b483f429c47342" />
    <bindingRedirect oldVersion="2.0.0.0-10.9.9.9" newVersion="2.102.3.2" />
    </dependentAssembly>
    </assemblyBinding>
    </runtime>
    6. Save the following as a *.reg file and run it...
    Windows Registry Editor Version 5.00
    [HKEY_LOCAL_MACHINE\SOFTWARE\Wow6432Node\Microsoft\MSDTS\Setup\WizardPath]
    @="C:\\Progra~2\\Microsoft SQL Server\\90\\DTS\\Binn\\DTSWizard.exe"
    7. Save the following as a *.bat file...
    rem Progra~2 is short path name for "Program Files (x86)" and works around an Oracle client bug that doesn't like the ()'s in the path
    start /B "C:\Progra~2\Microsoft Visual Studio 8\Common7\IDE" "C:\Progra~2\Microsoft Visual Studio 8\Common7\IDE\devenv.exe"
    8. Save the following as a *.bat file...
    start /B "C:\Progra~2\Microsoft SQL Server\90\Tools\binn\VSShell\Common7\IDE" "C:\Progra~2\Microsoft SQL Server\90\Tools\binn\VSShell\Common7\IDE\SqlWb.exe"
    9. Create a shortcut that points to each of the batch files you just created and only launch BIDS or SSMS from those *.bat files...
    10. create your packages, etc., as normal, but you will not be able to fully test them from BIDS or SSMS because they are 32 bit running in 64 bit o/s...
    11. Create a job with a line similar to the following as a cmdexec type of job...
    "C:\Progra~2\Microsoft SQL Server\90\DTS\Binn\DTExec.exe" /FILE "C:\SomeWidget.dtsx" /MAXCONCURRENT " -1 " /CHECKPOINTING OFF
    12. your done...
    Best Regards, from a Las Vegas Bohemian...

  • Error in adding PRIMARY Data file of SQL Server 2005

    Hi Experts,
    I am using SQL 2005 as back end database for SAP ECC6. Due to Space problem in PRD data volume, i have added PRDDATA4.ndf file in SSMS ( SQL Server Management Studio) and saved. But, suddenly i got one more idea and extended the file size of exisiting PRDDATA1.mdf, PRDDATA2.ndf & PRDDATA3.ndf immediately i removed the PRDDATA4.ndf file before starting the service. But, while seeing in SAP using ST04 it showing the PRDDATA4.ndf with error message of file missing. What i have to do now? How can i delete the information showing in ST04. But it is not showing in SSMS.
    Please help how to solve this.
    Regards,
    B.Sudharsan

    Hi,
    Following solution might help you to resolve the issue.
    Solution1:
    Issue happens because universe connection to the Database is not responding and the Data Source Name is not configured to write SQL Server 2005 database.
    Resolution
    Created new DSN connection under SYSTEM DSN to SQL Server 2005 Database.
    Import the universe from the designer.
    Go to File->Parameters.
    Edit Connection.
    On the Database Middleware Page expand SQl Server 2005->Select SQL Server 2005 driver.
    Change the DSN to new DSN connection
    Test the connection.
    Save the Universe and Export it to the repository.
    Solution2:
    I am taking Oracle as an example over here.
    This can be caused when the SQLNET.ora is located in a remote location using mapped drives.
    Resolution
    To resolve,
    check the registry [HKLM\SOFTWARE\ORACLE] you should see a folder called HOME0
    look for any paths that contain a mapped drive partition
    Change the mapped drive letter to a UNC path (
    server\folder)
    Open a command prompt and do a tnsping using the tnsname and verify the location of the SQLNET.ora or TNSNAMES.ora
    If the path is not updated to the new path
    Reboot the server
    Regards,
    Sarbhjeet Kaur

  • OraOledb for 64-bit, Linked Servers and SQL Server 2005 issues

    Our environment is : SQL Server 2005, Windows Server 2003, 64-bit and 32-bit operating systems.
    Problem on 64-bit operating system box: (32-bit works fine).
    I am trying to access Oracle 10g database using linked server from our SQL Server 2005. In case of number fileds i got the following error:
    Invalid data for type "numeric".
    After going through one of the posting in this forumn i was able to resolve the problem by converting those column values to char while querying and then converting them back to numeric type on SQL server side.
    But today i ran into another problem. There is a VARCHAR2 column. I was able to retrieve the data yesterday for that column but today i am getting a blank recordset. If i exclude the colum from the query then i am getting all the rows.
    I am querying against a view and it has got a number of columns whose data type is VARCHAR2.
    Again the problem is on 64-bit operating system only. We have a 32-bit operating system on which i am able to retrieve the data including this column. I looked at the data and everything looks OK. No funny characters etc.
    I tried workarounds like using cast, to_char, checking for nulls etc., Nothing works.
    Any help is greately appreciated. Thanks.

    Did you find a resolution for this? We have similar problem. Set up a linked server in SQL 2006 to Oracle (running on Windows 64-bit) Linked server works and views I had set up were working but they added some new data in the Oracle test database I am using and now I get errors on one of the views.
    Error I am getting on the view is "Cannot initialize the data source object of OLE DB Provider "OraOLDEDB.ORacle" for the linked server"
    If I fine tune my queries to find the specific table or view that is at issue, then I get the error "inconsistent metadata for a column"

  • OraOledb, Linked Servers and SQL Server 2005 issues

    Some issues I've come across when using SQL Server 2005 (and SQL Server Express 2005), the OraOLEDB provider (10.2.0.1) and a linked Oracle database (8.1.7.4.0 64-bit)
    1) You must set the OraOledb.Oracle\AllowInProcess value to 1 to allow the OraOledb provider to run in SQL Server's process. Without doing this I receive an 'unspecified error' from the OLE DB provider when attempting to run a query
    2) When running a ' select * ' query across a linked server using the provider, I receive the following error: Msg 9803, Level 16, State 1, Line 1
    Invalid data for type "numeric". I can, however, select all of the columns by name and the query completes (no error). Sometimes the 'select *' query returns a few rows before the error, sometimes it doesn't. The Microsoft Provider for Oracle does not have this problem

    Okay... I've got a Win2K3 Std Ed server (x64) running 64-bit SQL Server 2005 Enterprise Edition. I've installed the Oracle 10g 10.2.1 full client and admin tools, added two named services via the NetConfig assistant, and successfully set up (and tested) a connection via the ODBC Administrator to an Oracle database.
    Now... when I try to create a new connection manager in SQL Server 2005 Integration Services, the OLEDB provider for Oracle can't be found, and when I try to manually add an underlying OLEDB connection to the database, SQL Server reports:
    Test connection failed because of an error in initializing provider. The 'OraOLEDB.Oracle.1' provider is not registered on the local machine.
    Does anyone know what I need to do to see my ODBC Server data connections in SQL Server 2005 (64 bit)? I don't have this issue on my 32-bit SQL Server 2005 servers.

  • Storing the file in to the sql server 2005 table

    hi,every body...
    in my appliaction i need to store the files.
    i have searched the net but only i found as BLOB(Binary L:arge OBject) is the solution,
    how to use this blob...technique..
    can any one send some sample of code to explain.......
    how to insert a file and how to retrive the inserted file from the SQL SERVER 2005...
    thanks in advance..

    yes, ur rite thse two methods are available..
    but i didnt understand that..the method setBinaryStream() accepts 3 parameters
    the last parameter which i know that.. size of the file of int type rite?
    can u give some more information regarding this..
    thanks ..

  • Issue in rebuilding index after restoring database from sql server 2005 to 2012

    Hi All,
    We have recently restored a database named DBA_admin from our sql server 2005 instance to sql server 2012 and we have added the database to out weekly maintenance plan. After adding the database to maintenance plan the job is failing every weekend due to
    below error message.
    Error message from maintenance plan history:
    Executing the query "ALTER INDEX [PK_lsttype] ON [dbo].[lstAssign..." failed with the following error: "Invalid object name 'admin.dbo.databaselog'.
      ALTER_INDEX - dbo.PK_lsttype". Possible failure reasons: Problems with the query, "ResultSet" property not set correctly, parameters not set correctly, or connection not established correctly.
    Below is the code within the maintenance plan which the rebuild index task is using for database dba_admin:
    USE [dba_admin]
    GO
    ALTER INDEX [PK_lstType] ON [dbo].[lstType] REBUILD PARTITION = ALL WITH (PAD_INDEX = OFF, STATISTICS_NORECOMPUTE = OFF, SORT_IN_TEMPDB = OFF, ONLINE = OFF, ALLOW_ROW_LOCKS = ON, ALLOW_PAGE_LOCKS = ON, FILLFACTOR = 98, DATA_COMPRESSION = PAGE)
    Please let me know if any one knows how to fix this issue.
    Thanks in Advance.

    Sorry for the confusion.
    Here is what is happening.
    The query is failing while trying to rebuild index [Pk_lsttype] on table [lstAssign], both these objects belong to database
    admin.
    I have verified the old instance there is table databaselog within the database
    dba_admin, but in new SQL instance within the database dba_admin the table is not present.
    Even I'm not sure of what this table databaselog has to do with rebuilding indexes on table
    lstAssign within the admin database.
    I have manually tried to execute the alter index command within the database
    admin and it is failing with the below error.
    ALTER_TABLE - dbo.Query_pnb_sms
    Msg 208, Level 16, State 1, Procedure ddlDatabaseTriggerLog, Line 39
    Invalid object name 'dba_admin.dbo.databaselog'.

  • Connect to SQL Server 2005 Express MDF file

    Post Author: ims_user
    CA Forum: Data Connectivity and SQL
    Is it possible to connect directly to a SQL Server 2005 Express MDF file? Every time I select SQL Native Client it expects me to put in a server name etc, but there is no where to put in a file name or connection string. My application connects to this file directly using this in the connection string
    Data Source=.\SQLEXPRESS;AttachDbFilename=
    I now need to report on the data it contains.
    Many thanks.

    Post Author: ims_user
    CA Forum: Data Connectivity and SQL
    Is it possible to connect directly to a SQL Server 2005 Express MDF file? Every time I select SQL Native Client it expects me to put in a server name etc, but there is no where to put in a file name or connection string. My application connects to this file directly using this in the connection string
    Data Source=.\SQLEXPRESS;AttachDbFilename=
    I now need to report on the data it contains.
    Many thanks.

Maybe you are looking for