Data Mover login error (SQL 2008R2 as backend DB)

Hi,
This question was asked in this thread -> Continued Discussion - Steps to Upgrade PT8.49 to 8.50 Manually
The difference between problem in that thread and mine is that I have WIN08R2 SQL08R2 Peopletools9.52 and Peoplesoft HRMS 9.0 installed. I was on the steps of creating the databases manually with the sql scripts and that was also preformed.
I also have different ConnectID: people and AccessID: peoplesa with respective passwords. I try login in with the AccessID and I get the below error.
File: SQL Access ManagerSQL error. Stmt #: 2 Error Position: 0 Return: 8077 - Invalid SNAC client version (minimum is 10.00.1763) (SQLSTATE PS077)
Any suggestions?
-Vishal

Hi there,
I upgraded the SNAC client to 1033\x64\sqlncli.msi from the Microsoft and still I get the same Login error in the DataMover.
I checked the registry location --> HKEY_Local_Machine\Software\Microsoft\Microsoft SQL Server Native Client 10.0\Current Version = 10.51.2500.0
The Data Mover login error says that it requires minimum is 10.00.1763 so i guess the one I have should work fine?
Any suggestions?
-Vishal

Similar Messages

  • Login Error SQL Plus in Oracle9i

    I installed completely Oracle9i on my system. I have to make databases, make forms and generate reports for my scool project. I cannot login to sql Plus using scott tiger . I think the problem is with HOST STRING. Can anyone help please.
    Thank you
    Sharif Vohra

    it would help if you post the error message(s), if any, that you are receiving while connecting.
    Did you install the database part (or the client tools only), and did you also create a
    starter (or customized) database while installing Oracle?

  • Sql Developer MSSql Migration, Online data move, image to long raw mapping

    Hello,
    my task is to migrate a MSSql Server Database to Oracle 10.2.0.4 with SQL Developer Version 2.1.1.64
    and JTDS JDBC Driver oracle.sqldeveloper.thirdparty.drivers.sqlserver 11.1.1.58.17.
    Capture, convert and online data move works except data mapping from MSSql image to Oracle long raw dataype.
    I need to map MsSql image to oracle blob, to achieve online data move!
    is Sql Developer online migration MsSql image to oracle long raw possible?
    I know that long raw is outdated, but i'm dba and not responsible for the software part.
    What options should i try? offline data move? can MS Bulk Copy Program export image data and SQL*Loader import into long raw?
    Any experience or tipp?
    Thanks
    Michael

    Hi user12132314 ,
    I do not see any 'long raw' options on our current SQLServer 2005 data type mapping. (Raw is a separate 2000 byte data type).
    The usual route for SQLDeveloper is to go to Blob.
    Online - this is simple & automated.
    Offline - this actually goes via hex output in bcp to sqlldr, read in as CLOB which is then encoded to blob. (This is automated)
    (Online 'Copy to Oracle' option, right click on table, may be of use if advanced features not included in the select such as defaults are not required, it copies over a select * from table, including blobs)
    To move you from blob to long raw after migration can be done (easily enough if Oracle database >=10g so the stream does not have to be in smaller 'chunks')
    http://asktom.oracle.com/pls/asktom/f?p=100:11:3938270166267830::::P11_QUESTION_ID:702825000306
    I can look into this if you want to go down this route.
    Note however that long raw has drawbacks and blob is preferred, see
    Oracle® Database SQL Language Reference
    11g Release 2 (11.2)
    Part Number E17118-04
    Data Types
    LONG Data Type
    for example one restriction is:
    A table can contain only one LONG column.
    -Turloch
    SQLDeveloper Team

  • Problems Generating Data Move Scripts in v1.5.1

    Hi, I'm having problems when trying to generate data move scripts in SQL Dev 1.5.1 to carry out an off-line data load. I'm carrying out a migration from Sybase to Oracle and the database I'm working on has over 400 tables in it. I have successfully captured and migrated all the tables into the resp. models and have generated and created the DDL for the converted model. However, when I request the data move scripts to be generated I'm only getting ctl files created for the 1st 49 tables. Also, there is no oracle_ctl.sh script created. Also, no post_load.sql script is produced only a pre_load.sql script.
    I've got 3 databases to migrate and on the 2nd database I only get the data move scripts created for ths first 86 tables and there are over 250 tables in it.
    It appears to have worked better for the 3rd database which is much much smaller than the 1st two databases having only 59 tables in it. This time all the files were produced as expected. However, it's really the 1st two larger databases that are my priority to get migrated.
    I've tried changing the preferences within Migration/Generation Options to from 'One Single File' to 'A File per Object' but it makes no difference. I would prefer everything in one file but can work round that.
    Ideally, I'd like to generate all the ctl files for a database in one go so that I can group edit them and would prefer the tool to create the oracle_ctl.sh script to call all the ctl scripts for me rather than having to hand build it. I'm puzzled as to why the tool only creates ctl files for some of the tables contained within a converted model. it looks like it is not completing the job in these cases as it also doesn't create all the scripts that it is supposed to create either. It doesn't give out any error messages and the screen looks no different at completion to when it works successfully in the case of the very small database.
    Anybody had this problem or can suggest how to fix it ?
    Thanks all.

    Send me you phone number to [email protected]
    We'll help sort this out.
    Barry

  • Cannot login to App designer using PS id while able to login in data mover.

    Hi,
    I am installing peopletool 8.50 (HRMS). I ran all the sql scripts successfully. Ran the DMS script as well for demo database.
    When I am trying to login in App designer using PS/PS i'm getting the following error..
    File: SQL Access ManagerSQL error. Stmt #: 2 Error Position: 0 Return: 404 - ORA-01017: invalid username/password; logon denied
    while I am able to login in data mover in boot strap. Can anybody pls help me out?

    yes I am using Oracle 11g.
    and man.. thanks a lot for your help. that was really awesome post :-)
    People do say true.. here only I can find the GODs of Oracle.
    Thanks Again
    Gaurav
    Edited by: Gaurav_laconic on Mar 12, 2013 6:13 PM

  • Error: Could not continue scan with nolock due to data movement, DBCC proccache will clear the probelm

    SQL Server: 2008 R2 SP2
    Before describing my problem, I have gone via the forum, there is no view or functions inside my stored procedure
    When running a particular stored procedure inside crystal report, the error " Could not continue scan with nolock due to data movement" comes once every few weeks. After I clear the query cache plan, it works again for few weeks and the problem
    comes again. During these few weeks, there is no restart or query plan clearing.
    If I run the stored procedure inside SSMS, where the SQL statement is copied and pasted from SQL profiler during crystal report run, there is no error.
    I discovered running in SSMS and crystal report generate 2 different query plans even I copied the SQL from SQL profiler, I have actually saved the query plans. Unfortunately, this forum does not accept attachments, or otherwise I will post my query plans
    here.
    There is one thing I notice about the query plan is during nested loop operation, there is a warning "no join predicate". I don't use any views or UDF in the statement, nor did I use pre-1992 ANSI join syntax. However, I did use table variables.
    My guess is whether this will cause " Could not continue scan with nolock due to data movement", after I clear the cache, I run crystal report again, and I look at the plan again, the "nested loop no join predicate" warning is gone.
    Running this stored procedure took 1 second maximum, even when this error is popping up, it pop up within 1 second.
    DBCC checkdb has been run
    The same stored procedure running by crystal report in a SQL 2008 (non r2) live environment has no problems, so I am thinking this is R2 specific problems.
    The "nested loop no join predicate" error SQL statment is below, no views, no udf, but table variables
    INSERT @ChequeAccount
    SELECT        PS.PaySummaryID, PS.EmployeeID, PS.CostCentreID,
                (PS.GrossPay    + PS.LumpSumA + PS.LumpSumB    + PS.LumpSumD+ PS.LumpSumE+ PS.ETP+ PS.PaymentsAfterTax    - PS.DeductionsAfterTax  
     - PS.Tax- PS.ETPTax    + PS.TaxRebate) * -1 AS Amount,
                CGLM.GLAccountID
    FROM Pay_Summary PS JOIN Input_Sheet ISH ON PS.InputSheetID = ISH.InputSheetID  AND  ISH.PayrollID = @binPayrollID   
    AND PS.PaySummaryID NOT IN (SELECT PaySummaryID FROM @ChequeAccount)
    JOIN Payroll P ON P.PayrollID = ISH.PayrollID AND P.EmployerID = @binEmployerID
    JOIN CustomGLFixMapping CGLM ON CGLM.EmployerID = P.EmployerID AND CustomGLFixMappingNameID = 1 AND CGLM.CostCentreID IS NULL

    The error Could not continue scan with nolock due to data movement can occur when you use the NOLOCK table hint, or use the command SET TRANSACTION ISOLATION LEVEL READ UNCOMMITTED. That is, so-called dirty reads. The error is not related to the
    query plan per se, but when scanning a table, the storage engine will use an IAM scan rather than following the clustered index. If there is simultaneous activity, the storage engine may detect this and abort the operation to avoid returning incorrect data.
    Or it may not detect it, and return uncommitted data or fail to return committed data.
    All of these effects are transitory and they will not show up when you are alone on the system, only when there is concurrent activity in one or more of the tables in the query.
    Using dirty reads is a risky business for the reasons explained above, and it takes careful analysis to understand whether you can live with the errors you can get from a particular query. The error about data movement can be handled: trap the error and
    resubmit the query. But what about spurious incorrect results?
    If you believe locking to be a problem, you should consider setting the database to READ_COMMITTED_SNAPSHOT
    and take out all use of READ UNCOMMITTED/NOLOCK. When the database is in READ_COMMITTED_SNAPSHOT, readers read from the snapshot and only see committed data without blocking writers. This has some other effects like requiring a bigger tempdb,
    and there is a risk for other types of concurrency errors, but they tend to be smaller risks.
    I discovered running in SSMS and crystal report generate 2 different query plans even I copied the SQL from SQL profiler,
    This is because SSMS by default runs with SET ARITHABORT ON. I discuss this in more detail in this article on my web site:
    http://www.sommarskog.se/query-plan-mysteries.html
    However, as I said, this problem is not related to the query plan as such, although some query plans are more susceptible to this error than others. (All plans are suscpeitble to produce incorrect results).
    Erland Sommarskog, SQL Server MVP, [email protected]

  • SQL developer online data move

    Hi,
    I am using SQL developer 2.1 to migrate tables from Sybase 12 database to oracle 11g.
    I have used online data move option for moving sybase data into oracle tables, but even after data move is completed not all rows have been moved from sybase tables to corresponding oracle tables. Some rows are missing but still there is no error messege being displyed, how to find out what's going wrong.
    Is there any log file that is being generated in case of oracle sql developer?
    Please help me with this.

    Hi Mandark,
    This is a weakness of SQL Developers online datamove.
    It doesn't have logging and you cannot tweak the data move itself.
    It very basic SELECT * FROM -> INSERT INTO ... under the covers.
    If some rows are failing to be inserted,but some rows are being inserted, in the same table.
    Then my guess is that the data in some rows doesn't "fit" into the target row/column.
    And the datatype mapping used to convert the tables/columns was not good enough.
    Ex: Sybase column of VARCHAR(10000) mapped to Oracle VARCHAR2(4000).
    Some data which is greater than 4000 chars.
    Which causes the insert to fail for those rows with >4000 chars. But ok for other rows.
    The solution (in this example) is to map Sybase VARCHAR(100000) to Oracle CLOB.
    You can specify the data type mapping in the Convert page of the migration wizard.
    Another workaround is to use Copy To Oracle.
    http://dermotoneill.blogspot.com/2010/11/copy-to-oracle.html
    This is a very simple data move feature again. But it works slightly differently than Online Data Move.
    It has no logging or tweaking either, but may do the job "online" for you.
    You should temporarily turn off any check constraints, triggers, keys on each table you want to move data with using this method.
    If all those fail, then the only sure fire way would be to use Offline Data Move.
    Regards,
    Dermot
    SQL Developer Team.

  • Sql 2005 login error

    Post Author: dmccahanpppa
    CA Forum: General
    We are migrating from JD Edwards on AS400 to Microsoft Dynamics GP on SQL 2005. I can create and refresh reports without problem using sa as the login account in Crystal. When I use any of the users created in Dynamics GP (as they have to for users to access GP) I get a login error from the SQL server. This is using the same ODBC connection and I have taken some of the users and elevated there permission on the GP database and/or overall sql server to even dbo and sysadmin and still cannot get any login other than sa to work. Any sql GURUs out there that can tell me what I'm missing?
    Thanks

    Post Author: dmccahanpppa
    CA Forum: General
    I have tried creating ODBC connections (and also found a business objects article on OLE DB and tried that) with SQL Authentication and also with windows. I have made sure my users have usernames in GP that are all lower case name and password. The only thing that is getting me close is to use windows authentication on the odbc connection, set the odbc connection to a specific sql database, and give the users windows read permissions to that database. As a domain admin this gives complete access. With the users the login error goes away but I only see stored procedures and not the tables or views. I'm trying to find the minimum access to give the users in windows to allow windows authentication to work.
    What I don't understand is why this should be so difficult to find the correct answer on. I have the same issue on a very plain install (everything on the server, sql, gp, and crystal) with just microsoft's test company and data as I do the production system (gp and sql on one server and citrix client and crystal on another). I setup the test server (a couple of times blew the whole thing away and rebuilt following gp procedures) and consultants built the production environment. You would think Microsoft (since they blow their marketing horn about how Crystal Reports works with GP) or Business Objects would have the user and odbc setup for this documented.
    Ok...so I'm a bit delusional....

  • SQL 4060 Login Error

    <p>We have recently released a couple of new reports, but have discovered that although the reports run fine when launched in WEBi, whenever we try to refresh the report we fail with a SQL 4060 Login failed error message. This occurs in WEBi and BO. The reports run fine when logged into a &#39;Supervisor&#39; level account.</p><p>We&#39;ve reviewed user permissions, but can&#39;t seem to find the cause of this issue.</p><p> Any ideas?</p>

    Post Author: dmccahanpppa
    CA Forum: General
    I have tried creating ODBC connections (and also found a business objects article on OLE DB and tried that) with SQL Authentication and also with windows. I have made sure my users have usernames in GP that are all lower case name and password. The only thing that is getting me close is to use windows authentication on the odbc connection, set the odbc connection to a specific sql database, and give the users windows read permissions to that database. As a domain admin this gives complete access. With the users the login error goes away but I only see stored procedures and not the tables or views. I'm trying to find the minimum access to give the users in windows to allow windows authentication to work.
    What I don't understand is why this should be so difficult to find the correct answer on. I have the same issue on a very plain install (everything on the server, sql, gp, and crystal) with just microsoft's test company and data as I do the production system (gp and sql on one server and citrix client and crystal on another). I setup the test server (a couple of times blew the whole thing away and rebuilt following gp procedures) and consultants built the production environment. You would think Microsoft (since they blow their marketing horn about how Crystal Reports works with GP) or Business Objects would have the user and odbc setup for this documented.
    Ok...so I'm a bit delusional....

  • I am trying to setup a data source for postgresql with cf10 update 14. i get a time out error retruned in less than 2 seconds. I setup a data source for MS SQL 2005 no problem.

    I need a little help, what have I done wrong or not done?
    Here is the error message that is returned;
    Connection verification failed for data source: Archuleta_xxxxxx
    java.sql.SQLException: Timed out trying to establish connection
    The root cause was that: java.sql.SQLException: Timed out trying to establish connection
    any ideas would be appreciated.

    I see this error in your output:
    2014-11-26 10:55:23,583 ERROR [ThemeAutoDeployer]
    java.io.FileNotFoundException: /tmp/liferay/com/liferay/portal/deploy/dependencies/liferay-theme.tld (Too many open files)
    I'm not across what the EBS recommended setting for this is, or if there is one.  But try running:
    ulimit -n
    ... and if the number is low, edit /etc/security/limits.conf , add some entries for increased "soft nofile" and "hard nofile" and run sysctl -p.  See Linux & Java tips: Too many open files .

  • Error in one of data mover scripts during campus solution installation

    Hi everybody,
    This is my first attempt to install one of peoplesoft products
    I am installing HCM 9.0 on windows 2008 64 bit, oracle 11g
    I am now in task called 7A-16-9: Updating PeopleTools System Data
    I ran pt849tls.dms successfully
    but pt850tls.dms failed with error:
    File: Data MoverSQL error. Stmt #: 0 Error Position: 25 Return: 904 - ORA-00904: "PT_RETENTIONDAYS": invalid identifier
    Failed SQL stmt:UPDATE PS_PRCSSYSTEM SET PT_RETENTIONDAYS=RETENTIONDAYS
    Error: SQL execute error for UPDATE PS_PRCSSYSTEM SET PT_RETENTIONDAYS=RETENTIONDAYS
    Is there a script that I have missed?
    I followed the instructions step by step
    Thanks for your help,,,

    I highly appreciate your efforts,
    I ran rel849un.sql and rel850un.sql successfully (I checked the corresponding log files without errors)
    table description is:
    CREATE TABLE "SYSADM"."PS_PRCSSYSTEM"
    (     "OPSYS" VARCHAR2(1 CHAR),
         "RETENTIONDAYS" NUMBER(*,0),
         "PRCSPURGENEXTDTTM" TIMESTAMP (6),
         "RECURDTTM" TIMESTAMP (6),
         "PRCSPURGERECURNAME" VARCHAR2(30 CHAR),
         "PURGEPRCSFILES" NUMBER(*,0),
         "ARCH_PROCESSED" VARCHAR2(1 CHAR),
         "PRCSSYSLOADOPT" VARCHAR2(2 CHAR),
         "VERSION_UPDATED" VARCHAR2(1 CHAR))
    Best Regards,,,

  • Error -50 after failed data move to samsung hybrid internal drive.

    The original data move front external WD 3tb USB drive fails with error -36.  Then attempting to write any data to that drive dives a -50 error. Only cure I have found is to re-partition the drive. Fixing permissions, verifying drive, repairing drive, deleting .DS files,, etc. Nothing works but a partition recreation. Any clues? 2009 MBP, 6gig ram, 256 g Samsung 840 pro SSD primary drive. 1tb Samsung hybrid as secondary drive.

    That particular error happens on occasion after switching a TM drive from one Mac to another, erasing your TM disk/partition, or attaching a new TM drive with the same name as an old one.
    You can usually fix this by simply re-selecting your TM drive in TM Preferences > Change Disk.
    Since your backups are now working, it's likely ok.

  • Sql date time insert error

    Hi i am having trouble entering a date in to a sql stored
    procedure.
    first i convert the date to the users current datetime using
    <cfscript>
    userTZ="#qGetClubDetails.LocalTime12#"; // or whatever tz ID
    you need
    tz=createObject("component","timeZone");
    userDateTime=tz.castFromServer(today,userTZ);
    </cfscript>
    then i need to insert this into a stored procedure using
    <cfstoredproc procedure="xsp_InsertSmsRecords"
    datasource="#application.ds#">
    <cfprocparam type="In" cfsqltype="CF_SQL_DATETIME"
    dbvarname="UsersDate" value="#CREATEODBCDATETIME(userDateTime)#"
    null="No">
    </cfstoredproc>
    but i am getting a error of
    Macromedia][SQLServer JDBC Driver][SQLServer]Error converting
    data type varchar to datetime.
    any ideas what i need to do?

    quote:
    Originally posted by:
    craiglaw98
    ok i have tried that with no luck, i did a cfdump on
    userDateTime and i got
    {ts '2008-01-21 05:16:04'}
    any ideas what i need to do to insert this to sql server
    Looks ok to me. To troubleshoot this, try using cfquery for
    your insert instead of a stored procedure.

  • System.Data.SqlClient.SqlException (0x80131904): SQL Server detected a logical consistency-based I/O error: incorrect checksum

    Hello,
    I am migrating client's database to my hosted environment, the client runs
    SQL Server 2008 R2, and we run SQL Server 2014.
    I am stuck on restoring one log file, which left below error message when I restore it:
    Restoring files
      DBFile_20150401030954.trn
    Error Message: System.Data.SqlClient.SqlException (0x80131904): SQL Server detec
    ted a logical consistency-based I/O error: incorrect checksum (expected: 0x512cb
    ec3; actual: 0x512dbec3). It occurred during a read of page (1:827731)
    in databa
    se ID 49 at offset 0x000001942a6000 in file 'F:\MSSQL12.INSTANCE9\MSSQL\Data\Ody
    sseyTSIAKL_Data.mdf'.  Additional messages in the SQL Server error log or system
     event log may provide more detail. This is a severe error condition that threat
    ens database integrity and must be corrected immediately. Complete a full databa
    se consistency check (DBCC CHECKDB). This error can be caused by many factors; f
    or more information, see SQL Server Books Online.
    RESTORE LOG is terminating abnormally.
       at System.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolea
    n breakConnection, Action`1 wrapCloseInAction)
       at System.Data.SqlClient.SqlInternalConnection.OnError(SqlException exception
    , Boolean breakConnection, Action`1 wrapCloseInAction)
       at System.Data.SqlClient.TdsParser.ThrowExceptionAndWarning(TdsParserStateObj
    ect stateObj, Boolean callerHasConnectionLock, Boolean asyncClose)
       at System.Data.SqlClient.TdsParser.TryRun(RunBehavior runBehavior, SqlCommand
     cmdHandler, SqlDataReader dataStream, BulkCopySimpleResultSet bulkCopyHandler,
    TdsParserStateObject stateObj, Boolean& dataReady)
       at System.Data.SqlClient.SqlDataReader.TryConsumeMetaData()
       at System.Data.SqlClient.SqlDataReader.get_MetaData()
       at System.Data.SqlClient.SqlCommand.FinishExecuteReader(SqlDataReader ds, Run
    Behavior runBehavior, String resetOptionsString)
       at System.Data.SqlClient.SqlCommand.RunExecuteReaderTds(CommandBehavior cmdBe
    havior, RunBehavior runBehavior, Boolean returnStream, Boolean async, Int32 time
    out, Task& task, Boolean asyncWrite, SqlDataReader ds)
       at System.Data.SqlClient.SqlCommand.RunExecuteReader(CommandBehavior cmdBehav
    ior, RunBehavior runBehavior, Boolean returnStream, String method, TaskCompletio
    nSource`1 completion, Int32 timeout, Task& task, Boolean asyncWrite)
       at System.Data.SqlClient.SqlCommand.RunExecuteReader(CommandBehavior cmdBehav
    ior, RunBehavior runBehavior, Boolean returnStream, String method)
       at System.Data.SqlClient.SqlCommand.ExecuteScalar()
       at OnBoardingTrnRestore.FileRestorer.FileIsRestored(SqlConnection connection,
     String sourceBackupFileName, String sourceDatabaseName, String destinationDatab
    aseName, String databaseName, String strFileName) in c:\Dev\WiseCloud\OnBoarding
    TrnRestore\OnBoardingTrnRestore\FileRestorer.cs:line 223
    ClientConnectionId:6f583f98-9c23-4936-af45-0d7e9a9949ea
    Finished
    I have run restore filelistonly and verifyonly in both client and my box, and both worked well:
    XXX_Data D:\Program Files (x86)\XXX\Data\YYY_Data.mdf
    restore filelistonly:
    D PRIMARY
    70922469376 35184372080640
    1 0
    0 00000000-0000-0000-0000-000000000000
    0 0
    0 512 1
    NULL 279441000014839500242
    66CC6D17-575C-41E5-BB58-3FB4D33E288C
    0 1 NULL
    XXX_Log E:\Program Files (x86)\XXX\Log\YYY_Log.ldf
    L NULL
    31985762304 35184372080640
    2 0
    0 00000000-0000-0000-0000-000000000000
    0 0
    0 512 0
    NULL 0
    00000000-0000-0000-0000-000000000000 0
    1 NULL
    restore verifyonly:
    Attempting to restore this backup may encounter storage space problems. Subsequent messages will provide details.
    The path specified by "D:\Program Files (x86)\XXX\Data\YYY_Data.mdf" is not in a valid directory.
    Directory lookup for the file "E:\Program Files (x86)\XXX\Log\YYY_Log.ldf" failed with the operating system error 3(The system cannot find the path specified.).
    The backup set on file 1 is valid.
    Dbcc checkdb also confirmed there is no error in client sql server database.
    dbcc checkdb (XXX, NOINDEX) 
    Results as follows:
    DBCC results for 'XXX'.
    Warning: NO_INDEX option of checkdb being used. Checks on non-system indexes will be skipped.
    Service Broker Msg 9675, State 1: Message Types analyzed: 17.
    Service Broker Msg 9676, State 1: Service Contracts analyzed: 9.
    Service Broker Msg 9667, State 1: Services analyzed: 7.
    Service Broker Msg 9668, State 1: Service Queues analyzed: 7.
    Service Broker Msg 9669, State 1: Conversation Endpoints analyzed: 0.
    Service Broker Msg 9674, State 1: Conversation Groups analyzed: 0.
    Service Broker Msg 9670, State 1: Remote Service Bindings analyzed: 0.
    Service Broker Msg 9605, State 1: Conversation Priorities analyzed: 0.
    DBCC results for 'sys.sysrscols'.
    There are 23808 rows in 350 pages for object "sys.sysrscols".
    DBCC results for 'BMBoard'.
    There are 0 rows in 0 pages for object "BMBoard".
    DBCC results for 'CusOutturnHeader'.
    There are 0 rows in 0 pages for object "CusOutturnHeader".
    CHECKDB found 0 allocation errors and 0 consistency errors in database 'XXX'.
    DBCC execution completed. If DBCC printed error messages, contact your system administrator.
    We use ftp to transfer the log files from the another country to here, I confirmed the file's MD5(checksum) is same as the copy in client's environment, we have tried to upload the file a couple of times, same outcome.
    The full backup is fine, I have successfully restored a series log backups prior to this file.
    Has anyone seen this issue before? It is strange, all information indicates the file is good,but somehow it goes wrong.
    Below are the version details:
    Client: 
    MICROSOFT SQL SERVER 2008 R2 (SP2) - 10.50.4000.0 (X64) 
    My environment:
    Microsoft SQL Server 2014 - 12.0.2480.0 (X64) 
    Jan 28 2015 18:53:20 
    Copyright (c) Microsoft Corporation
    Enterprise Edition (64-bit) on Windows NT 6.3 <X64> (Build 9600: )
    Thanks,
    Albert

    Error Message: System.Data.SqlClient.SqlException (0x80131904): SQL Server detec
    ted a logical consistency-based I/O error: incorrect checksum (expected: 0x512cb
    ec3; actual: 0x512dbec3). It occurred during a read of page (1:827731)
    in databa
    se ID 49 at offset 0x000001942a6000 in file 'F:\MSSQL12.INSTANCE9\MSSQL\Data\Ody
    sseyTSIAKL_Data.mdf'.  Additional messages in the SQL Server error log or system
     event log may provide more detail. This is a severe error condition that threat
    ens database integrity and must be corrected immediately.
    Hi Albert,
    The above error message usually indicates that there is a problem with underlying storage system or the hardware or a driver that is in the path of the I/O request. You can encounter this error when there are inconsistencies in the file system or if the database
    file is damaged.
    Besides other suggestions, consider to use the
    SQLIOSim utility to find out if these errors can be reproduced outside of regular SQL Server I/O requests and change
    your databases to use the PAGE_VERIFY CHECKSUM option.
    Reference:
    http://support.microsoft.com/en-us/kb/2015756
    https://msdn.microsoft.com/en-us/library/aa337274.aspx?f=255&MSPPError=-2147217396
    Thanks,
    Lydia Zhang
    Lydia Zhang
    TechNet Community Support

  • SQL Insert Statement Data Type Mismatch Error

    I am doing a very simple web application that has a Microsoft Access database as the data source. I have been able to sucessfully create update and query statements using parameters but am having issues with an insert statement. I am using JSTL 1.1.2
    The following code creates the data type mismatch error.
    <sql:update
         sql="insert into tblTtoF(TFToolID,TFFeatID) values(?,?)">
            <sql:param value='$(ID}'/>
         <sql:param value='${feature}'/>
            </sql:update>The table has NUMBER as the data type for both of these fields and the variables I am feeding into it are both numbers. If I hard code the first number into the sql statement then it works. I have tried swapping the variables around and as long as the first one is hard coded the parameter for the second one works no matter which is first or second.
    However I can get the following code to work, which of course leaves me vulnerable to sql injection attacks which is not really a good thing.
    <sql:update>
         insert into tblTtoF(TFToolID,TFFeatID) values('<c:out value="${ID}"/>','<c:out value="${feature}"/>')
            </sql:update>So I am just looking for any suggestions as to why my first piece of code doesn't work seeing as it is the simplest of SQL statements and the most standard syntax.
    Thanks

    I changed it to the following
         <c:set var="featurenew" value="${0 + feature}"/>
         <c:set var="IDnew" value="${0 + param.toolID}"/>
              <sql:update
              sql="insert into tblTtoF(TFToolID,TFFeatID) values(?,?)">
              <sql:param value='$(IDnew}'/>
              <sql:param value='${featurenew}'/>
              </sql:update>And got the following error in the localhost.log
    31/07/2006 09:31:41 org.apache.catalina.core.StandardWrapperValve invoke
    SEVERE: Servlet.service() for servlet jsp threw exception
    java.sql.SQLException: SQL Exception : [Microsoft][ODBC Microsoft Access Driver]Optional feature not implemented
         at sun.jdbc.odbc.JdbcOdbcPreparedStatement.setObject(JdbcOdbcPreparedStatement.java:1437)
         at sun.jdbc.odbc.JdbcOdbcPreparedStatement.setObject(JdbcOdbcPreparedStatement.java:1072)
         at sun.jdbc.odbc.JdbcOdbcPreparedStatement.setObject(JdbcOdbcPreparedStatement.java:1063)
         at org.apache.taglibs.standard.tag.common.sql.UpdateTagSupport.setParameters(UpdateTagSupport.java:254)
         at org.apache.taglibs.standard.tag.common.sql.UpdateTagSupport.doEndTag(UpdateTagSupport.java:156)
         at org.apache.jsp.dataUpdated_jsp._jspx_meth_sql_update_1(dataUpdated_jsp.java:975)
         at org.apache.jsp.dataUpdated_jsp._jspx_meth_c_if_0(dataUpdated_jsp.java:879)
         at org.apache.jsp.dataUpdated_jsp._jspx_meth_c_forEach_0(dataUpdated_jsp.java:680)
         at org.apache.jsp.dataUpdated_jsp._jspService(dataUpdated_jsp.java:151)
         at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:97)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
         at org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:332)
         at org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:314)
         at org.apache.jasper.servlet.JspServlet.service(JspServlet.java:264)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
         at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:252)
         at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:173)
         at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:213)
         at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:178)
         at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:126)
         at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:105)
         at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:107)
         at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:148)
         at org.apache.coyote.http11.Http11AprProcessor.process(Http11AprProcessor.java:833)
         at org.apache.coyote.http11.Http11AprProtocol$Http11ConnectionHandler.process(Http11AprProtocol.java:639)
         at org.apache.tomcat.util.net.AprEndpoint$Worker.run(AprEndpoint.java:1285)
         at java.lang.Thread.run(Thread.java:595)
    I have also tried the following in the past with no luck
    <fmt:parseNumber value="${ID}" type="number" var="IDnew"/>
    AND......
    <sql:query
       sql="select TFToolID from tblTtoF where TFToolID = ?"
       var="toolresults">
       <sql:param value="${ID}"/>
    </sql:query>
    <c:forEach var="getID" items="${toolresults.rows}">
         <c:set var="theID" value="${getID.TFToolID}"/>
    </c:forEach>
    AND when that didn't work, added this....
    <fmt:parseNumber value="${theID}" var="IDnew"/>

Maybe you are looking for

  • Upgraded to iTunes 8, iPod no longer recognized

    Hello! I have a HP laptop running Vista Home and everything was working fine until I upgraded to version 8 of iTunes. Now, my laptop does not even recognize my iPod at ALL, not in Explorer, My Computer or iTunes. USB ports are working fine, I have an

  • 2.0 dd movie will it sound like proper 5.1 using up

    hi i just want to know if you have a dolby digital movie in 2.0 if i used the upmix on my gigaworks g500 will it sound like proper 5. or will it not sound as good thanks.

  • Security Filter Error

    Hi When i refresh my security filter i am getting an error saying cache size limit reached. can anybody tell me wer do i adjust my security filter chache size Thanks Abe Edited by: user11884591 on Sep 10, 2009 11:21 AM

  • "Duplicate" Capability on "2nd Monitor"

    I'd like to see a "Duplicate" view added to the current "Loupe /Grid / Compare / Survey" options that would send the same image to a "2nd monitor". This would provide the ability to project the same image using a different monitor calibration profile

  • Upgrading to larger HD - can I restore everything using Time Machine?

    I plan on upgrading my existing 120GB HD to a 500GB; moving all existing apps and data onto it, plus consolidating from an external HD. At the moment, I have everything installed and set up just how I want it, and don't look forward to doing this all