OLEDB Fixed Query Comments Create Errors through PCo 2.1

Testing PCo using OLEDB.  Everything is working fine except that existing Fixed queries with comments keep throwing syntax errors when executed through PCo. 
Anyone else seen this, and if so what should we use to replace "--" for comment lines?
The source system has no issues with these comment markers, and they worked fine in UDS 4.0.

Fixed in SP06.

Similar Messages

  • Error while creating asset through AS91

    Hi All,
    When i am trying to create asset through AS91 for a new company i am getting error AY 252 " Current fiscal year 2007 has to be after transfer daet 31.12.2007. In this company the Asset accounting has been recently customised but when i am trying for the upload of asset master data i am getting the above given error.
    Please tell what should be done to remove this error.
    Thanks & regards,
    Prachi

    Check the transfer date for asst data transfer...make sure the current fiscal yr is after the transfer date
    Asset accounting>Asset data transfer>Parameters for data transfer>date specifications>Specify transfer date/last closed fiscal yr.

  • How create the SMQ2 error through program ?

    Hi all,
    How create the SMQ2 error through program ?
    Any idea ... plsease share.
    Regards,
    Srikanth

    Hi juturusrikanth ,
    which one is the SMQ2 error?
    [FAQ's, intros and memorable discussions in the ABAP General Forum|FAQ's, intros and memorable discussions in the ABAP General Forum;
    [How to post code in SCN, and some things NOT to do...|How to post code in SCN, and some things NOT to do...;
    Regards
    Clemens

  • Error while creating hierarchy through " Create Hierarchy field" option

    Hi All,
    while trying to create hierarchy through Create Hierarchy field option i am facing the following error
    I am giving the following inputs
    Hierarchy field name : XYZ
    Parent field                : Level1
    Child Field                  : Level2
    Node name field        : Level3
    Cannot find the parent ID value "abc" in Field ID Level2 of Table " Hierarchy"
    FYI
    The source hierarchy is in Columns format with Level1(Parent)---->Lever2(Child of Level1) and Level3(Child of level2)
    Kindly guide me where i am going wrong
    Thanks in Advance
    Regards
    Sharma

    Hi Sharma,
    Your input file should have 3 columns for hierarchy one which will store level 1 information,2nd will have corresponding level 2 information and 3rd will have corresponding level 3 info.
    Now in Import Manager go to Partition,select 1st ,2nd and 3rd column.Donot combine the fields. Now map the partition fields with the target hierarchy field.Perform the value mapping and import.
    Or
    YOu can have one column in the excel file.with values L1>L2>L3(split by some delimiter).
    In import manager go to source hierechy pany select the Column and from context menu select option split into hierechy as per the delimiter.
    Now map the parttion field and target field ,do value mapping and import.
    Regards,
    Neethu Joy

  • HT3275 how do i fix this error - The backup disk image "/Volumes/My Passport/EDS MAC.sparsebundle" could not be created (error 1).

    When trying to use Time Capsule on my passport I get this message
    The backup disk image “/Volumes/My Passport/EDS MAC.sparsebundle” could not be created (error 1).

    That should be a C9 or C17 error.. see
    http://pondini.org/TM/Troubleshooting.html
    Try rebooting everything.. as something has lost connection.. perhaps.
    Make sure the disk is actually available in finder.. and try mounting it.

  • SQL 2012 SP1 - How to determine a query that causes Error 8623 in SQL Log: The query processor ran out of internal resources and could not produce a query plan. This is a rare event...

    We are getting multiple 8623 Errors in SQL Log while running Vendor's software.
    How can you catch which Query causes the error?
    I tried to catch it using SQL Profiler Trace but it doesn't show which Query/Sp is the one causing an error. 
    I also tried to use Extended Event session to catch it, but it doesn't create any output either.
    Error:
    The query processor ran out of internal resources and could not produce a query plan. This is a rare event and only expected for extremely complex queries or queries that
    reference a very large number of tables or partitions. Please simplify the query. If you believe you have received this message in error, contact Customer Support Services for more information.
    Extended Event Session that I used;
    CREATE EVENT SESSION
        overly_complex_queries
    ON SERVER
    ADD EVENT sqlserver.error_reported
        ACTION (sqlserver.sql_text, sqlserver.tsql_stack, sqlserver.database_id, sqlserver.username)
        WHERE ([severity] = 16
    AND [error_number] = 8623)
    ADD TARGET package0.asynchronous_file_target
    (SET filename = 'E:\SQLServer2012\MSSQL11.MSSQLSERVER\MSSQL\Log\XE\overly_complex_queries.xel' ,
        metadatafile = 'E:\SQLServer2012\MSSQL11.MSSQLSERVER\MSSQL\Log\XE\overly_complex_queries.xem',
        max_file_size = 10,
        max_rollover_files = 5)
    WITH (MAX_DISPATCH_LATENCY = 5SECONDS)
    GO
    -- Start the session
    ALTER EVENT SESSION overly_complex_queries
        ON SERVER STATE = START
    GO
    It creates only .xel file, but not .xem
    Any help/advice is greatly appreciated

    Hi VK_DBA,
    According to your error message, about which query statement may fail with error message 8623, as other post, you can use trace flag 4102 & 4118 for overcoming this error. Another way is looking for queries with very long IN lists, a large number of
    UNIONs, or a large number of nested sub-queries. These are the most common causes of this particular error message.
    The error 8623 occurs when attempting to select records through a query with a large number of entries in the "IN" clause (> 10,000). For avoiding this error, I suggest that you could apply the latest Cumulative Updates media for SQL Server 2012 Service
    Pack 1, then simplify the query. You may try divide and conquer approach to get part of the query working (as temp table) and then add extra joins / conditions. Or You could try to run the query using the hint option (force order), option (hash join), option
    (merge join) with a plan guide.
    For more information about error 8623, you can review the following article.
    http://blogs.technet.com/b/mdegre/archive/2012/03/13/8623-the-query-processor-ran-out-of-internal-resources-and-could-not-produce-a-query-plan.aspx
    Regards,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

  • Sparsebundle could not be created (error 45)

    I just bought my first ever Mac a little over 30 days ago and also purchased a 1TB Seagate FreeAgent GoFlex Home network storage system. I immediately installed the GoFlex software that included a full integration into Time Machine. This worked perfectly until about 4 days ago when all of a sudden it stopped backing up. I'm getting an error that reads: Time Machine could not complete the backup "...sparsebundle could not be created (error 45)".
    I can still access the drive outside of the TM program (although not the ATM folder), so the drive itself perfectly installed and connected. I tried rebooting computer and hard drive; that didn't work. I tried re-selecting the drive from within TM; that didn't work either. I'm not really sure what to do...and am pretty pi$$ed at all those "I'm a Mac vs. I'm a PC" commercials right now because this is the same kinda stuff that got me to leave the Windows based systems in the first place. UGH...so WHY is my $4,000 Mac supposed to be any better that an equivelent $2,500 PC?????? Any suggestions on a fix would be greatly appreciated.

    +I'm running 10.6.7, last I checked that was later then 10.5.6.+
    Try reading a little more carefully. The server has to be a Time Capsule or a Mac running 10.5.6 or later, not the client. Your NAS is neither, so your configuration is unsupported, and predictably, it doesn't work.
    +Mac lends their name and brand to the Seagate product, so I must therefore assume your comments are are ill-informed.+
    Apple has nothing to do with the Seagate product. Seagate, not Apple, claims that it's compatible with Time Machine, and Seagate, not Apple, is responsible for that claim. If you think the claim is misleading, you have a point, and if you can't get satisfaction from Seagate I suggest you consult a lawyer.
    For the benefit of anyone finding this topic, the Seagate NAS is *not safe to use with Time Machine*.

  • Processing a cursor of 11,000 rows and Query completed with errors

    So I have 3rd party data that I have loaded into a SQL Server Table. I am trying to determine if the 3rd party Members reside in our database by using a cursor and going through all 11,000 rows...substituting the #Parameter Values in a LIKE statement...trying
    to keep it pretty broad. I tried running this in SQL Server Management Studio and it chunked for about 5 minutes and then just quit. I kind of figured I was pushing the buffer limits within SQL Server Management Studio. So instead I created it as a Stored
    Procedure and changed my Query Option/Results and checked Discard results after execution. This time it chunked away for 38 minutes and then stopped saying
    Query completed with errors. I did throw a COMMIT in there thinking that the COMMIT would hit and free up resources and I'd see the Table being loaded in chunks. But that didn't seem to work.
    I'm kind of at a loss here in terms of trying to tie back this data.
    Can anyone suggest anything on this???
    Thanks for your review and am hopeful for a reply.
    WHILE (@@FETCH_STATUS=0)
    BEGIN
    SET @SQLString = 'INSERT INTO [dbo].[FBMCNameMatch]' + @NewLineChar;
    SET @SQLString = ' (' + @NewLineChar;
    SET @SQLString = ' [FBMCMemberKey],' + @NewLineChar;
    SET @SQLString = ' [HFHPMemberNbr]' + @NewLineChar;
    SET @SQLString = ' )' + @NewLineChar;
    SET @SQLString = 'SELECT ';
    SET @SQLString = @SQLString + CAST(@FBMCMemberKey AS VARCHAR) + ',' + @NewLineChar;
    SET @SQLString = @SQLString + ' [member].[MEMBER_NBR]' + @NewLineChar;
    SET @SQLString = @SQLString + 'FROM [Report].[dbo].[member] ' + @NewLineChar;
    SET @SQLString = @SQLString + 'WHERE [member].[NAME_FIRST] LIKE ' + '''' + '%' + @FirstName + '%' + '''' + ' ' + @NewLineChar;
    SET @SQLString = @SQLString + 'AND [member].[NAME_LAST] LIKE ' + '''' + '%' + @LastName + '%' + '''' + ' ' + @NewLineChar;
    EXEC (@SQLString)
    --SELECT @SQLReturnValue
    SET @CountFBMCNameMatchINSERT = @CountFBMCNameMatchINSERT + 1
    IF @CountFBMCNameMatchINSERT = 100
    BEGIN
    COMMIT;
    SET @CountFBMCNameMatchINSERT = 0;
    END
    FETCH NEXT
    FROM FBMC_Member_Roster_Cursor
    INTO @MemberIdentity,
    @FBMCMemberKey,
    @ClientName,
    @MemberSSN,
    @FirstName,
    @MiddleInitial,
    @LastName,
    @AddressLine1,
    @AddressLine2,
    @City,
    @State,
    @Zipcode,
    @TelephoneNumber,
    @BirthDate,
    @Gender,
    @EmailAddress,
    @Relation
    END
    --SELECT *
    --FROM [#TempTable_FBMC_Name_Match]
    CLOSE FBMC_Member_Roster_Cursor;
    DEALLOCATE FBMC_Member_Roster_Cursor;
    GO

    Hi ITBobbyP,
    As Erland suggested, you can compare all rows at once. Basing on my understanding on your code, the below code can lead to the same output as yours but have a better performance than cursor I believe.
    CREATE TABLE [MemberRoster]
    MemberKey INT,
    FirstName VARCHAR(99),
    LastName VARCHAR(99)
    INSERT INTO [MemberRoster]
    VALUES
    (1,'Eric','Zhang'),
    (2,'Jackie','Cheng'),
    (3,'Bruce','Lin');
    CREATE TABLE [yourCursorTable]
    MemberNbr INT,
    FirstName VARCHAR(99),
    LastName VARCHAR(99)
    INSERT INTO [yourCursorTable]
    VALUES
    (1,'Bruce','Li'),
    (2,'Jack','Chen');
    SELECT * FROM [MemberRoster]
    SELECT * FROM [yourCursorTable]
    --INSERT INTO [dbo].[NameMatch]
    --[MemberNbr],
    --[MemberKey]
    SELECT y.MemberNbr,
    n.[MemberKey]
    FROM [dbo].[MemberRoster] n
    JOIN [yourCursorTable] y
    ON n.[FirstName] LIKE '%'+y.FirstName+'%'
    AND n.[LastName] LIKE '%'+y.LastName+'%'
    DROP TABLE [MemberRoster], [yourCursorTable]
    If you have any question, feel free to let me know.
    Eric Zhang
    TechNet Community Support

  • How to create WWWBinaryFileAsset through programming

    Hi, Could anybody provide me steps to create WWWBinaryFileAsset through programming.
    I followed below steps but i am getting error,
    MutableRepository pubRepository = getPublishingRepository();
    // pub repository is pointing to /atg/epub/file/SecuredPublishingFileRepository
    wwwBinaryFileAsset = pubRepository.createItem(filesProperties.getWwwBinaryFileAssetItemDescriptorName());
         if (wwwBinaryFileAsset != null) {
         wwwBinaryFileAsset.setPropertyValue(filesProperties.getFilenamePropertyName(), imageName);
         wwwBinaryFileAsset.setPropertyValue(filesProperties.getParentFolderPropertyName(), parentFolder);
         wwwBinaryFileAsset.setPropertyValue("content", new File("D:\\media\\images\\hmc\\Penguins.jpg"));
         pubRepository.addItem(wwwBinaryFileAsset);
    Thanks,
    Raj

    Hi,
    I am able to create project and adding binary file asset to the project with code as i copy/pasted in this query, after creation, also i am able to see the asset in BCC editor as well.
    I can say asset looks like as it was created in BCC editor, means there are nodifferences in assest properties which created through programming from asset created in BCC editor.
    I am not sure what i am missing. But only the problem is, i am getting the error when deploying project from BCC, error stack trace is,
    /atg/epub/DeploymentServer 14000002:Production:atg.deployment.adapter.DistributedDeploymentAdapter.pollTillDataApplied(DistributedDeploymentAdapter.java:2588) Polling Until Data Appplied
    /atg/deployment/DeploymentManager Starting deployment with id: 14000002
    /atg/deployment/DeploymentManager Executing deployment in LOCAL mode
    /atg/deployment/DeploymentManager item = fileMarker:mark7800001 cause = java.lang.IllegalArgumentException: Negative initial size: -1
    /atg/deployment/DeploymentManager at java.io.ByteArrayOutputStream.<init>(ByteArrayOutputStream.java:57)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileDeploymentCommandImpl.getUncompressedInputStream(FileDeploymentCommandImpl.java:398)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileDeploymentCommandImpl.getInputStream(FileDeploymentCommandImpl.java:470)
    /atg/deployment/DeploymentManager at atg.deployment.file.DeploymentProtocolImpl.sendFileDeploymentCommand(DeploymentProtocolImpl.java:913)
    /atg/deployment/DeploymentManager at atg.deployment.file.DeploymentProtocolImpl.sendCommand(DeploymentProtocolImpl.java:179)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileWorkerThread.runCommand(FileWorkerThread.java:716)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileWorkerThread.processMarkerForAddUpdatePhase(FileWorkerThread.java:441)
    /atg/deployment/DeploymentManager at atg.deployment.DeploymentWorkerThread.processMarkerPhase(DeploymentWorkerThread.java:521)
    /atg/deployment/DeploymentManager at atg.deployment.DeploymentWorkerThread.run(DeploymentWorkerThread.java:300)
    /atg/deployment/DeploymentManager message = Deployment Failed time = Wed Jun 20 09:12:50 IST 2012 atg.deployment.DeploymentFailure@93d841 java.lang.IllegalArgumentException: Negative initial size: -1
    /atg/deployment/DeploymentManager at java.io.ByteArrayOutputStream.<init>(ByteArrayOutputStream.java:57)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileDeploymentCommandImpl.getUncompressedInputStream(FileDeploymentCommandImpl.java:398)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileDeploymentCommandImpl.getInputStream(FileDeploymentCommandImpl.java:470)
    /atg/deployment/DeploymentManager at atg.deployment.file.DeploymentProtocolImpl.sendFileDeploymentCommand(DeploymentProtocolImpl.java:913)
    /atg/deployment/DeploymentManager at atg.deployment.file.DeploymentProtocolImpl.sendCommand(DeploymentProtocolImpl.java:179)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileWorkerThread.runCommand(FileWorkerThread.java:716)
    /atg/deployment/DeploymentManager at atg.deployment.file.FileWorkerThread.processMarkerForAddUpdatePhase(FileWorkerThread.java:441)
    /atg/deployment/DeploymentManager at atg.deployment.DeploymentWorkerThread.processMarkerPhase(DeploymentWorkerThread.java:521)
    /atg/deployment/DeploymentManager at atg.deployment.DeploymentWorkerThread.run(DeploymentWorkerThread.java:300)
    /atg/deployment/DeploymentManager
    /atg/deployment/DeploymentManager Unexpected error occurred. DeploymentWorkerThread terminated prematurely. atg.deployment.DistributedDeploymentException: Exceeded allowable error count (0)
    /atg/deployment/DeploymentManager at atg.deployment.DeploymentWorkerThread.processMarkerPhase(DeploymentWorkerThread.java:626)
    /atg/deployment/DeploymentManager at atg.deployment.DeploymentWorkerThread.run(DeploymentWorkerThread.java:300)
    Thanks.
    Edited by: 938890 on Jun 19, 2012 9:09 PM

  • Query engine failed error for crytal report refreshing to new params in jsp

    Using licensed weblogic 8.1 server in production mode. Weblogic workhsop has inetegrated supprot for crystal reports 9. Using standalone report & accessing sql server through odbc, i got result for different parameters passed.
              Problem Area : Passing parameter to have specific report causes unexpected query engine failed error in com.crystaldecisions.report.web.viewer.CrystalViewer class. If viewer.refresh method is commented, then static (already saved) report is displayed through jsp. But using viewer.refresh method for dynamic report genertaion for new parameters through jsp gives above error.

    Hello
              I'm experiencing the same problem. Please let me know if you've any solution.

  • Write empty value through PCO

    Hi,
    I have a tag on my OPC server with datatype String.
    When I write an empty value to the tag through PCo I always get the error: "Invalid tag value for x".
    Writing a space works but this is not good because the PLC will interpret this differently.
    How can I write an empty value to a tag?
    I am using Kepware OPC and PCO 15.0.0.1959
    Thanks

    If i comment out.write(value) it prints fine to file. below it saves "test" at every columns.
    So that narrow it to this line "String value = (String)custTable.getValueAt(j, k);"
    for (int j = 0; j < rows; j++){
              for (int k = 0; k < columns; k++){
              String value = (String)custTable.getValueAt(j, k);
                   // out.write(value);
                    out.write("test");
                    out.newLine();
                    out.close();getValueAt dont return any data at all when it meet an empty column?

  • How Can I Fix A QuickTime Codec Error?

    How Can I Fix A QuickTime Codec Error?
    I have several movies from 2008 that are labeled as .mov but I now get a QuickTime error saying "The document "XXXXXX" could not be opened. A required Codec isn't available."
    These movies ran in QuickTime just fine until about a month ago.
    All the folders in which these movies were stored in 4 folders in 2 folders in one folder. All folders at each level were encrypted with Stuffit Deluxe a version circa 2008 and decrypted with a newer Stuffit Deluxe 10.x. Other movies that are NOT .mov and in with the .mov movies still run.
    If I click on the Thumbnail Preview "It" runs.
    How can I fix this?
    TOP image is the Error message.
    BOTTOM image is the Thumbnail Preview info.
    I've checked out "Tell Me More" and have not found a solution.

    How Can I Fix A QuickTime Codec Error?
    The basic work flow is to determine which codecs were used to create the non-playing file and then check to see if these formats are supported by your current system codec component configuration.
    If the files will open in one of the QT player app, the open the inspector window to see which codecs were used to create the original file as seen here...
    or here in the "Format" entry...
    Or, if you have QT 7 keyed for "Pro" use, you can open the "Properties" window to gain even more information as demonstrated here...
    if you file will not open in either of the QT Player apps, then you can try opening the Finder "Get Info" window for the file,
    use a third-party media player that will open the file, or
    use a dedicated media information utility to learn this information....
    Once you know what codecs were used to create the file, post back with this information so we can deterine if the file should be natively supported by your Mac OS X install, requires an Apple Pro editing codec normally installed by the editor app, or requires a third-party codec which the user must supply.
    These movies ran in QuickTime just fine until about a month ago.
    All the folders in which these movies were stored in 4 folders in 2 folders in one folder. All folders at each level were encrypted with Stuffit Deluxe a version circa 2008 and decrypted with a newer Stuffit Deluxe 10.x. Other movies that are NOT .mov and in with the .mov movies still run.
    Your information does not indicate whether or not the files actually played on the current system or were transferred from a different system which was properly configured for playbac, whether or not you recently upgraded your operating system/modified your codec component configuration by adding or deleting any codecs, whether your compound "Stuffing" of the files created corruption, etc.
    As to MOV files the selves, this extension just implies the data was placed in the QT generic file container and the actually data could be of any compression format thathat was compatible with the codec component configuration for the system on which the files were created and could include natively supported system codecs, pro editing codecs, and/or third-party proprietary or hybrid codecs. (I.e., this is why we may need more information regarding the individual non-playing files.)
    If I click on the Thumbnail Preview "It" runs.
    f you are referring to QuickLook (QL) then the technology is different than QuickTime (QT). In the former case, the system is playing representative data but not actually opening the file as QT does. Thus, there is always the possibility that the problem is in the file structure (atom) information. For instance, suppose someone merely changed the MOV file extension in the Finder rather than actually copying data to a real MOV file container. Then it is possible that the QT player app may be searching for a codec that is not supposed or expected to be supported in the MOV file container and, not finding it, issues the erroneous codec error message—especially since you did not inticate the source of the file you are trying to play or, in some cases, which QT player is now not playing the file.
    In short, there are several possibilities here and not enough specific information.

  • "Error in creating error file" message

    Tried to install Palm Desktop 6.2 that came with my new Centro and at the end the message "Error in creating error file" keeps popping up numerous times.
    What can be the reason?
    Post relates to: Centro (AT&T)

    I was also experiencing the "error in creating error file" message multiple times during installation on my XP Pro system. I googled the error message and found this post plus a reference to a problem with riched32.dll) in in another post. But that didn't help. After quite a bit of troubleshooting on my own, I found the problem. I had moved the "My Documents" folder under "D:\users" to simplify my backups. Apparently the Palm installer needs the "My Documents" folder on the C: drive. I temporarily put "My Documents" back into it's default location under "C:\Documents and Settings" and the Palm installer ran just fine with no error message. I then moved everything back under "D:\users" including the folder "Palm OS Desktop", that was created by the Palm installer and everything has been working fine since.
    Thought I'd share this with others, as it's not an obvious fix.
    Post relates to: Palm TX

  • The backup disk image "/Volumes/FW1000HD/MacBkPro.sparsebundle" could not be created (error 1).

    I added an 1TB external usb hard disk to my time capsule as the existing 1TB hard disk is nearly full (less than 12 GB left on time capslule hard disk).  Initially, I connected external hard disk to my MacBookPro (late 2009 Intel machine) and erased most of the data on that disk. I added the disk to time machine's backup disks thru the time machine window in system preferences. Now I supposedly have both hard disks, the one in the time capsule and the new usb hard disk, available for backup. Time Machine immediately backed up my system, which consists of my MacBookPro hard disk (about 96.8 GB) and another external usb hard disk which contains about 500 GB of data, on the new extenal hard disk.Time Machine did not attempt to create a new sparsebundle at that time on my new external backup disk.  This backup was done while the new backup hard disk and the data disk were both connected to the MacBookPro and the Time Capsule was directly connected by ethernet cable to the MacBookPro. I then disconnected the two hard disks from the MacBookPro and reconnected them both via usb to the Time Capsule. Now, hopefully, the external usb backup hard disk can be written to via wi-fi thru the usb connection thru the Time Capsule? I successfully use the data disk (which contains my full iTunes Library) from the Time Capsule all of the time.
    Time Machine has continued to backup small backups of a few MB to a few hundred MB to the  Time Capsule hard disk. Time Machine has attempted at least twice that I know of to set up a sparsebundle on the new external backup hard disk, but it has failed both times with the following message:
    The backup disk image “/Volumes/FW1000HD/MacBkPro.sparsebundle” could not be created (error 1).
    Before the two attempts that I witnessed to create the sparsebundle on the new backup drive, I got the following message:
    The identity of the backup disk “FW1000HD” has changed since the previous backup. 
    I did not change the identity of the FW1000HD, but it originally was connected directly to the MacBookPro when it was backed up. Could this be the problem? And if so, how do I fix this, so I can use this hard disk for back up from the wi-fi connection to Time Capsule?
    Gary Olson

    I'm not sure this is at the root of your problems, but Time Machine manages backups to a local disk differently than to a network disk.  The Time Machine backup written to the disk while it was connected locally can't be used for network backups.

  • Query to create a view using the following tables

    Hi,
    I am struggling to write the proper query to create a view using three tables.
    I would really be thankful if anyone can help me.
    I am pasting the following script to create the required tables and also insert appropriate data in each of the table.
    create table HR.STUDENT_DETAILS (
    STUDENT_ID NUMBER(10),
    STUDENT_NAME VARCHAR2(50),
    DOB DATE,
    SEX CHAR(1),
    ACTIVE CHAR(1),
    CONTACT_NUMBER NUMBER(20),
    primary key(STUDENT_ID)
    create table HR.PAYMENT_HEADER (
    PAY_HEADER_ID NUMBER(10),
    DOC_NUM NUMBER(5),
    MONTH_NAME VARCHAR2(10),
    primary key(PAY_HEADER_ID)
    create table HR.PAYMENT_DETAILS (
    HEADER_ID NUMBER(10),
    LINE_ID NUMBER(10),
    STUDENT_ID NUMBER(10),
    CUM_AMOUNT NUMBER(10),
    primary key(HEADER_ID,LINE_ID)
    INSERT ALL
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1112,'James','17-JUN-05','M','Y',23674378)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1113,'David','21-SEP-05','M','Y',24565457)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1114,'Michael','13-JAN-06','M','Y',25436784)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1115,'Joseph','03-JAN-06','M','Y',23435673)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1116,'Juliet','21-MAY-05','F','Y',23234527)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1117,'Monica','25-JUN-05','F','Y',24873245)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1118,'William','05-FEB-05','M','Y',23623245)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1119,'Karen','07-FEB-06','F','Y',26757543)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1120,'Erika','17-AUG-05','F','Y',25435465)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1121,'Noah','16-AUG-05','M','Y',23457645)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1122,'Angelina','28-SEP-05','F','Y',26456787)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1123,'Gabriela','30-SEP-05','F','Y',29767543)
    INTO HR.STUDENT_DETAILS (STUDENT_ID,STUDENT_NAME,DOB,SEX,ACTIVE,CONTACT_NUMBER) VALUES (1124,'Sofia','07-MAR-06','F','Y',27656578)
    SELECT * FROM DUAL;
    INSERT ALL
    INTO HR.PAYMENT_HEADER (PAY_HEADER_ID,DOC_NUM,MONTH_NAME) VALUES (305,1,'JAN')
    INTO HR.PAYMENT_HEADER (PAY_HEADER_ID,DOC_NUM,MONTH_NAME) VALUES (306,2,'FEB')
    INTO HR.PAYMENT_HEADER (PAY_HEADER_ID,DOC_NUM,MONTH_NAME) VALUES (307,3,'MAR')
    INTO HR.PAYMENT_HEADER (PAY_HEADER_ID,DOC_NUM,MONTH_NAME) VALUES (308,4,'APR')
    INTO HR.PAYMENT_HEADER (PAY_HEADER_ID,DOC_NUM,MONTH_NAME) VALUES (309,5,'MAY')
    SELECT * FROM DUAL;
    INSERT ALL
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (305,12,1112,40)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (305,13,1113,40)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (305,14,1114,40)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (306,15,1112,80)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (306,16,1113,80)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (306,17,1114,80)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (306,18,1115,40)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (306,19,1116,40)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (307,20,1112,120)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (307,21,1113,120)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (307,22,1114,120)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (307,23,1115,80)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (307,24,1116,80)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (307,25,1117,40)
    INTO HR.PAYMENT_DETAILS (HEADER_ID,LINE_ID,STUDENT_ID,CUM_AMOUNT) VALUES (307,26,1118,40)
    SELECT * FROM DUAL;
    The above table STUDENT_DETAILS stores the details of all the students and each student has a unique student id.
    Another table PAYMENT_HEADER saves the payment details as a document for each month which has a header id.
    In the PAYMENT_DETAILS table, payment details are stored for the students (who made the payment for that month). This table does not save the data for a student if he has not paid on that month. This table is connected to the PAYMENT_HEADER table through a header Id.
    The view should be such that when I pass the document number, it shall show the names of all the students, but shall show payment amount only for those students who had paid under the selected document number, for other it should show null.
    I tried the following query:
    select * from (
    select sd.student_name,sd.DOB,sd.sex,sd.contact_number,pd.doc_num,pd.month_name,pd.cum_amount
    from hr.student_details sd left join
    (select hdr.doc_num,hdr.month_name,det.student_id,det.cum_amount
    from hr.payment_header hdr, hr.payment_details det
    where hdr.pay_header_id = det.header_id) pd on sd.student_id = pd.student_id)
    But when I pass the filtering values like document number, it does not show some students with null values.
    select * from (
    select sd.student_name,sd.DOB,sd.sex,sd.contact_number,pd.doc_num,pd.month_name,pd.cum_amount
    from hr.student_details sd left join
    (select hdr.doc_num,hdr.month_name,det.student_id,det.cum_amount
    from hr.payment_header hdr, hr.payment_details det
    where hdr.pay_header_id = det.header_id) pd on sd.student_id = pd.student_id)
    where doc_num = 1 or doc_num is null;
    My requirement is that, every time I use a document number in the where clause for the view it should show the paid amount against the student names who has paid on that month and for all other student it should null. Total student number is 13. So every time it should show 13 students only even when I pass document No 2 or 3.
    Seeking your help.
    Regards
    Hawker

    Hi Frank,
    as per your advice, I am omitting the 'DOB', 'GENDER','ACTIVE','CONTACT_NUMBER' fields from the 'STUDENT_DETAILS' table.
    I shall create separate tables and insert the desired output from the view for each where clause.
    First let us create three tables each for the desired out put:
    1) For the desired output I want to see when I pass doc_num as 1
    create table HR.SDT_PAY_DET_DOC_ONE (
    STUDENT_ID NUMBER(10),
    STUDENT_NAME VARCHAR2(50),
    DOC_NUM NUMBER(5),
    C_AMOUNT NUMBER(10),
    primary key(STUDENT_ID)
    2) For the desired output I want to see when I pass doc_num as 2
    create table HR.SDT_PAY_DET_DOC_TWO (
    STUDENT_ID NUMBER(10),
    STUDENT_NAME VARCHAR2(50),
    DOC_NUM NUMBER(5),
    C_AMOUNT NUMBER(10),
    primary key(STUDENT_ID)
    3) For the desired output I want to see when I pass doc_num as 3
    create table HR.SDT_PAY_DET_DOC_THREE (
    STUDENT_ID NUMBER(10),
    STUDENT_NAME VARCHAR2(50),
    DOC_NUM NUMBER(5),
    C_AMOUNT NUMBER(10),
    primary key(STUDENT_ID)
    Now I shall insert values in each of the above three tables:
    INSERT ALL
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1112,'James',1,40)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1113,'David',1,40)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1114,'Michael',1,40)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1115,'Joseph',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1116,'Juliet',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1117,'Monica',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1118,'William',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1119,'Karen',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1120,'Erika',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1121,'Noah',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1122,'Angelina',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1123,'Gabriela',null,null)
    INTO HR.SDT_PAY_DET_DOC_ONE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1124,'Sofia',null,null)
    SELECT * FROM DUAL;
    INSERT ALL
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1112,'James',2,80)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1113,'David',2,80)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1114,'Michael',2,80)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1115,'Joseph',2,40)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1116,'Juliet',2,40)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1117,'Monica',null,null)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1118,'William',null,null)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1119,'Karen',null,null)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1120,'Erika',null,null)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1121,'Noah',null,null)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1122,'Angelina',null,null)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1123,'Gabriela',null,null)
    INTO HR.SDT_PAY_DET_DOC_TWO(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1124,'Sofia',null,null)
    SELECT * FROM DUAL;
    INSERT ALL
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1112,'James',3,120)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1113,'David',3,120)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1114,'Michael',3,120)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1115,'Joseph',3,80)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1116,'Juliet',3,80)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1117,'Monica',3,40)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1118,'William',3,40)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1119,'Karen',null,null)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1120,'Erika',null,null)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1121,'Noah',null,null)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1122,'Angelina',null,null)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1123,'Gabriela',null,null)
    INTO HR.SDT_PAY_DET_DOC_THREE(STUDENT_ID,STUDENT_NAME,DOC_NUM,C_AMOUNT) VALUES (1124,'Sofia',null,null)
    SELECT * FROM DUAL;
    Thanks & Regards
    Hawker

Maybe you are looking for

  • Works fine with Safari but not with Internet Explorer

    Hi, I just created a website using iWeb, and it looks great in Safari in my Mac and iphone, but when I try to open it in Vaio using Internet Explorer, I get some problems. Sometimes the page just don't open, and when it opens, the fonts are not the s

  • Lowering my bandwidth (Qos?) setup help needed

    In order to better decide if a lower bandwidth from my provider will suffice I'm trying to find out if there's a way I can configure my router to limit bandwidth to a certain speed, so as to simulate what this lower speed will feel like. Having read

  • Sql 2008 nested case statement

    I have a question about understanding nested case statements in sql server 2008:  The sql looks like the folloiwng:  select numberofcases   from inventory   where inventory_cnt > 1000   (when select top 1     from inventory      where  inventory_cnt

  • Data Load Best Practice.

    Hi, I needed to know what is the best way to load the data from a source. Is the SQL load the best way or using data files better? What are the inherent advantages and disadvantages of the two processes? Thanks for any help.

  • How to check 10.2.0.4 system requirement

    Following are the system requirements for the installation of patch 6810189 in AIX 5.3 64-bit: 1. IY73570: IBM C++ Runtime Environment Components for AIX PTF My Issue: ====== My AIX administrator installed the patch and got the message of successful