Best way to show the following data?

Hi Experts,
I am designing a validation solution in java webdynpro.
I need to process 1000 records against certain logic and need to show the errors in the interface..Each record has more than 30 columns. I would like to get your suggestions on,
1. How to display 1000 records with nearly 30 columns? A normal table with scrolling is enough or i need to split into 2 or 3 tabs based on some grouping and show these 1000 records?
2. What would be the best way to show the error messages on these records?, Is it advisable to show the error messages on a tooltip? Is it possible to show a tooltip for each record in a table?
Any suggestions would be really helpful for my work...
Thanks in advance..
Regards,
Senthil
Edited by: Senthilkumar.Paulchamy on Jun 23, 2011 9:18 PM

Hi,
Depending on your data would use a TableColumnGroup to group your 30 columns under a common header (displaying only the most significant data)
Expanding the row would then show all 30 columns' data
To display the errors, I would display the common header rows in bold red to make them stand out from the non-erroneous rows.
Furthermore, I would add a toggle button, to alternate between 'all records' and 'erroneous records only'
Best,
Robin

Similar Messages

  • Is there a way to show the end dates of the previous month stand out?

    is there a way to show the end dates of the previous month in bold or make it stand out? so 30,31 and 1st of this month don't all look the same? For example the 30,31st get a regular text since they are a part of the the last month and this month gets bold dates? Its hard to read the dates anyways.

    Pics and more
    I may receive some form of compensation, financial or otherwise, from my recommendation or link.
    5:05 PM Sunday; January 4, 2009

  • Best way to show movie release date by year under "Movies" section

    I am a proud owner of a new 2nd gen Apple TV. It does everything I hoped. I am using IDentity 2 to capture the movie Metadata and import into a new m4v file. All data is trnsferred into iTunes. The question: What is the best way to have the release year show in Apple TV? I currently have a description, actor, director, time, etc showing. The Metadata is taken from IMDB, so the release date is listed in the video description when I "get info" in iTunes. As you can imagine, I don't want to manually input the release year in every movie file name. Any suggestions?

    No asnswer. I can't find an answer on the web either. How many more years till Apple fixes this?

  • What's the best way to model the following requirements in a data model?

    I need a data model to represent the following scenario but i just can't get my head around it.
    Class A has many Class B's. Class B's can have many Class C's and vice versa. So ultimately the relationship for the above 3 classes is as follows:
    Class A -- (1 : M) --> Class B --- (M : M) ---> Class C
    And Class C must know about its parent referenced (Class B) and also same applies to Class B where it needs to know who owns it (Class A).
    Whats the best way to construct this? In a Tree? Graph?
    Or just simply:
    Class A has list of Class B and Class B has list of Class C.
    But wouldn't this be tight dependencies?

    Theresonly1 wrote:
    Basically:
    A's owns multiple B's (B's need to know who owns it) BUT a B can be only owned by one A
    B's owns multiple C's AND C's can be owned by multiple B's (this is a many-many relationship, right?) - Again; (C's need to know who owns it)I'd reckon that you'd need some references. First, figure out the names of each tier of class. I would say maybe A is a principal/school, B's are teachers (because typically teachers only teach under one principal/ in one school), and C's are students (because teachers can have multiple students, and each student has multiple teachers). So now that you have the names, make some base classes. If I understand your problem correctly, A's really don't need to know who they own, but B's need to know who owns them, so A's don't need to have much of anything. B's have a reference to the A that owns them, but not much else because they don't need to know who they own. C's own nothing, but they are owned by multiple B's, so they have an Array of refrences to each of the B's that own them. I'd use an ArrayList, considering each could have a different amount of B's, but you could do it with an array if you tried. I'll leave it up to you how you implement everything, but I'll give you some guides to how I might do it:
    public class Principal{
    public class Teacher{
          public Principal owner;
          public Teacher(Principal owner){
                this.owner=owner;
    public class Student{
          public Teacher[] owners;
          public Student(Teacher owner...){
                owners=owner;
          public void addOwner(Teacher newOwner){
                //basically copy and paste the old array into a temporary one
                //with an extra spot, add newOwner to the end of that, then
                //make the owner array refrence that array.
    }In Student, I'm pretty sure thats how you allow an undetermined number of parameters be used, and I'm pretty sure that it comes out as an array. I hope this helps you!

  • Best way to write the following query

    Hi ,
    I've the following table structures and data..
    And i've written the query to get the records which are greater than BBB-
    But could you please hint me to write in simpler way.
    create table obj (ob_id )
    as select 1 from dual union all
    select 2 from dual union all
    select 3 from dual union all
    select 4 from dual union all
    select 5 from dual union all
    select 6 from dual
    create table og_dt (or_id , rt_cd,rt_ct_cd)
    AS SELECT 1 ,'B','BRID' FROM DUAL UNION ALL
       SELECT 1 ,'B','BRD' FROM DUAL UNION ALL
       SELECT 2 ,'BB-','ACR' FROM DUAL UNION ALL
       SELECT 2 ,'BB-','AQCR' FROM DUAL UNION ALL
       SELECT 3 ,'BBB','QYRE' FROM DUAL UNION ALL
       SELECT 4 ,'BB+','TUR' FROM DUAL UNION ALL
       SELECT 5 ,'BBB-','KUYR' FROM DUAL
    create table rt_srt (srt_ord,rt_cd,rt_ct_cd)
    as select 50 ,'B','VID' FROM DUAL UNION ALL
       SELECT 50 ,'B','BRD' FROM DUAL UNION ALL
       SELECT 40 ,'BB-','ACR' FROM DUAL UNION ALL
       SELECT 41 ,'BB-','AQCR' FROM DUAL UNION ALL
       SELECT 30 ,'BBB','QYRE' FROM DUAL UNION ALL
       SELECT 33 ,'BB+','TUR' FROM DUAL UNION ALL
       SELECT 20 ,'BBB-','KUYR' FROM DUAL
    select distinct
       from obj,og_dt,rt_srt
      where obj.ob_id=og_dt.or_id
      and og_dt.rt_cd = rt_srt.rt_cd
      and og_dt.rt_ct_cd=rt_srt.rt_ct_cd
    and rt_srt.srt_ord > all (select rt_srt.srt_ord from rt_srt
    where rt_cd='BBB-' I've used the table rt_srt twise in the above query
    Could you please hint to write it in the simple way.
    Thank you

    Here are execution plans for 3 possible solutions (incuding one you posted). Second & third solutions assume rt_srt.srt_ord is not null:
    SQL> explain plan for
      2  select  distinct *
      3    from  obj,
      4          og_dt,
      5          rt_srt
      6    where obj.ob_id = og_dt.or_id
      7      and og_dt.rt_cd = rt_srt.rt_cd
      8      and og_dt.rt_ct_cd = rt_srt.rt_ct_cd
      9      and rt_srt.srt_ord > all (
    10                                select  rt_srt.srt_ord
    11                                  from  rt_srt
    12                                  where rt_cd = 'BBB-'
    13                               )
    14  /
    Explained.
    SQL> @?\rdbms\admin\utlxpls
    PLAN_TABLE_OUTPUT
    Plan hash value: 3210303028
    | Id  | Operation              | Name   | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT       |        |     7 |   504 |    16  (25)| 00:00:01 |
    |   1 |  HASH UNIQUE           |        |     7 |   504 |    16  (25)| 00:00:01 |
    |   2 |   MERGE JOIN ANTI NA   |        |     7 |   504 |    15  (20)| 00:00:01 |
    |   3 |    SORT JOIN           |        |     7 |   385 |    11  (19)| 00:00:01 |
    |*  4 |     HASH JOIN          |        |     7 |   385 |    10  (10)| 00:00:01 |
    |*  5 |      HASH JOIN         |        |     7 |   238 |     7  (15)| 00:00:01 |
    PLAN_TABLE_OUTPUT
    |   6 |       TABLE ACCESS FULL| OBJ    |     6 |    78 |     3   (0)| 00:00:01 |
    |   7 |       TABLE ACCESS FULL| OG_DT  |     7 |   147 |     3   (0)| 00:00:01 |
    |   8 |      TABLE ACCESS FULL | RT_SRT |     7 |   147 |     3   (0)| 00:00:01 |
    |*  9 |    SORT UNIQUE         |        |     1 |    17 |     4  (25)| 00:00:01 |
    |* 10 |     TABLE ACCESS FULL  | RT_SRT |     1 |    17 |     3   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       4 - access("OG_DT"."RT_CD"="RT_SRT"."RT_CD" AND
    PLAN_TABLE_OUTPUT
                  "OG_DT"."RT_CT_CD"="RT_SRT"."RT_CT_CD")
       5 - access("OBJ"."OB_ID"="OG_DT"."OR_ID")
       9 - access("RT_SRT"."SRT_ORD"<="RT_SRT"."SRT_ORD")
           filter("RT_SRT"."SRT_ORD"<="RT_SRT"."SRT_ORD")
      10 - filter("RT_CD"='BBB-')
    Note
       - dynamic sampling used for this statement (level=2)
    31 rows selected.
    SQL> explain plan for
      2  select  distinct *
      3    from  obj,
      4          og_dt,
      5          rt_srt
      6    where obj.ob_id = og_dt.or_id
      7      and og_dt.rt_cd = rt_srt.rt_cd
      8      and og_dt.rt_ct_cd = rt_srt.rt_ct_cd
      9      and rt_srt.srt_ord > (
    10                            select  max(rt_srt.srt_ord)
    11                              from  rt_srt
    12                              where rt_cd = 'BBB-'
    13                           )
    14  /
    Explained.
    SQL> @?\rdbms\admin\utlxpls
    PLAN_TABLE_OUTPUT
    Plan hash value: 3391900174
    | Id  | Operation              | Name   | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT       |        |     1 |    55 |    14  (15)| 00:00:01 |
    |   1 |  HASH UNIQUE           |        |     1 |    55 |    14  (15)| 00:00:01 |
    |*  2 |   HASH JOIN            |        |     1 |    55 |    10  (10)| 00:00:01 |
    |   3 |    MERGE JOIN CARTESIAN|        |     2 |    68 |     6   (0)| 00:00:01 |
    |*  4 |     TABLE ACCESS FULL  | RT_SRT |     1 |    21 |     3   (0)| 00:00:01 |
    |   5 |      SORT AGGREGATE    |        |     1 |    17 |            |          |
    PLAN_TABLE_OUTPUT
    |*  6 |       TABLE ACCESS FULL| RT_SRT |     1 |    17 |     3   (0)| 00:00:01 |
    |   7 |     BUFFER SORT        |        |     6 |    78 |     3   (0)| 00:00:01 |
    |   8 |      TABLE ACCESS FULL | OBJ    |     6 |    78 |     3   (0)| 00:00:01 |
    |   9 |    TABLE ACCESS FULL   | OG_DT  |     7 |   147 |     3   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       2 - access("OBJ"."OB_ID"="OG_DT"."OR_ID" AND
                  "OG_DT"."RT_CD"="RT_SRT"."RT_CD" AND
    PLAN_TABLE_OUTPUT
                  "OG_DT"."RT_CT_CD"="RT_SRT"."RT_CT_CD")
       4 - filter("RT_SRT"."SRT_ORD"> (SELECT MAX("RT_SRT"."SRT_ORD") FROM
                  "RT_SRT" "RT_SRT" WHERE "RT_CD"='BBB-'))
       6 - filter("RT_CD"='BBB-')
    Note
       - dynamic sampling used for this statement (level=2)
    30 rows selected.
    SQL> explain plan for
      2  select  distinct obj.*,
      3                   og_dt.*,
      4                   rt_srt.srt_ord,
      5                   rt_srt.rt_cd,
      6                   rt_srt.rt_ct_cd
      7    from  obj,
      8          og_dt,
      9          (
    10           select  t.*,
    11                   max(case rt_cd when 'BBB-' then srt_ord end) over() max_srt_ord
    12             from  rt_srt t
    13          ) rt_srt
    14    where obj.ob_id = og_dt.or_id
    15      and og_dt.rt_cd = rt_srt.rt_cd
    16      and og_dt.rt_ct_cd = rt_srt.rt_ct_cd
    17      and rt_srt.srt_ord > max_srt_ord
    18  /
    Explained.
    SQL> @?\rdbms\admin\utlxpls
    PLAN_TABLE_OUTPUT
    Plan hash value: 998396165
    | Id  | Operation             | Name   | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT      |        |     7 |   476 |    11  (19)| 00:00:01 |
    |   1 |  HASH UNIQUE          |        |     7 |   476 |    11  (19)| 00:00:01 |
    |*  2 |   HASH JOIN           |        |     7 |   476 |    10  (10)| 00:00:01 |
    |*  3 |    HASH JOIN          |        |     7 |   238 |     7  (15)| 00:00:01 |
    |   4 |     TABLE ACCESS FULL | OBJ    |     6 |    78 |     3   (0)| 00:00:01 |
    |   5 |     TABLE ACCESS FULL | OG_DT  |     7 |   147 |     3   (0)| 00:00:01 |
    PLAN_TABLE_OUTPUT
    |*  6 |    VIEW               |        |     7 |   238 |     3   (0)| 00:00:01 |
    |   7 |     WINDOW BUFFER     |        |     7 |   147 |     3   (0)| 00:00:01 |
    |   8 |      TABLE ACCESS FULL| RT_SRT |     7 |   147 |     3   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       2 - access("OG_DT"."RT_CD"="RT_SRT"."RT_CD" AND
                  "OG_DT"."RT_CT_CD"="RT_SRT"."RT_CT_CD")
       3 - access("OBJ"."OB_ID"="OG_DT"."OR_ID")
    PLAN_TABLE_OUTPUT
       6 - filter("RT_SRT"."SRT_ORD">"MAX_SRT_ORD")
    Note
       - dynamic sampling used for this statement (level=2)
    27 rows selected.
    SQL> SY.
    Edited by: Solomon Yakobson on May 7, 2012 4:46 PM

  • What is the best way to capture the Form data Temporarly

    Hi,
    I have a form with lot of validations.
    Suppose user entered some of the fields and he decided to fill rest of the field later on.
    So he now want to save the record temporarly without applying any validation and next time when user logs in, the form should be populated with those invalid data and then he continues to fill up the rest of the fields, now once he SUBMIT the page..all the business validations should apply and record is commited.
    What is the best approach to acheive this??
    Have any one has the done this kind of requirement.
    Thanks,
    Deepak

    This is a little more complex as well because you'll need to save this data in a "temporary" table. Not temporary like global temporary, but just a staging table that has no constraints. In a real application, your table would not have NOT NULL constraints, but your user may not use them.
    Along with the previous post, you may need to follow this logic:
    1) If the save as draft button is clicked, you save to the table with no constraints.
    2) When someone tries to navigate to the form, you first check the "draft" table and if there is a record there, pull it. Otherwise pull from your production table.
    3) When the person clicks the commit/submit/etc button, save the data back to your "draft" table, copy the information over to the production table, and delete the temporary record.
    It's a lot of coordination honestly - my client asked for the same thing and when I explained the complexity involved for a form that only had 5 or 6 fields, they agreed. And when I say complex, I tell them it will cost them more money, yet provide little benefit. That usually solves it quickly.

  • How we can show the xml data without any attribute in a flex tree (without having any label field) ?

    how we can show the following data in a flex tree (without having any
    label field) ?
    <?xml version="1.0"?>
    <rootNode>
            <childNode>
            < subchildNode >
                    <valueNode>1000</valueNode >
                    < valueNode >999-888-777</valueNode >
                    < valueNode >STORTZ</valueNode >
                    < valueNode >PAM STORTZ</valueNode >
                    < valueNode >88 ST. MORTON ROAD</valueNode>
            </subchildNode>
            </childNode >
    </rootNode >

    as is - no how
    wrap those xml chunk with a class having clear public properties like:
    public class dataRecord {
         protected var xml:XML;
         function dataRecord(xml:XML) {
              this.xml = xml;
         public function get id ():String {
              return XML(xml.descendants("valueNode")[0]).toString();
         // and so on
    simpliest way is to change xml structure to use attributes than wrapping that thing with AS code, it's time consuming and non efficient.

  • I have lightroom 5.7. Now I have apple TV to connect my Mac to the TV scree. I wish to do a slide show on the TV. However, on the Mac, using ITunes to share photos with the TV, I cannot locate all photo files. What is the best way to save the slide show a

    I have lightroom 5.7. Now I have apple TV to connect my Mac to the TV scree. I wish to do a slide show on the TV. However, on the Mac, using ITunes to share photos with the TV, I cannot locate all photo files. What is the best way to save the slide show and put it where the Mac sharing can find it? So far, I made on one folder in Lightroom, put some photos there and I found them easily. Can this be done with a slide show? Please help quickly! Also worried that the photos I put on the new folder are hard to find afterwards, when the show is done. Where do they go when I delete from from the new folder?I am not alone

    Not that I'm aware of. You just export JPEG copies to a folder that you can point iTunes to. For instance, I have created a folder in my Pictures folder called Apple TV. And within that folder I have other folders of pictures that I can choose from in iTunes to share with Apple TV. But there doesn't seem to be any way to share a Lightroom slideshow. If you have laid to create a video file that would probably work. Apple TV is a little clunky in my opinion. Some things are a little more difficult to do now than they were a while back. I probably haven't provided you with much help, but just keep experimenting and I think you will figure it out.

  • What is the best way to kill/stop a data load?

    Hi.
    What is the best way to kill/stop a data load?
    I have a data load from my QA R/3 system that is extracting 115.000.000+ records. The problem is that the selection in the function module used in the data source does not work, and the problem was not detected because of the nature of the data on the development system.
    I could kill processes owned by my background user (on both R/3 and BW) but I risk killing other loads, and sometimes the job seems to restart if I just try to kill processes. If I remove transactional RFCs in SM58 the load does not terminate; I only skip one or more datapackages. I have also tried to change the QM-status in the monitor to red, but that does not stop the load either...
    So isn't there a nice fool-proof way of stopping a dataload?
    Best regards,
    Christian Frier

    Hi,
    There r 2 ways to kill the job.
    One is using transation RSMO locate the job and display the status tab double click on the yellow light that is shown on the line total, a pop will come 'set overall status ' is displayed select the desired status that is red and save it. Then return to the monitor page and select the header tab double ckick on the data target right click and then goto 'manage',there should be request sitting there probably with yellow lights , highlight the line with the faulty request click the delete button then click refresh button.
    Second is goto SM37 and click on the active selection and enter the jobname and then click excute the particulr job should appear highlight the jobname then click on the stop iconthat appears on the taskbar( 3 rd from left)
    hope it is clear.
    Regards-
    Siddhu

  • I am moving from PC to Mac.  My PC has two internal drives and I have a 3Tb external.  What is best way to move the data from the internal drives to Mac and the best way to make the external drive read write without losing data

    I am moving from PC to Mac.  My PC has two internal drives and I have a 3Tb external.  What is best way to move the data from the internal drives to Mac and the best way to make the external drive read write without losing data

    Paragon even has non-destriuctive conversion utility if you do want to change drive.
    Hard to imagine using 3TB that isn't NTFS. Mac uses GPT for default partition type as well as HFS+
    www.paragon-software.com
    Some general Apple Help www.apple.com/support/
    Also,
    Mac OS X Help
    http://www.apple.com/support/macbasics/
    Isolating Issues in Mac OS
    http://support.apple.com/kb/TS1388
    https://www.apple.com/support/osx/
    https://www.apple.com/support/quickassist/
    http://www.apple.com/support/mac101/help/
    http://www.apple.com/support/mac101/tour/
    Get Help with your Product
    http://docs.info.apple.com/article.html?artnum=304725
    Apple Mac App Store
    https://discussions.apple.com/community/mac_app_store/using_mac_apple_store
    How to Buy Mac OS X Mountain Lion/Lion
    http://www.apple.com/osx/how-to-upgrade/
    TimeMachine 101
    https://support.apple.com/kb/HT1427
    http://www.apple.com/support/timemachine
    Mac OS X Community
    https://discussions.apple.com/community/mac_os

  • What is the best way to mimic the data from production to other server?

    Hi,
    here we user streams and advanced replication to send the data for 90% of tables from production to another production database server. if one goes down can use another one. is there any other best option rather using the streams and replication? we are having lot of problems with streams these days they keep break and get calls.
    I heard about data guard but dont know what is use of it? please advice the best way to replicate the data.
    Thanks a lot.....

    RAC, Data Guard. The first one is active-active, that is, you have two or more nodes accessing the same database on shared storage and you get both HA and load balancing. The second is active-passive (unless you're on 11.2 with Active Standby or Snapshot Standby), that is one database is primary and the other is standby, which you normally cannot query or modify, but to which you can quickly switch in case primary fails. There's also Logical Standby - it's based on Streams and generally looks like what you seem to be using now (sort of.) But it definitely has issues. You can also take a look at GoldenGate or SharePlex.

  • Best Way to port the data from one DB to another DB using Biztalk

    Hi,
    please suggest best way to move the data from one db to another DB using biztalk.
    Currently I am doing like that, for each transaction(getting from different source tables) through receive port, and do some mapping (some custom logic for data mapping), then insert to target normalized tables(multiple tables) and back to update the status
    of transaction in source table in sourceDB. It is processing one by one.
    How/best we we can do it using  bulk transfer and update the status. Since it has more than 10000 transaction per call.
    Thanks,
    Vinoth

    Hi Vinoth,
    For SQL Bulk inserts you can always use SQL Bulk Load
    adapter.
    http://www.biztalkgurus.com/biztalk_server/biztalk_blogs/b/biztalksyn/archive/2005/10/23/processing-a-large-flat-file-message-with-biztalk-and-the-sqlbulkinsert-adapter.aspx
    However, even though a SQL Bulk Load adapter can efficiently insert a large amount of data into SQL you are still stuck with the issues of transmitting the
    MessageBox database and the memory issues of dealing with really large messages.
    I would personally suggest you to use SSIS, as you have mentioned that records have to be processed in specific time of day as opposed to when the
    records are available.
    Please refer to this link to get more information about SSIS: http://msdn.microsoft.com/en-us/library/ms141026.aspx
    If you have any more questions related to SSIS, please ask it in
    SSIS 
    forum and you will get specific support.
    Rachit

  • What is the best way to export the data out of BW into a flat file on the S

    Hi All,
    We are BW 7.01 (EHP 1, Service Pack Level 7).
    As part of our BW project scope for our current release, we will be developing certain reports in BW, and for certain reports, the existing legacy reporting system based out of MS Access and the old version of Business Objects Release 2 would be used, with the needed data supplied from the BW system.
    What is the best way to export the data out of BW into a flat file on the Server on regular intervals using a process chain?
    Thanks in advance,
    - Shashi

    Hello Shashi,
    some comments:
    1) An "open hub license" is required for all processes that extract data from BW to a non-SAP system (including APD). Please check with your SAP Account Executive for details.
    2) The limitation of 16 key fields is only valid when using open hub for extracting to a DB table. There's no such limitation when writing files.
    3) Open hub is the recommended solution since it's the easiest to implement, no programming is required, and you don't have to worry much about scaling with higher data volumes (APD and CRM BAPI are quite different in all of these aspects).
    For completeness, here's the most recent documentation which also lists other options:
    http://help.sap.com/saphelp_nw73/helpdata/en/0a/0212b4335542a5ae2ecf9a51fbfc96/frameset.htm
    Regards,
    Marc
    SAP Customer Solution Adoption (CSA)

  • Hi I would like to show a presentation written on my macbook air on my television what is the best way to connect the two ?

    Hi I would like to show a presentation written on my macbook air on my television what is the best way to connect the two ?

    Much depends on the available connectivity  of your television.
    If it has HDMI input available, all you will need is a minidisplayport to HDMI adapter, and an HDMI cable long enough to place the Mac in the desired proximity to the TV. The adapter is avaialble from the Apple Store online.
    Provided the HDMI input on the television supports sound, and the model of your MBA supports it as well, make suree that your adapter is also supportive of sound, otherwise you will need to run separate cabling from the headphone jack of the MBA to TV sound input or external speakers.
    This is a somewhat specific answer to a question without any requisite detail. You don't indicate the model or type of inputs on your TV, nor do you indicate the model/generation of MBA you have.

  • Best way to check the data

    I have a made some changes to an existing view. The change is adding a join and a new column from new table. I have created a version2 with all these changes. I want to make sure that same data exists in both the views. what is the best way to test the data.

    Bad wording, the query results could be:
    No rows selected. Same number of rows on both views and same data (considering view1 structure and view without the new column)
    At least one row selected. If the count(*) on both was different, rows returned will be at least the difference. If the count was the same, then every returned row means that there's no equivalent row on view2 and a data difference may exist in at least one of the fields, so You have to find the equivalent row on view2 to compare.
    On the query sintaxis, as view2 hava a new column and I guess it's not a constant value, You have to specify every column for the view1, and the same columns for view2, so new field isn't included and compared.

Maybe you are looking for