Stroke weight on table doesn't match non-table strokes?

Hi,
I've experienced this anomoly in CS2 and CS3.
If I set a table stroke to 0.25, and place a "non-table" stroke (eg. a line or paragraph rule) close to this table for comparison, also set to 0.25, then export to PDF and print the result, the table stroke is noticeably heavier than the other strokes. On screen (both in Indesign and Acrobat) the lines appear identical - the difference is only apparent in the print. The effect is more noticeable on strokes set to a colour other than 100% black (probably because a dot pattern is used to print rather than solid toner).
To compensate for this, I find setting the other strokes to about 0.35 gives the same appearance as a table stroke set to 0.25.
This is quite odd, isn't it?
It could be an issue with the printer (Fuji-Xerox DocuColor 5065), but I'd appreciate it if someone could check and see if they get the same result.
Thanks,
Pete.

Thanks for the reply, Scott.
This is a universal problem - all table strokes are slightly heavier than they should be.
It doesn't matter what the stroke weight is set to - it's just most noticeable with a fine stroke such as 0.25.
After some more tests, I've found that PDFs coming out of Indesign (all flavours), prints directly from Indesign, and artwork printed to Postscript from Indesign and then run through Distiller display this problem. However, PDFs opened and then resaved from Illustrator print with the table strokes at the correct weight.
So I'm concluding that the problem is with the Indesign PDF/printing engine.
Can anyone add anything further? Is this bug documented?

Similar Messages

  • How to insert the data if the data in the two tables doesn't match.

    Hi,
    I have requirement like,I have to insert the data by comparing the two tables based on the date field.If they both are equal then the data shouldn't be inserted if not the data has to be inserted.
    Can some one help me on this ASAP.
    Regards,
    Sudha
    sudha

    Hi,
    Thanks for the reply,i would like to know the way that we can design the packages
    sudha
    You have multiple ways of implementing this
    1. Using Lookup Task
    http://www.sqlis.com/sqlis/post/Get-all-from-Table-A-that-isnt-in-Table-B.aspx
    2. using SCD wizard
    http://www.bimonkey.com/2009/07/the-slowly-changing-dimension-transformation-part-1/
    3. using hashing
    http://visakhm.blogspot.in/2014/06/ssis-tips-implementing-scd.html
    Please Mark This As Answer if it solved your issue
    Please Mark This As Helpful if it helps to solve your issue
    Visakh
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • "Create Data Warehouse Tables"  doesn't create any tables ?

    We are attempting to install and configure OBIA 7.9.6 for Peoplesoft 8.9. The steps to install and configure Informtica and DAC and their respective repositories are completed. However, Step 4.12 to Create Data Warehouse Tables on Oracle database, although apparently successful (returns no error) does not create any tables or other objects in the datawarehouse schema. Apparently, "oracle_bi_dw.ctl" is being generated with NO DDL definitions only a two line header to the effect
    [DDL File]
    Version = 1Has anyone encountered a similar issue and resolved it ? (We have an SR open with Oracle Support).
    NOTE : ddlimp does connect succesfully to the database -- I created a Database ON LOGON trigger to catch and trace logons by the datawarehouse schema.
    Hemant K Chitale
    Edited by: Hemant K Chitale on Nov 20, 2009 11:54 AM

    Closing this thread.
    A Reinstall worked.

  • Error: The decapsulated inner packet doesn't match the negotiated policy in the SA

    I upgraded my ASA from 8.2(1) to 8.4(3) as I wanted to try to get Android devices to properly connect via VPN.
    After some effort, I was able to get the Android devices to connect via VPN.  However, my syslog server has a number of errors recorded that look this this:
    %ASA-4-402116: IPSEC: Received an ESP packet (SPI= 0x1E76EFA6, sequence number= 0x1F0) from x.x.x.x (user= testuser) to y.y.y.y.  The decapsulated inner packet doesn't match the negotiated policy in the SA.  The packet specifies its destination as z.z.z.z, its source as a.a.a.a, and its protocol as tcp.  The SA specifies its local proxy as y.y.y.y/255.255.255.255/udp/42246 and its remote_proxy as x.x.x.x/255.255.255.255/udp/0.
    Digging further, it seems this error might be due to a NAT issues with the VPN connections.  VPN previously worked with Cisco's VPN client on Windows, though I did not test to see if that is no longer working.  However, I made no changes in the config, except for those related to additions needed to support L2TP.  With the below config, Android clients can connect to the ASA and access the internal network, but they cannot connect to external addresses.  I'm at a loss.
    The addresses used in the config: 192.168.1.0/24 are on the internal LAN and 192.168.3.0/24 are addresses assigned to VPN clients.
    I noted in the config this line:
    access-list inside_nat0_outbound extended permit ip any 192.168.3.0 255.255.255.0
    The access list is not referenced anywhere, though it was referenced in the 8.2(1) config like this:
    nat (inside) 0 access-list inside_nat0_outbound
    I'm not sure what else changed, but I've looked over the config and I just cannot see what the issue might be.  I'm hoping somebody might be able to point out my error.
    Here's the config file (at least the parts that might be of interest):
    : Saved
    ASA Version 8.4(3)
    interface Vlan1
    nameif inside
    security-level 100
    ip address 192.168.1.1 255.255.255.0
    interface Vlan2
    nameif outside
    security-level 0
    ip address dhcp setroute
    boot system disk0:/asa843-k8.bin
    object network obj-192.168.3.0
    subnet 192.168.3.0 255.255.255.0
    object network obj_any
    subnet 0.0.0.0 0.0.0.0
    access-list outside_access_in extended permit icmp any interface outside time-exceeded
    access-list outside_access_in extended permit icmp any interface outside echo-reply
    access-list outside_access_in extended permit icmp any interface outside unreachable
    access-list outside_mpc extended permit ip any interface outside
    access-list inside_mpc extended permit ip 192.168.1.0 255.255.255.0 any
    access-list testVPN_splitTunnelAcl extended permit ip 192.168.1.0 255.255.255.0 any
    access-list inside_nat0_outbound extended permit ip any 192.168.3.0 255.255.255.0
    ip local pool VPN-Pool-1 192.168.3.1-192.168.3.254 mask 255.255.255.0
    ip verify reverse-path interface outside
    nat (inside,any) source static any any destination static obj-192.168.3.0 obj-192.168.3.0 no-proxy-arp
    object network obj-192.168.3.0
    nat (outside,outside) dynamic interface
    object network obj_any
    nat (inside,outside) dynamic interface
    access-group outside_access_in in interface outside
    crypto ipsec ikev1 transform-set ESP-AES-256-SHA esp-aes-256 esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-AES-256-MD5 esp-aes-256 esp-md5-hmac
    crypto ipsec ikev1 transform-set ESP-DES-SHA esp-des esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-3DES-SHA esp-3des esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-DES-MD5 esp-des esp-md5-hmac
    crypto ipsec ikev1 transform-set ESP-AES-192-MD5 esp-aes-192 esp-md5-hmac
    crypto ipsec ikev1 transform-set ESP-3DES-MD5 esp-3des esp-md5-hmac
    crypto ipsec ikev1 transform-set ESP-AES-128-SHA esp-aes esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-AES-192-SHA esp-aes-192 esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-AES-128-MD5 esp-aes esp-md5-hmac
    crypto ipsec ikev1 transform-set ESP-AES-128-SHA-TRANSP esp-aes esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-AES-128-SHA-TRANSP mode transport
    crypto ipsec ikev1 transform-set ESP-3DES-SHA-TRANSP esp-des esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-3DES-SHA-TRANSP mode transport
    crypto dynamic-map SYSTEM_DEFAULT_CRYPTO_MAP 10 set ikev1 transform-set ESP-AES-128-SHA-TRANSP ESP-3DES-SHA-TRANSP
    crypto dynamic-map SYSTEM_DEFAULT_CRYPTO_MAP 65535 set pfs
    crypto dynamic-map SYSTEM_DEFAULT_CRYPTO_MAP 65535 set ikev1 transform-set ESP-AES-128-SHA ESP-AES-128-MD5 ESP-AES-192-SHA ESP-AES-192-MD5 ESP-AES-256-SHA ESP-AES-256-MD5 ESP-3DES-SHA ESP-3DES-MD5 ESP-DES-SHA ESP-DES-MD5
    crypto map outside_map 65535 ipsec-isakmp dynamic SYSTEM_DEFAULT_CRYPTO_MAP
    crypto map outside_map interface outside
    crypto ikev1 enable outside
    crypto ikev1 policy 5
    authentication pre-share
    encryption aes
    hash sha
    group 2
    lifetime 86400
    crypto ikev1 policy 10
    authentication pre-share
    encryption aes-256
    hash sha
    group 2
    lifetime 86400
    group-policy testVPN internal
    group-policy testVPN attributes
    wins-server value 192.168.1.8
    dns-server value 192.168.1.8 192.168.1.4
    vpn-idle-timeout none
    vpn-tunnel-protocol ikev1
    split-tunnel-policy tunnelspecified
    split-tunnel-network-list value testVPN_splitTunnelAcl
    default-domain value test.us
    group-policy testVPNnsl2tp internal
    group-policy testVPNnsl2tp attributes
    wins-server value 192.168.1.8
    dns-server value 192.168.1.8 192.168.1.4
    vpn-idle-timeout none
    vpn-tunnel-protocol l2tp-ipsec
    group-policy testVPNns internal
    group-policy testVPNns attributes
    wins-server value 192.168.1.8
    dns-server value 192.168.1.8 192.168.1.4
    vpn-idle-timeout none
    vpn-tunnel-protocol ikev1
    username testuser password PASSWORD encrypted privilege 15
    username testuser2 password PASSWORD nt-encrypted privilege 15
    tunnel-group DefaultRAGroup general-attributes
    address-pool VPN-Pool-1
    default-group-policy testVPNnsl2tp
    tunnel-group DefaultRAGroup ipsec-attributes
    ikev1 pre-shared-key *****
    tunnel-group DefaultRAGroup ppp-attributes
    authentication ms-chap-v2
    tunnel-group testVPN type remote-access
    tunnel-group testVPN general-attributes
    address-pool VPN-Pool-1
    default-group-policy testVPN
    tunnel-group testVPN ipsec-attributes
    ikev1 pre-shared-key *****
    tunnel-group testVPNns type remote-access
    tunnel-group testVPNns general-attributes
    address-pool VPN-Pool-1
    default-group-policy testVPNns
    tunnel-group testVPNns ipsec-attributes
    ikev1 pre-shared-key *****
    tunnel-group testVPNnsl2tp type remote-access
    tunnel-group testVPNnsl2tp general-attributes
    address-pool VPN-Pool-1
    default-group-policy testVPNnsl2tp
    tunnel-group testVPNnsl2tp ipsec-attributes
    ikev1 pre-shared-key *****
    tunnel-group testVPNnsl2tp ppp-attributes
    authentication ms-chap-v2
    One last question: in order to get the connection from Android to work, I was forced to use "tunnel-group DefaultRAGroup".  Is that actually a limitation, or did I make an error that forced that requirement?  I wanted to use "tunnel-group testVPNnsl2tp".
    Thanks!

    Chris,
    This is still a bit off the mark.  I think I might be confusing the issue by including some of the VPN configuration that I had previously installed and working (e.g., two other VPN tunnel groups with split tunneling on one of them).  Let's just remove that stuff from consideration.  I actually tested the current configs just to see if they are working since the upgrade.  testVPN is working with the split tunneling, but testVPNns (no-split tunneling) does not allow external access.  I guess there is a NAT config issue there, too, but not sure what it is, yet.  I've not investigated that closely.
    I want to solve one problem at a time, though I understand there are some interdependencies.
    What I'd like to focus on right now is just the L2TP VPN connection.
    From what I've been able to understand from the documentation, what I need are these lines:
    crypto ipsec ikev1 transform-set ESP-AES-128-SHA-TRANSP esp-aes esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-AES-128-SHA-TRANSP mode transport
    crypto ipsec ikev1 transform-set ESP-3DES-SHA-TRANSP esp-des esp-sha-hmac
    crypto ipsec ikev1 transform-set ESP-3DES-SHA-TRANSP mode transport
    crypto dynamic-map SYSTEM_DEFAULT_CRYPTO_MAP 10 set ikev1 transform-set ESP-AES-128-SHA-TRANSP ESP-3DES-SHA-TRANSP
    crypto map outside_map 65535 ipsec-isakmp dynamic SYSTEM_DEFAULT_CRYPTO_MAP
    crypto map outside_map interface outside
    crypto ikev1 enable outside
    group-policy testVPNnsl2tp internal
    group-policy testVPNnsl2tp attributes
    wins-server value 192.168.1.8
    dns-server value 192.168.1.8 192.168.1.4
    vpn-idle-timeout none
    vpn-tunnel-protocol l2tp-ipsec
    tunnel-group DefaultRAGroup general-attributes
    address-pool VPN-Pool-1
    default-group-policy testVPNnsl2tp
    tunnel-group DefaultRAGroup ipsec-attributes
    ikev1 pre-shared-key P74bmqL6rT40bl5
    tunnel-group DefaultRAGroup ppp-attributes
    authentication ms-chap-v2
    crypto ikev1 policy 5
    authentication pre-share
    encryption aes
    hash sha
    group 2
    lifetime 86400
    I still want to assign the IP addresses to VPN clients out of 192.168.3.0/24.
    The tricky part is understanding exactly what NAT rules to insert and to avoid that error message I'm getting relatred to the encapsulated packets.  I tried to introduce the commands you had, but it's missing stuff that I would need for L2TP/IPSec (e.g., "mode transport").  I also don't think I want "pfs group5".  The above config "works" in that I get connected -- all negotiation is done.  It's just that packets from the VPN client are not able to go out to the Internet and I'm seeing that encapsulation error messages when I try to send a packet.
    Paul

  • Image in Lightroom doesn't match exported image viewed in Photoshop

    After all my hard work in Lightroom, I export the image to a folder on my hard drive. When I open this exported image in Photoshop and view it alongside the Lightroom version of its "twin" in Lightroom, there is a significant difference. Why is this? I want to use these images in my books, but my time is wasted if I end up with anything other than what I created in Lightroom. Shouldn't these images match exactly?

    Thanks Jao and Michael...the color management idea makes a lot of sense.
    Adding to this idea, when I open the Pshop file I get a message about the doc having an embedded profile that doesn't match current RGB working space, and the usual 3 options: Use the embedded profile, Convert color to the working space, or discard embedded profile. No matter which option I choose, none of the opened docs match the Lightroom version.
    The color balance is fine, but the colors aren't as vibrant. The Pshop photo is dull and lifeless compared to the Lightroom version. I used Clarity which added subtle brightness, and Vibrance to crank up the color in LR which added wonderful dimensionality and punch to the landscape scene, but the Pshop version has dulled these back down.

  • How to value non-table items in Forms

    Portal 3.0.7.6.2 on NT
    I am trying to display non-table items on a Form. I want to do things like: Display the Credit Card Type (non-table item) based on the Credit Card Number (table column). Use a foreigh key (table column) to retrieve related data for display/information only on the Form. The doesn't appear to be p_session variables for these 'added' Form items so I can't use the p_session functions. Any ideas??? I'd prefer a PL/SQL solution.

    Hi Michael,
    As Dmitry said that values for non table form field cannot be displayed completely using pl/sql, u can use some pl/sql and some javascript to achieve this .
    Say, take the example of the Emp table where I would like to display the department name for the deptno. Here I'll use pl/sql to retrieve the value of department name using the deptno and use javascript to display the results in the new field added to the form.
    Steps :
    1> Create a form say, on the emp table
    2> In the field formatting section, add a new item of the type text box and add a javascript event onFocus and give the foll javascript function call for the same :-
    javascript:getval(this)
    3> In the additional pl/sql section, for the "after displaying page ..." part we'll add the following pl/sql code :-
    declare
    l_dno number(4) default null;
    l_desc varchar2(100);
    begin
    l_dno := p_session.get_value_as_NUMBER
    (p_block_name => 'DEFAULT',
    p_attribute_name => 'A_DEPTNO',
    p_index => 1);
    if l_dno is not null then
    begin
    select dname into l_desc
    from scott.dept
    where deptno = l_dno;
    exception
    when others then
    null;
    end ;
    end if;
    htp.p('<script>');
    htp.p('function getval(ele){'&#0124; &#0124;chr(10)&#0124; &#0124;
    'ele.value = "'&#0124; &#0124;nvl(l_desc,' ')&#0124; &#0124;'";}');
    htp.p('</script>');
    end;
    4> Run the form and perform a query. Place the cursor on the desc field after the page reloads with results. The department name would come up.
    Hope this helps.

  • Column count doesn't match value count at row 1, unknown number of columns

    Hi,
    I am making a program to read data from excel files as the above and store them in tables. I have managed to read all the data from excel files as a string and store them in a table.
    ID Name Salary
    50 christine 2349000
    43 paulina 1245874
    54 laura 4587894
    23 efi 3456457
    43 jim 4512878
    But in my project I have several other files that have same cell that are blank as the above example
    ID Name Salary
    50 christine 2349000
    43 paulina
    laura 4587894
    23 3456457
    43 jim 4512878
    and when i ran the same program i get this exception :
    SQLException: Column count doesn't match value count at row 1
    SQLState: 21S01
    VendorError: 1136The code for creating the table and inserting the values is above:
    private static String getCreateTable(Connection con, String tablename,
                        LinkedHashMap<String, Integer> tableFields) {
                   Iterator iter = tableFields.keySet().iterator();
                   Iterator cells = tableFields.keySet().iterator();
                   String str = "";
                   String[] allFields = new String[tableFields.size()];
                   int i = 0;
                   while (iter.hasNext()) {
                        String fieldName = (String) iter.next();
                        Integer fieldType = (Integer) tableFields.get(fieldName);
                        switch (fieldType) {
                        case Cell.CELL_TYPE_NUMERIC:
                             str = fieldName + " INTEGER";
                             break;
                        case Cell.CELL_TYPE_STRING:
                             str = fieldName + " VARCHAR(255)";
                             break;
                        case Cell.CELL_TYPE_BOOLEAN:
                             str = fieldName + " INTEGER";
                             break;
                        default:
                             str = "";
                             break;
                        allFields[i++] = str;
                   try {
                        Statement stmt = con.createStatement();
                        try {
                             String all = org.apache.commons.lang3.StringUtils.join(
                                       allFields, ",");
                             String createTableStr = "CREATE TABLE IF NOT EXISTS "
                                       + tablename + " ( " + all + ")";
                             System.out.println("Create a new table in the database");
                             stmt.executeUpdate(createTableStr);
                        } catch (SQLException e) {
                             System.out.println("SQLException: " + e.getMessage());
                             System.out.println("SQLState:     " + e.getSQLState());
                             System.out.println("VendorError:  " + e.getErrorCode());
                   } catch (Exception e)
                        System.out.println( ((SQLException) e).getSQLState() );
                        System.out.println( e.getMessage() );
                        e.printStackTrace();
                   return str;
              private static void fillTable(Connection con, String fieldname,
                        LinkedHashMap[] tableData) {
                   for (int row = 0; row < tableData.length; row++) {
                        LinkedHashMap<String, Integer> rowData = tableData[row];
                        Iterator iter = rowData.entrySet().iterator();
                        String str;
                        String[] tousFields = new String[rowData.size()];
                        int i = 0;
                        while (iter.hasNext()) {
                             Map.Entry pairs = (Map.Entry) iter.next();
                             Integer fieldType = (Integer) pairs.getValue();
                             String fieldValue = (String) pairs.getKey();
                             switch (fieldType) {
                             case Cell.CELL_TYPE_NUMERIC:
                                  str = fieldValue;
                                  break;
                             case Cell.CELL_TYPE_STRING:
                                  str = "\'" + fieldValue + "\'";
                                  break;
                             case Cell.CELL_TYPE_BOOLEAN:
                                  str = fieldValue;
                                  break;
                             default:
                                  str = "";
                                  break;
                             tousFields[i++] = str;
                        try {
                             Statement stmt = con.createStatement();
                             String all = org.apache.commons.lang3.StringUtils.join(
                                       tousFields, ",");
                             String sql = "INSERT INTO " + fieldname + " VALUES (" + all
                                       + ")";
                             stmt.executeUpdate(sql);
                             System.out.println("Fill table...");
                        } catch (SQLException e) {
                             System.out.println("SQLException: " + e.getMessage());
                             System.out.println("SQLState: " + e.getSQLState());
                             System.out.println("VendorError: " + e.getErrorCode());
                   }To be more specific the error it in the second row where i have only ID and Name in my excel file and only these i want to store. The third row has only Name and Salary and no ID. How i would be able to store only the values that i have and leave blank in the second row the Salary and in the third row the ID? Is there a way for my program to skip the blanks as empty value?
    Edited by: 998913 on May 9, 2013 1:01 AM

    In an unrelated observation, it appears you are creating new database tables to hold each document. I don't think this is a good idea. Your database tables should be created using the database's utility program and not programmatically. The database schema should hardly ever change once the project is complete.
    As a design approach: One database table can hold your document names, versions, and date they were uploaded. Another table will hold the column names and data types. Another table can hold the data (type for all data = String). This way, you can join the three tables to retrieve a document. Your design will only consists of those three tables no matter how many unique documents you have. You probably should seek the advice of a DBA or experienced Java developer on exactly how structure those tables. My design is a rough layout.

  • Getting Records Returned When Criteria Doesn't Match (Using One Query)

    I have made two queries which I would like to fold into one query if possible. The first query includes all grp_id's. The second query updates grp_id's with an (active) member count. This second query has criteria in it that grabs only grp_id's that have active members. This would be similar to say, for example, a left join situation where everything is returned from table A and only those matching from table B.
    The problem is that if I put these together, they have to be linked by grp_id, so obvioulsy, if I put these queries together, the only grp_id's I will get back are the ones that have active members in them, but as I mentioned, I need all grp_id's to be shown regardless of any active membership.
    So, here is the code that doesn't work to give a better idea of what I want.
    select g00.grp_id,
    count(m06.subs_ssn)
    from gmaster g00
    join m_elig m06 on
    g00.grp_id = m06.grp_id
    and m06.eff_date < sysdate
    and m06.exp_date > sysdate
    and m06.stat_code = 'ACTIVE'
    group by g00.grp_id
    Again, this will only give me counts for grp_id's with active members in them, whereas I want all grp_id's shown and those grp_id's without active members to have a zero next to the count..
    m06 gives the member data and g00 gives the grp_id data. The m06 eff_date and exp_date and stat_code as shown above determine active members..
    Thanks for any suggestions..

    I don't know why its not working. I did make an error in the intial posting of the data which was to put NOAM in grpid 'C'. I thought this might be my error, but I corrected it and am still getting the same results.
    I have repasted the DDL statements, the query I'm using, the results and also the data as it shows when I query the two tables. (my query is based off the tables created using the posted DDL).
    I also posted QUERY B. Which shows that the removal of the eff_dt and exp_dt criteria allows grp_id to be posted in the results, obviously showing 1 record as I no longer am defining an ACTIVE member by having criteria as posted in QUERY A.
    Is my LEFT OUTER JOIN statement written correctly?
    Also of note: I could not insert the data without the apostrophes in the dates. It gave me an error (Column not allowed here). Using the apostrophes eliminates this message and allows the data insertion. At least for now its not causing an issue as the data shows correct when I query it based a simple query of the dates.
    Using PL/SQL don't know if that makes a difference..?
    =======================================================
    DATA IN TBL: GRP_ID:
    1     A
    2     B
    3     C
    4     D
    DATA IN TBL: MBR
    1     MARK     A     Y     1/1/2011     7/1/2011
    2     MARK     A     Y     7/1/2011     1/1/2012
    3     MARTY     A     Y     1/1/2011     7/1/2011
    4     MARTY     A     Y     7/1/2011     1/1/2012
    5     FRANK     B     Y     1/1/2011     7/1/2011
    6     FRANK     B     Y     7/1/2011     1/1/2012
    7     MARY     B     Y     1/1/2011     7/1/2011
    8     MARY     B     Y     7/1/2011     1/1/2012
    9     JEAN     C     Y     1/1/2011     7/1/2011
    10     JEAN     C     Y     7/1/2011     1/1/2012
    11     NOAM     D     Y     1/1/2011     7/1/2011
    As you can see, Mark, Marty should have 1 active record in grp_id 'A' (total 2), Frank and Mary should have 1 active record each 'B' (total 2), Jean should have 1 active record in 'C' grp_id, total 1) and Noam should have no active records.
    ====================================
    QUERY A:
    SELECT
    g00.grpid,
    count(m00.grpid) mbr_count
    FROM
    grp_id g00
    left outer join mbr m00 on
    g00.grpid = m00.grpid
    WHERE
    m00.eff_dt < sysdate
    AND
    m00.exp_dt > sysdate
    GROUP BY
    g00.grpid
    QUERY A RESULTS
    1     A     2
    2     B     2
    3     C     1
    QUERY B
    SELECT
    g00.grpid,
    count(m00.grpid) mbr_count
    FROM
    grp_id g00
    left outer join mbr m00 on
    g00.grpid = m00.grpid
    GROUP BY
    g00.grpid
    QUERY B RESULTS:
    1     A     4
    2     B     4
    3     C     2
    4     D     1
    =================================================
    DDL STATEMENTS
    create table mbr (mbr_name varchar(10), grpid varchar(1), stat varchar(1), eff_dt date, exp_dt date)
    insert into mbr values ('MARK', 'A','Y', '01-jan-2011', '01-jul-2011')
    insert into mbr values ('MARK', 'A','Y','01-jul-2011', '01-jan-2012')
    insert into mbr values ('MARTY', 'A','Y','01-jan-2011', '01-jul-2011')
    insert into mbr values ('MARTY', 'A','Y', '01-jul-2011', '01-jan-2012')
    insert into mbr values ('FRANK', 'B','Y','01-jan-2011', '01-jul-2011')
    insert into mbr values ('FRANK', 'B','Y','01-jul-2011', '01-jan-2012')
    insert into mbr values ('MARY', 'B','Y','01-jan-2011', '01-jul-2011')
    insert into mbr values ('MARY', 'B','Y','01-jul-2011', '01-jan-2012')
    insert into mbr values ('JEAN', 'C','Y','01-jan-2011', '01-jul-2011')
    insert into mbr values ('JEAN', 'C','Y','01-jul-2011', '01-jan-2012')
    insert into mbr values ('NOAM', 'D','Y','01-jan-2011', '01-jul-2011')
    commit

  • How to adjust path stroke weight?

    I'm unable to adjust the weight of a path. Here's a 40 sec video (no audio, 1920x1080) that demonstrates I know how to adjust the stroke weight on a path, followed by a demo of it not working when I try to adjust it:
    http://stephenbuck.org/adobe/il-point-width/point-width.html
    Here's what I do in the video:
    First, on a path that allows the change:
    Select an elipse.
    Increase the stroke weight of the elipse.
    Control-Z to undo.
    Decrease the stroke weight of the elipse.
    The above demonstrates I know how to adjust the path weight (at least I think it does).
    Next, on the path that doesn't allow the stroke width to be reduced:
    Select path.
    Increase stroke weight to demonstrate it increases.
    Control-Z to undo.
    Decrease the stroke weight to .25 to demonstrate it does not reduce in size.
    And that's the problem.
    What do I need to do to decrease the weight of the path I am trying to adjust?
    Thanks,
    Buck

    Hi Jacob.
    Yes, thank you and I've figured out what has happened. The layers panel included a layer titled "Meridians" and a sublayer titled "<Path>". Each item has an effect applied through the appearance panel, and that is what was causing the problem. I had been meatballing only the <Path> item in the layer panel, which is behind the Meridians layer item. Decreasing <Path> was not noticable because the Meridians layer item was the 5 point weight in front of it... I changed each item to contrasting colors, increased the weight of <Path> to 20pt, and then it was apparent that the Meridian layer effect was the item that needed to be resized.
    Somewhere in this process I mistakenly applied the same effect to the Layer and the Path, and that's what caused my little dilemma.
    Thanks to you and everyone else who participated in researching this problem. It's solved.
    Buck

  • Animate Stroke Weight of a vector graphic

    Hi. I'm new to adobe's programs and digital art as a whole, so bare with me. I've watched all of the tutorials on adobe's website to learn my way around and am excited to utilize these new tools. I'm trying to create an idea, but am having trouble figuring out how to do is this. I would like to create an animation of a vector graphic I created, but I would like the stroke weight to increase from .001 to about .3 or so. So It would be 300 frames. I imported my AI file to AE but there was no option to adjust the stroke weight in there. Is there an easy way to go about what I'm trying to do rather than save out 300 different files and organize them into a video? It's also composed of a complex pattern brush so I think the program is working pretty hard. It takes like 10 seconds or so for a big increase in stroke weight to load, but there is lots of detail so it makes sense I guess. Thanks for your time.

    Thank you. I made a simple vector to test your method. This is exactly what I was trying to do. However, I think my image is too complex for the program to handle. I waited 53 minutes for after effects to try and load the shapes before force quitting.. Do you think I need to just be more patient and give it a few hours? My other concern with that though is if it manages to load all the shapes if I'm going to be experiencing ultra lag trying to do any editing to the image once it's fully loaded. My image is metatron's cube composed of very tiny metatron cubes. I'd upload the file for someone else to experiment with but it doesn't look like these forums have attachment options. Any advice on this issue? I'm running a 2011 iMac with 16gb ram. The activity monitor only shows about 1-4gb of ram being used during this process.

  • Serial number doesn't match product

    My coworker downloaded Adobe Acrobat XI Standard on our computers recently and now whenever I try to open up a PDF file a pop up screen appears with the heading "Serial Number Upgrade" asking me to chose a product and type in the serial number I have tried matching every single product option with my serial number and none of them work. The screen does say my serial number is correct but it doesn't match any of the products. I'm not sure what to do at this point...?

    Ok, so what you are saying is we bought the upgraded Adobe Acrobat, but we would need to buy the previous version if we want the serial number to work?
    King & Wood, P.A.
    1701 Hermitage Blvd., Ste. 104
    (850) 580-7711
    (850) 205-4501 fax
    Privileged & Confidential: This email is intended for the named recipient(s) only and may contain information that is proprietary, privileged, confidential, or otherwise legally exempt from disclosure. If you are not the named addressee, you are not authorized to read, print, retain, copy, or disseminate this message or any part of it. Nothing in this email is intended to constitute a waiver of any privilege or the confidentiality of this message. If you have received this email in error, please notify the sender immediately and delete this message. Circular 230 Disclosure:To ensure compliance with Treasury Department regulations, we advise you that, unless expressly indicated, any federal tax advice contained in this message or any attachments cannot be used for the purpose of avoiding penalties imposed by the Internal Revenue Code, or (ii) promoting, marketing, or recommending to another party any matters addressed herein

  • Datawarehouse query detail report result doesn't match summary report.

    Hello,
    In our DW, we run a detail report query ( like below No. 1) , while in the summary report, in which the query is partially based on aggregated table T_AGG_UNC_GROSS (like below No.2), the query result doesn' match the query result of the details report. (supposed to match exactly)
    To pick out the problem why summary report (like below NO.3) doesn't match details report. I put two filters in the details query with the filter is "year 2007", filter location is "California")
    I can get 295 selected.
    while I put above two filters in summary report, get no row selected.
    I suspect aggregation table has some problem (wrong aggregated).
    Is there any method I can pick out the problem behind?
    thanks a lot in advance.
    --1.detail report
    select D1.c14 as c1,
    D1.c13 as c2,
    D1.c12 as c3,
    D1.c11 as c4,
    D1.c10 as c5,
    D1.c15 as c6,
    D1.c9 as c7,
    D1.c8 as c8,
    D1.c7 as c9,
    D1.c6 as c10,
    D1.c5 as c11,
    D1.c16 as c12,
    D1.c17 as c13,
    D1.c18 as c14,
    D1.c19 as c15,
    D1.c4 as c16,
    D1.c3 as c17,
    D1.c20 as c18,
    D1.c2 as c19,
    D1.c21 as c20,
    D1.c22 as c21,
    D1.c1 as c22
    from
    (select /*+ full(T278878) full(T144662) full(T144546) full(T144557) full(T144665) full(T155155) */ sum(T144442.FINAL_BILL_AMOUNT) as c1,
    sum(T144400.WRITE_OFF_AMOUNT) as c2,
    case when T144416.FMCI_IND = 'Y' then 'Yes-fMCI' else 'No-fMCI' end as c3,
    T144442.IS_FIOS_FLAG as c4,
    T314105.DATE_DESC as c5,
    T144442.CYCLES_DELINQUENT as c6,
    T144590.DATE_DESC as c7,
    T144564.DATE_DESC as c8,
    T144557.LOB_CODE as c9,
    case when T144442.IS_NEVER_PAY_FLAG = 1 then 'Yes-NP' else 'No-NP' end as c10,
    T144416.CLASS_OF_SERVICE as c11,
    T144416.ACCOUNT_NAME as c12,
    T144416.BTN as c13,
    T144416.ACCOUNT_NUMBER as c14,
    T144634.ACCOUNT_STATUS as c15,
    T144687.CSG_ID as c16,
    T441666.DISCONNECT_REASON_DESC as c17,
    T144612.LOB_CODE as c18,
    T144612.STATE_CODE as c19,
    T144640.DATE_DESC as c20,
    T144586.PRIMARY_CIC_CATEGORY as c21,
    T144586.SECONDARY_CIC_DESC as c22
    from
    V_D_STATE T144612 /* V_D_STATE_UNCGRO */ ,
    D_DISCONNECT_REASON T441666 /* D_DISCONNECT_REASON_UNCG */ ,
    D_CUSTOMER_SERVICES_GROUP T144687 /* D_CUSTOMER_SERVICES_GROUP_UNCGRO */ ,
    V_D_ACCOUNT_STATUS_FINAL T144634 /* V_D_ACCOUNT_STATUS_UNCGRO */ ,
    D_DATE T144640 /* D_DATE_UNCGRO */ ,
    V_D_CARRIER_CODES T144586 /* V_D_CARRIER_CODES_UNCGRO */ ,
    F_UNCOLLECTIBLES T144400 /* F_UNCOLLECTIBLES_UNCGRO */ ,
    D_ACCOUNT T144416 /* D_ACCOUNT_UNCGRO */ ,
    D_ACCOUNT_DETAIL T144442 /* D_ACCOUNT_DETAIL_UNCGRO */ ,
    D_LINE_OF_BUSINESS T144557 /* D_LINE_OF_BUSINESS_UNCGRO */ ,
    D_DATE T144564 /* D_DATE_UNCGRO_CONNECT */ ,
    D_DATE T144590 /* D_DATE_UNCGRO_FINAL */ ,
    D_DATE T314105 /* D_DATE_UNCGRO_DISCONNECT */
    where ( T144400.ORGANIZATION_ID = T144612.ORGANIZATION_ID and T144416.ACCOUNT_STATUS_ID = T144634.ACCOUNT_STATUS_ID and T144400.WRITE_OFF_DATE = T144640.DATE_ID and T144400.CIC_CODE_ID = T144586.CIC_CODE_ID and T144416.CUSTOMER_SERVICE_GROUP = T144687.CSG_CD and T144400.ACCOUNT_ID = T144442.ACCOUNT_ID and T144400.ACCOUNT_ID = T144416.ACCOUNT_ID and T144400.DISCONNECT_DATE = T314105.DATE_ID and T144400.WRITEOFF_REFERRAL_IND = 'WRO' and T144416.DISCONNECT_REASON = T441666.DISCONNECT_REASON and T144612.U_STATE_DESC = 'California - BUSINESS' and T144640.YEAR_DESC = 'Year 2007' and T441666.DISCONNECT_REASON_DESC <> 'G' and T144416.FINAL_BILL_DATE = T144590.DATE_ID and T144416.CONNECTION_DATE = T144564.DATE_ID and T144416.LOB_ID = T144557.LOB_ID and T441666.DISCONNECT_REASON_DESC <> 'H' )
    group by T144416.ACCOUNT_NAME, T144416.ACCOUNT_NUMBER, T144416.BTN, T144416.CLASS_OF_SERVICE, T144442.IS_FIOS_FLAG, T144442.CYCLES_DELINQUENT, T144557.LOB_CODE, T144564.DATE_DESC, T144586.PRIMARY_CIC_CATEGORY, T144586.SECONDARY_CIC_DESC, T144590.DATE_DESC, T144612.LOB_CODE, T144612.STATE_CODE, T144634.ACCOUNT_STATUS, T144640.DATE_DESC, T144687.CSG_ID, T314105.DATE_DESC, T441666.DISCONNECT_REASON_DESC, case when T144416.FMCI_IND = 'Y' then 'Yes-fMCI' else 'No-fMCI' end , case when T144442.IS_NEVER_PAY_FLAG = 1 then 'Yes-NP' else 'No-NP' end
    ) D1;
    295 rows selected.
    --2. aggregation table T_AGG_UNC_GROSS
    create table T_AGG_UNC_GROSS parallel nologging as
    SELECT a.account_id,a.ORGANIZATION_ID,a.UNC_LOB_ID,
    MAX (a.WRITEOFF_REFERRAL_IND) AS WRITEOFF_REFERRAL_IND,
    MAX (a.WRITE_OFF_DATE) AS WRITE_OFF_DATE,
    COUNT (1) AS NO_OF_UNC,
    MAX(a.DISCONNECT_DATE) as DISCONNECT_DATE,
    sum(a.WRITE_OFF_AMOUNT) as WRITE_OFF_AMOUNT
    FROM f_uncollectibles a
    GROUP BY a.account_id,a.ORGANIZATION_ID,a.UNC_LOB_ID;
    ---3. sumarry resport (query)
    select SUBSTR (f.WRITE_OFF_DATE, 5, 2)
    || '/'
    || SUBSTR (f.WRITE_OFF_DATE, 7, 2)
    || '/'
    || SUBSTR (f.WRITE_OFF_DATE, 0, 4)
    writeoff_date,
    'ALL CIC',
    o.U_STATE_DESC,
    s.U_ACCOUNT_STATUS,
    rl.RISK_LEVEL,
    d.Amount_Range_ID || ' - WR',
    f.WRITEOFF_REFERRAL_IND,
    fi.fios_flag_desc,
    DR.DISCONNECT_REASON || '-' || DR.DISCONNECT_REASON_DESC,
    b.BEHAVIOR_SCORE || ' - B',
    cs.csg_cd || ' - CG',
    CASE WHEN d.IS_NEVER_PAY_FLAG = 1 THEN 'Yes-NP' ELSE 'No-NP' END
    AS Neverpay_Ind,
    CASE WHEN u.FMCI_IND = 'Y' THEN 'Yes-fMCI' ELSE 'No-fMCI' END
    AS FMCI_IND,
    lob.DESCRIPTION,
    COUNT (f.Account_id) No_of_Accounts,
    SUM (d.NO_DENIAL_NOTICES) NO_DENIAL_NOTICES,
    SUM (d.CYCLES_DELINQUENT) CYCLES_DELINQUENT,
    sum(f.WRITE_OFF_AMOUNT) WRITE_OFF_AMOUNT ,
    sum(d.FINAL_BILL_AMOUNT) FINAL_BILL_AMOUNT
    FROM t_agg_unc_gross f,
    d_account u,
    d_account_detail d,
    D_DISCONNECT_REASON DR,
    d_writeoff_amount_range a,
    v_d_account_status_final s,
    D_BEHAVIOR_RANGE b,
    v_d_state o,
    d_Fios_Indicator fi,
    d_risk_level rl,
    d_customer_services_group cs,
    D_UNCOLLECTIBELS_LOB lob,
    D_DATE
    WHERE f.account_id = u.account_id
    AND u.account_id = d.account_id
    AND d.Amount_Range_ID = a.wo_range_id
    AND d.IS_FIOS_FLAG = fi.FIOS_FLAG
    AND u.Account_Status_ID = S.ACCOUNT_STATUS_ID
    AND u.RISK_LEVEL_ID = rl.risk_level_id
    AND o.organization_id = f.organization_id
    AND dr.DISCONNECT_REASON = u.DISCONNECT_REASON
    AND d.BEHAVIOR_SCORE12 = b.BEHAVIOR_SCORE
    AND u.CUSTOMER_SERVICE_GROUP = cs.csg_cd
    AND Dr.disconnect_reason NOT IN ('H', 'G')
    AND lob.LOB_UNC_ID =f.UNC_LOB_ID
    AND o.U_STATE_DESC = 'California - BUSINESS'
    AND D_DATE.YEAR_DESC = 'Year 2007'
    and  u.FINAL_BILL_DATE = D_DATE.DATE_ID
    --and u.CONNECTION_DATE = D_DATE.DATE_ID
    GROUP BY SUBSTR (f.WRITE_OFF_DATE, 5, 2)
    || '/'
    || SUBSTR (f.WRITE_OFF_DATE, 7, 2)
    || '/'
    || SUBSTR (f.WRITE_OFF_DATE, 0, 4),
    'ALL CIC',
    o.U_STATE_DESC,
    s.U_ACCOUNT_STATUS,
    rl.RISK_LEVEL,
    d.Amount_Range_ID || ' - WR',
    f.WRITEOFF_REFERRAL_IND,
    fi.fios_flag_desc,
    DR.DISCONNECT_REASON || '-' || DR.DISCONNECT_REASON_DESC,
    b.BEHAVIOR_SCORE || ' - B',
    cs.csg_cd || ' - CG',
    lob.DESCRIPTION,
    CASE WHEN d.IS_NEVER_PAY_FLAG = 1 THEN 'Yes-NP' ELSE 'No-NP' END,
    CASE WHEN u.FMCI_IND = 'Y' THEN 'Yes-fMCI' ELSE 'No-fMCI' END;
    no row selected.
    Edited by: ROY123 on Jul 23, 2010 12:17 PM

    fixed.

  • CS6: Object Find/Change-Stroke weight?

    Hi all,
    First time using the find/change Object query in CS6.
    I'd like to search for and decrease a specific stroke weight on all frames in my document (in this case they are all graphic frames, but I'm unable to get it to work)
    I've tried searching for all frames, text frames, graphic frames, within the Document, but the query is not finding any of the bad strokes. (I watched a Lynda.com tutorial and it looked so easy, but I can't figure out why it isn't searching for them. (none of the objects are locked)
    Is there an attribute I need to uncheck in the Find Object Format Options to ensure it only searches of the stroke? Thanks for any input.
    Here's an example:

    Thank you! That worked.

  • Setting stroke weight of adiacent cell

    Hello,
    I would like to know if there's a way to do this:
    I have a series of tables which have to be formatted in a specific way. Unfortunately, with the table formatting options I'm able to do only a part of these formattings; so the question is: how can I set the stroke weight of a cell border based on an adiacent cell?
    See attached image for explanation
    Hope someone can help me!
    Thanks in advance

    Hi,
    since nobody gave me answer, I answered myself with this JavaScript:
    var myDoc = app.activeDocument;
    var totTxtFrames = myDoc.textFrames;
    for(i=0; i<totTxtFrames.length; i++){
        myTxtFrame = totTxtFrames[i];
        totTables = myTxtFrame.tables;
        for(j=0; j<totTables.length; j++){
            myTable = totTables[j];
            totRows = myTable.rows;       
            for(k=0; k<totRows.length; k++){
                myRow = totRows[k];
                myRowIndex = myRow.index;
                totCells = myRow.cells;
                for(n=0; n<totCells.length; n++){
                        myCell = totCells[n];
                        if(myCell.rowSpan > 1){
                            myCellSpan = myCell.rowSpan;
                            if(myCell.bottomEdgeStrokeWeight == 0.5){
                                    totRows.item(myRowIndex + myCellSpan - 1).bottomEdgeStrokeWeight = 0.5;
                        else{
                            myCellSpan = myCell.rowSpan;
                            if(myCell.bottomEdgeStrokeWeight == 0.5){
                                    totRows.item(myRowIndex + myCellSpan - 1).bottomEdgeStrokeWeight = 0.5;
    Hope might help someone with the same problem
    Cheers!

  • Why are my stroke weights not scaling?

    I'm new to InDesign. Just installed it and playing for the first time.
    I'm having issues with scaling the stroke weight of vector objects. I create an object or path, but when I go to scale it, the weight doesn't scale too. I've made sure that the 'adjust stroke weight when scaling' is ticked.
    What am I doing wrong?
    Really appreciate any help I can get on this.
    Thanks

    There have been several discussions reporting the same problem.
    Re: Photos with iOS8.1 Event and Face out of sequence
    There is no known fix. Send feedback to Apple to request a better photo sync, that will respect the sorting of faces albums.
    http://www.apple.com/feedback/iphoto_ios.html

Maybe you are looking for