Clearing hash value in GSS for a answer group

Hi,
I have a GSS rule pointing to an answer group with 2 answers. The rule uses hashed as the predictor to load balance requests between the 2 answers configured. The query for GSS comes from 2 clients. I have an issue where GSS is returning the same answer to queries from both the sources. Although both the answers are available, it seems to prefer only one answer. If I change the load balancing algorithm from hashed to Round robin, it starts returning the IP address of both the answers in round robin. The moment I change it back to Hashed, it returns only one IP address to both the sources. I suspect that the hash table is corrupted and may need to be cleared. Is there any way to do this or fix this issue? Any help on this would help.
Regards
Ralph

Hi Ralph,
When the GSS uses the hashed balance method, elements of the client's DNS proxy IP address
and the requesting client's domain are extracted to create a unique value, referred to as
a hash value. The unique hash value is attached to and used to identify a VIP that is
chosen to serve the DNS query.
The use of hash values makes it possible to "stick" traffic from a particular requesting
client to a specific VIP, ensuring that future requests from that client are routed to the
same VIP. This type of continuity can be used to facilitate features, such as online
shopping baskets, in which client-specific data is expected to persist even when client
connectivity to a site is terminated or interrupted.
The GSS supports the following two hashed balance methods. You can apply one or both
hashed balance methods to the specified answer group.
• By Source Address—The GSS selects the answer based on a hash value created from the
source address of the request.
• By Domain Name—The GSS selects the answer based on a hash value created from the
requested domain name.
http://www.cisco.com/en/US/docs/app_ntwk_services/data_center_app_services/gss4400series/v1.2/configuration/guide/Intro.html#wp1102666
For even load distribution you can use leastconns or round-robin
Regards,
Siva

Similar Messages

  • Clear text value not allowed for credentials in config.xml

    I am in the process of migrating our J2EE application from Weblogic 8.1 to Weblogic 10 MP1 and have converted config.xml to the new schema. When starting up in production mode, I see the following error which causes the server to shutdown:
    <Server failed. Reason: [Management:141266]Parsing Failure in config.xml: java.lang.IllegalArgumentException: In production mode, it's not allowed to set a clear text value to the property: CredentialEncrypted of SecurityConfigurationMBean>
    In 8.1 it seems that it would fill in the credentials based on the username and password in boot.properties so it was possible for our installation program to leave these entries blank. However, when starting in 10.0 it doesn't do this. Below is an excerpt from config.xml.
    <security-configuration>
    <name>xyz</name>
    <realm>
    <sec:authentication-provider xsi:type="wls:default-authenticatorType"></sec:authentication-provider>
    <sec:authentication-provider xsi:type="wls:default-identity-asserterType">
    <sec:active-type>AuthenticatedUser</sec:active-type>
    </sec:authentication-provider>
    <sec:role-mapper xsi:type="wls:default-role-mapperType"></sec:role-mapper>
    <sec:authorizer xsi:type="wls:default-authorizerType"></sec:authorizer>
    <sec:adjudicator xsi:type="wls:default-adjudicatorType"></sec:adjudicator>
    <sec:credential-mapper xsi:type="wls:default-credential-mapperType"></sec:credential-mapper>
    <sec:cert-path-provider xsi:type="wls:web-logic-cert-path-providerType"></sec:cert-path-provider>
    <sec:cert-path-builder>WebLogicCertPathProvider</sec:cert-path-builder>
    <sec:user-lockout-manager></sec:user-lockout-manager>
    <sec:security-dd-model>Advanced</sec:security-dd-model>
    <sec:combined-role-mapping-enabled>false</sec:combined-role-mapping-enabled>
    <sec:name>myrealm</sec:name>
    </realm>
    <default-realm>myrealm</default-realm>
    <credential-encrypted></credential-encrypted>
    <web-app-files-case-insensitive>os</web-app-files-case-insensitive>
    <compatibility-connection-filters-enabled>true</compatibility-connection-filters-enabled>
    <enforce-strict-url-pattern>false</enforce-strict-url-pattern>
    </security-configuration>
    Is anyone else trying to do this?
    Henry

    Hi henry,
    While running a particular instance in the production mode(wls 10.x),prior to upgrading the domain you need to upgrade the security provider(all the mbeans ) you use in the previous versions and later on if you follow the domain upgrade you may find a solution for this issue...

  • Differences between sql id and hash value

    Hello All,
    I have read the following from Tom's web site. Ask Tom &amp;quot;What is a child cursor&amp;quot;
    Dear Tom,
    Appreicate your services to the oracle community.
    What is the column combinations which is leading to SQL_ID.  is SQL_ID = hash_value + address ?  Regards Lalitha
    Followup   August 28, 2012 - 1pm UTC:
    sql id = hash( sql statement )
    hash value is maintained for backward compatibility, sqlid is a new, better hash.
    I am just wondering what did he mean when he said "hash value is maintained for backward compatibility, sqlid is a new, better hash."?
    If both of them are the output of hash function, could someone please explain clearly what is the differences between SQL_ID and HASH_VALUE?
    Thanks for your help.

    SQL_ID is just a fancy representation of hash value | Tanel Poder&amp;#039;s blog: IT &amp;amp; Mobile for Geeks and Pro…
    Regards
    Jonathan Lewis

  • Hash value in report

    Hi sdn,
    I'm getting hash value in report for one infoobject,that is for 0customer .Could you pls let me know how to remove that particular problem.
    thanks
    R

    Hi
    So u have records as follows:
    ckf            Customer Text       Value
    Base         # Not Assigned      -59,286.60
    value
    So as for the above...if there is no data as the JR.. said,it will display in that row as # or Not Assigned.
    1.So if u want dont want to display it & then u can EXCLUDE that # values in ur FILTER as other guy says.
    2.Do as per JR..
    3. U cannot totally remove tht frm ur report...becoz..where there is dependent char. which has data
    Hope this helps.

  • Who will generate hash value

    Hi Everyone,
    I need some information about hash value getting generated for every new SQl
    I think that PGA will generate hash value every time
    thanks in advance
    Shareef

    912856 wrote:
    Hi Everyone,
    I need some information about hash value getting generated for every new SQl
    I think that PGA will generate hash value every time
    PGA would "generate" the value? Isn't it that the hash-values are generated by hash functions ? If you want information about how the parsing works, please read,
    http://docs.oracle.com/cd/E11882_01/server.112/e16638/optimops.htm#i21299
    And,
    http://docs.oracle.com/cd/E11882_01/server.112/e25789/sqllangu.htm#CNCPT1740
    Aman....

  • How to clear alert when a null value is returned for a metric column?

    Hey all,
    Recently, I've encountered a problem about clearing alert when a null value is returned for a metric column. From Oracle official doc, it seems that NO_CLEAR_ON_NULL could be used for this propose. However, it doesn't.
    Here is the line from doc, you can see when NO_CLEAR_ON_NULL is set as FALSE, the null value will clear alert, and the default value for NO_CLEAR_ON_NULL is FALSE.
    NO_CLEAR_ON_NULL: This attribute is used to control severity clearing when a null value is returned for a metric column. It defaults to FALSE with the behavior that a null value ends up clearing previous alert severities. With a TRUE value for this attribute, null values will be skipped in severity evaluations without clearing the severity.
    TRUE | FALSE (default)
    STATELESS_ALERTS: This attribute if set to TRUE indicates to EM that alerts on this
    column will not have corresponding clears. This allows the UI to decide whether to
    allow users to manually clear alerts on this column.
    TRUE | FALSE (default)Here is my code for the metric,
    -- meta definition
      <Metric NAME="Flushing_Table_Overtime" TYPE="TABLE" HELP="NO_HELP">
        <Display>
          <Label NLSID="mmd_ip">Flushing Table Overtime</Label>
        </Display>
        <TableDescriptor>
          <ColumnDescriptor NAME="State" TYPE="STRING" IS_KEY="FALSE">
            <Display>
              <Label NLSID="mmd_rc_State">State</Label>
            </Display>
          </ColumnDescriptor>
          <ColumnDescriptor NAME="Time_Seconds" TYPE="NUMBER" IS_KEY="FALSE">
            <Display>
              <Label NLSID="mmd_fto_Time_Seconds">Flush Table consume Seconds</Label>
            </Display>
          </ColumnDescriptor>
        </TableDescriptor>
        <ExecutionDescriptor>
          <GetTable NAME="MysqlProcessAggregate"/>
          <GetView NAME="v_flushing_tables" FROM_TABLE="MysqlProcessAggregate">
            <Column NAME="State" />
            <Column NAME="Time_Seconds" />
            <Filter COLUMN_NAME="State" OPERATOR="EQ">Flushing tables</Filter>
          </GetView>
        </ExecutionDescriptor>
      </Metric> 
    -- coll definition
      <CollectionItem NAME="Flushing_Table_Overtime">
        <Schedule>
          <IntervalSchedule INTERVAL="10" TIME_UNIT="Min" />
        </Schedule>
        <MetricColl NAME="Flushing_Table_Overtime">
          <Condition COLUMN_NAME="Time_Seconds"
                     CRITICAL="30" OPERATOR="GT"
                     MESSAGE="There is a Flushing Table command has exceeded %critical_threshold% seconds."
                     MESSAGE_NLSID="msg_Flushing_Table_Overtime" />
        </MetricColl>
      </CollectionItem>If this feature can not be achieved by coding XML, then is there a way to clear alert with Oracle Package? I have confirm neither sysman.em_severity.delete_current_severity nor sysman.em_severity.clear_alerts could do it.
    Would anyone give me a hint?
    Thanks in advance!
    Best wishes,
    Satine

    > I am executing in this manner
    var RESULT BOOLEAN;
    BUT I am unable to create variable with boolean data type.
    WARNING. Do not confuse SQL*Plus with PL/SQL.
    There is no variable command in PL/SQL. PL/SQL does support the boolean data type.
    There is a variable command in SQL*Plus. It allows one to define bind variables. It is limited in the data types it supports. It does not support boolean as a bind variable type.
    SQL*Plus is a CLI (Command Line Interface) client tool. It has a very limited vocabularly, allowing you to (primarily):
    a) configure its local environment (e.g. size of the terminal ito number of characters per lines)
    b) configure its bevahiour (e.g. spooling data to a file)
    c) defining substitution and bind variables
    And that's it. It is not PL/SQL. It is not SQL. It takes the PL/SQL and SQL you enter, and submit that to the Oracle server. (it does some basic parsing of the data to substitute variables and bind variables where applicable)
    Do not confuse this CLI client tool with the server side PL/SQL or SQL languages.

  • Calculating hash values for really big files

    I am using the following code to calculate the hash values of files
    public static String hash(File f, String algorithm)
                throws IOException, NoSuchAlgorithmException {
            if (!f.isFile()) {
                throw new IOException("Not a file");
            RandomAccessFile raf = new RandomAccessFile(f, "r");
            byte b[] = new byte[(int) raf.length()];
            raf.readFully(b);
            raf.close();
            MessageDigest messageDigest = MessageDigest.getInstance(algorithm);
            messageDigest.update(b);
            return toHexString(messageDigest.digest());
        }Now the problem is, for really big files, 100 MB or over, I get an OutOfMemoryError.
    I have used the -Xms and -Xms options to increase the JVM heap size, and untimately made it to work. However, I think this is lame and there is also a limit to the -Xmx option.
    Is there any other way I can calculate the hash values of these really big files?
    Thanks a lot in advance.

    why do u open the file the way u do ?
    why to u upload ALL the file AT ONCE into the memory ?
    i would do it like this:
    FileInputStream fis = new FileInputStream (f);
    int fileSize = f.available();
    byte buffer[] = new byte[1000];
    MessageDigest messageDigest = MessageDigest.getInstance(algorithm);
    for(int read = 0;read < fileSize;read +=1000;)
    if(fis.available() > 1000)
    fis.read(buffer, read, 1000);
    else if(fis.available() > 0)
    fis.read(buffer, read, fis.available());
    else
    break;
    messageDigest.update(b);
    fis.close();
    return toHexString(messageDigest.digest());

  • Problem w. hash-value calculation for CLOB with DBMS_UTILITY.GET_HASH_VALUE

    Hello Oracle-Experts,
    I had to calculate hash-values for a corrorponding CLOB-field (see my post 'Buffer to small ORA-22835 error after migration from 9i to 10g' in forum 'database general').
    I calculate the hash-values with the DBMS_UTILITY.GET_HASH_VALUE, e.g:
         SQL> SELECT DBMS_UTILITY.GET_HASH_VALUE(LPAD('X',3998,'X'),1,POWER(2,30)) FROM dual;
         DBMS_UTILITY.GET_HASH_VALUE(LPAD('X',3998,'X'),1,POWER(2,30))
         1053896858
    Because the calculation failed with 10g I had taken a closer look at this function and realised the following results:
    VALUE                                             DBMS_UTILITY.GET_HASH_VALUE(VALUE,1,POWER(2,30))
    LPAD('X',3997,'X') 557754150
    LPAD('X',3998,'X')      1053896858
    LPAD('X',3999,'X')          888036750
    LPAD('X',4000,'X') 162062978
    LPAD('X',4001,'X')          162062978
    LPAD('X',4002,'X') 162062978
    LPAD('X',10000,'X') 162062978
    It seems to me that I can't use this function for clob-values with a length greater than 4000 characters because of collisions. Maybe someone with experience
    can give me a hint to handle this problem. Worst case i had to write my own CLOB_2_HASH function.
    TIA + Best regards
    Matthias

    Yeah, the 4000-byte limit would be a factor in working CLOBs. I if you only had 4000 bytes you would not need a CLOB in the first place.
    If you don't find something better then writing your own function might be the thing to do.

  • I want to clear the values for the controls

    Hi all,
            After Adding Process, i want to clear the values for the controls in the form.
    Any one help me.

    If you use 2005, you can use the Formdata Event and catch the FORM_DATA_ADD event... In there you can clear the data

  • Create hash value for clob column ?

    Hi,
    does anybody know a way to calculate a hash value for a clob column (9i) ?
    DBMS_UTILITY.GET_HASH_VALUE could only handle varchar2(4000).
    Thank you!
    Regrads
    Michael
    Message was edited by:
    mseiwert

    I can't reproduce it on my 10.2.0.4.0. CTL file:
    load data
    INFILE *
    Replace into table samp
    fields terminated by ","
    trailing nullcols
    no,
    col1 Char(100000000) ,
    col2 Char(100000000) enclosed by '"' and '"'
    BEGINDATA
    1,asdf,"assasadsdsdsd""sfasdfadf""sdsdsa,ssfsf"
    2,sfjass,"dksadk,kd,ss""dfdfjkdjfdk""sasfjaslaljs"Loading:
    SQL> Create table samp
      2  (
      3  no number,
      4  col1 clob,
      5  col2 clob
      6  );
    Table created.
    SQL> host sqlldr scott/tiger control=c:\temp\samp.ctl log=c:\temp\samp.log
    SQL> select * from samp
      2  /
            NO
    COL1
    COL2
             1
    asdf
    assasadsdsdsd"sfasdfadf"sdsdsa,ssfsf
             2
    sfjass
    dksadk,kd,ss"dfdfjkdjfdk"sasfjaslaljs
            NO
    COL1
    COL2
    SQL> SY.

  • Plan hash value for two queries!

    Hi,
    DB : Oracle 11g (11.2.0.3.0)
    OS: RHEL 5
    I have two question:
    1. Can two queries have same plan hash value? I mean I have below two queries:
    SELECT /+ NO_MERGE */ MIN(payor.next_review_date)*
    * FROM payor*
    * WHERE payor.review_complete = 0*
    * AND payor.closing_date IS NULL*
    * AND payor.patient_key = 10;*
    and
    SELECT  MIN(payor.next_review_date)
    * FROM payor*
    * WHERE payor.review_complete = 0*
    * AND payor.closing_date IS NULL*
    * AND payor.patient_key = 10;*
    When I tried to review the execution plan for both queries, the plan hash value remain same. Does it mean that execution plan for both queries are same? If yes, then how Oracle understands or changes the execution plan based on hint. If no then what plan hash value represents?
    2. If the execution plan with hint and without hint is same except for a given query except no.of rows and bytes. Does it mean that query with less rows and bytes scanned is better?
    Thanks in advance
    -Onkar

    Hi,
    there are two different things. One is EXPLAIN PLAN, which is how the optimizer thinks the query will be executed. It contains some estimates of cost, cardinalities etc. There is also EXECUTION PLAN. It also contains all this information regarding the optimizer estimates, but on the top of that, it also contains information about actual I/O incurred, actual cardinalities, actual timings etc.
    So if a hint is changing optimizer estimates, but the plan stays the same, then impact on query's performance is zero.
    If the actual numbers are changing, this is probably also irrelevant to the hint (e.g. you can have less physical reads because more blocks are found in the buffer cache the second time you're running the query, or you less work because you don't have to parse the statement etc.).
    Actually, most of optimizer hints don't affect optimizer estimates; rather, they try to get the optimizer to use a certain access method or a certain join order etc. regardless of the cost. So you must be talking about such hints as cardinality, dynamic_sampling etc. If that's not the case -- please clarify, because this means that something wrong is going on here (e.g. an INDEX hint may work or it may fail to work, but if it fails, optimizer estimates shouldn't change).
    Best regards,
    Nikolay

  • Generating MD5 hash value for any specific flat file

    Hi experts,
    I am developing a program that will generate flat files and also I should generate the MD5 Hash value for each and every flat files. My question is how can I generate the MD5 hash value for the generated .txt files.
    Thanks in advance
    Shabir

    You can use functions
    MD5_CALCULATE_HASH_FOR_CHAR for text file
    MD5_CALCULATE_HASH_FOR_RAW for binary

  • Looking for sha-1 hash values for the Windows 10 Enterprise x64 & x86 ISO's

    When I downloaded the ISO's, I either missed the hash values or they weren't there. Still can't fine them. I'd sure like to validate the ISO's before trying to install.
    Can anyone point me to their locations or post them here?
    Mahalo

    Maybe aboodi86 just ran a sha1 generator on the downloads.
    For what it's worth, I got the exact same value listed above for my x64 download, so it's legit.  I didn't download the x86 version.
    -Noel
    Detailed how-to in my eBooks:  
    Configure The Windows 7 "To Work" Options
    Configure The Windows 8 "To Work" Options

  • Ora_hash - Same hash value for different inputs (Urgent)

    Hi,
    Trying to use ora_hash to join between tables but i noticed that in some cases, when working on different input value the ora_hash function generates same results.
    select ora_hash('oomeroe03|6NU3|LS006P|7884|1|17-JUL-13 13.18.22.528000|0005043|'),ora_hash('GSAHFFXTK|GCQ3|A6253S|12765|1|17-JUL-13 17.26.26.853000|0136423|')
    from dual
    Output value : 1387341941
    Oracle version is 11gR2.
    Thanks

    Why would anyone limit the hash distribution to three buckets ?
    However, one must understand that the default seed is 0.  So one input repeated gets the same hash value unless the seed is changed.
    SQL> select ora_hash(rn) , rn
      2  from
      3  (Select rownum as rn from dual connect by level < 11)
      4  order by rn;
    ORA_HASH(RN)         RN
      2342552567          1
      2064090006          2
      2706503459          3
      3217185531          4
       365452098          5
      1021760792          6
       738226831          7
      3510633070          8
      1706589901          9
      1237562873         10
    10 rows selected.
    SQL> l
      1  select ora_hash(rn) , rn
      2  from
      3  (Select rownum as rn from dual connect by level < 11)
      4* order by rn
    SQL> /
    ORA_HASH(RN)         RN
      2342552567          1
      2064090006          2
      2706503459          3
      3217185531          4
       365452098          5
      1021760792          6
       738226831          7
      3510633070          8
      1706589901          9
      1237562873         10
    10 rows selected.
    SQL> /
    ORA_HASH(RN)         RN
      2342552567          1
      2064090006          2
      2706503459          3
      3217185531          4
       365452098          5
      1021760792          6
       738226831          7
      3510633070          8
      1706589901          9
      1237562873         10
    10 rows selected.
    SQL>
    Hemant K Chitale

  • Delete hash value row in Bex report output

    Hi All,
    Can anybody suggest how can i remove hash value(not assigned) row in Bex report out put.
    my report looks like this:
    rows
    project position
      (hierarchy)
      WBS element
        (hierarchy)
        Order
    For each Project position and WBS element hierarchies, i did supress the " not assigned" tick in hiearchy attributes in RSH1.
    but in colums i have budget as one of the column.
      this budget comes from project position nodes to last node of WBS element but not in order, for order it will be blank but after this order a new row coming up with bold" Not assigned" with that budget amount against this not assigned.
    My problem is how can i get rid of this not assinged entire row. I search so many links in SDN, they talk about just to replace the not assigned  to some other value as blank or something like that.
    Please suggest me how can i get rid of this.
    i did tried to restrict Order with "#" but then the whole budget column getting blank.
    Regards
    Robyn.

    HI Arun,
    As i mentioned earlier, i have gone through these links, they just talk about replacing # value or Not Assigned  to some other value or blank.
    my issue is i need to get rid of that entire row.
    when i drop proj def hierarchy ,then wbs hierachy starts, sometimes at the end of wbs hierarchy i get row with #
    and sometimes after wbs hierarchy ,then order row then # row is coming.
    even after i tick supress unassigned nodes in both Projdef and wbs hierarchies.
    i am not supposed to simply replace #( not assigned) symbol or text with some other symbol or text ( thats what they discussing in these links).
    i created that macro they said in those links but its just deleting # symbol to blank not the whole row.
    hope i made my issue clear.
    i have seen so many other links as well
    [Re: Bex macros]
    Regards
    Robyn.

Maybe you are looking for

  • HP ENVY m6 Notebook PC overheats several times per day

    My laptop overheats and crashes CONSTANTLY-several times per day!  I have vacuumed all of the vents in back, I took it to the Geek Squad, who sold me a chill pad that helps minimally, I have started putting ice packs on the lower left corner, I don't

  • Unable to send or receive text messages

    I just activated my Droid Razr and I'm unable to send or receive text messages.  Sending and receiving phone calls and emails works fine.  Any ideas?

  • Can I use USB for Samsung LD190 monitor?

    Just curious... I need a second monitor and this monitor can use USB too, but not sure how to make it work with my mac mini, anyone??

  • Imovie hd low quality when export in "full quality"

    Hi i made one hour of a movie with Imovie hd. When i exported to "full quality" is not equal to the original! I imported photos from iphoto, all in hi definition and i inserted audio track and Ken b. effects. I have to view this movie on the screen d

  • Time Capsule and Final Cut Studio

    Is there a way to have Time machine and Time Capsule halt operations while Final Cut Studio components are in use? I've had a couple of annoying hangs.