NLS Not Returning through NES Passthrough

Hello,
We are having a problem with the Natural Language Search on
RoboHelpPro within RoboEngine. Since RoboEngine only support IIS
and our client requires us to use the Sun One WebServer, we have
configured it as a reverse proxy and to pass through requests for
help. This works great except for one thing. When you access the
NLS and enter a text phrase to search, it doesn't return. It
appears that RoboEngine and IIS is unable to respond back, so
eventually it times out with an error. On the actual server where
RoboEngine is installed it works fine. The question we are trying
to get answered is how the search is working. When you press the
search button, is the request getting posted and then how is
RoboEngine communicating back to the browser? With this information
we are hoping to further understand how to configure our
environment. The search is clearly not responding back the same way
as when you click on a page with help, because it has to generate
the list of found items and serve that back to the browser. Any
help understanding this would be greatly appreciated.
Joe Blasko

Dear user2057059,
Could you specify more details about your question:
- Tables structure (columns, indexes, constraints)
- Could you post your query with execution plan
20 M rows it is not a big database especially for your hardware.
In my example:
CPU: Intel Core 2 Duo CPU 2.33 GHz,
RAM: 4 GB DDR2
HDD: 100 Gb SATA-II
OS: Fedora 8 x64 (Linux 2.6.23.1-42.fcb)
+
Oracle TimesTen 7.0.5.0.0 (64 bit Linux)
Command> select count(*) from accounts;
Query Optimizer Plan:
  STEP:                1
  LEVEL:               1
  OPERATION:           TblLkSerialScan
  TBLNAME:             ACCOUNTS
  IXNAME:              <NULL>
  INDEXED CONDITION:   <NULL>
  NOT INDEXED:         <NULL>
< 30000000 >
1 row found.
average time - 1,920321 sec, direct connection. Regards,
Gennady

Similar Messages

  • Autodiscover does not return UserSettingName.GroupingInformation for Office 365 through EWS

    Hi,
    I am using the following code to access GroupingInformation using AutoDiscover in Office 365. 
    AutoDiscoverService.GetUserSettings(Mailbox, UserSettingName.ExternalEwsUrl, UserSettingName.GroupingInformation);
    This returns ExternalEwsUrl while GroupingInformation is blank. I have even tried using EWS editor and this property is not returned there as well. I have tried for user as well as room resources. 
    Looks like these properties were added to the latest version and are not available in earlier versions of EWS. I am using the latest version of this API
    Can anyone please suggest a solution to this?

    Hi,
    Based on my research, the GroupingInformation property contains a value that is used to group the user's mailbox to maintain affinity when subscribing to notifications across multiple mailboxes. Mailboxes that have the same value and use the same Exchange
    Web Services (EWS) endpoint can be grouped together to maintain affinity:
    http://msdn.microsoft.com/en-us/library/office/dn529149(v=exchg.150).aspx
    Since it's Exchange online server, I recommend you ask for more professional help on our Exchange online forum:
    http://social.technet.microsoft.com/Forums/msonline/en-US/home?forum=onlineservicesexchange
    Thanks,
    Angela Shi
    TechNet Community Support

  • TimesTen Queries not returning

    We are evaluating TimesTen IMDB Cache as an option to improve our performance of queries that do aggregate operations across millions of rows.
    My environment is a Windows Server 2003 R2 X64 with Intel Xeon X5660 16 CPU cores, and 50G of RAM. The database is Oracle Enterprise 11.2.0.2.
    I have the following DSN parameters.
    DataStore Path + Name : H:\ttdata\database\my_ttdb
    Transaction Log Directory : H:\ttdata\logs
    Database Character Set: WE8MSWIN1252
    First Connection:
    Permanent Data Size: 26000
    Temporary Data Size: 1600
    IMDB Cache:
    PassThrough :1
    Rest of the parameters are default.
    I have created 2 read only cache groups and 1 asynchronous write-through cache group.
    The first read only cache group is on a table A with 108 rows , the second read only cache group is on table B with 878689 rows and table C is the fact table with 20.5 million rows.
    I have loaded these cache groups. Now I am trying to do join queries across these tables that do aggregation and group by on some measures in table C.
    I have seen using dssize that the data has been loaded in permanent data area.
    These queries execute in Oracle in around 5s and my expectation was that these would return in sub-seconds.
    But these queries are not returning back at all even after hours. I have looked at the query plan and I do not have any lookups which say that they are not indexed.
    I have even tried simpler join queries without any aggregation. Even those get stuck. The only queries that I have been able to run are select * from tables.
    What may be wrong in this setup/configuration ? How do I debug what is causing the problem.
    Thanks,
    Mehta

    Dear user2057059,
    Could you specify more details about your question:
    - Tables structure (columns, indexes, constraints)
    - Could you post your query with execution plan
    20 M rows it is not a big database especially for your hardware.
    In my example:
    CPU: Intel Core 2 Duo CPU 2.33 GHz,
    RAM: 4 GB DDR2
    HDD: 100 Gb SATA-II
    OS: Fedora 8 x64 (Linux 2.6.23.1-42.fcb)
    +
    Oracle TimesTen 7.0.5.0.0 (64 bit Linux)
    Command> select count(*) from accounts;
    Query Optimizer Plan:
      STEP:                1
      LEVEL:               1
      OPERATION:           TblLkSerialScan
      TBLNAME:             ACCOUNTS
      IXNAME:              <NULL>
      INDEXED CONDITION:   <NULL>
      NOT INDEXED:         <NULL>
    < 30000000 >
    1 row found.
    average time - 1,920321 sec, direct connection. Regards,
    Gennady

  • I'm having all sorts of connection issues on a iPhone i brought in the uk 6 months ago. am i able to use a genius bar abroad as I'm not returning to the uk?

    i brought a iphone 5 back in feb/march and since then its slowley getting worse and worse, my wifi hardly pickes up any signals and when it does they drop out all the time. my girlfriends iphone4 picks up and has no trouble holding onto the wifi, whilst sitting at the same table mine shows no signals. Im not returning to the uk any time soon so i want to know if its possible to visit an apple store for replacement abroad? im only 6 months into my apple care so i wouldnt think this should pose an issue. any help would be fantastic
    many thanks

    they will not repair the device in the us you must send it back to the country of purchase or set up repair through them the uk apple car number is as follows (44) 0844 209 0611*** you should be able to set up a mail repair through them

  • Error: Expression must not return a multiline value

    Hi all,
    I am doing a file to file scenario. My input file has more than one record. I have to validate for each of these records and send back the erroneous records in a file to the source folder.
    I am using BPM for this. In my BPM, i have some multiline container elements. When i try to activate my BPM, i am getting an error saying: <b>Expression must not return a multiline value.</b>
    Can anybody tell me why this error is coming? Also i want to know what type of mapping i have to do to split my source file.
    Regards,
    Divija.

    "Can anybody tell me why this error is coming? Also i want to know what type of mapping i have to do to split my source file."
    Go through the following blogs which give step-bystep approach for split mapping using BPM:-
    1. /people/sudharshan.aravamudan/blog/2005/12/01/illustration-of-multi-mapping-and-message-split-using-bpm-in-sap-exchange-infrastructure
    2. /people/narendra.jain/blog/2005/12/30/various-multi-mappings-and-optimizing-their-implementation-in-integration-processes-bpm-in-xi
    Also, you might want to look at this, where a BPM is not required..i think you can avoid a BPM altogether if you have very little validation..
    /people/jin.shin/blog/2006/02/07/multi-mapping-without-bpm--yes-it146s-possible
    Regards,
    Sushumna

  • BAPI PO GETDETAIL NOT RETURNING PREQ_NO IN PO_ITEMS

    When calling BAPI_PO_GETDETAIL with ITEMS checked "X", the table PO_ITEMS does not return the value of PREQ_NO (Purchase Requisition Number) for the corresponding PO.  When viewing the PO through ME23n the Purchase requisition Number is visable on the screen for each line item.
    This is not happening on all PO only certain PO documents, is their a condition that this depends on?

    Bill
    I do not have any problem for this document type.
    Can you place a break-point at line439 of include: LMEWPF04 and check.
    As itz a normal transfer of EKPO entry to the output table. It shouldnt be a problem.
    Do re-check if you are using the right PO when displaying in ME23(N) and when using the BAPI: BAPI_PO_GETDETAIL.
    Regards
    Eswar

  • Workflow not returning document details after update

    Hi,
    I have implemented a custom BPEL workflow process which is triggered upon a document being created or updated in Content Services. The BPEL process uses embedded java to obtain the filename of the document which triggered the workflow (much in the same way as done in the Content Services Web Services Development Kit). This works fine in the GUI version of content services, and using Oracle Drive works when a file is created, however the filename is not retrieved when updating a file using Oracle Drive (even though the workflow is initiated). As far as I can make out the Java Content Services API Workflow Manager is not returning any items when the file is updated, does anyone know why this might be?
    Many Thanks
    Dave

    Please ignore the above as the problem is not actually what I originally thought was happening. It turns out that any file which is updated through Oracle Drive does not trigger the BPEL workflow for the Upload Document workflow operation. The workflow that was being triggered for which I could not get details for was a temp file that was being created when browsing Oracle Drive before overwriting a file, the workflow could not get the temp files document details as the file was deleted before it could do so. Does anyone know a way of getting the workflow to be triggered upon the update of a file in Oracle Drive?
    Many Thanks
    Dave

  • After updating to iOS 8.3 on my iPhone 6, Blu Logic bluetooth calls with a charging phone in my car do not go through the car's speakers.  Siri's message now adds, "....calling _______on speaker."

    Blu Logic bluetooth calls using "Hey, Sir" do not go through my car's speakers since updating to iOS 8.3.  They are routed through the iPhone 6 speaker after Siri says, "....calling ________ on speaker."  The "on speaker" term was never used before.  The Blu Logic bluetooth unit is an aftermarket device installed in a 2005 Buick LaCrosse.  It worked perfectly until the iOS 8.3 update.

    OldGnome: 
    Blu Logic seems to keep a virtually invisible support presence on the internet.  I did return to the aftermarket installer of the unit.  He told me pretty much the same as you.  Apple expects the equip. manufacturers to keep up with iOS.  It's a bit of a Catch-22.  Perhaps iOS 9 will exterminate the bugs!!
    In the meantime, the only thing I've really lost is hands-free dialing for bluetooth calls in my car.

  • RFC Call not returning in normal exec;retunrs in debugger

    Hi all
    i m sorry for posting it twice,due to my ignorance
    i first posted it in wrong forum
    i had a requirement to invoke a RFC Async and recieve the results in my program with a condition that the name and parameters of RFC are determined runtime.
    To meet this requirement, i generated a the RFC call and return form code dynamically and called that code from my program via a subroutine.
    The problem i am facing is that, whenever i execute my program in debug mode the RFCs are called and the parameters are returned( perform on end of task edition is used)
    via subroutine, but when i execute the code via F8 the program goes in indefinite wait for return parameters.I am calling 2-3 RFcs at same time, and all of them are on local host.
    the Functions are working well and the RFC destination is also working well.
    But this amalgum of my program and dynamic calls is not going through.
    I read in documentation that if we change the program context or the roll area then the results are not recieved by main program. by the way i am using WAIT to receive the results.
    For the moment if i consider, my program context or roll area is getting changed because of dynamic code generation then also the dynamically generated code is the one containing the RFC call and its revieve subroutine, therefore there shall be no problem with that, the other thought i am getting is that in debugger implict commit work happens, is the program working because of that in debugger?? i dont know....
    Can any one help me with this, in the above long paragraph if anyone gets confused..i can again try to restate my problem.
    One more thing..i have checked the ST22 and SM50 for any abnormalities, couold not find one.
    thanks in advance
    praveen

    WAIT statement also triggers a DB commit like debugger.
    Are you using WAIT UNTIL <condition>?
    Regards
    Sridhar

  • OCIStmtExecute does not return immediately when client is busy.

    Hi.
    I'm testing a very busy multi-threaded client server that consistently generates
    a large number of simple queries through oci. The problem is that, when the
    server(client) is busy, OCIStmtExecute does not return immediately in
    non-blocking mode.
    Of course I have set non-blocking mode and OCIStmtExecute does return
    OCI_STILL_EXECUTING immediately when the server is not busy. But
    when log rotation occurs which concatenates a large text file (~500MB)
    onto an even larger text file (up to several giga bytes), or when I simply copies
    or concatenates large text files manually, OCIStmtExecute returns very slowly.
    (roughly about after 100~200ms)
    However, while log rotation takes place, everything else including other oci
    calls that come before OCIStmtExecute (prepare, define) return fast. So
    for me it really seems that only OCIStmtExecute becomes extremely slower
    when local server (especially the disk) is busy.
    Is there any way to let OCIStmtExecute immediately return all the time?
    Thanks in advance.

    Yes, I knew that OCIStmtExecute would be the only function that causes such
    delay and that was why I traced that call. And so far, I checked several times
    what happens at the exact moment on the server but everything was ok.
    Actually OCIStmtExecute becomes slower exactly when crontab-ed log rotate
    occurs so I think this delay must be a client-side problem for now.
    This server is quite busy and has to respond fast, so it is important to
    guarantee fast response time while a small number of timeout losses are tolerable.
    But as OCIStmtExecutes' first OCI_STILL_EXECUTING return takes hundreds of
    ms it has become more like a blocking call and currently I cannot find any way to do what I want.
    So now everytime such thing happens, the thread waits
    quite long, after the first OCI_STILL_EXECUTING return
    the time difference exceeds timeout limit, and the thread
    calls OCIBreak() and OCIReset() and returns.

  • My emails are not coming through to my blackberry phone....

    My emails are not coming through to my blackberry phone. I think it is because I changed my password to my email?

    Hi there!
    If you changed your email password, then you need to change the stored credentials within BIS. See this demo for how to access your BIS configuration:
    http://na.blackberry.com/eng/support/blackberry101/setup.jsp#tab_tab_email
    If that does not work, then please return to these forums (http://supportforums.blackberry.com/rim/) and register an account so that we can provide you with direct assistance.
    Cheers!
    Occam's Razor nearly always applies when troubleshooting technology issues!
    If anyone has been helpful to you, please show your appreciation by clicking the button inside of their post. Please click here and read, along with the threads to which it links, for helpful information to guide you as you proceed. I always recommend that you treat your BlackBerry like any other computing device, including using a regular backup schedule...click here for an article with instructions.
    Join our BBM Channels
    BSCF General Channel
    PIN: C0001B7B4   Display/Scan Bar Code
    Knowledge Base Updates
    PIN: C0005A9AA   Display/Scan Bar Code

  • External Content Typ: "Connection manager did not return valid connection."

    Hi,
    I created an external content typ in the sharepoint designer which takes my custom WCF Service as a data source. I sucessfully added all my required operations and created a new list based on the new content typ. At first I had a security issue which I got
    resolved by editing my BDC configuration in the Central Administration. Now I face a new issue. When I open my external list the BDC Service gives me the following error message:
    "Connection manager did not return valid connection."
    Has someone a solution for this kind of message?
    Thanks,
    Eric

    Hi,
       I got the same error "Connection manager did not return valid connection.". In the error log it shows
    09/17/2010 12:21:35.28  w3wp.exe (0x13FC)                        0x0F60 Business Connectivity Services Business Data                
     a5t8 Medium   Chose 'PassThrough' auth mode cb34d6ff-663a-4872-9022-21782a7be122
    09/17/2010 12:21:35.28  w3wp.exe (0x13FC)                        0x0F60 Business Connectivity Services Business Data                
     a5uo Medium   Initialized WcfSystemUtility for LobSystemInstance with Id '144'. cb34d6ff-663a-4872-9022-21782a7be122
    09/17/2010 12:21:35.28  w3wp.exe (0x13FC)                        0x0F60 Business Connectivity Services Business Data                
     a5up Medium   Wcf Service AuthN Mode: PassThrough  cb34d6ff-663a-4872-9022-21782a7be122
    09/17/2010 12:21:35.28  w3wp.exe (0x13FC)                        0x0F60 Business Connectivity Services Business Data                
     f8dk Medium   Checking for current connection to SystemInstance: NewCustomer cb34d6ff-663a-4872-9022-21782a7be122
    09/17/2010 12:21:35.28  w3wp.exe (0x13FC)                        0x0F60 Business Connectivity Services Business Data                
     a5s9 Medium   Creating ClientBase object with type : BCSServiceProxy.WCFCustomerClient cb34d6ff-663a-4872-9022-21782a7be122
    09/17/2010 12:21:35.29  w3wp.exe (0x13FC)                        0x0F60 Business Connectivity Services Business Data                
     a5tz Unexpected System.MissingMethodException: Method not found: 'Void Microsoft.BusinessData.SystemSpecific.Wcf.IBdcClientBase.set_ActAs(Boolean)'.     at Microsoft.SharePoint.BusinessData.SystemSpecific.Wcf.WcfConnectionManager.ConfigureCredentials(Type
    clientBaseType, Object clientBase)     at Microsoft.SharePoint.BusinessData.SystemSpecific.Wcf.WcfConnectionManager.CreateClientBase()     at Microsoft.SharePoint.BusinessData.SystemSpecific.Wcf.WcfConnectionManager.GetConnection()    
    at Microsoft.SharePoint.BusinessData.SystemSpecific.Wcf.WcfSystemUtility.ExecuteStatic(IMethodInstance methodInstance, ILobSystemInstance lobSystemInstance, Object[] args, IExecutionContext context) cb34d6ff-663a-4872-9022-21782a7be122
    Can you help me on this?
    Thanks in advance
    Thybu

  • Why does my function not return anything when I create as a schema object

    I have user ABC who owns several tables some of which have foreign key constraints.
    I have user XYZ that has been granted access to all tables owned by user ABC.
    When I create a function as user XYZ using following I get no return when I issue:
    select XYZ.ztm_tables_depended_on('ABC', 'A_TABLE_OWNED_BY_ABC') from dual :
    Please see after function definition.
    CREATE OR REPLACE FUNCTION ZTM_TABLES_DEPENDED_ON(p_Owner VARCHAR2, p_Table_Name VARCHAR2) RETURN VARCHAR2 IS
      CURSOR C1 IS
      SELECT OWNER, CONSTRAINT_NAME, R_OWNER, R_CONSTRAINT_NAME
      FROM   ALL_CONSTRAINTS
      WHERE  OWNER           = p_Owner
      AND    TABLE_NAME      = p_Table_Name
      AND    CONSTRAINT_TYPE = 'R'
      ORDER  BY OWNER, CONSTRAINT_NAME, R_OWNER, R_CONSTRAINT_NAME;
      v_Referenced_Owner       VARCHAR2(31);
      v_Ret_Val                VARCHAR2(4000);
      FUNCTION CONSTRAINT_TABLE_NAME(p_Owner VARCHAR2, p_Constraint_Name VARCHAR2) RETURN VARCHAR2 IS
        CURSOR C1 IS
        SELECT TABLE_NAME
        FROM   ALL_CONSTRAINTS
        WHERE  OWNER           = p_Owner
        AND    CONSTRAINT_NAME = p_Constraint_Name;
        v_Ret_Val ALL_CONSTRAINTS.TABLE_NAME%TYPE;
      BEGIN
        OPEN  C1;
        FETCH C1 INTO v_Ret_Val;
        CLOSE C1;
        RETURN v_Ret_Val;
      END;
    BEGIN
      FOR R IN C1 LOOP
        IF (R.OWNER <> R.R_OWNER) THEN v_Referenced_Owner := R.R_OWNER || '.';
        ELSE                           v_Referenced_Owner := NULL;
        END IF;
        v_Ret_Val := v_Ret_Val || ', ' || v_Referenced_Owner || CONSTRAINT_TABLE_NAME (R.R_OWNER, R.R_CONSTRAINT_NAME);
      END LOOP;
      RETURN LTRIM(v_Ret_Val, ', ');
    END;
    But, if I embed the function within an anonymous block as follows, I get results:
    DECLARE
      CURSOR C1 IS
      select owner, table_name
      FROM   all_tables where owner = 'ABC';
      FUNCTION ZTM_TABLES_DEPENDED_ON(p_Owner VARCHAR2, p_Table_Name VARCHAR2) RETURN VARCHAR2 IS
        CURSOR C1 IS
        SELECT OWNER, CONSTRAINT_NAME, R_OWNER, R_CONSTRAINT_NAME
        FROM   ALL_CONSTRAINTS
        WHERE  OWNER           = p_Owner
        AND    TABLE_NAME      = p_Table_Name
        AND    CONSTRAINT_TYPE = 'R'
        ORDER  BY OWNER, CONSTRAINT_NAME, R_OWNER, R_CONSTRAINT_NAME;
        v_Referenced_Owner       VARCHAR2(31);
        v_Ret_Val                VARCHAR2(4000);
        FUNCTION CONSTRAINT_TABLE_NAME(p_Owner VARCHAR2, p_Constraint_Name VARCHAR2) RETURN VARCHAR2 IS
          CURSOR C1 IS
          SELECT TABLE_NAME
          FROM   ALL_CONSTRAINTS
          WHERE  OWNER           = p_Owner
          AND    CONSTRAINT_NAME = p_Constraint_Name;
          v_Ret_Val ALL_CONSTRAINTS.TABLE_NAME%TYPE;
        BEGIN
          OPEN  C1;
          FETCH C1 INTO v_Ret_Val;
          CLOSE C1;
          RETURN v_Ret_Val;
        END;
      BEGIN
        FOR R IN C1 LOOP
          IF (R.OWNER <> R.R_OWNER) THEN v_Referenced_Owner := R.R_OWNER || '.';
          ELSE                           v_Referenced_Owner := NULL;
          END IF;
          v_Ret_Val := v_Ret_Val || ', ' || v_Referenced_Owner || CONSTRAINT_TABLE_NAME (R.R_OWNER, R.R_CONSTRAINT_NAME);
        END LOOP;
        RETURN LTRIM(v_Ret_Val, ', ');
      END;
    BEGIN
      FOR R IN C1 LOOP
        DBMS_OUTPUT.PUT_LINE(ztm_tables_depended_on(R.Owner, R.Table_Name));
      END LOOP;
    END;
    Any ideas what is happening here?

    Any ideas what is happening here?
    Justin explained the probable reason.
    See the 'How Roles Work in PL/SQL Blocks' section of the database security doc for the details
    http://docs.oracle.com/cd/E25054_01/network.1111/e16543/authorization.htm#i1007304
    How Roles Work in PL/SQL Blocks
    The use of roles in a PL/SQL block depends on whether it is an anonymous block or a named block (stored procedure, function, or trigger), and whether it executes with definer's rights or invoker's rights.
    Roles Used in Named Blocks with Definer's Rights
    All roles are disabled in any named PL/SQL block (stored procedure, function, or trigger) that executes with definer's rights. Roles are not used for privilege checking and you cannot set roles within a definer's rights procedure.
    The SESSION_ROLES view shows all roles that are currently enabled. If a named PL/SQL block that executes with definer's rights queries SESSION_ROLES, then the query does not return any rows.
    Roles Used in Named Blocks with Invoker's Rights and Anonymous PL/SQL Blocks
    Named PL/SQL blocks that execute with invoker's rights and anonymous PL/SQL blocks are executed based on privileges granted through enabled roles. Current roles are used for privilege checking within an invoker's rights PL/SQL block. You can use dynamic SQL to set a role in the session.
    See that line starting with 'All roles are disables in any named PL/SQL block'?

  • Javascript not parsing through in an HTML data set (table)

    I have a spry dataset (html table) working perfectly on my site.  I have an 'info' column with some HTML, an image, and youtube embed.  All of these parse correctly and show on the page, but for some reason my google adsense (javascript code) is not parsing through at all.  I'm using Spry 1.6.1.
    Here is my code:
    <script type="text/javascript">
    <!--
    var PrankDialerLibrary = new Spry.Data.HTMLDataSet("SpryAssets/library.html", "pdLibrary", {sortOnLoad: "hits", sortOrderOnLoad: "descending"});
    pdLibrary.setColumnType("preview", "html");
    pdLibrary.setColumnType("info", "html");
    pdLibrary.setColumnType("filename", "html");
    //-->
    </script> ...          <div class="MasterDetail">
                <div spry:region="PrankDialerLibrary" class="MasterContainer">
                  <div class="MasterColumn" spry:repeat="pdLibrary" spry:setrow="pdLibrary" spry:hover="MasterColumnHover" spry:select="MasterColumnSelected" align="left">{preview}  {name}</div>
                </div>
                <div spry:detailregion="pdLibrary" class="DetailContainer">
                  <div class="DetailColumn">{info}</div>
                  <div class="DetailColumn">{filename}</div>
                </div>
                <br style="clear:both" />
              </div>
    Thanks!

    Treytor72 wrote:
    Thanks again for your expertise, as I can see you've helped quite a bit around here!
    I've tried adding that quick and dirty workaround to the page ( http://www.prankdialer.com/beta/popular.php ), but it seems to have no affect.  I may not be fully understanding it, but I don't believe your other options will work in this particular situation because the google adsense code (what is being omitted in my case)
    is placed exactly where the ad should appear.  I want it to show up right underneath the youtube embed for each item in the data table.  Thus, it doesn't belong in the <head> of the page.
    Thank you again for your time, and let me know if by any chance you would have any use for my site, PrankDialer.com.  I'd be more than happy to give you credits for your help!
    Hello,
    If I inspect the generated DOM of the page, I do see that the script have been correctly injected in to your page. So we have succeeded so far. You have placed it at a correct position. We could still try the modified setInnerHTML function. I searched the forum and found a improved version of it. (Less work for me )
    Spry.Utils.setInnerHTML = function(ele, str, preventScripts)
    if (!ele)
    return;
    ele = Spry.$(ele);
    var scriptExpr = "<script[^>]*>(.|\s|\n|\r)*?</script>";
    ele.innerHTML = str.replace(new RegExp(scriptExpr, "img"), "");
    if (preventScripts)
    return;
    //bodyMatches is used to store the script tags that have body. For example, <script type='text/javascript'> some JS </script>
    var bodyMatches = str.match(new RegExp(scriptExpr, "img"));
    if (bodyMatches)
    var numMatches = bodyMatches.length;
    for (var i = 0; i < numMatches; i++)
    var s = bodyMatches[ i ].replace(/<script[^>]*>[\s\r\n]*(<\!--)?|(-->)?[\s\r\n]*<\/script>/img, "");
    //make sure that we are not create a script tag if there is no body to it.
    if(s.length > 0)
    //alert(s);
    var scriptEle= document.createElement("script");
    scriptEle.text = s;
    ele.appendChild(scriptEle);
    //now we begin taking care of the script includes (those that have a src attribute defined)
    var srcScript = "<script[^>].*src.*>(.|\s|\n|\r)*?</script>";
    var srcMatches = str.match(new RegExp(srcScript, "img"));
    if(srcMatches)
    var numMatches2 = srcMatches.length;
    var srcAttrib = "[\s]*src=[\'\"](.*\.js)[\'\"]";
    for(var j=0; j<numMatches2; j++)
    // when the below statement executes, m should have the src attrib and value:
    // src='sompath.js' as a string.
    var m = srcMatches[j].match(new RegExp(srcAttrib, "img")).toString();
    //now we strip out the src= part. this will leave us with just the path along with the double quotes.
    var n = m.replace(/^[(?=src=)]*/img, "");
    if(n.length > 0)
    var scriptEle2 = document.createElement("script");
    //get rid of the quotes as well.
    //This is done to take care of single quotes too so that it will work if they defined either:
    // <script src='somefile.js' type="text/javascript></script>
    // <script src="somefile.js" type="text/javascript></script>
    scriptEle2.src=n.replace(/[\"\']/img, "");
    scriptEle2.type="text/javascript";
    ele.appendChild(scriptEle2);
    But!
    There's always a but, even if it would work. You might face a even bigger issue. The Google script uses document.write. This method should only be used once the main page is still loading. If you execute this function while the page is done with loading. It will replace the whole content of your page with what is beeing writen.
    For example if it does:
    document.write('Hello');
    Your whole place would be replaced and only showing the word "Hello".
    A solution to this would be to use a iframe to display the ads.

  • ALUI Gateway Not Returning Scripts for Subset of Users

    We have a problem where the ALUI gateway is not returning some .NET scripts for a subset of users. We have the ALUI 6.5 portal and our using the .NET accelerator 3.1.
    The situation is that this subset of users request one of our portal pages via https, which then reaches through our firewall to our remote server which is running the .NET portlet. The .NET page is served and returned to the users correctly and quickly, but this particular subset of users do not see the result rendered in their browsers for about 3 minutes. A view html source in the browser, as well as tools like Fiddler, show the page is indeed in the browser, but it is stuck trying to request some .NET scripts, and only displays the page when those requests timeout.
    The .NET scripts that are problems are both WebResource.axd and ScriptResource.axd, which in some cases are in our .NET portlets because of the .NET framework itself, but in other cases they are there only because of the ALUI portal itself, when it munges the .NET portlet to handle multiple server forms and validators and such. These .axd scripts are gatewayed so that the client browser requests them through the ALUI gateway, which in turn requests them through our firewall to our remote server -- which always serves these scripts correctly and quickly according to the IIS logs. The problem seems to lie in the ALUI gateway, as it is receiving these scripts correctly and quickly, but it is not returning them to this subset of users. Instead the ALUI gateway seems to be processing for about 3 minutes, and eventually returns an html error page, which of course the client never sees since it is expecting javascript, but we can capture the error page via Fiddler and its just telling us there was a timeout -- the client browser just notes that there is a javascript error.
    The really bizarre part is that this only happens for a subset of users, which amounts to about 20% of our users. There are 2 things that delineate these users that we have found so far. First, these users have email addresses that are 27 - 30 characters long, and the email address is our login id. Note that both shorter and longer email addresses are OK, so there is not some limit to email addresses like this might sound like at first. Secondly, these users have to be in a particular branch of our ldap store, which means they are replicated across to the portal in a particular group. We can move these "bad" users into another branch of our ldap store and once they are replicated to the portal then they work fine, and then if we move them back they return to not working. We cannot find any other difference in our ldap branches or in the corresponding ALUI groups, plus its only the ones in that particular branch with the email lengths in that very specific range.
    The gatewayed requests for these scripts vary by user since the PTARGS in the gatewayed request include the integer userid, but that does not seem to matter because we can have a "good" user successfully request the script with a "bad" user's id, and we can have a "bad" user fail to successfully request the script with a "good" user's id. That seems to point to maybe the authentication cookie being the differentiating factor that determines whether or not a gatewayed request for one of these script files will succeed or fail. So far we have only seen the problem with these particular .net axd scripts, but that may simply be because we don't have many, if any, other scripts or resources that need to be gatewayed since we usually put resources on our imageserver -- these being different because .NET and/or the ALUI portal puts these references in there for us whether we like it or not. Long-term we can re-architect our .NET portlets to not get have these axd scripts, although as mentioned earlier, we also see the ALUI portal put these axd scripts in our portlets as part of their munging process -- so that is not in our control completely. We do need to test if this subset of users can successfully request other gatewayed resources or not -- this is actually the first time I thought of that test case, so all I can say right now is its axd scripts that we know are problems, but it may or may not be a bigger problem.
    One last comment, as we appear to have found a work-around, but it does not make sense at all, and its not our preferred solution, so we still very much believe there is a problem elsewhere -- most likely in the ALUI gateway, but possibly somehow related to authentication that we do not understand. Our work-around that so far seems to work is to make our remote server be accessed via https instead of http -- which matches the way the client browsers call our portal (https). Again that first doesn't make sense, since this is only a problem for a small subset of users -- obviously calling our remote server via http works successfully for all other users, so its not just is a simple case that we must match protocols or it won't work. We also use http successfully for our calls to the remote server for portlets that are Java, although its possible that they don't have any gatewayed resources. But we also would just prefer to not use https for our internal calls in our own network as there is no need for the extra overhead -- and by the way our dev and qa environments do use http even for these .NET portlets and do not have the same problem. What's different in our production environment? The only things that should be different are that we have multiple portal servers and multiple remote servers that are load balanced (not sure that's the right term for how the remote servers are combined) -- and of course we have a firewall between them that does not exist in dev or qa.
    So we would very much appreciate any thoughts on this, whether you've seen something like it before, or just have some additional insight into the gateway and/or authentication process that seems to be the issue.
    Thanks, Paul Wilson

    We've ran into this problem when using the Microsoft ReportViewer control. In our case, we found that the portal gateway malformed the urls containing webresource.axd, so the browser was unable to get the correct address to the files. Note that there are usually multiple links to the axd files, they return different resources depending on the query string they get.
    To solve the problem, we ended up with a bit of a hack solution, but it works well. We extracted the resources we needed from the ReportViewer control's assembly using Reflector, and then published them on the image server. The next piece was to override the Render method of the page that hosted the control. In our custom version of Render, we parsed the html of the page, and replaced the contents of the src= elements with pt:images// links. These processed just fine in the portal's transformer, and our resources started showing up.
    Our Render looks something like the following code sample. The "HACKReportViewerControlPortalImageGatewayFix" class has all of the code to do the parsing. In this case, it is specific to the report viewer, because it has some special considerations for parsing the urls. My bet if that your code will be quite custom as well. Therefore, I've not included this piece of code. The important piece below is the invocation of MyBase.Render, which tells the page to render all of it's contents. Once that method is done, all of the HTML for the page is in the writer. The ModifyImageTags method then parses the html, doing the necessary replacements. Finally, the modified html is written to the page's writer, so it can be output following the normal .net processes. Also note that when parsing for urls to replace, don't do all of them, just look for the ones containing axd.
    (VB.NET)
    Protected Overrides Sub Render(ByVal writer As System.Web.UI.HtmlTextWriter)
              Dim fixer As New HACKReportViewerControlPortalImageGatewayFix
              MyBase.Render(fixer.GetWriter)
              writer.Write(fixer.ModifyImageTags())
    End Sub
    This works great for images. However, if you are dealing with javascripts, I'm not sure if this will work for you - as some .NET controls send different scripts depending on the browser. For example, in IE, you get more buttons on the toolbar for the ReportViewer, so you get more javascript too. When using FF, you get less buttons, and less script. We didn't have a problem with the scripts, so we haven't needed to solve this one.
    As for timing, this type of solution doesn't take much to put together. You are really just doing some string parsing and replacements. If you are a regex ninja, it's probably even easier. We had our solution working in a day or two.
    An added benefit to this solution is that you are putting less bytes through the portal's gateway, and sending that traffic to the image server instead.

Maybe you are looking for

  • I updated my iPhone 4 and lost the most of my music, which I have bought on the itunes store. And now I can't get it back. What to do?

    Not all of it is gone. Some of it stayed and there is a couple, where the music is gone but the booklet stayed. When I try to buy it again, it says that I already have bought it and that I should try the availeble transferring purchases. I have done

  • "Rolling" interference on DVD player

    I've run out of ideas so hoping somebody here can think of something I haven't tried. The office just bought a new DVD player (Panasonic DVD-S53) and VHS (JVC HM-DT100) to transfer video/audio to FCP. Both replaced previous models that needed to be r

  • Please help me return Lightroom product

    I bought lightroom for $149 at Best Buy and I want to return it. Best buy says I have to get a refund from Adobe. I did not activate the key, only used the 30 day free trial. How do I get my money back?

  • Exterior Hard Drive for storing files

    On my system I've backed up my files on my (K) drive.  I've just recently backed my files again.  (had been a while) The free space on the drive is getting low. Should I erase what is on that drive and do a fresh new back up or leave it be. How do I

  • Cisco ASA 5510 Specs

    In Cisco ASA Firewall 5510 does the feature content filter come built in? Posted by WebUser Allyson Buscemi from Cisco Support Community App