GW Business Data Provider Problem

Hello everyone,
My company wants to implement SAP NG and we are trying trial version of NG 2.0. We have completed most of the steps of an RFC based GW consumption model.
We are having problem using and testing this consumption model.
When we first created consumption model, we get the URL:
http://*gateway.contoso.com*:8000/sap/opu/sdata/sap/USERMAINTAINRFC/?sap-client=001&$format=xml
When we pasted it in a browser, it couldn't find the server. So, we replaced server (gateway.contoso.com) with our (GW) system server (like 10.xxx.xx.67) Then we pasted modified URL and get the required XML. Similarly we can get metadata of this XML.
BUT, when we try to send request via browsers(Fİrefox, IE) or JAVA we get this error:
<error><code>/IWFND/CM_IFL/003</code>
<message xml:lang="en">GW Business Data Provider failed
</message><innererror>GW Business Data Provider failed.</innererror></error>
We used below code for the request:
http://10.xxx.xx.67:8000/sap/opu/sdata/sap/USERMAINTAINRFC/zya_user_rfcCollection
(zya_user_rfcCollection is one of the existing collections in XML)
We guessed that the problem is about the connection between SAP box and back-end system but not sure about that.
So, how can we fix that?
Thanks.
Kerim
Edited by: kerimmkilic on Jan 25, 2012 12:35 PM

Hi Pramila,
  Could you please check if your service is configured properly. Generally one gets this error when the data provider is not configured against the model provider . You can view your configiration in spro.
This is the help link which might be helpful which has all the configuration information
http://help.sap.com/saphelp_gateway20sp05/helpdata/en/a1/bdaae2b2a94dd5bcb987638f5337e4/frameset.htm
Further you may go to "Settings for OData Channel Service Development on Backend System".
and check about your service.
I hope this will help you.
Thanks,    
Mohan

Similar Messages

  • GW Business Data Provider failed

    Hiie,
    By following the below link i'm able to subscribe for workflow scenario,
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/b01f6a02-5464-2e10-bd8b-efa13d7f20b6?overridelayout=t…
    but while trying to fetch workflow task collection system is raising the below error :
    <?xml version="1.0" encoding="utf-8" ?>
    - <error xmlns="http://schemas.microsoft.com/ado/2007/08/dataservices/metadata"> 
    <code>/IWFND/CM_IFL/003</code>  
    <message xml:lang="en">GW Business Data Provider failed</message>  
    <innererror>GW Business Data Provider failed.</innererror>  
    </error>
    I have referred the below link:
    https://scn.sap.com/thread/2090627
    waiting for suggestions ..,

    lokeswar reddy byni
    Good to hear you have solved it by yourself. Please share the soltuion here so that it would be helpful for others.
    Rgrds,
    Jitendra

  • Data provider problem in custom item renderer

    I have a complex, custom item renderer for a list. I add
    items that I extracted from an xml to the data provider using the
    IList interface. But when displaying the list, the items are all
    screwed up. Each rendered item has some parts which are initialized
    as different components depending on the values from the xml. This
    initialization is called in the item renderer for the
    creationComplete event.
    The weird thing is that when I output the dataProvider to
    check its values, some of the items have internal uids sometimes
    and sometimes they don't. If I output the dataProvider right after
    I add the items to it, none of them get internal uids. But from the
    initialize method, some of them do and some don't.
    To make things weirder, sometimes, as I scroll up and down
    the list, the dynamic components get all switched up. I'm either
    having a problem with internal uids or with the creation policies
    for lists. Or it's probably some simpler mistake I have yet to see.
    Anyone have any idea where the problem could lie? Any help is
    greatly appreciated.

    Any successful render must:
    1) override the set data property of the component
    Further, best practice is to store any data you need in the
    override set data(), and call invalidateProperties(). Then do the
    actual work in an override commitProperties() function.
    The framework is smart about when to call commitProperties
    efficiently. set data() gets called much more often.
    Tracy

  • I bought new i phone 5 3 days back, after that asked for new update which is 6.1.4, after updating my iphone i am not able to use cellular data services. I called up data provider, they says its the problem with new software update. There is no option add

    I bought new i phone 5 3 days back, after that asked for new update which is 6.1.4, after updating my iphone i am not able to use cellular data services. I called up data provider, they says its the problem with new software update. There is no option add APN. Now when i switch to safari its showing you are not subscribed for cellular data. But I am able to use data on other phone.
    Will you please help me in this regard?
    Another issue, since i bought my new iphone there is dust inside back main camera.
    Your advises are highly appreciated.

    Hey Shaiju isac,
    I'd take a look at the following article, it'll guide you though steps to you troubleshoot cellular data issues on your iPhone:
    iPhone: Troubleshooting a cellular data connection
    http://support.apple.com/kb/ts3780
    Cheers,
    David

  • Outlook 2013 - Calendar issues: Suggestions cannot be provided because free/busy data could not be retrieved

    I noticed recently that I cannot see anymore free/busy information for my own organization colleagues.
    I'm using Outlook 2013 (latest updates) upon Exchange 2010 (SP ?).
    The suggestion DO work on my account using OWA.
    Other users w/Outlook 2010 have no problems.
    I tried to run "outlook /cleanfreebusy" but that switch isn't supported anymore in outlook 2013.
    Anyone can help ?
    Thanks

    Hi,
    I didn't say two but three places.
    Actually, my wording is a bit confusing. In terms of where you go to remove them, it is indeed "Program and features" and "see installed updates". Now within there, I had the KB listed several times (3) under different products (like "Microsoft Outlook 2013"
    and "Microsoft Office Professional Plus 2013") and I had to uninstall the KB from all three entries.
    Note that, as per Robert Sparnaaij (http://www.howto-outlook.com, http://www.msoutlook.info)
    : "I’m aware of the issues but they are not as black and white as the article(s) suggest(s) and not that widespread either to do a general recommendation to uninstall
    a security update. In case of the IMAP issues, they are actually wrong.", and "It was said to
    be fixed with Path Tuesday for December (December 10) but unfortunately somehow the Outlook 2013 update missed the cut. It got published on December 13 after all but only was approved a little over 24 hours ago as KB2850061. However, at the moment, only the
    downloads are available but not yet the KB article nor are they currently being offered via Microsoft Update so unfortunately, the issue will drag on a bit longer. I’ve just published an article about it here:
    http://www.howto-outlook.com/news/update-2013-12-ol2013.htm".
    Lastly, "The KB article of the November update confirms that it fixes the Autodiscover connectivity issues with Exchange 2007:  
    If you are using Outlook to connect to a Microsoft Exchange Server 2007 mailbox:
    You receive an error message that resembles either of the following when you try to configure Automatic Replies (Out of Office): Your automatic reply settings cannot be
    displayed because the server is currently unavailable. Try again later.
    You cannot retrieve Free/Busy data for calendar scheduling.
    Add-ins that use the Account.SmtpAddress property no longer work.
    These features rely on an underlying
    Autodiscover technology. After you install this security update, Autodiscover may fail for Exchange 2007 configurations. Therefore, Outlook features that rely
    on Autodiscover will also stop working. To resolve these issues, install update 2850061."
    Hope this helps.

  • Problem with PCo 2.1 and ODBC data provider

    I am trying to connect MII 12.0 to an Aspen IP.21 data historian using PCo as the universal data server. Configuration steps so far:
    1) On the machine where PCo is installed, I have created an ODBC data source using Aspen's ODBC driver for SQL+.
    2) In the PCo management console I have configured a source system using the Microsoft OLEDB Provider for ODBC Drivers, and selected the ODBC connection that I created. I click "Test Connection" and get a successful result.
    3) In the PCo management console I have configured a destination system that points to an MII server. I put in the appropriate server/port/version/credentials, and click "Test Connection", and get a successful result.
    4)  In the PCo management console I have configured an agent instance connecting the source from step 2 and the destination from step 3.
    When I start the agent instance, I get the red box with a white X in it indicating that the connection was not successful. When I go to the error logs I see:
    Error     .     6     6076     RS1630IP21T01     Host     failed to create ConnectivityAgentProxy     General PCo Fault: The .Net Framework Data Provider for OLEDB (System.Data.OleDb) does not support the Microsoft OLE DB Provider for ODBC Drivers (MSDASQL). Use the .Net Framework Data Provider for ODBC (System.Data.Odbc)
    So, I installed the .NET data provider for ODBC to try to use that instead of the OLEDB provider for ODBC. However, when I try to reconfigure the source system, I do not see the .NET provider as one of my options. This is after a reboot, and starting/stopping all of the different PCo services.
    Any thoughts experts??

    Install PCo 2.2 on a server that has network access to your IP.21 server, and that your destination server (and SAP MII server in my case) has access to.
    In the PCo2.2 Management Console:
    In the "Source Systems" section:
    Create a source system of type "IP21 Agent"
    On the Server Settings tab, provide the server name of your IP.21 server
    In the "Agent Instances" section
    Create an agent instance, picking the source system you just created
    In the Query Ports tab, under port type, select the type of system you will be communicating to. In my case, this is SAP MII, but you may be using something different.
    Also under the query ports tab, you can enter enter a port number if you like. I just accept the default, which is 9000.
    If you are going to "pull" data from the IP.21 server, this is all you need. If you want to "push" data to a destination system from your PCo server, you will need to set it up under the Destination Systems section. In my case, I am doing a data pull, so I haven't done much with destination systems.
    Not sure what your destination system is, but in my case it is MII. For MII here are the basic steps:
    Create a new UDC data server in the MII menu under Data Services -> Data Servers.
    Set the IP address of the data server to the IP address of your PCo server.
    Set the port number of the data server to the port number you set up in the agent instance.
    Make sure the agent instance is started on your PCo server.
    Create a new business logic transaction containing a Tag query, and configure the tag query to use the new MII data server you just created to query the tag(s) you are interested in.

  • I have a problem with a chart refreshing after its bound data provider changes.

    I have a problem with a chart refreshing after its bound data
    provider changes. After the subsetHighLowData.refresh() is called I
    would expect the chart to redraw, but it still looks the same. I
    even inspected the debugger to see if the slice had taken place and
    it had.
    Any ideas coding friends?

    "nikos101" <[email protected]> wrote in
    message
    news:gib6aa$prr$[email protected]..
    >I have a problem with a chart refreshing after its bound
    data provider
    >changes.
    > After the subsetHighLowData.refresh() is called I would
    expect the chart
    > to
    > redraw, but it still looks the same. I even inspected
    the debugger to see
    > if
    > the slice had taken place and it had.
    >
    > Any ideas coding friends?
    >
    > subsetHighLowData.source =
    >
    highLowData.source.slice(slider.values[0],slider.values[1]);
    Just use the ArrayCollection's removeItemAt, and it should
    handle the event
    dispatching for you.
    HTH;
    Amy

  • Data provider row set problem

    I am having a problem with CachedRowSetDataProvider and the associated CachedRowSet.
    I have placed a drop down onto my design and then I drag over a table from my database. Creator binds the CachedRowSetDataProvider to the dropdown but there is no CachedRowSet associated with the DataProvider. I'm cornfused?

    In the Outline window, select the node for the data provider. The node should appear near the botton of the page's section (before the RequestBean1 section).
    After you select the node for the data provider, look in the Properties window.
    What is the value of the cachedRowSet property?
    Is this the first time that you have used a cachedRowSetDataProvider?
    If not, have you been able to use it successfully before? If yes, in this project or a different project?

  • Oracle data provider not appearing in MS Business Intelligence reporting

    Oracle Data provider does appear is VS 2005 server explorer but NOT as a data source in the reporting project type.
    This is in oracle ODAC beta 11
    Is this supported?

    I believe your developer encounters a hang when using the Microsoft reports wizard with MS OracleClient. I don't know why this is happening, but this issue would need to be resolved by Microsoft. They build both the reports wizard and MS OracleClient. They would need to identify what in their code is causing the problem. If they find out the underlying cause is a call they are making into Oracle, then Oracle will fix it.
    Oracle builds ODP.NET, which I haven't seen work with the reports wizard query builder. But I haven't fully investigated why. There may be additional configuration needed to wire these together.
    If you would like to discuss this in detail, you can email me at alex.keh [at] oracle.com

  • BPM process to manage business data vs Business Data in RDBMS

    Hi all,
    I have so far seen BPM as a pure business process tool rather than a data management one even thought BPM provides for managing data. If in case, we have a nice Business Process, which also collects plenty of related business data, what would be the suggestion?
    1. To store Business Data in RDBMS and provide BPM with just enough info. for the process.
    (or)
    2. To store Business Data in BPM Business Catalog & do away with RDBMS (ofcourse BPM uses RDBMS for dehydration )
    In our project, we are discussing this and some points in favour of approach (1) are:
    i. For web based applications with multiple UI forms to collect data, storing data in RDBMS is way better on performance than accessing/storing in BPM business catalog
    ii. Data Retention of Business Data in BPM needs special consideration which may lead to dumping data eventually in RDBMS
    iii. UI frameworks help build UIs quick with a known data model rather than over APIs exposed by BPM
    iv. Reporting over an RDBMS data model is easier than over a Business Catalog in BPM
    Are these points valid or is approach (2) having other advantages to consdier?

    I think I'm just backing up what you had on your original post, but here's what we typically tell customers when this comes up.
    This has long been a best practice recommended by Oracle.  In Oracle’s Performance Tuning for Oracle Business Process Management Suite 11g document ,  on page 17 it states:
    "Minimize the amount of data stored in the process instance. Obviously, there is a tradeoff between the cost of storing data locally compared to storing keys and retrieving data for use within the process, which needs to be considered.
    A reasonable starting point would be to model the process state to hold only values that are needed to control the process flow and keys to get any other (external) data on an ‘as needed’ basis. If retrieval is too frequent/slow, or the systems holding that data are not always available, then move more data into the process."
    You touched on this, but decoupling the process payload and the underlying data for these reasons:
    1. The underlying data and the processes typically have different lifecycles and need to be independent of one another
    There is a need to maintain each at different times
    They are typically modified by different groups of people with differing skills
    The data stored in a database is typically the “source of truth” that sometimes must be able to be accessed and easily manipulated by applications outside of Oracle BPM; if stored as process instance data, instead of SQL extracting data from a database, the outside applications would need to access it through Oracle BPM APIs they are not necessarily familiar with
    2. Lightweight process data persistence improves performance
    The underlying message contract between the process instance and the engine that persists the payload should leverage key values where possible (think primary keys / relational keys from classic DBMS design patterns), rather than defining instance variables for every data element.  The performance of the Oracle BPM engine is improved and the data for the instances are rendered faster.
    The process instance is carrying the necessary process payload, rather than a bloated payload.  Only the information germane to the current activity should be retrieved and rendered.  This allows the application server to run more efficiently.
    At each step in the process, the process payload is hydrated and then dehydrated (read from the engine’s underlying database tables and then written back to the tables).  If this information is stored in an external database, there is no need for the overhead of this hydration and dehydration of large amounts of data to occur.
    At each step in the process, if stored externally in a separate database outside of Oracle BPM, only the data required is read and / or updated when it is required to do so.
    3. Decoupling helps speed development
    Oracle BPM was built with the Decoupled Model View Controller (MVC) pattern in mind
    One of its strengths is the architecture‘s business services layer that can make the source of the data transparent.  Given a single key value stored in the process instance payload, services can be invoked from the process and the human steps in the process that represent the “real source of truth” that the business needs.
    The MVC pattern’s model layer assumes that given the process’s key value, it is then possible to easily access underlying business data from a variety of sources including databases, EJBs and web services.   Although storing all of the information inside the process payload can be considered one of the model’s business service sources, the overhead of using this in production systems is not recommended.
    Once exposed, the business services can be  reused by any business process needing the information.
    User interfaces created with Oracle’s Application Development Framework (ADF) have out-of-the-box components and operations that take advantage of this MVC pattern.  Some examples of these out-of-the-box patterns that do not have to be programmed include:
    Database table information can easily displayed using Next and Previous that automatically retrieve the next or previous sets of rows
    Similarly, scrolling in a table with many rows up and down renders data automatically
    Both server and client side validations and rules
    Database dropdowns and cascading dropdowns 
    Forms automatically created with Master / Detail patterns
    4. Decoupling reduces the complexities arising from data synchronization
    When orchestrating various external systems into a process, care must be given to account for “Systems of Record” and the purview these systems have over data values
    Decoupling process instance data so that only key values are in the payload allow the Systems of Record to continually update the subservient element values without fear of stagnant data in the process
    Participants in the process receive the most current data values when dealing with process instances
    When data objects span several process instances, finding and updating data is easier if stored in a databaseExample: Process instance based on Orders. Several process instances may involve orders for a single customer. When the order changed, no problem, just find the process instance using correlations and updated it. When customer info changed, you need to synchronize any number of process instances. Placing the data in an RDBS made the solution simple. Simply updated the customer tables and all orders now have the latest info. No need to find related process instances and update them.
    Some BPM events don’t carry sufficient information and need enrichment to process events. With data stored in the payload, there is no easy way to enrich the event data. This is especially true with ACM events. Events in ACM do not have instance information. Storing data in the database will prevent a costly work around using a dedicated process and correlations to get the info needed.
    5. Decoupling facilitates the data capture for reporting and archiving
    Keeping data in the BPM payload takes away the option to do custom reporting (outside of BAM) and archiving of business data.
    Storing data in the RDBMS makes possible to create custom reports (outside of BAM) which would be not be possible or hard to do if all data lived in the BPM payload. Also, if you wanted to capture custom data changes or progressions thru a BPM and/or BPEL instance, RDBMS tables have a clear advantage over payload information. Payload (in most cases) would not have the data progression captured, and also reading data progression from the logs is not a recommended option.
    Many organizations have data retention policy, which requires data to be archived and be accessible. Archiving and data accessibility is very limited if data is stored in the BPM payload.
    6. The need for process Intelligence goes beyond the instance life cycle
    Instances get cleaned up from process database and many organizations are interested in not only keeping the business data but also all the BPM related intelligence related to audit trail, KPIs etc.  BAM data getting to  BI cubes is one of the ways to ensure that intelligence lives on, but viewing process audit maps, audit trails and knowing what attachments were part of the process is a very common use-case. For the latter, the common patterns used are the use of UCM or other ECM products to store the correlated set of documentation that can be brought together in a Webcenter like content portal for historical research and auditability purposes. That, couple with an application database strategy to keep correlated application data would paint the full picture for the business users.
    Hope this helps,
    Dan

  • Data provider internal error(-3000)  during DataAdapter.Fill

    I get this error sporadically.
    From other messages on this forum, it seems others are having similar problems. It appears the Oracle staff
    is still struggling to identify and fix (or did I miss a post?)
    I hope the next release can include some more informative diagnostic error messages. (no judgment implied, just a suggestion).
    But I need this to work, so I'm posting my
    info as well in hopes it will help.
    Sorry for the length, but I think all the detail may be needed to diagnose.
    I have been using ODP.Net for weeks without trouble
    until recently. I just added a CLOB field to my table yesterday, and
    the problem started. I don't know whether that's cause or coincidence.
    Possible workarounds I will try:
    1) Explicit select of columns rather than select *
    2) Retry once on failure after short delay
    Below, I'm including the exception message from the ASP.Net page.
    It occurs at the same spot in my code, but often
    it works fine. Now, an hour later, I cannot repeat the problem.
    I've edited the "Source Error" section to
    provide more of the code, so you can see the full context.
    I've also pasted the stored procedure called so you can see what that is about.
    I need this to work, so feel free to contact me if you need further information
    to diagnose this in order to fix it.
    I go live on March 7, so if I don't get a fix soon and can't find
    a workaround, I'll just make a lame error message like "Site is very busy, please try again".
    Even though the site is NOT busy. But having them try again seems to work.
    And I don't want to tell them why I'm asking them to try again <grin>.
    I am thinking of putting a try/catch block and sleeping for a second and then trying one more
    time in my code. I'll let you know if that is an effective work-around.
    David Kreth Allen
    University of Minnesota
    Carlson School of Management
    [email protected] remove XYZ to send emails
    612-625-0386
    <WebPageError>
    Server Error in '/ScholarshipApplicationWeb' Application.
    Data provider internal error(-3000)
    Description: An unhandled exception occurred during the execution of the current web request. Please review the stack trace for more information about the error and where it originated in the code.
    Exception Details: Oracle.DataAccess.Client.OracleException: Data provider internal error(-3000)
    Source Error:
         OracleCommand cmd = new OracleCommand();
         cmd.Connection = new OracleConnection(
              ApplicationEnvironment.current.database.OracleConnectionString);
         cmd.CommandType = CommandType.StoredProcedure;
         cmd.CommandText = "pkg_batch.get_details";
         cmd.Parameters.Add("batch_id",OracleDbType.Int32,
              batchId,ParameterDirection.Input);
         cmd.Parameters.Add("batchCursor",OracleDbType.RefCursor,
              DBNull.Value,ParameterDirection.Output);
         cmd.Parameters.Add("majorsCursor",OracleDbType.RefCursor,
              DBNull.Value,ParameterDirection.Output);
         cmd.Parameters.Add("detailCursor",OracleDbType.RefCursor,
              DBNull.Value,ParameterDirection.Output);
         OracleDataAdapter da = new OracleDataAdapter(cmd);
         DataSet ds = new DataSet("batches");
         da.Fill(ds);
         DataTable dt = ds.Tables[0];
         dt.TableName = "batch";
    Source File: C:\Projects\ScholarshipSolution\ScholarshipLibrary\Data\ScholarshipApplicationBatch.cs Line: 329
    Stack Trace:
    [OracleException: Data provider internal error(-3000)]
    Oracle.DataAccess.Client.OracleException.HandleErrorHelper(Int32 errCode, OracleConnection conn, IntPtr opsErrCtx, IntPtr opsSqlCtx, Object src, String procedure, String[] args)
    Oracle.DataAccess.Client.OracleDataReader.NextResult() +1259
    System.Data.Common.DbDataAdapter.FillNextResult(IDataReader dataReader) +98
    System.Data.Common.DbDataAdapter.FillFromReader(Object data, String srcTable, IDataReader dataReader, Int32 startRecord, Int32 maxRecords, DataColumn parentChapterColumn, Object parentChapterValue) +261
    System.Data.Common.DbDataAdapter.Fill(DataSet dataSet, String srcTable, IDataReader dataReader, Int32 startRecord, Int32 maxRecords) +129
    Oracle.DataAccess.Client.OracleDataAdapter.Fill(DataSet dataSet, String srcTable, IDataReader dataReader, Int32 startRecord, Int32 maxRecords) +516
    Oracle.DataAccess.Client.OracleDataAdapter.Fill(DataSet dataSet, Int32 startRecord, Int32 maxRecords, String srcTable, IDbCommand command, CommandBehavior behavior) +290
    System.Data.Common.DbDataAdapter.Fill(DataSet dataSet) +38
    CSOM.ScholarshipLibrary.Data.ScholarshipApplicationBatch.getDataById(Int32 batchId) in C:\Projects\ScholarshipSolution\ScholarshipLibrary\Data\ScholarshipApplicationBatch.cs:329
    CSOM.ScholarshipLibrary.Data.ScholarshipApplicationBatch.GetById(Int32 batchId) in C:\Projects\ScholarshipSolution\ScholarshipLibrary\Data\ScholarshipApplicationBatch.cs:290
    CSOM.ScholarshipApplicationWeb.BatchSummaryControl.Edit_Click(Object sender, CommandEventArgs e) in c:\inetpub\wwwroot\ScholarshipApplicationWeb\BatchSummaryControl.cs:68
    System.Web.UI.WebControls.LinkButton.OnCommand(CommandEventArgs e) +110
    System.Web.UI.WebControls.LinkButton.System.Web.UI.IPostBackEventHandler.RaisePostBackEvent(String eventArgument) +115
    System.Web.UI.Page.RaisePostBackEvent(IPostBackEventHandler sourceControl, String eventArgument) +18
    System.Web.UI.Page.RaisePostBackEvent(NameValueCollection postData) +138
    System.Web.UI.Page.ProcessRequestMain() +1244
    Version Information: Microsoft .NET Framework Version:1.0.3705.288; ASP.NET Version:1.0.3705.288
    </WebPageError>
    <PackageBody>
    PROCEDURE GET_DETAILS(
    p_batch_id IN number,
    batchCursor OUT pkg_batch.refcur,
    majorsCursor OUT pkg_batch.refcur,
    applicationsCursor OUT pkg_batch.refcur
    IS
    BEGIN
    OPEN batchCursor FOR SELECT
    * FROM scholarshipApplicationBatch
    where scholarshipApplicationBatch_id = p_batch_id;
    OPEN majorsCursor FOR SELECT
    * FROM major_response
    where batch_id = p_batch_id;
    OPEN applicationsCursor FOR SELECT
    * FROM schol_detail
    where batch_id = p_batch_id;
    END GET_DETAILS;
    </PackageBody>

    One workaround that seems to work is to retry in a loop.
    Sample code is pasted below.
    Obviously it may still fail, but less often.
    I'll report any further changes that affect
    the behavior in hopes it will assist the Oracle staff in
    fixing it.
    <CODE>
    try
    da.Fill(ds);
    catch (OracleException excp)
         const int maxTrys = 5;
         int failureCount = 1;
         bool succeeded = false;
         for (int i = 0;i<maxTrys;i++)
              try
                   System.Threading.Thread.Sleep(
                        TimeSpan.FromSeconds(1));
                   ds.Clear();
                   da.Fill(ds);
                   i = maxTrys;
                   succeeded = true;
              }//try
              catch
                   failureCount ++;
         }//for
         errorHandler.Report(
              String.Format(
              "Failure in da.Fill in ScholAppBatch. " +
              " Failed {0} times. Finally Succeeded = {1}.",
                                            failureCount.ToString(),
         succeeded.ToString()),
              excp);
    }//catch
    </CODE>

  • I can't believe I have to ask this...   I've minimized the "Data Provider" screen....

    Post Author: black
    CA Forum: General Feedback
    What an embarassing first post... I'm new to BusinessObjects (6.5.1) and half way through building a query I've minimized the "Data Provider" screen....  and can't restore it (can't even find it)...  it's obviously active (offscreen I guess) because everything else is now greyed out and is inactive....
    Tell me:    a)  I'm not the first person to do this...
            and b) How to get it back
    I'm unable to do anything until I have a solution to this since BO appears to save the position - if I close and re-open, or open a new query, I find myself in the same situation.
    I look forward (really) toyour suggestions!
    Thanks

    Post Author: jsanzone
    CA Forum: General Feedback
    Black,
    You're not the first person to do this, however, I've never had this problem before, only got some crib notes on the subject so this may or may not work:
    1. in regedit clear the key and all subfolders: HKEY_CURRENT_USER\Software\Business Objects\Suite 11.5\default\BusinessObjects\Application Preferences\BusinessReporter\Desktop Intelligence 2. Restart DeskI and all menus will be reset to original.

  • Render html in the Business Data web part (used for the BCS profile page)

    I am searching and indexing a webservice and pulling in external content, through SharePoint's BCS. I have used SharePoint Designer to make the connection and its a read only connection, we are not trying to write back to the external data source. The problem
    I have is that some of the columns are rendering the HTML as a string, with all of the HTML tags. For example, a column will render like this:
    &lt;p&gt; Onboarding, Recruiting &amp;amp; Talent Acquisition general information&lt;/p&gt;
    The profile page uses the business data web part and will accept custom XSL and I have modified the column that is rendering the html from:
    <xsl:value-of select="@u_cause" /> 
    to
    <xsl:value-of disable-output-escaping="yes" select="@u_cause" /> 
    The problem is that it does not make a difference the column will still render the HTML. Is there any other way to make the column render HTML?
    Joe Garcia

    Hi,
    According to your description, my understanding is that you want to render column value as the HTML format in the xslt list view.
    I suggest you can set the “disable-output-escaping” attribute like below, it will render as html:
    <xsl:value-of select="@MyColumn" disable-output-escaping="yes"/>
    Here are some detailed articles for your reference:
    http://sharepoint.stackexchange.com/questions/16089/unescape-html-from-list-column
    http://doitwithsharepoint.blogspot.com/2011/06/sharepoint-list-forms-display-html.html
    Thanks
    Best Regards
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Jerry Guo
    TechNet Community Support

  • How can I do for a row of a query be data provider for a variable?

    Hi friends, I have a problem !
    How can I do for a row of a query be data provider for a variable?
    I need that a value of variable be stored when the user select a row in a query. At the BPS we can do this configuring the variable selector in WIB, and in a WAB how I can do this ?
    Best regards,
    Gustavo Liberado

    In this case when I press the key to call other forms I need to wait for the response in the secondary form and then process the result.That is exactly what a "modal JDialog" (or JOptionPane) are used for.
    Try it. Create a short demo program. All you need is a JFrame with a single button to show the modal dialog. All you modal dialog needs is a single button to close the dialog. After you show the modal dialog add a System.out.println(...) statement in your code and you will see that it is not executed until the dialog is closed.
    Then once you understand the basics you add the code to your real program.
    If you need further help then you need to create a [Short, Self Contained, Compilable and Executable, Example Program (SSCCE)|http://homepage1.nifty.com/algafield/sscce.html], that demonstrates the incorrect behaviour.
    Don't forget to use the Code Formatting Tags so the posted code retains its original formatting. That is done by selecting the code and then clicking on the "Code" button above the question input area.

  • Best Practice on using and refreshing the Data Provider

    I have a �users� page, that lists all the users in a table - lets call it master page. One can click on the first column to of the master page and it takes them to the �detail� page, where one can view and update the user detail.
    Master and detail use two different data providers based on two different CachedRowSets.
    Master CachedRowSet (Session scope): SELECT * FROM UsersDetail CachedRowSet (Session scope): SELECT * FROM Users WHERE User_ID=?I want the master to be updated whenever the detail page is updated. There are various options to choose from:
    1. I could call masterDataProvider.refresh() after I call the detailDataProvider.commitChanges() - which is called on the save button on the detail page. The problem with this approach is that the master page will not be refreshed across all user sessions, but only for the one saving the detail page.
    2. I could call masterDataProvider.refresh() on the preRender() event of the master page. The problem with this approach is that the refresh() will be called every single time someone views the master page. Further more, if someone goes to next page (using the built in pagination on the table on master page) and clicks on a user to view its detail and then close the detail page, it does not keep track of the pagination (what page the user was when he/she clicked on a record to view its detail).
    I can find some work around to resolve this problem, but I think this should be a fairly common usage (two page CRUD with master-detail). If we can discuss and document some best practices of doing this, it will help all the developers.
    Discussion:
    1.     What is the best practice on setting the scope of the Data Providers and CahcedRowSet. I noticed that in the tutorial examples, they used page/request scope for Data Provider but session scope for the associated CachedRowSet.
    2.     What is the best practice to refresh the master data provider when a record/row is updated in the detail page?
    3.     How to keep track of pagination, (what page the user was when he/she clicked on the first column in the master page table), so that upon updating the detail page, we cab provide user with a �Close� button, to take them back to whaterver page number he/she was.
    Thanks
    Message was edited by:
    Sabir

    Thanks. I think this is a useful information for all. Do we even need two data providers and associated row sets? Can't we just use TableRowDataProvider, like this:
    TableRowDataProvider rowData=(TableRowDataProvider)getBean("currentRow");If so, I am trying to figure out how to pass this from master to detail page. Essentially the detail page uses a a row from master data provider. Then I need user to be able to change the detail (row) and save changes (in table). This is a fairly common issue in most data driven web apps. I need to design it right, vs just coding.
    Message was edited by:
    Sabir

Maybe you are looking for