Handling parameters when using BI Server as a Data Source

Hi,
Following an upgrade of our BI Publisher 10g content to 11g I've come across a syntax issue with 11g when it comes to the handling of parameters in BI Server Data Sets.
Under 10g, the documentation stated the following:
+"If your data source is the Oracle BI Server, use the following macro to handle the null:+
+{$ if ${sYear}='*'$}+
+{$elsif ${sYear}='2000' $}+
+where Year = :sYear+
+{$else $}+
+where Year = :sYear+
+{$endif$} "+
http://docs.oracle.com/cd/E10415_01/doc/bi.1013/e12187/T518230T518233.htm
Under 11g it does not seem to allow the use of macros defined in Data Sets and raises the following error when saving the query:
+"java.io.IOException: prepare query failed[nQSError: 43113] Message returned from OBIS. [nQSError: 27002] Near <$>: Syntax error [nQSError: 26012] ."+
If macros are no longer supported, what is the best approach to handle null value parameters (caused by all choices / nothing being selected)
as they need to be defined in the Data Set WHERE CLAUSE?
I've tried a rather clunky CASE WHEN statement which handles string parameters fine, but dates are another thing as null values are returned as 'null' text which causes TYPE mismatch errors.
Hopefully there is an elegant way to handle this in 11g (the 10g macros were very straight forward when dealing with BI Server)
Many Thanks

If what you are saying is true then I would highly recommend open a ticket with Oracle Support.
They should provide and alternative for handling NULLS in 11g.
regards
Jorge

Similar Messages

  • Error (IES 10901) (WIS 10901) when using Excel 2010 file as data source..

    Hello,
    I have a annoying issue with using an excel 2010 file as a data source for a universe. Here is the background information.
    Setup.
    Client: Laptop running Win XP 32bit, Client tools installed with SP2 v9 (latest)
    Server: Win 2008 x64, BOE 4.0 SP2 v9  (latest)
    Task:
    1. On Client, with information design tool, created Project, DB Connection, Data foundation using an excel source on a share
    server\share\data.xlsx
    2.. Published everything to Repository
    3. Try to load univerise in Web Intelligence via creating a new report. I Drag accross the dimensions. When i go to preview data recieve error
    Database error: [Microsoft][ODBC Excel Driver] External table is not in the expected format.. (IES 10901) (WIS 10901)
    4. If i change the excel file to a 2003 format *.xls it all works correctly. Issue being that 2003 cannot hold as much data as a 2010 file so i need to use 2010.
    Extra Notes:
    1. Both the client and server have the latest Access drivers http://www.microsoft.com/download/en/details.aspx?id=13255
    2. Both client and server have a system DSN setup using the 32bit ODBC, same DSN names to the network share.
    3. The data excel source sits on the server on a share.
    4. I am believing that the isssue is within the area of BOE as when you connect to the excel file via ODBC using crystal 2011, you need to right click on the connection and tick "System Tables" to be able to use the excel 2010 file. The BOE Information design tool does not seem to have this option??
    Any advice with this would be great?
    thanks
    david

    Issue Resolved:
    Install
    Access Database Engine 2010 Service Pack 1 on the server
    http://support.microsoft.com/kb/2460011
    Reboot server once completed.

  • Interactive Analysis error when using BW BEx query as data source

    Hello,
    When we try to create a new report based on a BW BEx with the "Interactive Analaysis" tool in the BO 4.0 Launchpad, for all the BEx queries created under one of our multiproviders, we get the following error:
    The DSL Service returned an error: com.businessobjects.mds.services.helpers.OlapUniverseHelper$DimensionsFromCubeFailedException: Cannot retrieve dimensions from cube Cartera de ventas (Ap)
         at com.businessobjects.mds.services.helpers.OlapUniverseHelper.createSapBusinessLayer(OlapUniverseHelper.java:444)
         at com.businessobjects.mds.services.helpers.OlapUniverseHelper.createBusinessLayer(OlapUniverseHelper.java:248)
         at com.businessobjects.dsl.services.workspace.impl.QueryViewAnalyzer.createOlapUniverse(QueryViewAnalyzer.java:286)
         at com.businessobjects.dsl.services.workspace.impl.QueryViewAnalyzer.createOlapUniverse(QueryViewAnalyzer.java:263)
         at com.businessobjects.dsl.services.datasource.impl.DirectOlapAccessUniverseProvider.provideUniverse(DirectOlapAccessUniverseProvider.java:97)
         at com.businessobjects.dsl.services.datasource.impl.DirectOlapAccessUniverseProvider.provideUniverseFromCms(DirectOlapAccessUniverseProvider.java:50)
         at
    With BEX queries from other multiproviders we're not getting this error and however.
    Moreover, if we try to create a new report with the "Analysis Edition for OLAP" tool in the BO Launchpad using these same problematic BEx queries, we don't get any errors with this other tool. The thing is that we need to create them with the Interactive Analaysis.
    Does anybody know why we're getting this error in the "Interactive Analysis" tool and how can we solve it?
    Any help would be very appreciated.
    Thank you and regards,
    Edited by: Juan Jose Catoira on Jul 28, 2011 2:46 PM
    Edited by: Juan Jose Catoira on Jul 28, 2011 2:47 PM

    Hey,
    Did you find out any solution for this. I am also facing the same issue.
    Thanks in advance.

  • Using Oracle BI Server as a Data Source for Third Party Tools

    I have been researching how to use the Oracle BI Server as a data source for two days now. I have not yet found a solution to my problem.
    We have a remote user who is pulling data from one of our Oracle databases using automated tools. This data base is essentially being replaced by Oracle Analytics. However, we do not want this user to get at the physical tables of Analytics, rather, we want the user to be able to connect to the Oracle BI Server itself, and issue Logical SQL to pull data.
    Here are the requirements which I need to be fulfilled:
    1) The ability for a remote user to connect to the BI Server as a data source.
    2) The remote user must be able to do this on a Unix box.
    3) The remote user must be able to automate the process of issuing SQL to the BI Server. (Command line interface is preferable)
    4) The solution needs to be easily implementable, that is, this solution will be implemented many times over.
    I found many articles explaining how to set up a Windows box to connect to the BI Server in this way. However, these articles outline how to use tools like Excel and Microsoft Query to pull data from the BI Server. Our current user is using SQLplus.
    If possible, I would like the user to continue using SQLplus as the tool for connecting to and retriving information from our BI Server. I have tested using SQLplus many different ways to no avail. I have also tried using other tools, such as Oracle SQL Developer.
    Is there a program out there that will fulfill the needs I have outlined above? Or, does someone know of a way to use SQLplus to connect to the Oracle BI Server as a data source? Or, is there a series of solutions which will fulfill the needs outlined above?
    Edited by: user11177366 on Jun 11, 2009 11:46 AM

    I agree it would be inefficient but only because you are using OBIEE for something it was not designed to do. You are basically asking for a "DB feed" and that should designed using PL/SQL, DB Links, Web Services or whatever is your entreprise standard to send data from one system to another. Using OBIEE for this seems a bad architecture decision to me.You are correct, the BI Server is not really intended to be queried like another Oracle Database remotely. However, the reason we are trying to do it this way is twofold:
    1) We want to enforce data level security.
    2) The BI Server has been fine tuned for high performance.
    Forcing the end users to go through the BI Server rather than directly to the physical tables or other sorts of systems built on top solves both of these issues.
    I think you are wrong. I don't see why SQLPlus should be able to query the BI Server. The BI Server can sit on top of many different DBs not just Oracle, even flat files etc. So you can't really expect the BI Server to able to translate all that into Oracle SQL output so you can query it using SQLPlus. The BI Server is not an Oracle RDBMS.You're right again. The BI Server isn't an Oracle RDBMS, and I have been wanting to treat it like one. From the data I have gathered SQLPlus can't query the BI Server.
    However, my research has led me to realize that the BI Server supports the JDBC API. Therefore, (I should have went down this route in the first place) one can use the JDBC driver provided by Oracle (bijdbc14.jar) to interface with the BI Server. There are many programs and applications out there that support the JDBC API and let us connect to BI Server the way we want to. Furthermore, these programs are mostly platform independent, as they are mostly written in Java.
    Lastly, from what I understand the NQCMD program can be installed on any platform as well (provided by Oracle). This is another option to connect to the BI Server as well.
    Thanks for your help and your comments as I went through this process of discovery.

  • UME using SAP R/3 as Data Source

    Hi,
    We are trying to set User authentication to SAP R/3 system, not load balanced system, on the User Management Configuration values: Client=501, Userid=sapjsf, Password=pwd, sys id=RS1, Group and Message server= blank, Application server= server.company.com, Sys. number=00, Max pool=10, Max wait=300000.
    When testing connection, I get this message:
    (System ID): com.sap.mw.jco.JCO$Exception: (101) RFC_ERROR_PROGRAM: 'mshost' missing
    (System ID & System Number): OK
    Is this an error? since our SAP R/3 is not a load balanced system.
    Did we miss any item for the setup, in dataSourceConfiguration_r3.xml? The SAPJSF "communication user" got the right sap role and authorizations.
    Portal version : EP6 SR1
    Regards
    Huzaifah

    Hi,
    If u want to Use The SAP R3 System as Data source u may
    do it from config tool if u got following message.
    WARNING! You are not allowed to select dataSourceConfiguration_r3.xml as active configuration file.
    (For Portal Patch less than SP13 u must download two data source file which is attached with note - 718383
    and upload it to portal which is described in the note)
    the following are the procedure which i apply ,
    Go to System Administration -> System Configuration ->UM  Configuration
    Now Do not change Data source from Here.
    Make sure  your data source is "Database Only"
    (dataSourceConfiguration_database_only.xml)
    Now enter the following value under SAP System Tab.
    Client : - Your sap system client
    User:-  Sap user
    password: - password
    System language:- your system language
    Application server: - Host name or IP of sap system
    System Number : -  SAP instance number
    Maximum Size of Connection Pool : -  As per req.
    Maximum Wait Time in Milliseconds :- 10000
    Now, save the changes and shutdown the portal server.
    Using Config Tool change the data source. Run the following
    <drive:\> usr\sap\<sid>\JC<instance number>\J2EE\configtool\configtool.bat
    (Make sure the portal system is shutdown)
    Under Cluster Data -> Global Server Configuration -> services -> com.sap.security.core.ume.services
    Now find the key: -  ume.persistence.data_source_configuration     
    The default was : - dataSourceConfiguration_database_only.xml
    change the value to :- dataSourceConfiguration_r3.xml     
    click on set and from flie-> apply
    Now restart the portal server ur data source changer to SAP R3 System
    Regards,
    Kaushal

  • Using COGNOS as an infoprovider/data source for WebI...?

    I saw a job posting today that listed a requirement that I am pretty sure is IMPOSSIBLE...
    WebIntelligence report experience with using COGNOS as an infoprovider/data source for BOBJ is an important skill.
    Data sources for client will be both COGNOS and SAP.
    Now, it wouldn't be the first time that a client spec asked for the IMPOSSIBLE- but it got me to wondering if maybe I am wrong in my understanding that COGNOS 8 still uses PowerPlay cubes that are 100% proprietary and can only be read with the CONGOS report-design tools.
    Is there actually a "supported" way to use COGNOS as an infoprovider/data source for BOBJ...?
    Maybe using some COGNOS equivalent of QAAS...?
    I have no intention of applying for this job - but would like to know if it is technically possible.

    It's been more than a week since I posted this question and nobody came up with a "YES" here or on the BOB Forums...so I am going to close this question as an IMPOSSIBLE request.
    Based on the current technology - COGNOS cubes can not be used as an infoprovider/data source for WebIntelligence.

  • Hardcoded BAM server parameters when using BAM sensors

    The problem we're having is that when a BPEL process has BAM sensors, the BAM server configuration parameters are hardcoded in bpel.xml and sensorAction.xml.
    This causes problems when the process is being deployed to a different instance than the development instance.
    I solved the problem for the bpel.xml file with the <customize> ant task, but haven't found a solution yet for sensorAction.xml.
    Any suggestions?
    Thanks,
    Arnold van Koppen

    See..when the BAM Server is down the corresponding data objects wont be available to you..During the design time wat u do is create a variable which contains the elements dat u wanna monitor in the BAM...IN BAM u create data objects ..then u map the schema from the BPEL to the BAM Data Objects ...U need a BAM Server connection which pulls up the metadata from the server ..If the BAM Server is down u wont get the BAM DOs...to keep a back up u can use DB as the sensor action so that the data from the sensor action is written to the DB..Using a DB Adapter i think the data from the Reports schema can be written to the OracleBAM schema ...so that even if the BAM Server is down the data is written to the OracleBAM schema

  • Missing Parameters when using .PrintToPrinter or .ExportToDisk

    I am creating a viewer that will set login information and allow the users to view, print or export reports.  VB.NET 2010 WinForms Crystal Engine 13.0.9.1312.  All reports are external designed by various programmers.
    The preview mode of my viewer is now working quite well (with the help of Don and Ludek most recently).  Anything I want to see on screen works great in the regular Crystal viewer.
    The problem is when I want to export to pdf or print to a printer without viewing on screen in the regular Crystal viewer.
    If there are no parameters in the reports, both exporting to pdf and printing to any printer selected works.
    When there is one or more parameters, it crashes and points to the line
    crReport.ExportToDisk(ExportFormatType.PortableDocFormat, strFile)
    or
    crReport.PrintToPrinter(1, True, 0, 0)
    with the message Missing Parameter Values.
    Through some searching (and yes I have checked on this site and Google for "Missing Parameter Values") I have found one big issue is that parameter values have to be set after the .ReportSource is set.  So before I changed my code, I never saw the Crystal Parameter prompt screen.  Now I see it but still get the crash (though once I have had it go to printer honouring the Parameter in a report, but I can't always reproduce it and never when sending to pdf).
    The reports may be designed by end users who are using my program.  I will not know what reports they are or if they have even put in parameters.  I will not know if they are date, string, number, discrete, range, multiple, etc.
    The original login code referred to below can be found here.
             'original login code here down to subreport login
             'rest of subroutine is just a Catch and End Try
             crViewer.ReportSource = Nothing
             crViewer.ReportSource = crReport
             'check if to be sent direct to printer and send all pages
             If Not autoforms Is Nothing AndAlso autoforms.ToPrinter Then
                crReport.PrintOptions.PrinterName = autoforms.Printer
                crReport.PrintToPrinter(1, True, 0, 0)   '-> Error Triggers Here
             End If
             'setup viewer options
             Dim intExportFormatFlags As Integer
             intExportFormatFlags = ViewerExportFormats.PdfFormat Or
                                     ViewerExportFormats.ExcelFormat Or
                                     ViewerExportFormats.ExcelRecordFormat Or
                                     ViewerExportFormats.XLSXFormat Or
                                     ViewerExportFormats.CsvFormat
             crViewer.AllowedExportFormats = intExportFormatFlags
             crViewer.ShowLogo = False
             crViewer.ToolPanelView = ToolPanelViewType.GroupTree
             'check if to be sent direct to pdf and send all pages
             If Not autoforms Is Nothing AndAlso autoforms.ToPDF Then
                Dim strPDFFilename As String = autoforms.PDFPath & "\" & autoforms.ReportName & "_" & autoforms.ReportType & ".pdf"
                'strPDFFilename = Replace(strPDFFilename, " ", "_")
                SendExport(strPDFFilename, "PDF")
             End If
             'if not preview then close form
             If Not autoforms Is Nothing AndAlso Not autoforms.Preview Then
                Me.Close()
                Exit Sub
             End If
    Private Sub SendExport(ByVal strFile As String, ByVal strExportFormat As String)
          'pass current parameter values to the pdf
          Dim crParamFieldDefs As ParameterFieldDefinitions
          Dim crParamFieldDef As ParameterFieldDefinition
          Dim crParamValues As New ParameterValues
          Dim x As Integer = crViewer.ParameterFieldInfo.Count
          crParamFieldDefs = crReport.DataDefinition.ParameterFields
          'loop through each parameter and assign it the current values
          For Each crParamFieldDef In crParamFieldDefs
             crParamValues = crParamFieldDef.CurrentValues
             crParamFieldDef.ApplyCurrentValues(crParamValues)
          Next
          Select Case strExportFormat
             Case Is = "PDF"
                crReport.ExportToDisk(ExportFormatType.PortableDocFormat, strFile)  '--> Error Triggers Here
             Case Is = "xls"
                crReport.ExportToDisk(ExportFormatType.Excel, strFile)
             Case Is = "xlsx"
                crReport.ExportToDisk(ExportFormatType.ExcelWorkbook, strFile)
          End Select
    End Sub
    This code doesn't work on my system, but my logic is as follows (as flawed as it is).  Look through the parameters, find the current values and then assign them to the parameters again just before export.
    With the code above, I am getting the enter parameter values screen in the viewer.  Therefore, my assumption is that the report now has a current value(s) for the parameter(s).  The loop will not require the program to know how may parameters there are or what the values or value types are.
    There are a couple of problems with this.  When I step through it, there are no current values, because the prompt doesn't come up with stepping through the code with F8.  Second problem is that it is finding the internal parameters used for subreport linking ({?PM-xxxxx}).  Third problem is that a lot of code I have looked at does not appear like it is assigning the reportsource to a viewer at all just to export or print to printer.  So is it necessary to do as I have done here?
    I am sorry this is so long and I appreciate anyone who has stuck to it this long.  I am at my wits end trying to figure out how to make this work.  I think there is a timing issue here but do not know how to resolve it.  I just want the report to trigger the prompts if there are parameters, then send the report screen, to pdf or to printer depending on the user's settings for the report (it may be any or all three options).  If you need more info, let me know.
    TIA, rasinc

    Hi
    I'd like to look at this from the "how parameters" work point of view. Then let's see if we can drill in on some of the details.
    The parameter prompt screen is driven by the viewer. E,g.; the screen will only be spawned on invocation of the viewer. If the viewer is not invoked (e.g.; direct print / export) the parameter prompt screen will not be displayed. Now, in your second post you say:
    With the code above, I am getting the enter parameter values screen in the viewer.
    This does not sound correct based on the above. E.g.; if the correct parameters are passed to the engine, the viewer should never be popping up the parameter screen. So to me it looks like the code for the parameter is not actually working. It looks like it is as there is no error as such. But as you get prompted, the engine is essentially saying; "I don't understand those parameters so as smart as I am, I'll use the parameter prompt screen to get them from the user on a silver platter".
    Ok. Now to some troubleshooting. I'd pick a report where I know the parameter type and acceptable values ahead of time. Don't use date or date / time params to start with as these have other complication. Use the parameter code to set the value(s) - preferably one value (I like to keep things simple). If the parameter is passed to the report engine correctly, there will not be a parameter prompt and the report will simply come up in the viewer. Once this works, add in a code for printing / exporting and comment out the viewer code. The report should print / export without any issues (no parameter prompt, no error). Repeat the above with a multi-parameter report.
    Then it will come to those date and date/time parameters. NULL handling, etc., etc. E.g.; this will not be trivial. And you will somehow need to know the parameter type (string / int / date) before you try to pass it into the report. More help:
    Sample apps
    vbnet_win_paramengine.zip
    vbnet_win_paramviewer.zip
    vbnet_win_multirangeparam.zip
    vbnet_win_rangeparameters.zip
    vbnet_win_sub_daterange_param_engine.zip
    vbnet_win_sub_daterange_param_viewer.zip
    The above samples are here: Crystal Reports for .NET SDK Samples - Business Intelligence (BusinessObjects) - SCN Wiki
    Helpful KBAs
    1688710 - When printing a Crystal report with parameters using 'PrintToPrinter' method, report does not prompt for parameter values
    1214110 - How to pass a Date/Time parameter to a report at runtime using VB .NET
    1496220 - Suppress the default parameter prompt when the parameter value is not passed through the code
    1893554 - How to pass a NULL value to a Crystal Reports optional parameter field using Visual Studio .NET
    1425379 - How to avoid being prompted for a parameter when refreshing a report running in a VS .NET application
    1461315 - Error; "Missing Parameter Values" when running a parameterized report using Crystal Reports SDK for VS .NET
    1215426 - Err Msg: "Missing parameter value" when printing or exporting in VS .NET
    And finally this:
    Crystal Reports for Visual Studio 2005 Walkthroughs
    (The above doc is good for all versions of CR and VS. It has some amazingly great "stuff" in it re. parameters, etc., and more.)
    - Ludek
    Senior Support Engineer AGS Product Support, Global Support Center Canada
    Follow us on Twitter

  • Handle leak when using ADO2.7 access oracle

    Ado version:2.7
    OS:Window2000 professional
    Oracle Server:9.2.0.1.0
    Client Provider:9.2.0.1.0
    Using Oracle OLEDB connection oracle
    when i execute such function in my test program continuously:
    Handle count of the program will increase.I know handle will increase for connection pool ,but it will stop increasing when handle count reach one value. i think this is a problem that ADO cann't match with Oracle OLEDB
    better.
    Any suggestion?
    Set mConOracle = New Connection
    mConOracle.CursorLocation = adUseClient
    mConOracle.ConnectionString = "Provider=OraOLEDB.Oracle.1;Password=123;Persist Security Info=True;User ID=abc;Data Source=aaa;OLE DB Services=-1; "
    mConOracle.Open
    mConOracle.Close
    Set mConOracle=Nothing

    Ok,
    I tried out connection pooling switching on/off. Also waited some minutes for the pool timeout. But nothing happened.
    So I made some further tests and this is what I found out:
    - The left handles are all semaphores
    - To generate a handle leak it is necessary to create an ADO object, call connect and do a database query. Actually some data must be transmitted over the network. Then close connection, release ADO object. -> 2-3 semaphore handles are left.
    - Handle leak occurs indepentendly of apartment or multi threaded
    - MDAC 2.7, MDAC 2.8 was used with no success
    However, we hat one piece of software here, which didn't have handle leaks. So I did some more investigation on that phenomena.
    This is what happened and could be used as a workaround until a fix is available:
    - I recognized, that if at least one connection (which has been really established once by the underlying components) has been opened and is kept open, it is possible to create more ADO objects, use them for database retrieval and release them again without having the handle count increased.
    - So the handle leak seems only to occur, if there is no more database connection in use and one of the underlying components tries to release everything. Each time this happens 2-3 semaphore handle are left.
    - Finally my current workaround is to create and open on initalization of my process an ADO object and the database connection and leave this object alone without touching it again until the process ends. For every database query in any thread I can now use a new ADO object, without having the handles increased and never released.
    So, this is the end for now. I hope this information can help somebody to figure out where the problem is eventually located and a fix could be released.
    Hopefully,
    Markus (writing under Uwes account)

  • Current Security Context Not Trusted When Using Linked Server From ABAP

    Hello,
    I am experiencing a head-scratcher of a problem when trying to use a Linked Server connection to query a remote SQL Server database from our R/3 system.  We have had this working just fine for some time, but after migrating to new hardware and upgrading OS, DBMS, and R/3, now we are running into problems.
    The target database is a named instance on SQL Server 2000 SP3, Windows 2000 Server.  The original source R/3 system was 4.7x2.00, also on SQL Server 2000 (SP4), Windows 2000 Server.  I had been using a Linked Server defined via SQL Enterprise Manager (actually defined when the source was on SQL Server 7), which called an alias defined with the Client Network Utility that pointed to the remote named instance.  This alias and Linked Server worked great for several years.
    Now we have migrated our R/3 system onto new hardware, running Windows Server 2003 SP1 and SQL Server 2005 SP1.  The application itself has been upgraded to ECC 6.0.  I performed the migration with a homogeneous system copy, and everything has worked just fine.  I redefined the Linked Server on the new SQL 2005 installation, this time avoiding the alias and referencing the remote named instance directly, and it tests out just fine using queries from SQL Management Studio.  It also tests fine with OSQL called from the R/3 server console, both when logged on as SAPServiceSID with a trusted connection, and with a SQL login as the schema owner (i.e., 'sid' in lowercase).  From outside of R/3, I cannot make it fail.  It works perfectly.
    That all changes when I try to use the Linked Server within an ABAP application, however.  The basic code in use is
    EXEC SQL.
       SET XACT_ABORT ON
       DELETE FROM [SERVER\INSTANCE].DATABASE.dbo.TABLE
    ENDEXEC.
    The only thing different about this code from that before the upgrade/migration is the reference to [SERVER\INSTANCE] which previously used the alias of just SERVER.
    The program short dumps with runtime error DBIF_DSQL2_SQL_ERROR, exception CX_SY_NATIVE_SQL_ERROR.  The database error code is 15274, and the error text is "Access to the remote server is denied because the current security context is not trusted."
    I have set the "trustworthy" property on the R/3 database, I have ensured SAPServiceSID is a member of the sysadmin SQL role, I've even made it a member of the local Administrators group on both source and target servers, and I've done the same with the SQL Server service account (it uses a domain account).  I have configured the Distributed Transaction Coordinator on the source (Win2003) system per Microsoft KB 839279 (this fixed problems with remote queries coming the other way from the SQL2000 system), and I've upgraded the system stored procedures on the target (SQL2000) system according to MS KB 906954.  I also tried making the schema user a member of the sysadmin role, but naturally that was disastrous, resulting in an instant R/3 crash (don't try this in production!), so I set it back the way it was (default).
    What's really strange is no matter how I try this from outside the R/3 system, it works perfectly, but from within R/3 it does not.  A search of SAP Notes, SDN forums, SAPFANS, Microsoft's KnowledgeBase, and MSDN Forums has not yielded quite the same problem (although that did lead me to learning about the "trustworthy" database property).
    Any insight someone could offer on this thorny problem would be most appreciated.
    Best regards,
    Matt

    Good news! We have got it to work. However, we did it in something of
    a backwards way, and I'm sure you'll laugh when you see how it was done. Also, the solution depends upon the fact that the remote server is still using SQL Server 2000, and so doesn't have quite so many restrictions placed upon it for distributed transactions and Linked Servers as SQL Server 2005 now does.
    At the heart of the solution is the fact that the Linked Server coming FROM the remote server TO our SAP system works fine. Finally, coupled with the knowledge that using DBCON on the SAP side to the remote server also does actually provide a connection (see Notes 323151 and 738371), we set up a roundabout way of achieving our goal. In essence, from ABAP, we set up the DBCON connection to the remote server, at which point all the Native SQL commands execute in the context of the remote server. From within that connection, we
    reference the tables in SAP via the Linked Server defined on the remote
    server, as if SAP were the remote server, selecting data from SAP and inserting it into the remote (but apparently local to this connection) tables.
    So, to spell it out, we define a Linked Server on the remote server pointing back to the SAP server as SAPSERV, with a SQL login mapping defined on the remote system pointing back to a SQL login in the SAP database. We also define a connection to the remote server from SAP using DBCON, using that remote SQL login for authentication.
    Then, in our ABAP code, we simply do something along the lines of
    exec sql.
       set connection 'REMOTE'
    endexec.
    exec sql.
       connect to 'REMOTE'
    endexec.
    exec sql.
       insert into REMOTE_TABLE
          select * from SAPSERV.SID.sid.SAP_TABLE
    endexec.
    exec sql.
       commit
    endexec.
    exec sql.
       disconnect 'REMOTE'
    endexec.
    This is, of course, a test program, but it demonstrated that it worked,
    and we were able to see that entries were appropriately deleted and inserted in the remote server's table. The actual program for use is a little more complex, in that there are about four different operations at different times, and we had to resolve the fact that the temp table SAP_TABLE was being held in a lock by our program, resulting in a deadly embrace, but our developer was able to work that out, and all is now well.
    I don't know if this solution will have applicability to any other customers, but it works for us, for now.
    SAPSERV, REMOTE, REMOTE_TABLE, and SAP_TABLE are, of course, placeholder names, not the actual server or table names, so as not to confuse anyone.
    Best regards,
    Matt

  • SP 2013 - Database Recovery Mode When Using SQL Server Mirroring

    Hello Community,
    Our DBA is configuring a High Availability Architecture for our SharePoint 2013 Farm.  Essentially he is using SQL Server Database mirroring with a clustered automatic failover using a witness server.  He has informed me that mirroring requires
    that the databases are set to Full Recovery Mode, but that several of the databases are set to Simple Recovery Mode, the databases are the following:
    SearchService_DB
    SearchService_DB_CrawlStore
    SearchService_DB_AnalyticsReportingStore
    SearchService_DB_LinksStore
    WSS_UsageApplication
    User_Profile_DB
    User_Sync_DB
    User_Social_DB
    However, when I checked the following article -
    http://technet.microsoft.com/en-us/library/cc678868.aspx - I see that Simple Recovery Mode is the default configuration for these databases.  So, finally, here is my question - will it be detrimental to these databases to set them to the Full Recovery
    Mode?
    Thanks!
    Tom
    Tom Molskow - Senior SharePoint Architect - Microsoft Community Contributor 2011 and 2012 Award -
    Linked-In - SharePoint Gypsy

    This really depends on your recovery objectives. The main reason for taking log backups, aside from truncating the log, is to be able to recover to a specific point-in-time. This is very helpful in mission-critical, highly transactional systems. I have not
    seen a SharePoint environment require a point-in-time restore of the content or application service database. So for this case, you can overwrite the LDF bakups. Just make sure that you can recover your databases appropriately and meet your recovery objectives
    on the farm level.
    As far as the log file growth is concerned, you are correct. If you cap the size of your LDF file to let's say 10 GB, regular log backups will truncate the log, thereby, having space for additional transaction log records. The only risk here is when your
    log backups are not frequent enough that the LDF file fills up before the next log backup runs to truncate the log. You run the risk of your database behaving as read-only and that will affect the site collections/applications using those databases
    Edwin Sarmiento SQL Server MVP | Microsoft Certified Master
    Blog |
    Twitter | LinkedIn
    SQL Server High Availability and Disaster Recover Deep Dive Course

  • Oracle giving Block corruption errors when using CDC for sending the data to SQL Server 2012

    Hello Friends,
    We are facing an error while using CDC with Oracle. It is a "Block corruption" error, which indicates at some level of data corruption. We ran RMAN validate command to scan the database for corruption but it returned with no errors, however he
    Alert Log in Oracle is still coming up with the following error. Has anyone experienced this error when using Oracle Standard Edition and SQL 2012 ?
    Trace file e:\app\pulse-ad\diag\rdbms\orcl\orcl\trace\orcl_ora_5992.trc
    Oracle Database 11g Release 11.1.0.7.0 - 64bit Production
    Windows Server 2003 Version V5.2 Service Pack 2
    CPU                 : 4 - type 8664, 4 Physical Cores
    Process Affinity    : 0x0000000000000000
    Memory (Avail/Total): Ph:6782M/24575M, Ph+PgF:12203M/30844M
    Instance name: orcl
    Redo thread mounted by this instance: 1
    Oracle process number: 151
    Windows thread id: 5992, image: ORACLE.EXE (SHAD)
    *** 2013-12-12 03:04:33.655
    *** SESSION ID:(1281.3832) 2013-12-12 03:04:33.655
    *** CLIENT ID:() 2013-12-12 03:04:33.655
    *** SERVICE NAME:(orcl) 2013-12-12 03:04:33.655
    *** MODULE NAME:(xdbcdcsvc.exe) 2013-12-12 03:04:33.655
    *** ACTION NAME:() 2013-12-12 03:04:33.655
    Lost-write detected for sequence 70856. The lost-write starts occurring in block 11193. The current block being validating is 12930.
    Block dump of the first lost-write block:
    Flag: 0x1 Format: 0x22 Block: 0x00002bb9 Seq: 0x000114bf Beg: 0x94 Cks:0x68ee
    Dump of memory from 0x0000000598D06C00 to 0x0000000598D06E00
    598D06C00 00002201 00002BB9 000114BF 68EE8094  [."...+.........h]
    598D06C10 00085BF1 0023BDA1 000DE19C 000DE19C  [.[....#.........]
    598D06C20 0000000C 00000000 2209160A 5BF10000  [..........."...[]
    598D06C30 3EB10502 00C0F5CA 0031BDA1 00010205  [...>......1.....]
    598D06C40 02B22C6A 038A6D69 00000001 00000000  [j,..im..........]
    598D06C50 4D554407 30373230 35BB0206 001100AE  [.DUM0270...5....]
    598D06C60 0001040A 000D000E 038A6D69 56B25735  [........im..5W.V]
    598D06C70 729C0003 E19C0001 000C0006 000D0006  [...r............]
    598D06C80 02BB0502 00C0F5CD 0023BDA1 000A0002  [..........#.....]
    598D06C90 00C00013 000000D0 00030201 56B25736  [............6W.V]
    598D06CA0 03890001 00000000 00000000 002E0105  [................]
    598D06CB0 FFFF0003 00C0F5CD 56B25736 3EB10003  [........6W.V...>]
    598D06CC0 FFFF0024 0014000C 000C0018 00120014  [$...............]
    598D06CD0 09CC0058 E75B0022 0009000F 00085BF1  [X...".[......[..]
    598D06CE0 0024BDA1 000DE19D 000DE19D 0000000C  [..$.............]
    598D06CF0 00000000 2309160A 5BF10000 3EB10502  [.......#...[...>]
    598D06D00 00C0F5CD 0020BDA1 00010205 02B22C72  [...... .....r,..]
    598D06D10 03900974 00000019 00000000 3030300A  [t............000]
    598D06D20 33303030 06323132 AE35BB02 0B441100  [0003212...5...D.]
    598D06D30 0001040A 000D000E 03900974 56B25736  [........t...6W.V]
    598D06D40 729C0003 E19D0011 000C0006 000D0006  [...r............]
    598D06D50 02BB0502 00C0F5CD 0024BDA1 00EA0002  [..........$.....]
    598D06D60 00270016 000001FC 00032C01 56B25736  [..'......,..6W.V]
    598D06D70 00000001 00000000 30393007 002E0105  [.........090....]
    598D06D80 FFFF0003 00C0F5CD 56B25736 00000003  [........6W.V....]
    598D06D90 FFFF0025 00140052 000C0018 00070035  [%...R.......5...]
    598D06DA0 0003000A 00070003 0001001D 00030001  [................]
    598D06DB0 00010001 00010001 00010001 00010001  [................]
    598D06DC0 00010001 00010001 00010001 00010001  [................]
    598D06DD0 00010001 00000001 00010001 00010001  [................]
    598D06DE0 00010001 00000014 09720174 00000022  [........t.r."...]
    598D06DF0 0009000F 00085BF1 0025BDA1 000DE19A  [.....[....%.....]
    Block dump of the current block being validating:
    Flag: 0x1 Format: 0x22 Block: 0x00003282 Seq: 0x000114c8 Beg: 0x0 Cks:0x312a
    Dump of memory from 0x0000000598DDFE00 to 0x0000000598DE0000
    598DDFE00 00002201 00003282 000114C8 312A8000  [."...2........*1]
    598DDFE10 50424703 31303607 34353335 69745319  [.GBP.6015354.Sti]
    598DDFE20 6E696C72 72502067 6375646F 4C207374  [rling Products L]
    598DDFE30 4E206474 C3025650 0380013D 0457454E  [td NPV..=...NEW.]
    598DDFE40 4E1E09C2 1E09C204 10C2024E 1E09C204  [...N....N.......]
    598DDFE50 09C2044E C2024E1E 03C30510 021B0929  [N....N......)...]
    598DDFE60 C3053DC3 0F192602 2602C305 C3050F19  [.=...&.....&....]
    598DDFE70 0C1A6203 5102C105 C2041F4E 044E1E09  [.b.....QN.....N.]
    598DDFE80 4E1E09C2 0410C202 4E1E09C2 1E09C204  [...N.......N....]
    598DDFE90 10C2024E 2903C305 78071B09 011D0B71  [N......)...xq...]
    598DDFEA0 BF020101 1FBF0215 4E018001 53014E01  [...........N.N.S]
    598DDFEB0 0723002C 0B0C7178 0A3C3C18 30303030  [,.#.xq...<<.0000]
    598DDFEC0 33373030 4D033337 47034255 36075042  [007373.MUB.GBP.6]
    598DDFED0 38333936 4E113331 2065776B 74616C50  [693813.Nkwe Plat]
    598DDFEE0 6D756E69 56504E20 0B0AC303 4E038001  [inum NPV.......N]
    598DDFEF0 C2045745 0459512E 59512EC2 5253C203  [EW...QY...QY..SR]
    598DDFF00 512EC204 2EC20459 C2035951 C3055253  [...QY...QY..SR..]
    598DDFF10 1B092903 0B0AC303 3C04C305 C3053239  [.).........<92..]
    598DDFF20 32393C04 4F08C305 C105114F 1F4E5102  [.<92...OO....QN.]
    598DDFF30 512EC204 2EC20459 C2035951 C2045253  [...QY...QY..SR..]
    598DDFF40 0459512E 59512EC2 5253C203 2903C305  [.QY...QY..SR...)]
    598DDFF50 78071B09 01190A71 C0030101 C0034709  [...xq........G..]
    598DDFF60 8001330A 4E014E01 002C5301 71780723  [.3...N.N.S,.#.xq]
    598DDFF70 3C180B0C 30300A3C 30303030 33373337  [...<<.0000007373]
    598DDFF80 42554D03 50424703 31304207 344C5131  [.MUB.GBP.B011QL4]
    598DDFF90 6F725020 63657073 614A2074 206E6170  [ Prospect Japan ]
    598DDFFA0 646E7546 64724F20 44535520 30302E30  [Fund Ord USD0.00]
    598DDFFB0 04C30331 03800133 0557454E 5B1603C3  [1...3...NEW....[]
    598DDFFC0 03C30521 04215B16 1F4004C3 1603C305  [!....[!...@.....]
    598DDFFD0 C305215B 215B1603 4004C304 03C3051F  [[!....[!...@....]
    598DDFFE0 031B0929 043304C3 4D245AC2 245AC204  [).....3..Z$M..Z$]
    598DDFFF0 02C3054D 040A1A18 494002C1 1603C305  [M.........@I....]
    *** 2013-12-12 03:05:07.984
    ** LOGMINER WARNING - Invalidated 6 LCRs **
    Complete dump of first invalid START LCR follows:
    ++  LCR Dump Begin: 0x0000000532C004E0 - CANNOT_SUPPORT
         op: 255, Original op: 3, baseobjn: 0, objn: 233316, objv: 1
         DF: 0x00000002, DF2: 0x00000000, MF: 0x00000000, MF2: 0x00000000
         PF: 0x40000001, PF2: 0x00002000
         MergeFlag: 0x00, FilterFlag: 0x00
         Id: 0, iotPrimaryKeyCount: 3, numChgRec: 4
         NumCrSpilled: 0
         RedoThread#: 1, rba: 0x0114c8.0001c6ce.00d4
         scn: 0x0003.56b593be, xid: 0x0008.00c.00100d85, pxid: 0x0008.00c.00100d85
         ncol: 0newcount: 0, oldcount: 0
         LUBA: 0x3.c109c0.c.15.38f64
    Thanks
    Dee

    Hi Dee,
    Thank you for your question.
    I am trying to involve someone more familiar with this topic for a further look at this issue. Sometime delay might be expected from the job transferring. Your patience is greatly appreciated.
    Thank you for your understanding and support.
    Regards,
    Mike Yin
    TechNet Community Support

  • HTTP Error with Status code 403 when using ABAP Server Proxies

    Hi All,
    I have been working on ABAP Client and Server proxies in my project. The weblogs by Ravikumar, Prateek Shah,  Siva Maranani & VijayaKumari were very helpful and thank them all for just helpful and guiding postings.
    Interfaces using Client proxies are working fine and have been tested successfully. But I'm facing bit problems with Server Proxies. The scenario is of File to SAP system via XI. Data is coming correct up to Technical routing in pipeline but at call adapter(for Server Proxy to SAP)in XI pipeline I receive Error:
    HTTP response contains status code 403 with the description Business Server Page (BSP) Error Error when sending by HTTP (error code: 403, error text: Business Server Page (BSP) Error)
    Previously I was receiving authentication and authorization related error of status code 401 and then in communication channel I replaced user profile with a user id having role SAP_XI_APPL_SERV_USER. The basis peron has created new user id with role SAP_XI_APPL_SERV_USER  only and that I'm using. I've tried creating a HTTP destination with the created user id and in test connection too receive same error message.
    Please help me out with your inputs and understanding on the topic.
    Thanks and Best Regards
    Prashant Rajani

    Hi Prashant,
    Please take a look at these threads..They deal with the same problem, and they have been resolved here.
    BSP Authorization error
    Service cannot be reached .  error code 403 the reason forbiddn
    403 Error when accessing the server
    Hope your issue gets resolved.
    cheers,
    Prashanth
    P.S Please mark helpful answers

  • Group folders when using Lion server machine as a client

    Hi,
    I want to use the server we use in a small lab as a client as well. Most of the collaborative work is done in a group folder, which works perfectly on all clients.
    However, when I log into the OD on that sever on the server itself (as a OD user) the logon procedure correctly finds that it should mount the group folder.
    However it also sees that the shared folder is a local folder, so it suggests (complains) that I should use the file locally instead.
    This is all wery well, but I have applications where the config items require a path to be specified. This path now differs between computers (server computer v.s. pure clients) which creats a problem for me.
    So, how do I make the server (loop) mount itself as if it was a pure client (and not complain about it)?
    Please help me.
    Fredrik

    Have you try this command ?
    $ sqlplus scott/tiger@scomachine
    -agun-
    null

  • Selecting data source in physical view in OBIEE 11G when using sql server

    Hi,
    I doing my physical view against a 2008 R2 SQL Server. Looks like the list goes upto 2005. Should I select 2005 in that case? As per image: http://i.imgur.com/3H7MO.png
    Cheers.
    I am using OBIEE 11G 32bit Windows.

    Hi,
    Yes can. either you can select as SQL 2005 or ODBC basic also can
    I hope your version less than obiee11.1.1.5.0, version 11.1.1.1.6.0 and above has update SQL 2008 in the data source selection list.
    refer,
    http://imgur.com/JSYoL
    Thanks
    Deva

Maybe you are looking for

  • Saved movie trailers crash iTunes 11.1.1.11

    iTunes 11.1.1.11 on my ThinkPad X230 with Win7 64bit crashes. I can click and watch a movie trailer in iTunes fine, but the same trailer downloaded and saved will crash iTunes if I click on the saved trailer. If I try to just play the saved trailer w

  • How to get vendor ID in Portal activity Report through PAR/Webdynpro

    Hi , I need to show Vendor ID in my Portal activity Report  along with other details .This vendor ID I need to fetch From Portal Database.Is there any way to code in PAR/Webdynpro or use any standard Template which will allow me to fetch Vendor ID of

  • Wifi connection AirPort Extreme with IP-camera

    I can't establish wifi connection Airport Extreme with IP-camera. At the same time cable connection is fine, but wifi not. Please, give me an advise.

  • Currency valuation when posting manual statements:

    Does anybody know why SAP System (R/3 Enterprise 4.7) is doing currency valuation during a manual statement posting? The account does not have valuation accounts at OBA1. The error occurs during the clearing process. I don't have exchange rate differ

  • Words containing national characters (hungarian) are not indexed

    I created a table data source that contains text in a CLOB field, and added it to the sample UltraSearch query application. I created few lines in this table and run the index building. The demo is now able to search in the clob field. If the text co