BW reporting against a relational database

I have a basic question on SAP BW, any help or advice is gratefully received.
Basically we have a requirement to generate and present reports through a web browser (Java application - JSPs, HTML etc, may be SAP Portal in the future).  The database is an SAP DB instance.
So my question is, is there some element of BW we can use for creating standard reports (real-time straight off SAP DB tables rather than through a cube/warehouse) that can be embedded into our web pages?  Effectively a bit like you would use Crystal Reports but better.
We would like to use SAP BW as it is something we envisage using more fully in the future, we are just not sure if it is suitable for our current requirement, if not we could use a J2EE reporting tool like Panscopic or JReport.
Thanks for any help.

Hi Dirk, Thanks for the tip.
I have defined the infosources and created a general remote cube with source system. No problem in activating the cube. But when I execute queries on the cube I get the following message:
1. Error occurred while determining target system for the logical system
2. System error in program SAPLRRK0 and form RSRDR;SRRK0F30-01-
Any ideas what could be wrong?
Thanks, Hari

Similar Messages

  • Web Analysis Report prompted for relational database login

    We have given access to the WA reports and everything works fine in general. One of the user trying to open same report(s) document(s) via Workspace and its getting prompted with relational database login? I have verified for the access and everything seems to be fine.
    I do not know the reason for this. Can anyone help me? Thanks.

    From the forums, i got the below MAXL commands to sync the user access :
    alter system resync sss; --> this command will sync all the users, groups.
    alter group <group name> sync security with all application; --> This can be used to individually sync the required group
    alter user <user name> sync security with all application; --> This can be used to individually sync the required user
    Will check it up with user on any change in her access.
    Thanks,
    Praveen.

  • Web Analysis Reports using a Relational Database

    I would like to create a Web Analysis report using data from an Oracle database. From Web Analysis, I try to create a new database connection. I select Oracle as the Driver type. In the Database Connection String field, I replace {HOST-NAME} with the Host information from the TNSNAMES.ORA file. I've confirmed the Port number is correct. When I test the connection, it fails. What am I missing? The format of the Database Connection String that is defaulting into the field is:
    jdbc:oracle:thin:@{HOST-NAME}:1521:ORCL
    Is there anything else I need to include besides the Host name and Port number, for example, do I need to include the SID?
    Thanks.
    Terri T.

    The ojdbc14.jar currently exists in the following directories:
    bea\weblogic81\server\lib
    oracle\product\10.2.0\client_1
    bea\weblogic81\server\ext\jdbc\oracle\920
    There is a ojdbc14_g.jar that exists in:
    bea\weblogic81\server\ext\jdbc\oracle\10g
    bea\weblogic81\server\ext\jdbc\oracle\920
    The classes12.jar file exists in:
    oracle\product\10.2.0\client_1\oui\jlib
    So from your response, I assume I need to copy classes12.jar to a directory under bea\weblogic81?
    Thanks.
    Terri T.

  • SQL Server Profiler does not capture cube processing queries against the relational database

    Hello, experts,
    This my first attempt to use SQL Server Profiler to find out why my cube processing is so slow. Here is what I did:
    My cube sources its data from a SQL Server database. While the cube processing is going, I launched the profiler connecting to the database engine (note: not the Analysis Services). I  choose the "Tuning" template and started the trace. However,
    I do not see any query activities (such as, SELECT DISTINCT ...) in the profiler trace window.
    Did I do something wrong, or do I have the wrong expectation?
    Thank you much in advance.

    Hi QQFA,
    Agree with Olaf.
    Please make sure that you choose the appropriate events to monitor the cube processing queries when using SQL Server Profiler.
    And I recommend you review the following links to get more information about tuning Analysis Services with SQL Profiler.
    Monitoring and Tuning Analysis Services with SQL Profiler:
    http://www.informit.com/articles/article.aspx?p=1745747
    Analysis Services Trace Events:
    http://msdn.microsoft.com/en-us/library/ms174867.aspx
    Thanks,
    Lydia Zhang

  • Web-Intelligence Reports against Access Database

    We are unable to run Web-Intelligence reports against Microsoft Access Database which resides on a network drive
    Able to create Unvierse, and Universe Connection to the Access Database through ODBC, but unable to run Web-I reports, when tried we get the following error
    A database error occured. The database error text is: [Microsoft][ODBC Microsoft Access Driver] '(unknown)' is not a valid path.  Make sure that the path name is spelled correctly and that you are connected to the server on which the file resides.. (WIS 10901)
    Did anyone tried to develop Web-I reports from Access database on a network drive, if so I request you to provide us details and your findings
    Thank you

    We run the designer on the same machine where BOBJ runs and also on other client machines, from disigner it works fine,
    The service account which is used to run BO Application Services has full control access to the network location and folder where the database resides
    We can run / refresh Desk-I and Web-I rich client but cannot run web-i reports against access database
    Thanks
    Sreekanth Pinnam
    Edited by: Sreekanth Pinnam on Jul 9, 2009 2:07 PM

  • Running reports stored in the database

    Hi guys,
    Is it possible to have Reports 6i produce reports that are
    stored in the database? Looking at the documentation, it would
    appear the report definition files (RDF) must be stored in the
    filesystem. I don't really want to be handing out filesystem
    access (managing and it) to people - is there a way around this?
    Thanks,
    denty.

    user5780461 wrote:
    Hi,
    Would anyone have any recommendations regrading running reports based on the data stored in an oracle database. We will soon go live with a project using an oracle DB. We do not wish to run reports directly from the production DB as it will impact performance. Have you actually measured the impact? If not, you are just making assumptions that -- lacking any other informatin -- appear to be invalid.
    Is there any recommended method/architecture for reporting suggested by oracle for this type of scenario?
    You make it sound like some unusual scenario. What you've described so far (running reports against an OLTP database) is just bread and butter operations for Oracle.
    One idea is too runs some scripts that will write the data to another database and from there we will runs out reporting queries. Are there any other potential solutions?
    And how does reading the data to insert into some other database cause less contention than reading data for a report?
    >
    Many thanks,
    Ro

  • Questions on OBIEE reporting against transactional Database

    Hello guys
    I have a situation where I have to bring in 5 tables from transactional database to do reporting on.. I looked at the tables, there are not typically dimension or fact tables. Each table contains information very specifically, at the same time all tables have numeric values that needed to be aggregated at the lowest level of that table..
    I can join all these 5 tables together knowing the key columns. However, this is basically joining 5 tables with fact measures... Since this is transactional database which we don't own, so I can't create views nor changing the table structures..
    I wonder how would I be able to configure such model in OBIEE?
    Any thoughts?

    Thank you for the reply..
    The specific situation I have here is:
    I have 4 tables that I am polling directly from the4 store's cash register (Oh yea, real real time), they are customer, receipt line, receipt header, inventory,
    The physical join is: they are customer-----> receipt line<-----receipt header <------------inventory...
    In the BMM layer, each of these 4 tables have columns that should be measures, such as customer.customer credit limit, receipt line. quantity, receipt header. reciept tendered, inventory.quantity on hand... They are more measure columns from each table, but I am just listing one from each..
    In this case if i set the aggregation rule on each column, then the BMM model won't a star schema anymore because every table can be fact..
    I thought about adding all the measure columns into 1 logical fact table, which then the logical fact will have 4 LTS: customer, Re Line, Re Header, Inv.. This will be fine if I am only dealing with 1 store... However, in reality there will be 50 stores with the same exact tables/structures but on different servesr. Therefore, the logical fact will end up with 4*50 = 200 LTS each has fragmentation defined "store number = NYC or something else"..
    Just wondering if combining all the measures in one logical table is the only way to model this design or not? What would be the best thing to do if in this situation I have to report against multiple data source of the same tables in 1 subject area?
    Please let me know your thoughts..
    Thank you

  • SQL ENTERPRISE: The edition of Reporting Services that you are using requires that you use local SQL Server relational databases for report data sources and the report server database

    The error below makes absolutely no sense! I'm using Enterprise Core...yet I'm being told I can't use remote data sources:
    w3wp!library!8!03/05/2015-19:08:48:: i INFO: Catalog SQL Server Edition = EnterpriseCore
    w3wp!library!8!03/05/2015-19:08:48:: e ERROR: Throwing Microsoft.ReportingServices.Diagnostics.Utilities.OperationNotSupportedException: , Microsoft.ReportingServices.Diagnostics.Utilities.OperationNotSupportedException: The feature: "The edition of Reporting
    Services that you are using requires that you use local SQL Server relational databases for report data sources and the report server database." is not supported in this edition of Reporting Services.;
    Really? This totally contradicts the documentation found here:
    https://msdn.microsoft.com/en-us/library/ms157285(v=sql.110).aspx
    That article says remote connections are completely supported.
    ARGH! Why does this have to be so difficult to setup?!?

    Hi jeffoliver1000,
    According to your description, you are using Enterprise Core edition and you are prompted that you can’t use remote data sources.
    In your scenario, we neither ignore your point nor be doubt with what you say. But actually we have met the case before that even though the SQL Server engine is Enterprise but the reporting services is still standard. So I would recommend you to find the
    actual edition of reporting services you are using. You can find Reporting Services starting SKU in the Reporting Service logs ( default location: C:\Program Files\Microsoft SQL Server\<instance name>\Reporting Services\LogFiles). For more information,
    please refer to the similar thread below:
    https://social.technet.microsoft.com/Forums/en-US/f98c2f3e-1a30-4993-ab41-acbc5014f92e/data-driven-subscription-button-not-displayed?forum=sqlreportingservices
    By the way, have you installed the other SQL Server edition before?
    Best regards,
    Qiuyun Yu
    Qiuyun Yu
    TechNet Community Support

  • Running a reports against different databases

    I have written a report that uses a ref cur in a database package to derive the main query. I've compiled the report and it runs fine.
    The problem is when I try to run it against a database other than the one it was compiled against (i.e. compile against DEV DB, test against TEST DB, finally migrate to PROD DB). Each of these databases are identical in terms of their objects but if I run the report against an environment it wasn't compiled against I get the following messages:
    REP-1437 Run time error in the PL/SQL development environment (DE).
    PDE-PSD001 Could not resolve reference to <Unknown Program Unit> while loading <unknown> <unknown>
    I noticed initially that the report and the database package had the same name, I though this may be confusing things so I renamed the stored proc but I still get the same results.
    I can get around the problem by either re-compiling it against each database or by getting the form to call the .rdf (rather than the .rep) so it is recompiled as and when required but this doesn't seem right.
    Anyone have any ideas?

    Hi Steve,
    You can try using the RECURSIVE_LOAD parameter. By default, it is set to YES. If you are running with the .rep, set it to NO on your command line. What this means, is that it will not try to recompile the plsql in your report.
    Hope that helps,
    Toby

  • Error (WIS 10901) while creating a webi report against SAP data source

    Hi,
    While creating a webi report against a universe created using Infocube/Bex query (basically SAP data source) get a error message -
    A database error occured. The database error text is: A runtime exception has occured. (License key check failed. Check that you are licensed to access SAP data sources). (WIS 10901).
    Any clues whether this is a license issue or something else?
    Thanks,
    -Purav.

    >
    George Pertea wrote:
    > Does your connection in the Universe test ok on the server if you are on Win? Did you install the Java Connector properly?
    I am having the same problem.  I checked and double checked everything, including deleting and recreating the Universe and the Connection.  No improvement.  On a whim, I deleted the hierarchy from the query on which the Universe was based, and recreated the Universe on the query without the hierarchy.  The problem went away.
    Now, the question is, why do I get the WIS 10901 error when a hierarchy is present in the query/Universe?

  • Insert XML file into Relational database model - no XMLTYPE!

    Dear all,
    How can I store a known complex XML file into an existing relational database WITHOUT using xmltypes in the database ?
    I read the article on DBMS_XMLSTORE. DBMS_XMLSTORE indeed partially bridges the gap between XML and RDBMS to a certain extent, namely for simply structured XML (canonical structure) and simple tables.
    However, when the XML structure will become arbitrary and rapidly evolving, surely there must be a way to map XML to a relational model more flexibly.
    We work in a java/Oracle10 environment that receives very large XML documents from an independent data management source. These files comply with an XML schema. That is all we know. Still, all these data must be inserted/updated daily in an existing relational model. Quite an assignment isn't it ?
    The database does and will not contain XMLTYPES, only plain RDBMS tables.
    Are you aware of a framework/product or tool to do what DBMS_XMLSTORE does but with any format of XML file ? If not, I am doomed.
    Cheers.
    Luc.
    Edited by: user6693852 on Jan 13, 2009 7:02 AM

    In case you decide to follow my advice, here's a simple example showing how to do this.. (Note the XMLTable syntax is the preferred approach in 10gr2 and later..
    SQL> spool testase.log
    SQL> --
    SQL> connect / as sysdba
    Connected.
    SQL> --
    SQL> set define on
    SQL> set timing on
    SQL> --
    SQL> define USERNAME = XDBTEST
    SQL> --
    SQL> def PASSWORD = XDBTEST
    SQL> --
    SQL> def USER_TABLESPACE = USERS
    SQL> --
    SQL> def TEMP_TABLESPACE = TEMP
    SQL> --
    SQL> drop user &USERNAME cascade
      2  /
    old   1: drop user &USERNAME cascade
    new   1: drop user XDBTEST cascade
    User dropped.
    Elapsed: 00:00:00.59
    SQL> grant create any directory, drop any directory, connect, resource, alter session, create view to &USERNAME identified by &PASS
    ORD
      2  /
    old   1: grant create any directory, drop any directory, connect, resource, alter session, create view to &USERNAME identified by &
    ASSWORD
    new   1: grant create any directory, drop any directory, connect, resource, alter session, create view to XDBTEST identified by XDB
    EST
    Grant succeeded.
    Elapsed: 00:00:00.01
    SQL> alter user &USERNAME default tablespace &USER_TABLESPACE temporary tablespace &TEMP_TABLESPACE
      2  /
    old   1: alter user &USERNAME default tablespace &USER_TABLESPACE temporary tablespace &TEMP_TABLESPACE
    new   1: alter user XDBTEST default tablespace USERS temporary tablespace TEMP
    User altered.
    Elapsed: 00:00:00.00
    SQL> connect &USERNAME/&PASSWORD
    Connected.
    SQL> --
    SQL> var SCHEMAURL varchar2(256)
    SQL> var XMLSCHEMA CLOB
    SQL> --
    SQL> set define off
    SQL> --
    SQL> begin
      2    :SCHEMAURL := 'http://xmlns.example.com/askTom/TransactionList.xsd';
      3    :XMLSCHEMA :=
      4  '<?xml version="1.0" encoding="UTF-8"?>
      5  <!--W3C Schema generated by XMLSpy v2008 rel. 2 sp2 (http://www.altova.com)-->
      6  <xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns:xdb="http://xmlns.oracle.com/xdb" xdb:storeVarrayAsTable="true">
      7     <xs:element name="TransactionList" type="transactionListType" xdb:defaultTable="LOCAL_TABLE"/>
      8     <xs:complexType name="transactionListType"  xdb:maintainDOM="false" xdb:SQLType="TRANSACTION_LIST_T">
      9             <xs:sequence>
    10                     <xs:element name="Transaction" type="transactionType" maxOccurs="unbounded" xdb:SQLCollType="TRANSACTION_V"
    >
    11             </xs:sequence>
    12     </xs:complexType>
    13     <xs:complexType name="transactionType" xdb:maintainDOM="false"  xdb:SQLType="TRANSACTION_T">
    14             <xs:sequence>
    15                     <xs:element name="TradeVersion" type="xs:integer"/>
    16                     <xs:element name="TransactionId" type="xs:integer"/>
    17                     <xs:element name="Leg" type="legType" maxOccurs="unbounded" xdb:SQLCollType="LEG_V"/>
    18             </xs:sequence>
    19             <xs:attribute name="id" type="xs:integer" use="required"/>
    20     </xs:complexType>
    21     <xs:complexType name="paymentType"  xdb:maintainDOM="false" xdb:SQLType="PAYMENT_T">
    22             <xs:sequence>
    23                     <xs:element name="StartDate" type="xs:date"/>
    24                     <xs:element name="Value" type="xs:integer"/>
    25             </xs:sequence>
    26             <xs:attribute name="id" type="xs:integer" use="required"/>
    27     </xs:complexType>
    28     <xs:complexType name="legType"  xdb:maintainDOM="false"  xdb:SQLType="LEG_T">
    29             <xs:sequence>
    30                     <xs:element name="LegNumber" type="xs:integer"/>
    31                     <xs:element name="Basis" type="xs:integer"/>
    32                     <xs:element name="FixedRate" type="xs:integer"/>
    33                     <xs:element name="Payment" type="paymentType" maxOccurs="unbounded"  xdb:SQLCollType="PAYMENT_V"/>
    34             </xs:sequence>
    35             <xs:attribute name="id" type="xs:integer" use="required"/>
    36     </xs:complexType>
    37  </xs:schema>';
    38  end;
    39  /
    PL/SQL procedure successfully completed.
    Elapsed: 00:00:00.00
    SQL> set define on
    SQL> --
    SQL> declare
      2    res boolean;
      3    xmlSchema xmlType := xmlType(:XMLSCHEMA);
      4  begin
      5    dbms_xmlschema.registerSchema
      6    (
      7      schemaurl => :schemaURL,
      8      schemadoc => xmlSchema,
      9      local     => TRUE,
    10      genTypes  => TRUE,
    11      genBean   => FALSE,
    12      genTables => TRUE,
    13      ENABLEHIERARCHY => DBMS_XMLSCHEMA.ENABLE_HIERARCHY_NONE
    14    );
    15  end;
    16  /
    PL/SQL procedure successfully completed.
    Elapsed: 00:00:00.26
    SQL> desc LOCAL_TABLE
    Name                                                                   Null?    Type
    TABLE of SYS.XMLTYPE(XMLSchema "http://xmlns.example.com/askTom/TransactionList.xsd" Element "TransactionList") STORAGE Object-rela
    ional TYPE "TRANSACTION_LIST_T"
    SQL> --
    SQL> create or replace VIEW TRAN_VIEW
      2  as
      3  select
      4    extractvalue(x.column_value,'/Transaction/TradeVersion/text()') tradeversion,
      5    extractvalue(x.column_value,'/Transaction//text()') transactionid
      6  from
      7    local_table,
      8    table(xmlsequence(extract(OBJECT_VALUE,'/TransactionList/Transaction'))) x
      9  /
    View created.
    Elapsed: 00:00:00.01
    SQL> create or replace VIEW TRAN_LEG_VIEW
      2  as
      3  select
      4    extractvalue(x.column_value,'/Transaction/TransactionId/text()') transactionid,
      5    extractvalue(y.column_value,'/Leg/Basis/text()') leg_basis,
      6    extractValue(y.column_value,'/Leg/FixedRate/text()') leg_fixedrate
      7  from
      8    local_table,
      9    table(xmlsequence(extract(OBJECT_VALUE,'/TransactionList/Transaction'))) x,
    10    table(xmlsequence(extract(x.column_value,'/Transaction/Leg'))) y
    11  /
    View created.
    Elapsed: 00:00:00.01
    SQL> create or replace VIEW TRAN_LEG_PAY_VIEW
      2  as
      3  select
      4    extractvalue(x.column_value,'/Transaction/TransactionId/text()') transactionid,
      5    extractvalue(y.column_value,'/Leg/LegNumber/text()') leg_legnumber,
      6    extractvalue(z.column_value,'/Payment/StartDate/text()') pay_startdate,
      7    extractValue(z.column_value,'/Payment/Value/text()') pay_value
      8  from
      9    local_table,
    10    table(xmlsequence(extract(OBJECT_VALUE,'/TransactionList/Transaction'))) x,
    11    table(xmlsequence(extract(x.column_value,'/Transaction/Leg'))) y,
    12    table(xmlsequence(extract(y.column_value,'/Leg/Payment'))) z
    13  /
    View created.
    Elapsed: 00:00:00.03
    SQL> desc TRAN_VIEW
    Name                                                                   Null?    Type
    TRADEVERSION                                                                    NUMBER(38)
    TRANSACTIONID                                                                   VARCHAR2(4000)
    SQL> --
    SQL> desc TRAN_LEG_VIEW
    Name                                                                   Null?    Type
    TRANSACTIONID                                                                   NUMBER(38)
    LEG_BASIS                                                                       NUMBER(38)
    LEG_FIXEDRATE                                                                   NUMBER(38)
    SQL> --
    SQL> desc TRAN_LEG_PAY_VIEW
    Name                                                                   Null?    Type
    TRANSACTIONID                                                                   NUMBER(38)
    LEG_LEGNUMBER                                                                   NUMBER(38)
    PAY_STARTDATE                                                                   DATE
    PAY_VALUE                                                                       NUMBER(38)
    SQL> --
    SQL> create or replace VIEW TRAN_VIEW_XMLTABLE
      2  as
      3  select t.*
      4    from LOCAL_TABLE,
      5         XMLTable
      6         (
      7            '/TransactionList/Transaction'
      8            passing OBJECT_VALUE
      9            columns
    10            TRADE_VERSION  NUMBER(4) path 'TradeVersion/text()',
    11            TRANSACTION_ID NUMBER(4) path 'TransactionId/text()'
    12         ) t
    13  /
    View created.
    Elapsed: 00:00:00.01
    SQL> create or replace VIEW TRAN_LEG_VIEW_XMLTABLE
      2  as
      3  select t.TRANSACTION_ID, L.*
      4    from LOCAL_TABLE,
      5         XMLTable
      6         (
      7            '/TransactionList/Transaction'
      8            passing OBJECT_VALUE
      9            columns
    10            TRANSACTION_ID NUMBER(4) path 'TransactionId/text()',
    11            LEG            XMLType   path 'Leg'
    12         ) t,
    13         XMLTABLE
    14         (
    15           '/Leg'
    16           passing LEG
    17           columns
    18           LEG_NUMBER     NUMBER(4) path 'LegNumber/text()',
    19           LEG_BASIS      NUMBER(4) path 'Basis/text()',
    20           LEG_FIXED_RATE NUMBER(4) path 'FixedRate/text()'
    21         ) l
    22  /
    View created.
    Elapsed: 00:00:00.01
    SQL> create or replace VIEW TRAN_LEG_PAY_VIEW_XMLTABLE
      2  as
      3  select TRANSACTION_ID, L.LEG_NUMBER, P.*
      4    from LOCAL_TABLE,
      5         XMLTable
      6         (
      7            '/TransactionList/Transaction'
      8            passing OBJECT_VALUE
      9            columns
    10            TRANSACTION_ID NUMBER(4) path 'TransactionId/text()',
    11            LEG            XMLType   path 'Leg'
    12         ) t,
    13         XMLTABLE
    14         (
    15           '/Leg'
    16           passing LEG
    17           columns
    18           LEG_NUMBER     NUMBER(4) path 'LegNumber/text()',
    19           PAYMENT        XMLType   path 'Payment'
    20         ) L,
    21         XMLTABLE
    22         (
    23           '/Payment'
    24           passing PAYMENT
    25           columns
    26           PAY_START_DATE     DATE      path 'StartDate/text()',
    27           PAY_VALUE          NUMBER(4) path 'Value/text()'
    28         ) p
    29  /
    View created.
    Elapsed: 00:00:00.03
    SQL> desc TRAN_VIEW_XMLTABLE
    Name                                                                   Null?    Type
    TRADE_VERSION                                                                   NUMBER(4)
    TRANSACTION_ID                                                                  NUMBER(4)
    SQL> --
    SQL> desc TRAN_LEG_VIEW_XMLTABLE
    Name                                                                   Null?    Type
    TRANSACTION_ID                                                                  NUMBER(4)
    LEG_NUMBER                                                                      NUMBER(4)
    LEG_BASIS                                                                       NUMBER(4)
    LEG_FIXED_RATE                                                                  NUMBER(4)
    SQL> --
    SQL> desc TRAN_LEG_PAY_VIEW_XMLTABLE
    Name                                                                   Null?    Type
    TRANSACTION_ID                                                                  NUMBER(4)
    LEG_NUMBER                                                                      NUMBER(4)
    PAY_START_DATE                                                                  DATE
    PAY_VALUE                                                                       NUMBER(4)
    SQL> --
    SQL> set long 10000 pages 100 lines 128
    SQL> set timing on
    SQL> set autotrace on explain
    SQL> set heading on feedback on
    SQL> --
    SQL> VAR DOC1 CLOB
    SQL> VAR DOC2 CLOB
    SQL> --
    SQL> begin
      2    :DOC1 :=
      3  '<TransactionList>
      4    <Transaction id="1">
      5      <TradeVersion>1</TradeVersion>
      6      <TransactionId>1</TransactionId>
      7      <Leg id="1">
      8        <LegNumber>1</LegNumber>
      9        <Basis>1</Basis>
    10        <FixedRate>1</FixedRate>
    11        <Payment id="1">
    12          <StartDate>2000-01-01</StartDate>
    13          <Value>1</Value>
    14        </Payment>
    15        <Payment id="2">
    16          <StartDate>2000-01-02</StartDate>
    17          <Value>2</Value>
    18        </Payment>
    19      </Leg>
    20      <Leg id="2">
    21        <LegNumber>2</LegNumber>
    22        <Basis>2</Basis>
    23        <FixedRate>2</FixedRate>
    24        <Payment id="1">
    25          <StartDate>2000-02-01</StartDate>
    26          <Value>10</Value>
    27        </Payment>
    28        <Payment id="2">
    29          <StartDate>2000-02-02</StartDate>
    30          <Value>20</Value>
    31        </Payment>
    32      </Leg>
    33    </Transaction>
    34    <Transaction id="2">
    35      <TradeVersion>2</TradeVersion>
    36      <TransactionId>2</TransactionId>
    37      <Leg id="1">
    38        <LegNumber>21</LegNumber>
    39        <Basis>21</Basis>
    40        <FixedRate>21</FixedRate>
    41        <Payment id="1">
    42          <StartDate>2002-01-01</StartDate>
    43          <Value>21</Value>
    44        </Payment>
    45        <Payment id="2">
    46          <StartDate>2002-01-02</StartDate>
    47          <Value>22</Value>
    48        </Payment>
    49      </Leg>
    50      <Leg id="22">
    51        <LegNumber>22</LegNumber>
    52        <Basis>22</Basis>
    53        <FixedRate>22</FixedRate>
    54        <Payment id="21">
    55          <StartDate>2002-02-01</StartDate>
    56          <Value>210</Value>
    57        </Payment>
    58        <Payment id="22">
    59          <StartDate>2002-02-02</StartDate>
    60          <Value>220</Value>
    61        </Payment>
    62      </Leg>
    63    </Transaction>
    64  </TransactionList>';
    65    :DOC2 :=
    66  '<TransactionList>
    67    <Transaction id="31">
    68      <TradeVersion>31</TradeVersion>
    69      <TransactionId>31</TransactionId>
    70      <Leg id="31">
    71        <LegNumber>31</LegNumber>
    72        <Basis>31</Basis>
    73        <FixedRate>31</FixedRate>
    74        <Payment id="31">
    75          <StartDate>3000-01-01</StartDate>
    76          <Value>31</Value>
    77        </Payment>
    78      </Leg>
    79    </Transaction>
    80  </TransactionList>';
    81  end;
    82  /
    PL/SQL procedure successfully completed.
    Elapsed: 00:00:00.01
    SQL> insert into LOCAL_TABLE values ( xmltype(:DOC1))
      2  /
    1 row created.
    Elapsed: 00:00:00.01
    Execution Plan
    Plan hash value: 1
    | Id  | Operation                | Name        | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | INSERT STATEMENT         |             |     1 |   100 |     1   (0)| 00:00:01 |
    |   1 |  LOAD TABLE CONVENTIONAL | LOCAL_TABLE |       |       |            |          |
    SQL> insert into LOCAL_TABLE values ( xmltype(:DOC2))
      2  /
    1 row created.
    Elapsed: 00:00:00.01
    Execution Plan
    Plan hash value: 1
    | Id  | Operation                | Name        | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | INSERT STATEMENT         |             |     1 |   100 |     1   (0)| 00:00:01 |
    |   1 |  LOAD TABLE CONVENTIONAL | LOCAL_TABLE |       |       |            |          |
    SQL> select * from TRAN_VIEW_XMLTABLE
      2  /
    TRADE_VERSION TRANSACTION_ID
                1              1
                2              2
               31             31
    3 rows selected.
    Elapsed: 00:00:00.03
    Execution Plan
    Plan hash value: 650975545
    | Id  | Operation          | Name                           | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT   |                                |     3 |   168 |     3   (0)| 00:00:01 |
    |   1 |  NESTED LOOPS      |                                |     3 |   168 |     3   (0)| 00:00:01 |
    |*  2 |   TABLE ACCESS FULL| SYS_NTGgl+TKyhQnWoFRSrCxeX9g== |     3 |   138 |     3   (0)| 00:00:01 |
    |*  3 |   INDEX UNIQUE SCAN| SYS_C0010174                   |     1 |    10 |     0   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       2 - filter("SYS_NC_TYPEID$" IS NOT NULL)
       3 - access("NESTED_TABLE_ID"="LOCAL_TABLE"."SYS_NC0000800009$")
    Note
       - dynamic sampling used for this statement
    SQL> select * from TRAN_LEG_VIEW_XMLTABLE
      2  /
    TRANSACTION_ID LEG_NUMBER  LEG_BASIS LEG_FIXED_RATE
                 1          1          1              1
                 1          2          2              2
                 2         21         21             21
                 2         22         22             22
                31         31         31             31
    5 rows selected.
    Elapsed: 00:00:00.04
    Execution Plan
    Plan hash value: 1273661583
    | Id  | Operation           | Name                           | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT    |                                |     5 |   560 |     7  (15)| 00:00:01 |
    |*  1 |  HASH JOIN          |                                |     5 |   560 |     7  (15)| 00:00:01 |
    |   2 |   NESTED LOOPS      |                                |     3 |   159 |     3   (0)| 00:00:01 |
    |*  3 |    TABLE ACCESS FULL| SYS_NTGgl+TKyhQnWoFRSrCxeX9g== |     3 |   129 |     3   (0)| 00:00:01 |
    |*  4 |    INDEX UNIQUE SCAN| SYS_C0010174                   |     1 |    10 |     0   (0)| 00:00:01 |
    |*  5 |   TABLE ACCESS FULL | SYS_NTUmyermF/S721C/2UXo40Uw== |     5 |   295 |     3   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       1 - access("SYS_ALIAS_1"."NESTED_TABLE_ID"="SYS_ALIAS_0"."SYS_NC0000800009$")
       3 - filter("SYS_NC_TYPEID$" IS NOT NULL)
       4 - access("NESTED_TABLE_ID"="LOCAL_TABLE"."SYS_NC0000800009$")
       5 - filter("SYS_NC_TYPEID$" IS NOT NULL)
    Note
       - dynamic sampling used for this statement
    SQL> select * from TRAN_LEG_PAY_VIEW_XMLTABLE
      2  /
    TRANSACTION_ID LEG_NUMBER PAY_START  PAY_VALUE
                 1          1 01-JAN-00          1
                 1          1 02-JAN-00          2
                 1          2 01-FEB-00         10
                 1          2 02-FEB-00         20
                 2         21 01-JAN-02         21
                 2         21 02-JAN-02         22
                 2         22 01-FEB-02        210
                 2         22 02-FEB-02        220
                31         31 01-JAN-00         31
    9 rows selected.
    Elapsed: 00:00:00.07
    Execution Plan
    Plan hash value: 4004907785
    | Id  | Operation            | Name                           | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT     |                                |     9 |  1242 |    10  (10)| 00:00:01 |
    |*  1 |  HASH JOIN           |                                |     9 |  1242 |    10  (10)| 00:00:01 |
    |*  2 |   HASH JOIN          |                                |     5 |   480 |     7  (15)| 00:00:01 |
    |   3 |    NESTED LOOPS      |                                |     3 |   159 |     3   (0)| 00:00:01 |
    |*  4 |     TABLE ACCESS FULL| SYS_NTGgl+TKyhQnWoFRSrCxeX9g== |     3 |   129 |     3   (0)| 00:00:01 |
    |*  5 |     INDEX UNIQUE SCAN| SYS_C0010174                   |     1 |    10 |     0   (0)| 00:00:01 |
    |*  6 |    TABLE ACCESS FULL | SYS_NTUmyermF/S721C/2UXo40Uw== |     5 |   215 |     3   (0)| 00:00:01 |
    |*  7 |   TABLE ACCESS FULL  | SYS_NTelW4ZRtKS+WKqCaXhsHnNQ== |     9 |   378 |     3   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       1 - access("NESTED_TABLE_ID"="SYS_ALIAS_1"."SYS_NC0000900010$")
       2 - access("SYS_ALIAS_1"."NESTED_TABLE_ID"="SYS_ALIAS_0"."SYS_NC0000800009$")
       4 - filter("SYS_NC_TYPEID$" IS NOT NULL)
       5 - access("NESTED_TABLE_ID"="LOCAL_TABLE"."SYS_NC0000800009$")
       6 - filter("SYS_NC_TYPEID$" IS NOT NULL)
       7 - filter("SYS_NC_TYPEID$" IS NOT NULL)
    Note
       - dynamic sampling used for this statement
    SQL>Out of interest why are you so against using XMLType...
    Edited by: mdrake on Jan 13, 2009 8:25 AM

  • Crystal Report against ECC 6.0 with SSO configured

    I created a simple report against the ECC 6.0 Article table and saved it to InfoView. I can run the report fine, but when others run it they get an error stating the logon parameters are incomplete. We are in the process of configuring single sign on and have created other Crystal Reports going against ECC 6.0 with Business Views where SSO turned on in the data connection. Everyone can refresh those reports with no problems. Is their some report setting that needs to be set for Crystal Reports using a data source directly against ECC 6.0 tables?

    We have multiple reasons for implementing SNC.
    The main reason we have enabled SNC is to simply the users experience while working in the SAP environment. We are going live this month with Plant Maintenance, SRM, Project System, General Ledge and Asset Management reporting cubes in our BW environment as well as in our ECC 6.0 environment. We have  set up SAP Portal and have created iviews to BOBJ folders and reports. Users are authenticated once by logging on to the Lan. Once logged on we would like them to seemlessly run SAP transactions and Business Objects reports through the SAP Portal based on their job function. Each user has security setup in BOBJ and SAP to allow access to transactions and reports appropriate for their job function.
    We have three types of SAP reports and data connectors.
    - One set of Crystal reports go against ECC 6.0 through Business Views with a Data Connection set up for SSO.
    - One set of WEBI reports goes against BW queries though Universe Connections set up for SSO.
    - One set of Crystal reports go directly against ECC 6.0 tables.
    We currently have two issues:
    1) The Crystal reports going directly against ECC 6.0 tables are returning the error
         Error - The database logon information for this report is either incomplete or incorrect.
    2) When we send a link to any report using OpenDocument if the user is not logged on to InfoView they get a rather long error message:
    Account Information not recognized: Active Directory Authentication failed to log you on. Please contact your system admin.... 
    Our hardware and server setup is a load balancer connected to a cluster of two Tomcat servers connected to two BOBJ servers.

  • Building reports around grid control database.

    As a part of Grid control support activity many times we need spend hours to verify if all the targets have been assigned rules and thresholds so that they can generate alerts well before the trouble occurs.
    So i was wondering is there any way by which which can write some script that will scan all the targets and find out which are not assigned to rules.
    I am not sure if i can access these information from grid control database.
    Can somebody help me in this ? is there any way to find out which rules have been assigned to which target or what is the threshold assigned to a particular targets by writing sql query / using oracle packages ?
    Appreciate help in this regard.

    Here is some sql I use for this purpose..
    (I comment out pieces and such to tweak what I am reporting against)
    Run this as sysman in your repository db.
    select
    target_type,
    target_name,
    metric_column Metric,
    metric_name || ' ' || key_value Name,
    decode(warning_threshold,' ',null, decode(warning_operator,'Equal To','=','Not Equal','!=','Greater Than','>','Less Than','<',warning_operator) || ' ' || warning_threshold) Warning,
    decode(critical_threshold,' ', null, decode(critical_operator,'Equal To','Not Equal','!=','=','Greater Than','>','Less Than','<',critical_operator) || ' ' || critical_threshold) Critical
    from mgmt$metric_collection m
    --where target_type = 'oracle_emd'
    --and  metric_column = 'HostInfovszGrowth'
    --and warning_threshold != '20'
    --where target_type = 'host'
    where-- metric_column like 'tnsPing'
    -- (warning_threshold != '80' and critical_threshold != '95')
    --and
    target_type = 'oracle_database'
    --and target_type = 'host'
    --and
    --target_type  = 'oracle_listener'
    --target_type = 'oracle_emd'
    --and metric_column like 'generic%'
    --and warning_threshold like '%ORA%'
    and target_name like 'QL%'
    and metric_column = 'pctUsed'
    order by 1,3,2

  • Can Oracle APEX access/report against an Oracle9i datbase?

    Can Oracle APEX access/report against an Oracle9i datbase? I know APEX has to be installed on Oracle 11g or 12c. I have a 11g environment to install it in, but I want to have ti read data from an existing 9i databasethat we are unable to upgrade at this time.

    APEX can create reports on any data that the database (which is running APEX) can access.
    This becomes a database question that is usually found in the General forum:
    How can an Oracle 11g database access data from an <place a name of a database here> database?
    One common answers:  Database Links.
    (Materialized Views and Golden Gate come to mind... but, again... these are database problems, not APEX problems.)
    I'll let others comment on the practicality of these solutions.
    MK

  • RELATIONAL DATABASE TABLES IN SAP BW

    hi
    friends please give me the information regarding the RELATIONAL DATABASE TABLES IN SAP BW?

    Hi,
    See the tables.
    InfoObjects Table:
    RSDIOBJ Directory of all InfoObjects
    RSDIOBJT Texts of InfoObjects
    InfoCube Tables:
    RSDCUBE Directory of InfoCubes
    RSDCUBET Texts on InfoCubes
    DSO Tables:
    RSDODSO Directory of all ODS Objects
    RSDODSOT Texts of all ODS Objects
    PSA Table:
    RSTSODS Directory of all PSA Tables
    For reports:
    RSRREPDIR
    Thanks & Regards,
    Sathish

Maybe you are looking for

  • Navteq sample data

    hi, I have problems during the navteq sampla data import (http://www.oracle.com/technology/software/products/mapviewer/htdocs/navsoft.html). I'm follwing exactly the instructions from the readme file. Import: Release 10.2.0.1.0 - Production on Thursd

  • Standard print program for PO (ME23N) Smartforms

    Hi Experts, I am create one smartform its working fine but i need to map this one with ME23N Tcode from there they take output. But i dnt knoe the standard print program. Pls help me Thanks in Advance purnaneelu

  • Add one field in ME2L T-code (program RM06EL00)

    hello all, I am new in sap. I want to add one field in program RM06EL00 ,tcode is ME2L and field name is zzstatus at last in report. i have check everything in program but i not able to find. please help me in which place(internal table) i have to ad

  • Problem meet when call exe in java.

    hi all, in my java project, i call an external exe . now the problem is that sometimes the exe is not fully finished. How could i confirm that the exe is finished before i start to use the output of this exe. i use following code to call the external

  • Obtain a controlling object substitution before message error KI235 appears

    Hello! I' m in trouble... I'd like to define a substitution rule that works out before message error KI235 appears in posting a document by using FB01 or FB50 trx. I've tried to do it by using OKB9 trx and inserting a default cost center. It is not a