2 data sources (oracle db) for the same report

Hi
I was wondering if I can have 2 sources (could be same or different instances). Example, I want to get information from one instance in the report trigger, and output of that i pass to the report (inside the query in the where clause).Is this possible with Oracle reports?
Any help appreciated.
Thx!

Hi Dan
Thanks for the response.
I need some information (in before report trigger) - from one oracle instance - and then pass this information as a parameter to the report - whose output will come from another oracle instance.
I haven't given it a try yet what you had recommended but will do so.
Thx!

Similar Messages

  • BAM Combo Chart showing data in 2 graphics for the same data object

    Hi all,
    I’m trying to show 2 views of the same data object in a Combo Chart (e.g. orders created today as a bar chart and orders created yesterday as line chart grouping by hour of day). It was pretty easy adding the same data object twice but the data type of the horizontal axis seems to be lost and it gets sorted as a simple string, leading to strange grouping of the hours with 05:00 representing 5 AM sided with 05:00 representing 5 PM. Also, I’m not able to format such fields using the format values tab.
    Is using the same data object two times in the same report an unsupported scenario?
    Thanks,
    Daniel

    why you want to add the same data object twice? Your use case can be achieved with single data object only.
    1) Create combo chart and select the data object.
    2) On "Choose data fields" page group by datetime field.
    3) You will see UI changed to 3 sections - left one with "Group by" section, middle with "Chart values" and right one with "Time Groups".
    4) In the right "Time Group" section, uncheck checkbox for continuous time series. Time Unit as hour, quantity as 1.
    5) In middle "Chart values" section ,select the fields that you want to display as chart values and respective chart types , in your case orders created today as a bar chart and orders created yesterday as line chart.
    5) Click Next and Finish.
    6) Then you can go to Value format tab, and change the format for the datetime field as timeunit and required format.

  • Data from different databases in the same report.

    Hi Everyone,
    I am trying to build a reconciliation report in which I need to show the data from the source and target, side by side.
    Source and target are both different databases, although being oracle only
    Whenever a new data model is created, it gets attached to a data source and in the report we need to choose a specific data model.
    Can we have data from different databases in the same report ?

    Yes, it is possible.
    One way Is to use dataTemplates. There you can make queries from any number of different databases (The max I have done is 5).
    It looks something like that:
    <dataTemplate name="NameOfTemplate">
    <dataQuery>
         <sqlStatement name="Q1" dataSourceRef="Connection1">
              <![CDATA[     select * from table1]]>
         </sqlStatement>
         <sqlStatement name="Q2" dataSourceRef="Connection2">
              <![CDATA[     select * from table2]]>
         </sqlStatement>
      </dataQuery>
      <dataStructure>
         <group name="RESULT1" source="Q1">
              <element name="P_FIRST_NAME" value="P_FIRST_NAME"/>
         </group>
         <group name="RESULT2" source="Q2">
              <element name="P_DATE" value="P_DATE"/>     
         </group>
      </dataStructure>
    </dataTemplate>dataStructure is very important when you get data from different places, if you don't define those elements, then only the result from Q1 is shown.
    The second possible way is to make as two different data models, each containing their own query and then set Main Data Set as concatenated SQL Data Source.
    Best of luck,
    Evelyn

  • Distribution working differently for the same report

    Hello all,
    I created a template with 3 reports, each of which opens for 4 plants.
    Then, when I schedule the offline distribution process, 2 of the reports are generated correctly for all plants, while the 3rd has a strange behavior. This 3rd report has 2 sheets and opens correctly for 2 of the plants, while for the other 2:
    - in one sheet, the formulas are not converted to offline mode (like _epmOfflineCondition_ ) giving an "#Error, no current connection." message and the sheet is not protected; however, the report data is correct.
    - in the other sheet, everything works ok and  the sheet is protected;
    Any thoughts? Thanks,
    Ana-Maria

    Thanks Ashok,
    but unfortunately the above points are not applicable to the issue indicated:
    1. Log off&log on has no effect since this problem appears on all PCs and always for the same data group.
    2. The data group is available in the model cube and master data.
    3. We are using the ADMIN user to execute the process so we have all authorizations (also, these are defined at Country level and only one Plant from a Country is not working).
    4. The formula is correct.
    5. Since the report executes correctly as a standalone report and works for other plants, I don't believe there is an issue with it.
    The issue still persists.
    Ana-Maria

  • Which Master data Source is available for the field VBUK-GBSTK (Doc Status)

    Hi
    Which Master data Source (Attr) is available for the field VBUK-GBSTK (Document Status).
    Please help me.
    Thanks
    Mannev.

    Hi
    Thanks for your reply.
    That is Transaction data source. I want master data source.
    Thanks,
    Mannev.

  • How do I get two independent data sources to graph on the same chart?

    I have one graph of streaming voltage and another graph of a simple boolean function which jumps from zero to one when a button is pushed. I'd like to have both lines of data display on one graph. The goal is to have an even marker for the voltage graph - pressing the boolean button would "mark" important data points during acquisition. How do I combine the two? (I tried putting both into an array and graphing that, but it seems to just average the two instead of plotting them separately.) Thanks.

    And yet another way...
    If you present an array of waveform data-types you can sidestep this issues of disimilar sample rates.
    Ben
    Ben Rayner
    I am currently active on.. MainStream Preppers
    Rayner's Ridge is under construction
    Attachments:
    Wform_chart.vi ‏40 KB

  • Why's the SQL format different for the same report on different PCs?

    I have 2 PCs running the same version of Crystal Reports (11.0.0.1994) using an identically configured system DSN with the exact same driver. When I view the SQL that CR is using on PC 1, part of it looks like this:
    " FROM   {oj (("PUB"."LOQEstimate" "LOQEstimate1" LEFT OUTER JOIN "PUB"."LOQComponent" "LOQComponent1" ON ("LOQEstimate1"."Estimate-ID"="LOQComponent1"."Estimate-ID") AND"
    On PC 2 it looks like this:
    "FROM   "PUB"."LOQEstimate" "LOQEstimate1", "PUB"."LOQComponent" "LOQComponent1", "PUB"."CustContact" "CustContact1", "PUB"."LOQSegment" "LOQSegment1"
    WHERE"
    It's the exact same report. Can someone tell me why the SQL is different?
    TIA for the help.

    Thanks for the reply, Patrick. Here are the answers you requested:
    1. Verify you are indeed using the same driver, and they are of the same version
    The driver on both PCs is the Progress OpenEdge 10.1A driver. The file is pgoe1021.dll, version 05.10.0037.
    2. Verify if you have the same database client install on both computer.
    Both use prowin32.exe, v10.1.1.1326
    If everything is the same, then it might be a registry key, and in this case, can you provide the following information:
    1. What database are you connecting to? ( Oracle, DB2, ... )
    Progress OpenEdge 10 RDMS (/community [original link is broken])
    2. What is the version of the database?
    10.1A
    3. What is the version of the database client?
    v10.1.1.1326
    4. How are you connecting to your database? ( Native, ODBC, OLEDB,... )
    ODBC
    5. If you are connecting via ODBC, what is the name of the driver used and it's version?
    The driver is the Progress OpenEdge 10.1A driver. The file is pgoe1021.dll, version 05.10.0037.
    6. What is the SQL query syntax do you prefer? ( PC 1, or the one generated on PC 2 )
    The SQL generated on PC1 doesn't return any records. The SQL generated on PC2 does. Records are preferable ;o)

  • How to load data using two maps for the same source file in Import Manager

    Hi,
    I am trying to load the source data to MDM using two map files, since one map is too big (creating memory error).
    However the first map data able to see after loading (SAP data) but when loading second map(nonSAP data) using the remote key to match the previously loaded SAP key ,even if MDM is processing the data it is not reflecting in data manager.
    Is there any reason for this.
    Appreciate your suggestion and tips on this.
    regrds,
    Reo

    Hi Reo,
    As ur requirement seems,In the second pass perform recording matching by select default import action as "<b>update null fields only"</b> for those data records which are matching and "<b>create</b> " for those data records which are not matching with previous data records from SAP system.
    If you define "update null fields only", then only the key of the new record from NON SAP will get appended to the existing record. And if you select "create" then it will create new records and will apply NON SAP key on it.
    Please check the remote key values using data manager.Key is applied in the records or not.
    Remember that import maps are client specific. Remote key comprises of both key and client system.So be sure that you are connecting to two different client systems.
    I have tried this and this is working fine on my system.
    Thanks,
    Shiv Prashant Dixit
    Thanks,

  • Unable to get another schema for the same Report!

    Hi All,
    I've created a report with data retrieved from three different schema's, but when I add a fourth schema with it's name before the query retrieved and run the report, it gives me that: table or view does not exist!
    Why does the report get the three schema's and refused the fourth one!? and how to solve the problem?
    Note: I am using reports 6i.

    HI,
    Make sure the connected user has sufficient privilege on that table.
    Regards,
    Manu.

  • How to generate different report view in teststand for the same execution?

    My requirment is to have two different view for the same report that gets generated during one execution using in TestStand OI(CVI)? One view would be summary view (displays only test step types, excluding action step types) and the other view is detail view that displays all the steps that are recorded. Data for all the steps needs to get recorded in database. Any suggestions are helpful!!
    CLD,CTD

    Hello.
    I think you would have to create two separate reports, one as summary report and one as a 'full' report  The viewers are just 'viewers' and simply display whatever you set to them.
    I had display a text report that was basically a summary report ( failures only ) and then display an HTML report that contained all steps were set to be recorded.  This required quite a bit of modifications on both the process model and the operator interface side.  The operator interface (in LabVIEW ) had to contain 2 separate UIReportView controls.
    Let me know if that is your desired direction and I could point you to some more ideas.
    Thanks,
    PH

  • Different Info Packages for the same data source.

    Hi Team,
    I have a question, which i asked before that but i ended up with different solution.
    I want to load 02.2002 to till date into ODS. Data source is 0FI_AR_10.
    My boss wants in this way .
    1) one infopackage for 02.2002 to 02.2005 with init.
    2) one infopack for 03.2005 to 07.2008 with init for the same datasource.
    3) one info pack with no selection criteira.
    What will be the result ?
    My guess is, Once we do init for the time period 02.2002 to 02.2005 ,
    the system with note the timepreiod for the Datasource.
    When i try to do Init for the second, the system will say init already exits for the data source. And it will not allow me to do init for the second selection.
    Correct me if  i am wrong.
    Regards,
    -Sonti-

    Your planned InfoPackages:
    - 1) one infopackage for 02.2002 to 02.2005 with init.
    - 2) one infopack for 03.2005 to 07.2008 with init for the same datasource.
    - 3) one info pack with no selection criteira.
    As the selection criteria of package 1) & 2) do not overlap, is it possible.
    For package 3), you need to do it as delta update instead of initialisation.  Package 3) with no selection criteria overlaps with package 1) & 2).
    Refer to the Help Documentation, which explains this quite clearly.
    Link: [Maintaining InfoPackage - Tab Page: Updating|http://help.sap.com/saphelp_nw70/helpdata/en/e3/e60138fede083de10000009b38f8cf/frameset.htm]
    >For selection criteria that do not overlap, several initialization criteria are possible for initializing the delta process when scheduling InfoPackages for data from an SAP system. This gives you the option of loading the relevant data for the delta process, step by step into the Business Information Warehouse. For example, you can load the data into BI for cost center 1000 in the first step, and the data for cost center 2000 in the second step.
    >A delta requested after several initializations, contains the super set of all the successful initial selections as a selection condition. This selection condition can then no longer be changed for the delta.
    Hope this help.

  • SCD - source sets where multiple changes exist for the same natural key

    I am using the Oracle SCD type2 method (end date old record and insert the new record) which was working fine until my source set contained multiple changes for the same natural key. Does anyone know of a way to handle this?
    Thanks

    Hello,
    I think, the way of handling multiply changes of NK is totally depend on business requirements for your project.
    There could be number of approaches for this. For example:
    - all changes but the last are considered active during one day, going subsequently right after previously existed SCD record;
    - all changes but the last are considered “active” depending on its occurrence within “fact” data (as far as you consider NK change based on some attribute set - probably some of attributes show themselves within “fact” data)
    - etc.
    So, make an agreement upon this with your customer – and ETL design decision will follow immediately.
    Sergey

  • Mix source data with different granularity into the same fact table?

    I have two transaction tables "Incident (157 columns)" and "Unit (70 Colums)". For every "Incident" that happens there could be one or more records in the "Unit" table.
    As part of my data mart design, I have merged both the tables into a single Fact "Incident Fact (227 Columns)" and inserted the records from both the tables with a join condition between them [incident.IN_NUM = Unit.IN_NUM].
    Is this correct, is my question? or am I mixing source data with different granularity in the same fact table. Appreciate your help.
    Best Regards
    Bees

    Bees,
    Are the measures from 'Incident' , repeated for a given incident where it has more than one record in the Unit table ? If so, then the sum(indicent.measure) will give an incorrect result, no ?
    What requirement is there to physically merge the tables together outside of OBIEE? With OBIEE you could have one logical 'fact' table to present to report users, which sourced from seperate Incidents and Units tables and would stop the incorrect aggregations occuring. A common modelling piece in the same way would be Order Headers and Order Lines, quite common in OBIEE to have a logical 'Orders' fact which contained both Order header measures and Order line measures, this translates to your Incidents -> Units relationship.
    To do what I've mentioned, is relatively straight forward, you need a 'Dim - Incident' with two levels, Incident and Unit, mapp the unique identifiers in as the level keys and then use these levels to set the content levels correctly in your 2 logical tables sources for logical 'Fact' , ie Incidents LTS at incident level, Units LTS as units level.
    Hope this helps, let us know if you get stuck.
    Cheers
    Alastair

  • Oracle Coherence*Web and BlazeDS: Multiple FlexSessions created for the same HttpSession

    Hi all,
    I have searched this forum and found a lot of good information from Alex Glosband and others about the infamous "Detected duplicate HTTP - based FlexSessions, generally due to the remote host disabling session cookies. Session coolkies must be enable to manage the client connection correctly." message.
    It seems, however, none of the cases are identical to ours. This is ours:
    - Resin 3.1.9
    - Oracle Coherence 3.7.1 with Coherence*Web (session replication)
    With this setup we get the "Detected duplicate HTTP..." message on the first attempt to use BlazeDS and on every subsequent call.  The same client and server code works fine in a local sessions setup.  With Coherence 3.3 (currently our production environment) it seems to occur less frequently, but still as frequent as it is a major issue for us.  It fails even with a single node using in-process distributed caching in our test setup (as well as with multi node out of process caching in our staging environment, for Coherence knowledgeable the resin app server runs with tangosol.coherence.session.localstorage=true in the first case and false in the second).
    Both the listener and message broker are mapped as "Coherence aware" in web.xml[1] so that they should use clustered sessions.
    We have been digging a bit and we found out that if we commented out lines 427 and 434 of flex.messaging.endpoints.BaseHTTPEndpoint from version 4.0.0.14931 it seems to mask the bug.  We added some logging in the setupFlexClient method and it seems that we get more or less a new FlexSession for each and every call - but they have the same cookie and thus underlying HttpSession. I.e. the list returned from flexClient.getFlexSessions() keeps growing. Thus we are not so keen on going to production with that memory leak and the above mentioned ugly hack of commenting out the detection of duplicates.
    We use request scope for the remote object, but could in theory use any scope as we do not really have any state on the object itself, it is all HttpSession state and return values that are key (logon is performed prior to doing the first blaze call, in pure forms and ajax, and it is not a timing issue in that regard we are seeing).
    Hope someone can shed some light on what can be happening. Is there any "reference testing"[2] or something when the FlexSessions are created that makes them being created as new? Where are they created?  We do not know the inner workings of the BlazeDS source, we just watched the call trace of the unwanted invalidation and found that to be line 427 of flex.messaging.endpoints.BaseHTTPEndpoint.
    Can we disable FlexSessions?  Since the flex and plain html parts of the app share the sessions, we always use FlexContext.getHttpRequest().getSession() anyway, never storing any state directly in the FlexSession or on the remote object. Or maybe there is a config option to help us with this detection (or creation) of multiple FlexSessions?
    Cheers and TIA,
    -S-
    [1] - For instance, this i the message broker servlet def:
    <servlet>
    <servlet-name>MessageBrokerServlet</servlet-name>
      <display-name>MessageBrokerServlet</display-name>
    <servlet-class>com.tangosol.coherence.servlet.api22.ServletWrapper</servlet-class>       
    <init-param>
    <param-name>coherence-servlet-class</param-name>
    <param-value>flex.messaging.MessageBrokerServlet</param-value>
    </init-param>       
    <init-param>
    <param-name>services.configuration.file</param-name>
    <param-value>/WEB-INF/flex/services-config.xml</param-value>
    </init-param>
    <load-on-startup>1</load-on-startup>
    </servlet>
    [2] - As you undertstand this is speculation based on pure air, but it could be that in Coherence there was a serialization/deserialization happening somehow that would break such a test?

    Just a quick update, it seems things are running in a stable fashion (and without visible memory leaks, just keeping the latest FlexSession) with these changes in BaseHTTPEndpoint:
         * Overrides to guard against duplicate HTTP-based sessions for the same FlexClient which will occur if the remote host has disabled session
         * cookies.
         * @see AbstractEndpoint#setupFlexClient(String)
        @Override
        public FlexClient setupFlexClient(String id) {
            log.debug("setupFlexClient start id " + id);
            FlexClient flexClient = super.setupFlexClient(id);
            // Scan for duplicate HTTP-sessions and if found, invalidate them and throw a MessageException.
            // A request attribute is used to deal with batched AMF messages that arrive in a single request by trigger multiple passes through this
            // method.
            boolean duplicateSessionDetected = (FlexContext.getHttpRequest().getAttribute(REQUEST_ATTR_DUPLICATE_SESSION_FLAG) != null);
            if (!duplicateSessionDetected) {
                List<FlexSession> sessions = flexClient.getFlexSessions();
                log.debug("Client has " + sessions.size() + " sessions.");
                int n = sessions.size();
                if (n > 1) {
                    int count = 0;
                    for (int i = 0; i < n; i++) {
                        if (sessions.get(i) instanceof HttpFlexSession)
                            count++;
                        if (count > 1) {
                            FlexContext.getHttpRequest().setAttribute(REQUEST_ATTR_DUPLICATE_SESSION_FLAG, Boolean.TRUE);
                            duplicateSessionDetected = true;
                            break;
            // If more than one was found, remote host isn't using session cookies. Kill all duplicate sessions and return an error.
            // Simplest to just re-scan the list given that it will be very short, but use an iterator for concurrent modification.
            int i = 0;
            if (duplicateSessionDetected) {
                List<FlexSession> sessions = flexClient.getFlexSessions();
                log.debug("Detected sessions from client: " + sessions);
                for (FlexSession session : sessions) {
                    if (session instanceof HttpFlexSession && i < sessions.size()) {
    //                    log.debug("----> sessionId: " + session.getId());
    //                    Enumeration e1 = session.getAttributeNames();
    //                    while (e1.hasMoreElements()) {
    //                        Object key = e1.nextElement();
    //                        log.debug("--------->" + key + "--------->" + session.getAttribute((String) key));
    //                    session.invalidate();
                        flexClient.sessionDestroyed(session);
                    i++;
                // Return an error to the client.
    //            DuplicateSessionException e = new DuplicateSessionException();
    //            e.setMessage(ERR_MSG_DUPLICATE_SESSIONS_DETECTED);
    //            throw e;
            return flexClient;
    It is not exactly beautiful (to say the least), but if it does the trick I might just be pragmatic enough to go with it... NB: I am of course not proposing this as a patch to this file or anything, it is just an ugly hack for our specific case, but maybe the information can help the BlazeDS team find the root cause making it incompatible with Coherence*Web.
    Will give it a test run on our staging servers.

  • Data not adding for the same Key up in Infocube

    Hi,
    My Dataflow goes like this:  Datasource  -> DSO -> Infocube.
    I have created 2 DSO's for Deliveries Header & Delivery Item Seperately which gets data from 2LIS_11_VCHDR & 2LIS_11_VCITM Datasources.  Both DSO's have Delivery Number & Fiscal Year variant as Key.  20 records are getting loaded to each DSO's seperately.  Both the DSO's have the same Delivery Numbers & Fiscal year Variant for this 21 Records in their respective Loads as given below:
                                     DSO 1                                                                  DSO 2
    S.No.       Deliver No.    Fiscal Year Var.                          Delivery No.       Fiscal Year Var.
       1               12345                     V3                                          12345                         V3
       2               28456                      V3                                          28456                        V3
       3               34986                      V3                                          34986                        V3
    Now, I am loading data from this 2 DSO's to one Infocube, Deliveries Infocube.  The problem is, after the Data load, when i go to the Contents of the Infocube, I could see 42 records.  As per Infocube, it should add up the values for the same key.  My Delivery numbers & Fiscal year vairant are the same in both the DSO's, so i should have got only 21 records.  Why is the Infocube not adding up the records in this case, am i missing anything here?
    Thanks & Regards,
    Murali

    Hi Murali,
    As you are loading data from two different sources, there are two different requests and that is why you are getting 42 records, if you want to check the data consistency then create a query on top of this infocube and then check the values of key figures for delivery no's. or verify your data through query.
    If you don't want to create query then you can use aggregation option from the display data of Infocube.
    If you compress both the requests then you will see only 21 records in infocube.
    Regards,
    Durgesh.

Maybe you are looking for