EPM 11.1.1.3 Drill back with Hyperion Planning and Financial Reporting

Hi,
I'm trying to drill back data with Planning thanks to FDM but it doesn't work.
In fact Planning server and FDM server are not installed on the same machine. My user is both admin on Planning application and FDM one.
When I drill back my data the url provided is the planning one and not FDM server one. I've set up application properties on Planning Application such as FDM_BASE_URL with
"http://servername/HyperionFDM/AuthorizedPages/IntersectionSummaryByLocation.aspx", and the restated my webserver, but it doesn't work.
I've the following errors :
Server Error in '/' Application.
The resource cannot be found.
Description: HTTP 404. The resource you are looking for (or one of its dependencies) could have been removed, had its name changed, or is temporarily unavailable. Please review the following URL and make sure that it is spelled correctly.
Requested URL: /HyperionFDM/AuthorizedPages/IntersectionSummaryByLocation.aspx
Best Regards
Eric

When you say drill back, are you saying you want to drill through to the source database or whatever it was you loaded the data from?
If so, my question is was this data loaded using FDM ?
Also I would run the validation tool in EPM Diagnostics and make sure everything is ok on the web end.
Not knowing exactly what your setup looks like, this is all I can suggest for now.

Similar Messages

  • Drill back from hyperion planning

    hi,
    We are extracting data from ERP and using ODI and we are loading into EPMA planning application. In this scenario can we do a Drill back from Planning Webform/HFR/Analyzer/Smartview to Oracle ERP system using ERP integrator? We don't have FDQM. Is it possible to drill back in our scenario without using FDQM and just using ERPi?
    Regards

    I thought that if you are using the ERP integrator as the source and then drill-back it will take you to the ERP landing page
    For example :- http://www.oracle.com/ocom/groups/public/@otn/documents/digitalasset/120599.gif
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Issue with Hyperion Planning 11.1.2 EPM

    Hi All,
    I am facing an issue with Hyperion Planning 11.1.2 which was working till yesterday. When i login into the and go to navigate --> Administor --> Dimension Library i am getting an error which iam not able to understand. kindly help me out. i checked all my EPM services and they are up and working and also i have restarted them but no use. Can someone help me out what should i do. I have also logged in http://<machine name>:19091/awb but getting the same below error.
    can someone help me out.
    org.apache.jasper.JasperException
         org.apache.jasper.servlet.JspServletWrapper.handleJspException(JspServletWrapper.java:510)
         org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:393)
         org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:314)
         org.apache.jasper.servlet.JspServlet.service(JspServlet.java:264)
         javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
         com.hyperion.awb.web.util.HttpFilter.doFilter(HttpFilter.java:19)
         com.hyperion.bpm.LocalizationServletFilter.doFilter(LocalizationServletFilter.java:69)
    root cause
    java.lang.NullPointerException
         com.hyperion.bpm.logon.CSSAuthenticate.authenticateSSOUser(CSSAuthenticate.java:334)
         com.hyperion.bpm.logon.CSSAuthenticate.getAuthenticatedUser(CSSAuthenticate.java:130)
         org.apache.jsp.index_jsp._jspService(index_jsp.java:182)
         org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:97)
         javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
         org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:332)
         org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:314)
         org.apache.jasper.servlet.JspServlet.service(JspServlet.java:264)
         javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
         com.hyperion.awb.web.util.HttpFilter.doFilter(HttpFilter.java:19)
         com.hyperion.bpm.LocalizationServletFilter.doFilter(LocalizationServletFilter.java:69)
    note The full stack trace of the root cause is available in the Apache Tomcat/5.5.17 logs.
    Edited by: [email protected] on Jul 18, 2009 10:38 PM

    Hi,
    First question, has anything changed since it worked like admin accounts.
    Reboot the machine.
    Next, just start the "Hyperion Foundation Shared Services - Web Application" service
    Can you log into Shared Services ?
    If you can, then start "Hyperion EPM Architect - Web Application" service
    Go to http://localhost:19091/awb/ Does the web page display ?
    If not then you it looks like EPMA needs reconfiguring.
    Next start "Hyperion EPM Architect - Process Manager" service, make sure the service starts and it should start another 4 services.
    Next refresh http://localhost:19091/awb/ and try and login.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Commit work and roll back with simple language and simple example

    hi guru
    commit work and roll back with simple language and simple example

    Hi,
    The statement COMMIT WORK completes the current SAP LUW and opens a new one, storing all change requests for the current SAP LUW in the process. In this case, COMMIT WORK performs the following actions:
    It executes all subroutines registered using PERFORM ON COMMIT.
    The sequence is based on the order of registration or according to the priority specified using the LEVEL addition. Execution of the following statements is not permitted in a subroutine of this type:
    PERFORM ... ON COMMIT|ROLLBACK
    COMMIT WORK
    ROLLBACK WORK
    The statement CALL FUNCTION ... IN UPDATE TASK can be executed.
    ROLL BACK:
    The statement ROLLBACK WORK closes the current SAP-LUW and opens a new one. In doing so, all change requests of the current SAP-LUW are canceled. To do this, ROLLBACK WORK carries out the following actions:
    1) Executes all subprograms registered with PERFORM ON ROLLBACK.
    2) Deletes all subprograms registered with PERFORM ON COMMIT.
    3) Raises an internal exception in the Object Services that makes sure that the attributes of persistent objects are initialised.
    4) Deletes all update function modules registered with CALL FUNCTION ...IN UPDATE TASK from the VBLOG database table and deletes all transactional remote Function Calls registered with CALL FUNCTION ... IN BACKGROUND TASK from database tables ARFCSSTATE and ARFCSDATA.
    5) Removal of all SAP locks set in the current program in which the formal parameter _SCOPE of the lock function module was set to the value 2.
    6) Triggers a database rollback, which also ends the current database-LUW.

  • HT201250 I accidentally didn't save a document on my desktop--a report.  Can I go back with time machine and recover my desktop and save the file?

    I accidentally didn't save a document on my desktop--a report.  Can I go back with time machine and recover my desktop and save the file?

    If you saved it at least once, then time machine should have a copy of the document at the state you last saved it.  I'm guessing tha's not the case however or the copy of the file would still be on your Desktop.
    If you never saved it a single time, yet closed the window and dismissed the warnings about an unsaved document, then you are out of luck as far as I know.

  • Oracle GL Integration with Hyperion Planning

    Dear All,
    I need to integrate Oracle GL with Hyperion Planning.
    My Question are:
    1) Without any adapters, we can do this integration(Using PL/SQL). It is Possible?
    2) Any document is available for intergration using the adapters(DIM/ODI).
    ODI - Using ODI, one way of integration is possible(Oracle GL to Planning)
    DIM - Using DIM, two way of integration is possible(Oracle GL to Planning & Planning to Oracle GL)
    can u guide me in this situtation.
    Thanks,
    PC

    H John,
    Thanks for the prompt reply.
    I am using Hyperion 9.3.1 with Oracle 10g. This is the first time, i am trying to do the integration with Oracle GL. My client dont have licenses for DIM, they are not ready to buy the licenses.
    There is anyway to load the essbase cubes through Oracle tables.
    John, This is my idea(Can u please suggest me on this). I am doing this in a test environment.
    1) From Oralce GL server, I am taking the Trial Balance values into flat files
    2) I am creating one Oracle table for TB_ORAGL(In Planning Server)
    3) In Planning server, i am importing the flat files(TB_ORAGL) to the Oracle
    4) From oracle table(TB_ORAGL), i am trying importing the data to essbase cubes using data load.
    Vice versa steps for Hyperion Planning to Oracle GL.
    can u guide me in this situtation.
    Thanks,
    PC
    Edited by: dwhpc on Oct 21, 2009 1:55 PM

  • FDM with Hyperion Planning

    Hi,
    We are new to FDM. Apologies in advance if the questions below are not appropriate for this forum.
    Environment: Oracle eBS, Hyperion Planning. Currently have a bespoke system for world wide finance consolidation - data from Oracle eBS (both Actuals / Budgets) are pumped into Finance consolidation system through CSV files generated in Oracle eBS. Not using HFM.
    Questions:
    1) Can Hyperion Planning data be pumped into FDM? How?
    2) Can FDM be used for performing validation on Planning data? (e.g. some dimension combination not valid etc)?
    3) How common is FDM in Planning environment? Are there many customers using FDM with Hyperion Planning?
    4) Is there any customer live with FDM Adapter for Oracle eBS?
    Thanks.

    Nick,
    Sorry if I am causing confusion here but allow me to clarify:
    1) HFM is not in use. Global consolidation engine is outside regional system - data needs to be sent to Global consolidation engine which is a non-HFM target.
    2) Before sending the data, validation needs to be performed on both Actual as well as Budget if data combination is valid.
    3) Oracle eBS does the validation for Acutals - so we are fine there.
    4) How do we ensure that Budget data (captured in Planning) is also valid before uploading to global consolidation engine?
    If Planning can not be a source to FDM, I can not perform validation on Budget data using FDM, right?
    So - what alternative do we have to perform validation on Budget data before it gets transmitted to global consolidation engine (non-HFM and outside our control)?
    Goal is - prior to transmitting data (both budget and actual), validation should be performed locally and only clean data should be transmitted for global consolidation engine.
    Hope this explains the situation.
    Once again many thanks for your response.

  • Try updating my ipad  2 ,came back with error 1602, and now my ipad is stuck with the itune and usb icon on the screen

    Try updating my ipad 2, pop up with error 1602 and now my ipad is stuck with the itune and USB icon on the screen. I tried to restart it, but screen still comes up the same.

    See if this page helps http://support.apple.com/kb/TS3694#error1602

  • Working with Grid Suppression in Financial Reporting for EPM 11

    Greetings:
    Does anyone know if it is possible in Financial Reporting Studio to have multiple grids with suppression so that if one grid has no data the grid below it moves into that grids spot on the report. Currently, I am able to suppress the grid if it contains no values however they stay in their fixed position.
    ie.
    Grid 1 (no values & will be suppressed)
    Grid 2
    Grid 3
    Would show in report preview:
    Grid 2
    Grid 3
    What I'm getting currently is:
    "blank" (suppressed grid)
    Grid 2
    Grid 3
    Thanks
    Bryan

    Just put all your grid possitions relative to another.

  • What causes ad-hoc to run with bad plan and its resolution

    One of our Company Application funcnalities uses the ad-hoc query that returns top first row from the result sets .There has been several times since this ad-hoc query ran with a bad
    execution in last 8 months .There is no specifice time and date or day when this query run with bad execution plan ,however whenever this query ran with a bad execution plan on SQL ,that happed always early in the morning , before 6am .The setting of database
    parameterization is set to forced .There is SQL maintenance reindexing job that runs each night at 2am that changes the recovedry model of our production database from Full to simple and changing back from simple to full after the reindexing operation .That
    causes the SQL plans to flush each night .The next day morning , when the application ad-hoc query runs on SQL first time , SQL generates a new plan for that query . Some strange reason , SQL genarates a good plan for that ad-hoc query most of the times ,but
    sometime this query picks up a bad plan .Whenever it happen,the application function becomes exetremly slow and it seems that the rest of the fuctionalities work fine meantime .
    Our application vender is aware of this issue and this issue will be fixed in the application next release .Developer is converting the ad-hoc query into stored procedure to fix this
    issue .Application new relase will not available for next 2-3 months ,meantime we still have to deal with the ad-hoc query bad execution plan issue . Usually ,when the ad-hoc query run with the bad execution plan , I retrieve the query bad plan and just remove
    the plan for that particular query from cache .Once the bad plan is removed from cache , the perfomance gets back to normal .
    As my understanding, this issue is related to parameter sniffing .but anyone thinks there are other factors responsible for this issue such as the network problem , the resource intensive
    report ran or something else .
    Has anyone have step by step troubleshooting steps that could help me with my investigation and what areas
     do I need to look into in order to investigate what could caused the ad-hoc query to pick up bad plan ?
    Any help will be apprecited .
    Daizy

    When we upgraded our Production Database to SQL server 2008R2 enterprise Edition. We ran into the transaction log file to grow too large during the reindexing job activities at night .To prevent the transaction log file to grow big , we had decided to change
    the recovery model of database from FULL to Simple before the reindexing job starts and the recovery model of Database is changed from Simple to Full after the reindexing job finishes . We also have the differential backup job setup right after the recovery
    model change back to full so the transaction log backup chain does not break.
    The transaction log chain breaks when you set the recovery to simple.
    Yes, if you take a diff backup, you can apply transaction logs from that point
    if the diff backup is good. If the diff backup is bad, you will lose all updates after you set recovery to simple. And, yes, when you hardware starts to deteoriate, you should not be surprise that the last backup is not good.
    Again, please stop setting recovery to simple, unless you have a very relaxed RPO in your SLA. Use bulk-logged recovery instead.
    By the way, you do take a t-log backup before you change the recovery mode, don't you?
     Changing the database recovery model does flush the SQL plans for specific Database. Please review this linkhttp://sqlblogcasts.com/blogs/steveh/archive/2010/09/02/changing-recovery-model-removes-plans-from-procedure-cache-for-a-database.aspx
    Hmmm. I don't have the time to verify this right now. Anyway, reindexing is good enough to flush the cache of most plans.
    I have pretty much all required info regarding this query , but what other areas or investigation do I need to do for the query parameter sniffing issue ?
    Have you researched what values the initial plan is sniffed for?
    As Erland said  adding space in query ,I don?t know adding space means just add additional space in the query so it hashes a different values .Please advice .
    Obviously, you need access to the application code. The other alternative is to set up an early morning job that runs after this first execution and which forces a recompilation; either by evicting this specific plan or more brutally runs sp_recompile on
    one of the tables in the query.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Thinking about Moves with my Plan and Future, Help Appreciated

    OK so this was the deal with my plan currently, is that right now im looking to save money on my phone bill every month and im being given the big run around with my Talk/Text/Data Promotion where i keep being told my data discount of 9.99 would come off and so far this month too it did not come off. The woman in Promotions department apparently fixed my account which not sure if she did and the problem with the whole thing is that different reps keep telling me i can only get my employee discount on my voice or the data discount which of course is wrong because the promo woman over the phone told me i can receive both and the loyalty reps over the phone keep giving me different stories and im tired of it. This was the reason i extended my contract for a bit to get this discount and its getting very frustrated about it!!
    The other thing was i was thinking about switching carriers when my contract is up because my bill has been pretty high the past couple of months and it makes no sense to early terminate my 2nd line because my ex gf upgraded to a smartphone and jacked up the fee. My mom hardly uses the 2nd line and i would like to cancel it but because its too high it makes no point to do that since both our contracts are up in a year. I was thinking about going to a basic phone to cut my data plan prices because the only cheapest verizon has is the 30 dollar plan. I don't use my data plan too much but i heard if i go to a basic phone and within 14 days i go back to a smartphone i can get my unlimited data back. So idk if this is true or not. But its gotten to the point that these smartphone plans and contracts are a bit much now. So if there is a Verizon rep that can chime in and see what can be done or discuss with me about options, that would be greatly appreciated. Otherwise i just had to vent about the issue with my current bill at this moment. So just was saying.

    To answer SuzyQ, my ex gf is off my account because of problems with her being on it (being late all the time, messing up my credit) and she wasn't choosing to fix her problems by doing and AOL because of her credit which Verizon wanted 400 and she didn't want to do it. Long story short, she lied to me about her AOL deposit amount just to stay on my account for a few more years (hence the early upgrade). So thats why shes no longer with my account.
    To answer AZS, your right my mom HAS a basic phone now and does not have data. My problem is that i have data on my plan which im trying to cut back on the cost of the plan and so far, i really wish Verizon would let me have that Talk/Text/Data promotion where it was 9.99 for the unlimited data and it would really help my bill alot. Of course i had a hard time getting it and one of the reps who i dealt with about getting it was very rude with me and she didn't care about satisfying the problem or the plan with me.
    But both lines have less than a year left so when the time comes, its time to move to something cheaper and better for me no contract wise. Unless anyone has something else to throw in their opinion of.

  • Problems with explain plan and statement

    Hi community,
    I have migrated a j2ee application from DB2 to Oracle.
    First some facts of our application and database instance:
    We are using oracle version 10.2.0.3 and driver version 10.2.0.3. It runs with charset Unicode 3.0 UTF-8.
    Our application is using Tomcat as web container and jboss as application server. We are only using prepared statements. So if I talk about statements I always mean prepared statements. Also our application is setting the defaultNChar property to true because every char and varchar field has been created as an nchar and nvarchar.
    We have some jsp sites that contains lists with search forms. Everytime I enter a value to the form that returns a filled resultset, the lists are performing great. But everytime I enter a value that returns an empty resultset, the lists are 100 times slower. The jsp sites are running in the tomcat environment and submitting their statements directly to the database. The connections are pooled by dbcp. So what can cause this behaviour??
    To anaylze this problem I started logging all statements and filled-in search field values and combinations that are executed by the lists described above. I also developed a standalone helper tool that reads the logged statements, executes them to the database and generates an explain plan for every statement. But now there appears a strange situation. Every statement, that performs really fast within our application, is now executed by the helper tool extremely slow. So I edited some jsp pages within our application to force an explain plan from there (tomcat env). So when I'm executing the same statement I'm getting with the exactly same code two completely different explain plans.
    First the statement itself:
    select LINVIN.BBASE , INVINNUM , INVINNUMALT , LINVIN.LSUPPLIERNUM , LSUPPLIERNUMEXT , LINVIN.COMPANYCODE , ACCOUNT , INVINTXT , INVINSTS , INVINTYP , INVINDAT , RECEIPTDAT , POSTED , POSTINGDATE , CHECKCOSTCENTER , WORKFLOWIDEXT , INVINREFERENCE , RESPONSIBLEPERS , INVINSUM_V , INVINSUMGROSS_V , VOUCHERNUM , HASPOSITIONS , PROCESSINSTANCEID , FCURISO_V , LSUPPLIER.AADDRLINE1 from LINVIN, LSUPPLIER where LINVIN.BBASE = LSUPPLIER.BBASE and LINVIN.LSUPPLIERNUM = LSUPPLIER.LSUPPLIERNUM and LINVIN.BBASE = ? order by LINVIN.BBASE, INVINDAT DESC
    Now the explain plan from our application:
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | 101 | 28583 | 55 (0)| 00:00:01 |
    | 1 | NESTED LOOPS | | 101 | 28583 | 55 (0)| 00:00:01 |
    | 2 | TABLE ACCESS BY INDEX ROWID| LINVIN | 93709 | 12M| 25 (0)| 00:00:01 |
    |* 3 | INDEX RANGE SCAN | LINV_INVDAT | 101 | | 1 (0)| 00:00:01 |
    | 4 | TABLE ACCESS BY INDEX ROWID| LSUPPLIER | 1 | 148 | 1 (0)| 00:00:01 |
    |* 5 | INDEX UNIQUE SCAN | PK_177597 | 1 | | 1 (0)| 00:00:01 |
    Predicate Information (identified by operation id):
    3 - access("LINVIN"."BBASE"=:1)
    filter("LINVIN"."BBASE"=:1)
    5 - access("LSUPPLIER"."BBASE"=:1 AND "LINVIN"."LSUPPLIERNUM"="LSUPPLIER"."LSUPPLIERNUM")
    Now the one from the standalone tool:
    | Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | 93773 | 25M| | 12898 (1)| 00:02:35 |
    | 1 | SORT ORDER BY | | 93773 | 25M| 61M| 12898 (1)| 00:02:35 |
    |* 2 | HASH JOIN | | 93773 | 25M| 2592K| 7185 (1)| 00:01:27 |
    | 3 | TABLE ACCESS BY INDEX ROWID| LSUPPLIER | 16540 | 2390K| | 332 (0)| 00:00:04 |
    |* 4 | INDEX RANGE SCAN | LSUPPLIER_HAS_BASE_FK | 16540 | | | 11 (0)| 00:00:01 |
    | 5 | TABLE ACCESS BY INDEX ROWID| LINVIN | 93709 | 12M| | 6073 (1)| 00:01:13 |
    |* 6 | INDEX RANGE SCAN | LINVOICE_BMDT_FK | 93709 | | | 84 (2)| 00:00:02 |
    Predicate Information (identified by operation id):
    2 - access("LINVIN"."BBASE"="LSUPPLIER"."BBASE" AND "LINVIN"."LSUPPLIERNUM"="LSUPPLIER"."LSUPPLIERNUM")
    4 - access("LSUPPLIER"."BBASE"=:1)
    6 - access("LINVIN"."BBASE"=:1)
    The size of the tables are: LINVIN - 383.692 Rows, LSUPPLIER - 115.782 Rows
    As you can see the one executed from our application is much faster than the one from the helper tool. So why picks oracle a completely different explain plan for the same statement? An why is a hash join much slower than a nested loop? Because If I'm right a nested loop should only be used when the tables are pretty small..
    I also tried to play with some parameters:
    I set optimizer_index_caching to 100 and optimizer_index_cost_adj to 30. I also changed optimizer_mode to FIRST_ROWS_100.
    I would really appreciated, if somebody can help me with this issue, because I'm really getting more and more distressed...
    Thanks in advance,
    Tobias
    Edited by: tobiwan on Sep 3, 2008 11:49 PM
    Edited by: tobiwan on Sep 3, 2008 11:50 PM
    Edited by: tobiwan on Sep 4, 2008 12:01 AM
    Edited by: tobiwan on Sep 4, 2008 12:02 AM
    Edited by: tobiwan on Sep 4, 2008 12:04 AM
    Edited by: tobiwan on Sep 4, 2008 12:06 AM
    Edited by: tobiwan on Sep 4, 2008 12:06 AM
    Edited by: tobiwan on Sep 4, 2008 12:07 AM

    tobiwan wrote:
    Hi again,
    Here ist the answer:
    The problem, because I got two different explain plans, was that the external tool uses the NLS sesssion parameters coming from the OS which are in my case "de/DE".
    Within our application these parameters are changed to "en/US"!! So if I'm calling in my external tool the java function Locale.setDefault(new Locale("en","US")) before connecting to the database the explain plans are finally equal.That might explain why you got two different execution plan, because one plan was obviously able to avoid a SORT ORDER BY operation, whereas the second plan required to run SORT ORDER BY operation, obviously because of the different NLS_SORT settings. An index by default uses the NLS_SORT = 'binary' order whereas ORDER BY obeys the NLS_SORT setting, which probably was set to 'GERMAN' in your "external tool" case. You can check the "NLS_SESSION_PARAMETERS" view to check your current NLS_SORT setting.
    For more information regarding this issue, see my blog note I've written about this some time ago:
    http://oracle-randolf.blogspot.com/2008/09/getting-first-rows-of-large-sorted.html
    Now let me make a guess why you observe the behaviour that it takes so long if your result set is empty:
    The plan avoiding the SORT ORDER BY is able to return the first rows of the result set very quickly, but could take quite a while until all rows are processed, since it requires potentially a lot of iterations of the loop until everything has been processed. Your front end probably by default only display the first n rows of the result set and therefore works fine with this execution plan.
    Now if the result set is empty, depending on your data, indexes and search criteria, Oracle has to work through all the data using the inefficient NESTED LOOP approach only to find out that no data has been found, and since your application attempts to fetch the first n records, but no records will be found, it has to wait until all data has been processed.
    You can try to reproduce this by deliberately fetching all records of a query that returns data and that uses the NESTED LOOP approach... It probably takes as long as in the case when no records are found.
    Note that you seem to use bind variables and 10g, therefore you might be interested that due to the "bind variable peeking" functionality you might potentially end up with "unstable" plans depending on the values "peeked" when the statement is parsed.
    For more information, see this comprehensive description of the issue:
    http://www.pythian.com/blogs/867/stabilize-oracle-10gs-bind-peeking-behaviour-by-cutting-histograms
    Note that this changes in 11g with the introduction of the "Adaptive Cursor Sharing".
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

  • Reverse engineer error with Hyperion Essbase and Planning

    We are getting the following error in ODI Operator (3rd step) when trying to reverse-engineer Hyperion Essbase and Planning models. com.microsoft.sqlserver.jdbc.SQLServerException: Cannot invoke a rollback operation when the AutoCommit mode is set to "true".
    Our ODI repository resides in SQL Server.. Any help is appreciated. Thanks.

    Hi Dmitriy,
    You should upgrade to the latest patch of ODI available on metalink.oracle.com 10.1.3.4.4.
    There was an issue in 10.1.3.4 ODI with the intial reverse when using SQL server. If you shutdown all the ODI UI's and then restart. And then reverse it should work correctly. But you have to make sure you shut down all ODI tools.
    Thanks,
    Denis

  • Pre-requisites for Drill-Down from Hyperion Planning to Oracle GL

    I am using Hyperion Planning 11.1.2 for maintaining budgets. I do comparison of actual vs budget in Hyperion. I export actual results from Oracle General Ledger (11.5.10) in flat file and upload to Hyperion planning using FDM. I maintain mapping in FDM for flat file and the Hyperion members.
    Q1) Is it possible to drill down actual numbers from Hyperion planning to Oracle General Ledger via FDM?
    Q2) If yes, what are prerequisites for using drill-down functionality?
    Can anyone please answer and refer some link/material that may help me in setting this up?
    Thanks in advance.

    Thanks
    But when i try to log in i m getting error bellow"
    Failure of server APACHE bridge: No backend server available for connection: timed out after 10 seconds or idempotent set to OFF."
    So do u have any other reference plz
    thanks

  • Replacing Oracle OFA with Enterprise Planning and Budget or Hyperion

    I'm looking at replacing OFA with EPB. Does this option still exist or has Oracle
    phased it out?

    Hi,
    Oracle Hyperion Planning is the way forward and is the budgeting and planning tool of choice for Oracle.
    So if you are going to make an investment I would go for Hyperion Planning.
    Cheers
    John
    http://john-goodwin.blogspot.com/

Maybe you are looking for