Queries in FDM

Hi, this is arvind , I have some queries in FDM
Q1. What is the difference between FM9i-G4-A.xml and FM2.9i-G4-B.xml adapters?
Q2. Whether we have the same .dll file to register both the ICP and Normal adapters, If No, then where we
could find the .dll file for ICP adapters?
Q3.When are we required to import the source Adapters?
Q4. Can you specify any 10 reasons why should a company go for HFM impementation.
Q5. What are the advantages of HFM over Excel.

I am using Windows server 2003 and Hyperion 11.1.1.3.
I an trying to configure FDQM.
For info purpose ( I have used CSS (Shared service) in Load Balancing Authentication provider)
I have configure all with username and password and domain name.
When I trying to login via http://hyperion/HyperionFDM/AuthorizedPages/LogonPage.aspx?ReturnUrl=%2fHyperionFDM%2fDefault.aspx
it's saying
Error: An error occurred logging on to the system.
Detail: The username or password is invalid or you do not have permission to create or add a new FDM application.
But when I trying to login without the Domain name, the login is accepted.
But after login when I'm trying to "
Verify that the list of accounts is from your Financial Management application."
its saying "ActiveX component can't create object: 'fdmFM11xG5C.clsFMAdapter'".
Please Help
Thnx for quick response
I have registered with FM11x-G5-C.xml , FM11i-G5-C.xml & upsFM40M.dll
is there any mistake.
it'll be higly helpful to let me know the reqd adapters name.
I am very new to this field.
Thanx
For Hyperion 11.1.1.3 FDQM which adapter files should I use?
Please let me know the .xml & .dll files name...
Can anybody support me giving the Link of OBE Hyperion FDQM 11.1.1.3.0 installation & config tutorials page?

Similar Messages

  • FDM & ERPI to FDMEE - Upgrade approach/challenges ?

    Hello,
    We are in FDM/ERPI 11.1.2.2.300 (used for loading data from EBS and file systems into HFM 11.1.2.2.300) and planning to upgrade (within the existing environment [in-place upgrade]) to 11.1.2.3. As we see that FDM/ERPI are integrated together as FDMEE in 11.1.2.3, we had big grey area on how FDMEE would function in contrast to disparate FDM/ERPI.
    We needed inputs from veterans or anyone who had experienced the upgrade to answer the follwing queries.
    Queries:
    [1] We have come to know that FDM and ERPI are integrated together as FDMEE in 11.1.2.3.
         [a] Since we have FDM and ERPI as separate components in 11.1.2.2.300, how should FDM and ERPI components be upgraded to FDMEE ? Any specific steps / approach that need to be followed ?
         [b] Are there any known issues / challenges involved in upgrading from disparate FDM & ERPI to FDMEE ?
         [c] How would ERPI artifacts (of 11.1.2.2.300) function in FDMEE - Are there any specific migration or any manual set-up required post-upgrade (to 11.1.2.3) to have ERPI artifacts (like source system, period mapping, location, import formats, data load mapping etc)      functioning in FDMEE ?
         [d] Similarly how would FDM artifacts (from 11.1.2.2.300) function in FDMEE ? Are there any specific migration or manual set-up required post-upgrade (to 11.1.2.3) to have FDM artifacts (like mapping, validation rules etc) functioning in FDMEE ?
    [2] Since only ODI 11.1.1.7 is compatible with 11.1.2.3 (from Compatability matrix), how do we upgrade ODI from 11.1.1.6 to 11.1.1.7 ? Are there any patch available to upgrade or should we uninstall ODI 11.1.1.6 and install ODI 11.1.1.7 ?
    Any insightful response will be helpful for us as it would help us to gain clarity/confidence and comfort in upgrading to 11.1.2.3.
    Regards,
    Sathish

    hello Sathish
    As you are upgrading from 11.1.2.2.300 to EPM 11.1.2.3 (within the existing environment [in-place upgrade]) , it's called Applying maintenance release EPM 11.1.2.3 on EPM 11.1.2.2
    [a] when you apply maintenance release , FDM 11.1.2.2 gets upgraded to FDM 11.1.2.3 ,
         and ERPi gets upgraded to FDMEE 11.1.2.3
    [b] i have not faced any issue while upgrading FDM
    [c] please check epm_install.pdf for EPM 11.1.2.3 for upgradation  task for FDM and FDMEE  (see page 257 / 258)
    regards,
    -DM

  • Need help in FDM data manipulations

    Hi,
    I am fairly new to Hyperion FDM and still learning its features with Planning. I have following scenario.
    I need to move my data from Dept ABC to Dept XYZ when the account = '10000'. I have both the ABC and XYZ depts in my Planning system.
    Is there a way I can accomplish this in FDM? I am looking at the document, but not sure where to look at it.
    If you can provide some insight on this that will be great.
    Thanks,
    NM

    Hi NM,
    Please see my response to your question on the Network 54 forum.
    FDM is really intended to perform data loads from a source system than move data around within the cube.
    Your requirement is somewhat ambiguous and it sounds like you want to move a specific piece of data that resides within the planning cube already rather than make all future data load like this. Presuming this is a one time activity you may consider using a calc script/business rule.
    You will find that may people use both forums and posting to one is a more effective way to rely on people answering your queries.
    Regards,
    John

  • FDM and Hyperion Financial Reporting

             Hi all,
    Is it possible to build Financial Reporting reports for FDM application? If it's possible, how to connect to FDM via FinRep?
    Thank you in advance.

    You can build custom reports to use custom SQL query.
    You could configure one report such as:
    Location     AccountA     AccountB     AccountC
    Location1     10000          12311          32423
    Location2     21311          50000          20000
    I would suggest you start with your SQL query (SQL Developer, MSSQL Studio, Toad...) and try to get data in your format.
    Then you can create a report and test this query in SQL Preview to see if it is working (sometimes SQL queries have some restrictions when being used in reports)
    Once you have your data in the SQL preview you can build your report to show it.
    HTH

  • Integartion of Oracle to FDM

    Hi All,
    I want to load data into FDM directly fom Oracle database through an Integration Script.
    I developed the script for integrating , but when i run the script i got an Error:in line
    Set rsAppend=DW.DataAccess.farstable(strworktablename)
    the Error says that the table doesn't exist
    Can any one guide me for which table its giving me the error.
    Plz guide me.
    Plz revert back to me in case of any queries.
    Thanks,

    Hi,
    I have tried replicating the Integration script at a client site as mentioned in the Admin Guide and its fails to execute giving me a data access error at the following statement:
    Set rsAppend = DW.DataAccess.farsTable(strWorkTableName)
    The section of the script is given below:
    Function CTIDB2Test(strLoc, lngCatKey, dblPerKey, strWorkTableName)
    'Hyperion FDM Integration Import Script:
    'Created By:     admin
    'Date Created:     2/19/2009 1:16:59 PM
    'Purpose: This import integration script connects to the xyz database
    ' and imports Actuals data from the generalLedgerMasterFile and glMasterAccount table into FDM
    Dim cnSS 'ADO Connection object
    Dim strSQL 'SQL String
    Dim rs 'Source System recordset
    Dim rsAppend 'FDM recordset
    Dim sPOVString
    Dim sPOVPeriod
    Dim sDataCol
    ' Initialise the ADO objects
    Set cnSS = CreateObject("ADODB.Connection")
    Set rs = CreateObject("ADODB.RecordSet")
    Set rsAppend = DW.DataAccess.farsTable(strWorkTableName)

  • Could I have CASE or IF statement in FMS queries?

    Is it possible to have case statements in FMS queries:
    For example:
    SELECT T0.[U_DepoistfeeON] case
          when T0.[U_DepoistfeeON] is NOT BLANK  then $[$38.111.160]='Deposit Fee'
          when T0.[U_DepoistfeeON] is BLANK  then  then $[$38.111.160]=BLANK
    end FROM OITM T0
    What is wrong with above query please? Thank you very much.
    I do not mind even if above query is doable with an IF statement in it.

    Hi Rahul, this is what I want -
    I have a user defined field attached to item master OITM. The field is called : U_DepoistfeeON
    The above field contains additional depositFee taxes for selling beer bottles.
    I have also created a new tax as part of Freight handling. While on Sales Order screen, the Freight (unhide first through forms field) drop down can have the new tax type selected automatically "Deposit Fee".
    Thus, if the line item is of beer type that has U_DepositfeeON, then Freight field should automaticlally pick the type "Deposit Fee".
    If U_DepositfeeON is zero then I would like the Freight field on Sales Order screen ($38.111.160) set to blank.
    I tried to achieve this through following using case statement.
    select T0.U_DepoistfeeON
    from oitm t0
    case
    when T0.U_DepoistfeeON 0
    then $http://$38.111.160='Deposit Fee'
    when T0.U_DepoistfeeON = 0
    then $http://$38.111.160=''
    else
    $http://$38.111.160=''
    end;
    Of course it doesn't work. Note: I know I typed in DepoistfeeON, the error is not due to that.
    Thanks.

  • How to Dene a Data Link Between Queries: Bind Variables

    This is an interesting topic and I cannot get it to work using Bind Variables.
    I have 2 queries: Q1 and Q2. Q2 needs c_id, account_code and account_type from Q1.
    Whe I run the data template below, I get only the data for Q1.
    Now people may argue that there is no data in Q2 for the relevant clause. So if I even remove the where clause in Q2 I still get no joy i.e Data appears for Q1 but not for Q2
    <dataTemplate name="FLCMR519_DATA_SET" description="Termination Quote Report">
         <parameters>
              <parameter name="cid" dataType="number" defaultValue="1"/>
              <parameter name="p_cln_id" dataType="number" defaultValue="62412"/>
         </parameters>
         <dataQuery>
              <sqlStatement name="Q1">
                   <![CDATA[SELECT qm.qmd_id,
    qm.contract_period,
    qm.quo_quo_id||'/'||qm.quote_no||'/'||qm.revision_no reference_no,
    qm.contract_distance,
    qm.mdl_mdl_id,
    q.qpr_qpr_id,
    q.quo_id,
    q.drv_drv_id,
    qm.revision_user username,
    pb.first_name||' '||pb.last_name op_name,
    pb.telephone_no,
    pb.facsimile_no,
    pb.email,
    q.c_id c_id,
    q.account_type account_type,
    q.account_code account_code,
    m.model_desc,
    ph.payment_description payment_head_desc,
    cl.fms_fms_id,
    cl.start_date,
    cl.end_date,
    cl.actual_end_date,
    cl.con_con_id,
    cl.cln_id,
    cl.term_qmd_id term_qmd_id,
    qm2.contract_period term_period,
    qm2.contract_distance term_distance
    FROM quotations q,
               quotation_models qm,
               contract_lines cl,
               personnel_base pb,
               models m,
               model_types mt,
               payment_headers ph,
               quotation_models qm2
    WHERE q.quo_id = qm.quo_quo_id
           AND cl.cln_id = :p_cln_id
           AND qm.qmd_id = cl.qmd_qmd_id
           AND qm2.revision_user = pb.employee_no (+)
           AND qm.mdl_mdl_id = m.mdl_id
           AND m.mtp_mtp_id = mt.mtp_id
           AND qm.payment_id = ph.payment_header_id
           AND qm2.qmd_id (+) = cl.term_qmd_id
    ]]>
              </sqlStatement>
              <sqlStatement name="Q2">
                   <![CDATA[SELECT ea.c_id,                  ea.account_type,ea.account_code,ea.account_name
    FROM external_accounts ea
                 WHERE ea.c_id = :c_id
                   AND ea.account_type = :account_type
                   AND ea.account_code = :account_code
    ]]>
              </sqlStatement>
         </dataQuery>
    </dataTemplate>

    Defining dataStructure section is mandatory for multiple queries.

  • How to use one WAD template for all the available queries

    Hi,
    I have created a WAD template... Now we have close to 25 queries which will use that template for displaying their output at Portal...
    Is there any way that i don`t have to create multiple copies of my WAD template...
    I.e. all queries would call that same template through portal...

    Hi,
    The Bex report uses the Standard Template 0ANALYSIS_PATTERN while it is executed.
    So if your sure that the look and feel of all the Reports is going to remain same, you don't need to even attach all these Queries to Templates in WAD.
    When you execute the BEx query, it would pick up the standard template and run it in portal.
    But if you want to customize the Report Template:
    1) You can change this standard template 0ANALYSIS_PATTERN  and customize it according to ur requrement in WAD.
    2) Or you can create a copy of this standard template and change it as required. Then change the template in spro (transaction RSCUSTV27) to point to this Z template.
    Let me know if you need more details.
    Regards,
    Forum

  • Can FDM handle more than one year at a time?

    Hi,
    I'm trying to load a multi-load file through FDM with the target being an Essbase cube and the source being a text file. The file is for 36 months which crosses 3 years. The problem is Oracle support is saying FDM can only handle one calendar year at a time and we need to load 3 separate files.
    The file imports and validates ok. And the export only pushes data to the first year '2010' but it doesn't error out.
    Any ideas?

    If you use an excel file as the source, you can load all 36 periods.
    You can also modify the adaptor to support this action but I would caution you that doing so will cause support to no longer support your application without requiring you to import a non customized adaptor. This also has consequences for when upgrading.
    Let me add, FDM support is the best in the business. I have never had a situation where what they told me was incorrect. What you are trying to accomplish would be a custom solution that should be implemented by an experienced FDM resource that understands the impact of the customization.
    Edited by: TonyScalese on May 6, 2010 3:10 PM

  • Error while Activating / running Bex queries/reports

    Hello
    I have following ERRORs while activating / running BEx reports/queries.
    1. While Activating BEx reports, I keel getting following ERRORS
    Object directory entry R3TR ELEM E341OIR6G715GUMGN0MRXP3MA does not exist
    Error when activating element BX7FOOKQB4588RROCBTC2YVZN
    BEx transport request 'BIDK901352' is not available or not suitable
    2. While running the BEx report/query in Bex Analyzer, I'm getting ERROR in entering the parameters, its NOT allowing / giving me option to select the parameters, for example for 0COSTCENTER, it is NOT displaying list costcenters to select before I can RUN that particular report.
    3. How can I trouble shoot ERRORs related to BEx and Portal connectivity.
    4. How would I setup the Variants for a Bex query/report ? Are these Variants for per user basis / per report basis. Can I have each variant for particular users OR all users must select their own variant each time when they RUN the report ? 
    Thanks, Sorry I have asked multiple questions, in jst one thread.
    BI

    Goto SE03 -->Search for Objects in Requests/Tasks  option , select the object type and execute it will give all the transports it has collected before ..
    Also you can use transport connection in RSA1 and select the query and find out all the objects in query locked by what request etc ..
    Thanks,
    Ravi

  • Error while transporting Queries

    Hi,
    I am getting this error while transporting queries!!
    Object '!ZTIC_UKA' (ELEM) of type 'Query' is not available in version 'A'
    Message no. RSO252
    Diagnosis
    You wanted to generate an object with the name '!ZTIC_UKA' (in transport request ELEM) of type 'Query' (TLOGO). This is, however, not available in the BW Repository database. It does not exist in the requested version A. If the version is 'D' then it is possible that an error arose during the delivery or installation. If the version is 'A' then the Object was either not created or not activated.
    System Response
    The object was not taken into account in the next stage of processing
    Thanks

    Hi Murali,
    YOu are trying to transport the element ZTIC_UKA'  of the query,may be an variable  which is not colelcted in the request.
    Try to transport the query through RSA1> Transport connection->Query-->give you query name -->Drag and Drop to Right hand side  and Click on collect all dependent objects and check whetehr they have collected in the same request.Make sure ur variables are collected.
    Tranport it now and Check.
    Rgds
    SVU123
    Edited by: svu123 on Mar 4, 2010 7:41 AM

  • Need procedure for creation of BW Roles, Assigning Queries,Publishing Roles

    Hi Experts,
      Could you please let me know the procedure for creation of BW Roles, Assigning Queries,Publishing Roles in Business Explorer (BEx - BW 3.5)
    Thanks in advance,
    Andy

    Hi,
    Creating BW Roles
    http://help.sap.com/saphelp_nw04/helpdata/en/52/6714b6439b11d1896f0000e8322d00/frameset.htm
    Assigning Queries
    After creating the query, save the query to a role from the query designer.
    Publishing Roles in Business Explorer
    https://websmp101.sap-ag.de/~sapdownload/011000358700002894802003E/HowToBIPortal1.pdf
    Hope this helps you..!
    -Pradnya

  • Problem using DECODE() function with a Query of Queries

    I
    posted
    on my blog about an issue I was having trying to use the PL/SQL
    DECODE() function with a Coldfusion Query of Queries. This function
    works fine when you query a database for information. However, when
    you query another query, it seems that CF doesn't recognize it. I
    got errors stating that it found a left parenthesis where it
    expected a FROM key word. Here is a simplified version of what I am
    trying to do:
    quote:
    <!--- Simulated query; similar to what I was calling from
    my database --->
    <cfscript>
    qOriginal = queryNew("Name,Email,CountryCode",
    "VarChar,VarChar,VarChar");
    newRow = queryAddRow(qOriginal, 5);
    querySetCell(qOriginal, "Name", "Joe", 1);
    querySetCell(qOriginal, "Email", "[email protected]", 1);
    querySetCell(qOriginal, "CountryCode", "AMER", 1);
    querySetCell(qOriginal, "Name", "Sally", 2);
    querySetCell(qOriginal, "Email", "[email protected]", 2);
    querySetCell(qOriginal, "CountryCode", "AMER", 2);
    querySetCell(qOriginal, "Name", "Bob", 3);
    querySetCell(qOriginal, "Email", "[email protected]", 3);
    querySetCell(qOriginal, "CountryCode", "ASIA", 3);
    querySetCell(qOriginal, "Name", "Mary", 4);
    querySetCell(qOriginal, "Email", "[email protected]", 4);
    querySetCell(qOriginal, "CountryCode", "EURO", 4);
    querySetCell(qOriginal, "Name", "John", 5);
    querySetCell(qOriginal, "Email", "[email protected]", 5);
    querySetCell(qOriginal, "CountryCode", "EURO", 5);
    </cfscript>
    <cfquery name="qCountries" dbtype="query">
    SELECT DISTINCT(CountryCode) AS CountryCode,
    DECODE(states, "AMER", "North America &amp; Canada",
    "EURO", "Europe &amp; Africa", "ASIA", "Japan &amp;
    Asia","") CountryName
    FROM qOriginal
    ORDER BY CountryCode
    </cfquery>
    <cfdump var="#qCountries#">
    <!--- ========== END OF CODE ========== --->
    So running this returned the following error:
    Query Of Queries syntax error.
    Encountered "(. Incorrect Select Statement, Expecting a
    'FROM', but encountered '(' instead, A select statement should have
    a 'FROM' construct.
    Does anybody know why this doesn't work? Is it just not
    supported? Please note that I have also tried to use the CASE()
    function instead of DECODE() and that resulted in basically the
    same error. For now I an looping over my distinct query with a
    switch statement and manually loading a new query with the data how
    I want it. But it would be a lot cleaner and less code to have the
    DECODE() to work. Thx!

    DECODE() is an Oracle function, not generic SQL. Q-of-Q is a
    very limited subset of SQL and lacks many functions and clauses
    available in standard SQL, especially what you may be used to using
    in your particular RDBMS.
    See
    Query
    of Queries user guide
    Phil

  • Report Developer Control Of Applying Hints to Analytics Queries

    There are numerous ways to apply hints to the queries generated by Analytics:
    - Table level
    - Join level
    - Evaluate calculation
    Each has its advantages and drawbacks.
    - Table level: applies the hint to every query that references the table.
    - Join level: applies the hint whenever the join is used in the query.
    - Evaluate: allows the report developer to include a hint, but can't control where Analytics decides to apply the hint.
    I propose another method for the report developer to apply hints, when needed, that uses join level hints. All the report developer
    does is add the hint column to the Answer or add a filter based on the hint column to the Answer to apply the hint.
    Setup
    NOTE: I suggest you do consistency checks along the way, especially before starting work in the next Layer, to be sure that all setup errors are resolved early.
    1) Start by defining a Logical SQL table in the Physical Layer using the following SQL: Select 1 Hint from dual
    2) Alias this table for each hint to be defined for report developer usage. As an example, alias the hint table, creating
    No Star and Parallel alias tables.
    3) Join each alias to the physical layer fact tables, using a complex join, where the hint could be applied. In the Join definition screen, put the hint in the HINT field and enter 1=1 for
    in the Expression section. Yes, we are creating a cartesian join between the hint and the other joining table. As the hint table always returns one and only one row, there
    is no effect on the rows returned by the query. For No Star, you
    put NO_STAR_TRANSFORMATION in the Hint field. For Parallel, you put PARALLEL(<physical table name>, default, default), where the physical table name
    is the name of the actual database table, not the name of the alias table (Analytics will put the alias in the place of the database table name
    when it generates the SQL). Additionally, for hints that have no parameters, you only need to join it
    to the Fact tables in a query and not necessarily the dimensions. If you include fields from multiple fact tables, the hint will be applied
    for each fact table. So, you may see the hint multiple times in the SQL (something like SELECT /*+ NO_STAR_TRANSFORMATION NO_STAR_TRANSFORMATION */ t00001.col1...)
    4) Add the hint alias tables to the BMM Layer.
    5) Rename the Hint field in each of the BMM hint tables to identify the hint being applied. For No Star, change the column name from Hint to No Star Hint. For Parallel,
    change the column name from Hint to Parallel Hint.
    6) Set the hint column as a key.
    7) Join the BMM hint tables to the appropriate fact tables, using a complex join.
    8) Define each hint table as a dimension.
    9) Set the Logical Level in the Content tab in each of the sources of the joined tables to use the Detail of the hint dimension.
    10) Create a folder in the Presentation Layer called Hints
    11) Place each BMM hint field into the Presentation Layer Hints folder.
    To apply a hint to your Answer, either add the Hint field to your Answer or create a filter where the Hint field is equal to/is in 1 (the number one). Check that the SQL generated
    contains the hint, in Answers, go into Administration, Session Manager, and view the log for the user (the user log level will need to have been set to 7 to see the SQL generated).
    Use of hints in more complex setups can be done by performing a setup of the hints that is parallel to the fact table setup. As an example, if you specify fragmentation content and a where
    clause in your BMM for your fact tables, you would setup parallel physical layer hint tables and joins, BMM objects and joins, fragmentation content, and where clauses based on the
    hint tables (only hint tables are referenced in the where clause).
    As any database person knows, hints can either help or degrade the performance of queries. So, taking the SQL of the pre-hint Answer and figuring out which hints give the best
    performance is suggested, prior to adding the hint fields/filters to the Answer.

    Hi Oliver,
    I would suggest you to have a look at the below WLST script which would give you the required report of the active threads and it would be send an email too.
    Topic: Sending Email Alert for Threads Pool Health Using WLST
    http://middlewaremagic.com/weblogic/?p=5433
    Topic: Sending Email Alert for Hogger Threads Count Using WLST
    http://middlewaremagic.com/weblogic/?p=5423
    Also you can use the below script in case of the stuck threads, this script would send you an email with the thread dumps during the issue occurred.
    Topic: Sending Email Alert For Stuck Threads With Thread Dumps
    http://middlewaremagic.com/weblogic/?p=5582
    Regards,
    Ravish Mody

  • FDM export to HFM - 2nd export did not clear all data from 1st export?

    Has anyone experienced exporting a replacement FDM file to HFM where the 2nd file did not contain an account that was present in the original file and that account/value in the 1st file then remained in the HFM trial balance and caused it to be out of balance? Does the replace export option not clear all the prior file data? The only unusual part of it might be that the problem account was derived using logic group script? Thanks.

    Thanks Hensen - Sure, it was easy enough to fix once I deduced what the problem was - although my solution was just to use the data clear function and reload. The real question is how to prevent it from happening again and it sounds like I can't do much about it - other than trying to train the users to watch out for do-over files that have fewer accounts than the original load?? Am I wrong to think that HFM should try to fully clear all previous data from a load file if there is a replacement or at least try to validate whether or not all prior accounts have been repopulated?? There's plenty of good validation in the validation step of FDM. Why not during the export step? If this makes sense, I'll put in another "enhancement request" to support.

Maybe you are looking for