ERPI: Data Loading problem Hyperion Planning & Oracle EBS

Hi
I am trying to load data from Oracle EBS to Hyperion Planning.
When i push data Zero rows are inserted in Target.
When i look at Table " SELECT * FROM TDATASEG "
It is showing me data but it is not comminting data in Target application.
The reason is Data difference in Source (EBS) and Target.
In my source Year is 2013 but in Target 'FY14' so for Entity Source is '21' but Target is '2143213'
Can you please let me know how to solve this issue?
can i place a Lookup table for this in EPRI.
i am using EPRI and ODI to push data.
Regards
Sher

Have you setup the data load mapping correctly to map the source value to the proper target value? Based on what you are describing it seems that the system generated * to * map is being used, if you are mapping a source to a different target, this needs to be added to the data load mapping.

Similar Messages

  • Outline load utility Data load error in Planning

    Hi,
    We are loading the data using OutlineLoad utility in Hyperion Planning 11.1.2.1,but data is not loading and the log file showing the below error.
    Unable to obtain dimension information and/or perform a data load: com.hyperion.planning.olap.NoAvailableOlapConnectionsException: Unable to obtain a connection to Hyperion Essbase. If this problem persists, please contact your administrator.
    Can you please why we are getting this error.
    Thanks.

    Hi John,
    I tried refresh the db and running the utility again, still its showing same error,
    and the log file showing the following.
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///2560/Info(1051164)
    Received login request from [::ffff:IP]
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///2560/Info(1051187)
    Logging in user [admin@Native Directory] from [::ffff:IP]
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///2788/Info(1051001)
    Received client request: List Applications (from user [admin@Native Directory])
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///11872/Info(1051001)
    Received client request: List Databases (from user [admin@Native Directory])
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///11232/Info(1051164)
    Received login request from [::ffff:IP]
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///11232/Info(1051187)
    Logging in user [admin@Native Directory] from [::ffff:IP]
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///2824/Info(1051001)
    Received client request: Get Application Info (from user [admin@Native Directory])
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///2560/Info(1051001)
    Received client request: Get Application State (from user [admin@Native Directory])
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///2788/Info(1051001)
    Received client request: Select Application/Database (from user [admin@Native Directory])
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///2788/Info(1051009)
    Setting application BUD_APP active for user [admin@Native Directory]
    [Sat Oct 15 16:52:23 2011]Local/ESSBASE0///11872/Info(1051001)
    Received client request: Logout (from user [admin@Native Directory])
    Thanks

  • ERPi Data load mapping Issue

    Hi,
    We are facing issue with ERPi data load mappings issue. Mapping file (txt file) has 36k records, whenever we are trying to load mappings, it's taking very long time, nearly 1 hour 30mins. but we want to reduce that time. is there any way to reduce data load mapping time??
    Hyperion verion: 11.1.2.2.300
    Please help, thanks in advance!!
    Thanks.

    Any one face the same kind of issue??

  • Data copy in Hyperion Planning taking long time

    Hi All,
    good morning.
    I am using data copy in hyperion planning (11 1 2 2) to copy from one scenario to other selecting account annotations and supporting details.
    The essbase part of copying data completed (checked sessions), but when I checked the job console in planning it still says processing for past 2 hrs.
    My java heap size for planning is 1.5GB and the backed database is SQL server
    my suspicion is that the issue was with backend sql. but I don't where to start.
    can anyone please guide me.

    I am working with version 11.1.2.1 and running into the same issue. The Supporting detail option works fine if the application has just a few details, let’s say 50 cells, but if we have around 500 cell with details then the copy process never ends, the details are not copied, and I have to restart the Planning service. As a workaround we are using a two steps process. In the first step we copy the just Essbase using the copy data functionality of Planning (with no data copy options enabled) or by using an Essbase calculation script. For the Supporting Details piece we use the Export for Edit functionality of LCM and export them to an XML file, and then we edit the XML file and change the source member name with the target member name, and finally we use the Import after Edit functionality of LCM. Of course if works only if the Planning application was deployed using EPMA.

  • Formatting header rows in data form in hyperion planning

    Hello Friends,
    I want to format header rows in data form in hyperion planning (formatting like Bold, Italic etc...) Is there any way to do that?

    I could be wrong. That's why i thought i should post this one.
    This is from 11.1.2.2.300 readme
    Customizing the Style of Header Cells in Forms
    Applying a bold style to all forms
    1.  From HyperionPlanning.war\custom, open HspCustom.css and insert this class code:
    customheaderStyle
    Font-family:<font>;
    Font-size:<size>;
    Font-weight:bold;
    For example, to use bolded garamond 14 as the style for header cells you would specify:
    customheaderStyle
    Font-family:garamond;
    Font-size:14;
    Font-weight:bold;
    You can also change the font and the font size in the same portion of code.
    Perform the tasks in "Registering Customized Skins in Applications".
    Applying a bold style to one form
    Determine the ID of the form to customize by running this query in which ‘<FORM_NAME’ is the name of the form:
    Select object_id AS FORM_ID,object_name AS FORM_NAME from hsp_object 
    where object_name like ‘<FORM_NAME>’
    For example, to modify a form called "1.30 Per Payor – Metrics”, you would run this query:
    Select object_id AS FORM_ID,object_name AS FORM_NAME from hsp_object where object_name like "1.30 Per Payor-Metrics"
    When the ID is returned, create a css class in HspCustom.css using the following format.  HspCustom.css is in HyperionPlanning.war\custom.
    “customheaderStyle_<FORM_ID>”
    Set font-weight in .customheaderStyle as bold as shown below. You can also change the font and the font size in the same portion of code.
    customheaderStyle
    Font-family:<font>;
    Font-size:<size>;
    Font-weight:bold;
    To use different styles on individual forms, create a css class for each form.
    so there is a way to create custom headers
    Regards
    Celvin
    http://www.orahyplabs.com

  • Oracle Hyperion Planning with EBS 11i

    hi all
    we have Oracle EBS 11i [Finance, HRMS, SCM, CRM] implemented and working fine...
    now we are going to select Oracle Hyperion Planning, for Budget and Planing;;
    Is Oracle Hypeion Planning Plus have integrator with EBS ...
    if not what are the recomended components required
    Thanks
    Abuzaid
    [email protected]

    Hi
    I not confident but I guess, Hyperion don't have integrator with OEBS
    We solved issue like this using ODI
    Regards
    Sasha

  • How exactly can I load TEXT DATA to a Hyperion Planning application?

    I have a data file that has TEXT data that I want to load to planning. I do not wish for this text data to be displayed in a Smart List, rather I want to use the new Planning data type "Text"
    In chapter 5 of the Planning Admin Guide in the Matrix for Data Load Tools, it says the Planning Outline Load Utility can load Text Data values.
    I can't make sense out of the the instructions to load data with this utility. Can someone please help me?
    -Chris

    Hi John,
    I like your idea of flipping the Data Load Dimension and the Driver Dimension. Unfortunately Planning won't let me set Period as a Data Load Dimension.
    I only get Account, CCCC, Employee, SCCC, and SAU_CAU_Facility (5 of the 11 dimensions in this application).
    So that's why I set the Data Load Dimension and the Driver Dimension as I did. It ran but with some kickouts on missing members. So 16 of the 19 records I sent processed and I understand why these three kicked out in the error file:_
    [Tue Jul 28 10:33:37 EDT 2009] Error occurred loading data record 1: Job_Description,Wrkforce,ENVIRONMENTAL COORDINATOR,"BegBalance, 71_71_CAU, GAK_SCCC, No_CCCC, Actual, FY09, Final, Emp_7340"
    [Tue Jul 28 10:33:37 EDT 2009] com.hyperion.planning.InvalidMemberException: The member Emp_7340 does not exist or you do not have access to it.
    [Tue Jul 28 10:33:37 EDT 2009] Error occurred loading data record 2: Job_Description,Wrkforce,EXECUTIVE ASSISTANT,"BegBalance, 71_71_CAU, G02_SCCC, No_CCCC, Actual, FY09, Final, Emp_4435"
    [Tue Jul 28 10:33:37 EDT 2009] com.hyperion.planning.InvalidMemberException: The member Emp_4435 does not exist or you do not have access to it.
    [Tue Jul 28 10:33:37 EDT 2009] Error occurred loading data record 4: Job_Description,Wrkforce,PRESIDENT & COO-YANKEE ENERGY SYSTEM,"BegBalance, 71_71_CAU, G02_SCCC, No_CCCC, Actual, FY09, Final, Emp_8747"
    [Tue Jul 28 10:33:37 EDT 2009] com.hyperion.planning.InvalidMemberException: The member Emp_8747 does not exist or you do not have access to it.
    [Tue Jul 28 10:33:38 EDT 2009]Planning Outline data store load process finished with exceptions: exceptions occured, examine the exception file for more information. 19 data records were read, 19 data records were processed, 16 were successfully loaded, 3 were rejected.
    The log file that loads tells some good news as well..._
    Successfully logged into "Wrkforce" application, Release 11.111, Adapter Interface Version 5, Workforce supported and not enabled, CapEx not supported and not enabled, CSS Version 3
    "Account" dimension properties and information:
    Account, Parent, Alias: Default, Valid For Consolidations, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Data Load Cube Name, BegBalance, Point-of-View, Account Type, Time Balance, Skip Value, Exchange Rate Type, Variance Reporting, Source Plan Type, Plan Type (Wrkforce), Aggregation (Wrkforce)
    No UDA's defined on "Account"
    "Period" dimension properties and information:
    Period, Parent, Alias: Default, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Type, Start Period, End Period, Aggregation (Wrkforce)
    No UDA's defined on "Period"
    "Years" dimension properties and information:
    Years, Parent, Alias: Default, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation
    No UDA's defined on "Years"
    "Scenario" dimension properties and information:
    Scenario, Parent, Alias: Default, Valid For Consolidations, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Start Year, Start Period, End Year, End Period, Exchange Table, Include BegBal, Process Management Enabled, Aggregation (Wrkforce)
    No UDA's defined on "Scenario"
    "Version" dimension properties and information:
    Version, Parent, Alias: Default, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Version Type, Process Management Enabled, Aggregation (Wrkforce)
    No UDA's defined on "Version"
    "CCCC" dimension properties and information:
    CCCC, Parent, Alias: Default, Valid For Consolidations, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Plan Type (Wrkforce), Aggregation (Wrkforce), Open_Closed_CCCC
    No UDA's defined on "CCCC"
    "Open_Closed_CCCC" attribute dimension (on base dimension "CCCC"). Attributes defined on the "Open_Closed_CCCC" dimension: Open_CCCC; Closed_CCCC;
    Open_Closed_CCCC, Parent, Alias: Default, Operation
    "Employee" dimension properties and information:
    Employee, Parent, Alias: Default, Valid For Consolidations, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Aggregation (Wrkforce)
    No UDA's defined on "Employee"
    "SCCC" dimension properties and information:
    SCCC, Parent, Alias: Default, Valid For Consolidations, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Aggregation (Wrkforce), Open_Closed_SCCC
    No UDA's defined on "SCCC"
    "Open_Closed_SCCC" attribute dimension (on base dimension "SCCC"). Attributes defined on the "Open_Closed_SCCC" dimension: Open_SCCC; Closed_SCCC;
    Open_Closed_SCCC, Parent, Alias: Default, Operation
    "SAU_CAU_Facility" dimension properties and information:
    SAU_CAU_Facility, Parent, Alias: Default, Valid For Consolidations, Data Storage, Two Pass Calculation, Description, Formula, UDA, Smart List, Data Type, Operation, Aggregation (Wrkforce), Open_Closed_SCF
    No UDA's defined on "SAU_CAU_Facility"
    "Open_Closed_SCF" attribute dimension (on base dimension "SAU_CAU_Facility"). Attributes defined on the "Open_Closed_SCF" dimension: Open_SCF; Closed_SCF;
    Open_Closed_SCF, Parent, Alias: Default, Operation
    Exchange Rates properties:
    Table, Description, To Currency, From Currency, Operation, Method, Historical, Beg Balance, Year, Period, Average, Ending
    UDA properties:
    Dimension, UDA, Operation
    Translation input file fields:
    Value, Driver Member, Point-of-View, Data Load Cube Name
    [Tue Jul 28 10:33:35 EDT 2009]Successfully located and opened input file "D:\Hyperion\products\Planning\bin\testload.csv".
    [Tue Jul 28 10:33:35 EDT 2009]Header record fields: Account, Data Load Cube Name, BegBalance, Point-of-View
    [Tue Jul 28 10:33:35 EDT 2009]Located and using "Account" dimension for loading data in "Wrkforce" application.
    [Tue Jul 28 10:33:38 EDT 2009]Load dimension "Account" has been unlocked successfully.
    [Tue Jul 28 10:33:38 EDT 2009]A cube refresh operation will not be performed.
    [Tue Jul 28 10:33:38 EDT 2009]Create security filters operation will not be performed.
    [Tue Jul 28 10:33:38 EDT 2009]Examine the Essbase log files for status if Essbase data was loaded.
    [Tue Jul 28 10:33:38 EDT 2009]Planning Outline data store load process finished with exceptions: exceptions occured, examine the exception file for more information. 19 data records were read, 19 data records were processed, 16 were successfully loaded, 3 were rejected.

  • ERPI data load error

    Hello,
    We are loading one month data from peoplesoft to planning application using ERPI 11.1.2.1. It takes about 1 hrs to load one month data.Is it normal? When looking at ODI PS_GL_LOAD_BALANCES_DATA is always successful but COMM_LOAD_BALANCES fails at the end with this error.Did anyone have this error before. Message in HyS9aifWeb-sysout file is:
    FROM AIF_HS_BALANCES
    WHERE LOADID = 132
    [EssbaseRuleFile] Locking rules file AIFData
    [EssbaseRuleFile] Successfully locked rules file AIFData
    [EssbaseRuleFile] Copying rules file OWBPLAND for data load as AIFData
    [EssbaseRuleFile] Unlocking rules file AIFData
    [EssbaseRuleFile] Successfully unlocked rules file AIFData
    [EssbaseRuleFile] The data rules file has been created successfully.
    [EssbaseRuleFile] Locking rules file AIFData
    [EssbaseRuleFile] Successfully locked rules file AIFData
    [EssbaseRuleFile] Load data into the cube by launching rules file...
    <Mar 1, 2012 1:09:30 PM PST> <Error> <WebLogicServer> <BEA-000337> <[STUCK] ExecuteThread: '5' for queue: 'weblogic.kernel.Default (self-tuning)' has been busy for "625" seconds working on the request "weblogic.servlet.internal.ServletRequestImpl@3bc2189[
    POST /aif/services/HPLService HTTP/1.1
    Content-type: text/xml;charset="utf-8"
    Accept: text/xml, multipart/related, text/html, image/gif, image/jpeg, *; q=.2, */*; q=.2
    Soapaction: "http://DEVAPPHPN001.indymacdev.biz:19000/aif/services/HPLServiceexecuteDataLoad"
    User-Agent: JAX-WS RI 2.1.7-b01-
    Content-Length: 638
    Connection: Keep-Alive
    Proxy-Client-IP: 10.205.101.108
    X-Forwarded-For: 10.205.101.108
    X-WebLogic-KeepAliveSecs: 30
    X-WebLogic-Force-JVMID: 2124105652
    ]", which is more than the configured time (StuckThreadMaxTime) of "600" seconds. Stack trace:
    Thread-176 "[STUCK] ExecuteThread: '5' for queue: 'weblogic.kernel.Default (self-tuning)'" <alive, suspended, priority=1, DAEMON> {
    jrockit.net.SocketNativeIO.readBytesPinned(SocketNativeIO.java:???)
    jrockit.net.SocketNativeIO.socketRead(SocketNativeIO.java:24)
    java.net.SocketInputStream.socketRead0(SocketInputStream.java:???)
    java.net.SocketInputStream.read(SocketInputStream.java:107)
    com.essbase.services.olap.main.main_direct.EssNetClient.readFrom(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssNetClient.readPackage(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssNetClient.receiveResponse2(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssNetClient.adNetReceiveResponse(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssAPIData._adImport(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssAPIData.ImportToEssbaseASO(Unknown Source)
    com.essbase.services.olap.main.main_direct.EssMAPIDir.BeginDataload(Unknown Source)
    com.essbase.server.framework.EssOlapMainService.BeginDataload(Unknown Source)
    com.essbase.server.framework.EssOrbPluginDirect.essMainBeginDataload(Unknown Source)
    com.essbase.api.session.EssOrbPlugin._invokeMainMethod(Unknown Source)
    com.essbase.api.session.EssOrbPlugin._invokeMethod2(Unknown Source)
    com.essbase.api.session.EssOrbPlugin._invokeMethod(Unknown Source)
    com.essbase.server.framework.EssOrbPluginDirect._invokeProtected(Unknown Source)
    com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
    com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
    com.essbase.api.session.EssOrbPlugin.essMainBeginDataload(Unknown Source)
    com.essbase.api.datasource.EssCube.beginDataload(Unknown Source)
    com.hyperion.aif.essbase.EssbaseRuleFile.executeDataRuleFile(EssbaseRuleFile.java:314)
    com.hyperion.aif.webservices.HPLService.executeDataLoad(HPLService.java:167)
    sun.reflect.GeneratedMethodAccessor7310.invoke(Unknown Source)
    sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    java.lang.reflect.Method.invoke(Method.java:575)
    org.apache.axis.providers.java.RPCProvider.invokeMethod(RPCProvider.java:397)
    org.apache.axis.providers.java.RPCProvider.processMessage(RPCProvider.java:71)
    org.apache.axis.providers.java.JavaProvider.invoke(JavaProvider.java:265)
    org.apache.axis.strategies.InvocationStrategy.visit(InvocationStrategy.java:32)
    org.apache.axis.SimpleChain.doVisiting(SimpleChain.java:113)
    org.apache.axis.SimpleChain.invoke(SimpleChain.java:78)
    org.apache.axis.handlers.soap.SOAPService.invoke(SOAPService.java:435)
    org.apache.axis.server.AxisServer.invoke(AxisServer.java:132)
    org.apache.axis.transport.http.AxisServlet.doPost(AxisServlet.java:586)
    javax.servlet.http.HttpServlet.service(HttpServlet.java:700)
    org.apache.axis.transport.http.AxisServletBase.service(AxisServletBase.java:325)
    javax.servlet.http.HttpServlet.service(HttpServlet.java:815)
    weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:224)
    weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:108)
    weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:206)
    weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:111)
    oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:299)
    oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:405)
    oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:94)
    oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:98)
    oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:70)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:86)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:25)
    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:55)
    weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3687)
    weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3681)
    weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:308)
    weblogic.security.service.SecurityManager.runAs(SecurityManager.java:116)
    weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2213)
    weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2135)
    weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1420)
    weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    weblogic.work.ExecuteThread.run(ExecuteThread.java:168)
    Any feedback is appreciated.Thank you for taking ime to read through.
    Regards.

    Hello,
    Part of the message you are getting is because the WebLogic process is a bit intuitive. It knows when something is running and waits. It also assumes anything running for a certain period of time is considered "stalled/stuck" and reports it. The time it waits is dependant on the Weblogic configuration.
    As for the time it takes to load data... that is very vague. Without knowing more about your environment that is very hard to state. We know nothing about the hardware, platform, dataset size, etc. You would need to review ODI to find out what is taking the long amount of time and try and replicate it outside of ERPi/ODI to see if you get the same results.
    Thank you,

  • Regarding ERPI Data Loading

    Dear All,
    I have few doubts on ERP Integrator.
    1) What are things required from Oracle GL to Planning for Data Loading using ERP Integrator? (Trail Balance is enough or we required some other file from Oracle GL)
    2) Is there any scheduling options available for Data loading using ERP Integrator?
    3) what is process for loading the data to Planning using ERP Integrator?
    4) How we load the data to Planning? (i.e. monthly load, hourly load)
    Anyone please guide me in this situation.
    Thanks,
    PC

    1) What are things required from Oracle GL to Planning for Data Loading using ERP Integrator? (Trail Balance is enough or we required some other file from Oracle GL)
    Assuming you have the right version of Oracle EBS, ERP Integrator queries the tables within the Oracle EBS database to get the appropriate information. In my case, the trail balance file was enough. Within the trail balance file you will have the appropriate dimension intersection (account, entity, period, etc.), the type of account (asset vs. liability, etc.) and finally the dollar amount.
    2) Is there any scheduling options available for Data loading using ERP Integrator?
    Yes. You can use FDQM to map and validate the data, then use the FDQM batch scheduler to load the data via command line or you can use the FDQM batch scheduler as well.
    3) what is process for loading the data to Planning using ERP Integrator?
    I'll try to do my best to summarize. (Assuming you are using FDQM) Create rules in ERPi -> Configure the adapters in the Workbench Client for the ERPi Rules -> Configure the FDQM Web Client to call the Adapters set in the Workbench Client -> Import the data into FDQM. Then from here you can call your command line automation for batching if you wish.
    4) How we load the data to Planning? (i.e. monthly load, hourly load)
    This depends on your business. Assuming you are going to load the data for budget and planning purposes then maybe your business is happy with a monthly load (and most of the time this is the case). An hourly load might be helpful if you deal with users that need up to date actuals. Loading hourly acutals data might be an overkill for a budget or planning application, but I have ran into situations where this is needed, but then find myself worried about speeding up the calculations after the data is loaded. Long store short you can load monthly or hourly.

  • Data load Error to Planning

    Hi All,
    I have created One Interface for loading data to Planning Application i have make the Account Dimension as Load dimension and Segment as Driver dimension (member is Audio)
    I have created one form for this now when i am running the interface its get successfully executed but at report statics it showing following problem
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 2, in ?
    Planning Writer Load Summary:
         Number of rows successfully processed: 0
         Number of rows rejected: 48
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:345)
         at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:169)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2374)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java:1615)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java:1580)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java:2755)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java:68)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2515)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:534)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:449)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1954)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:322)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:224)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:246)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:237)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:794)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:114)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
         at java.lang.Thread.run(Thread.java:619)
    I have find out a log file for it which says :
    Account,Data Load Cube Name,MP3,Point-of-View,Error_Reason
    Max Hits Per Day,Consol,Jan,"USD,Cannot load dimension member, error message is: java.lang.IllegalArgumentException: Unbalanced double quotes found at character 3 in the member list: "USD
    Avg Hits Per Day,Consol,Jan,"USD,Cannot load dimension member, error message is: java.lang.IllegalArgumentException: Unbalanced double quotes found at character 3 in the member list: "USD
    Can any buddy tell me what i did wrong. My Data file not contains any Double Quote mark with USD the POV is like this :---> USD,E01_0 ,Jan,Current,Working,FY11
    Thanks
    Gourav Atalkar
    Edited by: Gourav Atalkar(ELT) on May 3, 2011 4:06 AM

    As it is comma separated and all the members for the POV need to be together try enclosing it in quotes e.g "USD,E01_0 ,Jan,Current,Working,FY11"
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Automate dimension/data load update in Planning?

    Hi all,
    Without App Link and Translator (clients not buy), how to automate dimension/data load update in the planning? Any sample?
    Regards,
    Kenneth

    Hi John,
    I did the tests in Windows 2003 Server with Planning 9.3.1.
    1. 'bug 6829439: Task Flow is always set to active although the job has completed successfully'.
    I had a flow with 2 tasks. In this case when I launch the job it remains in a '4% complete' state because once the first task is completed, the systems doesn't put the task in 'completed' state and thus doesn't start the second one.
    2. 'bug 6785224: Manage TaskFlow does not show with interface tables'.
    When working with interface tables one cannot schedule taskflows.
    A colleague of mine working for another client has already opened up calls with Hyperion regarding this issues. Hyperion provided the bug numbers and informed us they will be fixed in version 9.5.
    3. Now I am testing ODI connectors on a Planning 9.2 systems and have some small problems (I've just opened a thread on this forum).
    Daniela S.

  • Data loading in Hyperion Web Forms

    Hi,
    I am new to Hyperion Planning technology.
    I have few doubts regarding Meta data and data loading.
    I am aware that we can load data into Essbase cubes and see it from Web Forms.
    Can we load data directly into webforms?
    Can we load meta data into essbase? or loading meta data through HAL to hyperion planning application?
    I understand that we can load data into Essbase and not clear whether we can load data into Hyperion Planning and also metadata..
    I request you to clarify me this asap.
    Thanks alot in advance for your help..
    Regards,
    Upendra.

    Hi,
    To answer some of your questions :-
    Can we load data directly into webforms?
    {color:#ff0000}The data in webforms is pulled directly from essbase, you can load the data into essbase or enter through the webforms or use something like Smartview.{color}
    Can we load meta data into essbase? or loading meta data through HAL to hyperion planning application?
    {color:#ff0000}If it is a planning application then the metadata has to go into planning first and is then pushed down to essbase, you can enter metadata directly into planning or use ODI or HAL (if you are licensed to use it) or if your brave then EPMA.{color}
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Network Error: Cannot Send Data (1042012) in Hyperion Planning

    I am using Hyperion Planning 11.1.2.1
    I'm getting this error when i'm trying to refresh my Planning database.
    Com.hyperion.planning.olap.EssbaseException: Network error [%s]: Cannot Send Data (1042012)
    Any suggestions?

    First make sure your essbase is up and running. If it is then:
    Increase the values for NETDELAY and NETRETRYCOUNT.
    Check the index cache size, data cache size, and data block size to make sure that they are within the recommended ranges.
    How to do it and how much to set, you will get this information in Essbase database administrator's guide:
    http://download.oracle.com/docs/cd/E17236_01/epm.1112/esb_dbag.pdf
    Also have a look at:
    Network Error message
    Cheers...!!!

  • Sorting a join loading into Hyperion planning

    Hi,
    i have Producr master data (1000+ data). I've to load into planning but i'm not able to understand how to sort (as SQL "order by 1 asc") a join of two table
    MASTER_PRODUCT, DESC_PROD.
    any suggestion?
    Thanks in advance
    DecaXD

    Hi,
    One method is customize the KM.
    If you right click "IKM SQL to Hyperion Planning" and select "Insert Option"
    Name the option something like "ORDER_OPTION" and set it to text.
    Now edit "IKM SQL to Hyperion Planning" and edit command "Load data into planning"
    In there should be
    sql= """select <%=odiRef.getPop("DISTINCT_ROWS") %> <%=odiRef.getColList("", "[EXPRESSION] [ALIAS_SEP] \u0022[COL_NAME]\u0022", ",", "", "INS and !TRG") %> from <%=odiRef.getFrom() %> where      (1=1) <%=odiRef.getFilter() %> <%=odiRef.getJrnFilter() %> <%=odiRef.getJoin() %> <%=odiRef.getGrpBy() %> <%=odiRef.getHaving() %>"""
    Add to the end - <%=odiRef.getOption("ORDER_OPTION")%>
    So it becomes
    sql= """select <%=odiRef.getPop("DISTINCT_ROWS") %> <%=odiRef.getColList("", "[EXPRESSION] [ALIAS_SEP] \u0022[COL_NAME]\u0022", ",", "", "INS and !TRG") %> from <%=odiRef.getFrom() %> where      (1=1) <%=odiRef.getFilter() %> <%=odiRef.getJrnFilter() %> <%=odiRef.getJoin() %> <%=odiRef.getGrpBy() %> <%=odiRef.getHaving() %><%=odiRef.getOption("ORDER_OPTION")%>"""
    Now in your interface and the flow section for the IKM you should have an Option "ORDER_OPTION"
    put in something like
    ORDER BY 1 ASC
    Give that a go.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Data load problem - BW and Source System on the same AS

    Hi experts,
    I’m starting with BW (7.0) in a sandbox environment where BW and the source system are installed on the same server (same AS). The source system is the SRM (Supplier Relationship Management) 5.0.
    BW is working on client 001 while SRM is on client 100 and I want to load data from the SRM into BW.
    I’ve configured the RFC connections and the BWREMOTE users with their corresponding profiles in both clients, added a SAP source system (named SRMCLNT100), installed SRM Business Content, replicated the data sources from this source system and everything worked fine.
    Now I want to load data from SRM (client 100) into BW (client 001) using standard data sources and extractors. To do this, I’ve created an  InfoPackage in one standard metadata data source (with data, checked through RSA3 on client 100 – source system). I’ve started the data load process, but the monitor says that “no Idocs arrived from the source system” and keeps the status yellow forever.
    Additional information:
    <u><b>BW Monitor Status:</b></u>
    Request still running
    Diagnosis
    No errors could be found. The current process has probably not finished yet.
    System Response
    The ALE inbox of the SAP BW is identical to the ALE outbox of the source system
    and/or
    the maximum wait time for this request has not yet run out
    and/or
    the batch job in the source system has not yet ended.
    Current status
    No Idocs arrived from the source system.
    <b><u>BW Monitor Details:</u></b>
    0 from 0 records
    – but there are 2 records on RSA3 for this data source
    Overall status: Missing messages or warnings
    -     Requests (messages): Everything OK
    o     Data request arranged
    o     Confirmed with: OK
    -     Extraction (messages): Missing messages
    o     Missing message: Request received
    o     Missing message: Number of sent records
    o     Missing message: Selection completed
    -     Transfer (IDocs and TRFC): Missing messages or warnings
    o     Request IDoc: sent, not arrived ; Data passed to port OK
    -     Processing (data packet): No data
    <b><u>Transactional RFC (sm58):</u></b>
    Function Module: IDOC_INBOUND_ASYNCHRONOUS
    Target System: SRMCLNT100
    Date Time: 08.03.2006 14:55:56
    Status text: No service for system SAPSRM, client 001 in Integration Directory
    Transaction ID: C8C415C718DC440F1AAC064E
    Host: srm
    Program: SAPMSSY1
    Client: 001
    Rpts: 0000
    <b><u>System Log (sm21):</u></b>
    14:55:56 DIA  000 100 BWREMOTE  D0  1 Transaction Canceled IDOC_ADAPTER 601 ( SAPSRM 001 )
    Documentation for system log message D0 1 :
    The transaction has been terminated.  This may be caused by a termination message from the application (MESSAGE Axxx) or by an error detected by the SAP System due to which it makes no sense to proceed with the transaction.  The actual reason for the termination is indicated by the T100 message and the parameters.
    Additional documentation for message IDOC_ADAPTER        601 No service for system &1, client &2 in Integration Directory No documentation exists for message ID601
    <b><u>RFC Destinations (sm59):</u></b>
    Both RFC destinations look fine, with connection and authorization tests successful.
    <b><u>RFC Users (su01):</u></b>
    BW: BWREMOTE with profile S_BI-WHM_RFC (plus SAP_ALL and SAP_NEW temporarily)
    Source System: BWREMOTE with profile S_BI-WX_RFCA (plus SAP_ALL and SAP_NEW temporarily)
    Someone could help ?
    Thanks,
    Guilherme

    Guilherme
    I didn't see any reason why it's not bringing. Are you doing full extraction or Delta. If delta extraction please check the extractor is delta enabled or not. Some times this may cause problems.
    Also check this weblog on data Load errors basic checks. it may help
    /people/siegfried.szameitat/blog/2005/07/28/data-load-errors--basic-checks
    Thanks
    Sat

Maybe you are looking for