Error while loading Data in HFM application through ODI
Hello All,
I am performing a integration from source(Flat file) to HFM application I have created a Integration to it and while loading the data i get the following error:-
org.apache.bsf.BSFException: exception from Jython:
Traceback (most recent call last):
File "<string>", line 3, in <module>
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:240)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
com.hyperion.odi.common.ODIHAppException: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:322)
at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:170)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2472)
at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:47)
at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:1)
at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2913)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2625)
at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:577)
at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:468)
at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2128)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:366)
at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:300)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:292)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:855)
at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)
at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
at java.lang.Thread.run(Thread.java:662)
Caused by: Traceback (most recent call last):
File "<string>", line 3, in <module>
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:240)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
com.hyperion.odi.common.ODIHAppException: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at org.python.core.PyException.fillInStackTrace(PyException.java:70)
at java.lang.Throwable.<init>(Throwable.java:181)
at java.lang.Exception.<init>(Exception.java:29)
at java.lang.RuntimeException.<init>(RuntimeException.java:32)
at org.python.core.PyException.<init>(PyException.java:46)
at org.python.core.PyException.<init>(PyException.java:43)
at org.python.core.Py.JavaError(Py.java:455)
at org.python.core.Py.JavaError(Py.java:448)
at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java:177)
at org.python.core.PyObject.__call__(PyObject.java:355)
at org.python.core.PyMethod.__call__(PyMethod.java:215)
at org.python.core.PyMethod.instancemethod___call__(PyMethod.java:221)
at org.python.core.PyMethod.__call__(PyMethod.java:206)
at org.python.core.PyObject.__call__(PyObject.java:397)
at org.python.core.PyObject.__call__(PyObject.java:401)
at org.python.pycode._pyx15.f$0(<string>:6)
at org.python.pycode._pyx15.call_function(<string>)
at org.python.core.PyTableCode.call(PyTableCode.java:165)
at org.python.core.PyCode.call(PyCode.java:18)
at org.python.core.Py.runCode(Py.java:1204)
at org.python.core.Py.exec(Py.java:1248)
at org.python.util.PythonInterpreter.exec(PythonInterpreter.java:172)
at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
... 19 more
Caused by: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:240)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java:175)
... 33 more
Caused by: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at com.hyperion.odi.hfm.ODIHFMUtil.validateRequiredProperties(ODIHFMUtil.java:87)
at com.hyperion.odi.hfm.ODIHFMMetadataLoader$OptionsMetadataLoad.validate(ODIHFMMetadataLoader.java:66)
at com.hyperion.odi.hfm.ODIHFMMetadataLoader.validateOptions(ODIHFMMetadataLoader.java:188)
at com.hyperion.odi.hfm.ODIHFMAppStatement.validateLoadOptions(ODIHFMAppStatement.java:168)
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:195)
... 38 more
Any help appreciated
Thanks,
Pratik
LKM looks fine. Sorry it was my bad the "data" at the end was not visible and you are right its IKM for integration
Log file that you have provided talks about these parameters but I doubt these are related to data
I was going through the link and it takes about the parameters CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE but they are related to metadata
Oracle Hyperion Financial Management
I have no idea. Will keep posted if I get to know anything
HTH
Amarnath
ORACLE | Essbase
Similar Messages
-
Error while loading data to hfm application
Hi
Friends i m totally new with this hfm, i created an application and i have loaded metadata,while loading data it gives error messsage as
Line: 2, Error: Invalid cell for Period January.
Actual; 2009; January; Periodic; Marina Beach; INR; Income; [ICP None]; Tea; [None]; [None]; [None]; 5000
Line: 3, Error: Invalid cell for Period July.
Budget; 2010; July; Periodic; White Field; INR; Income; [ICP None]; Cofee; [None]; [None]; [None]; 4800
please help me to debug this errorThe Data seems to be ok, so i suggest you log into your application and make sure the interception point is a data entry point. Basically in a grid view select the same members listed in the row of data. This will tell you what the issue is...
-
Integration Error in ODI while loading data to HFM application
Hello All,
I am performing a integration from source(Flat file) i.e example.csv to classic HFM application I have created an Integration and while loading the data with the all options provided, and if you see these Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory,which is getting as error as properties are not set
these properties are(as a options) is given in the IKM SQL to Hyperion Financial MetaData, but here I am using IKM SQL to Financial Management Data.
I get the following error:-
org.apache.bsf.BSFException: exception from Jython:
Traceback (most recent call last):
File "<string>", line 3, in <module>
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:240)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
com.hyperion.odi.common.ODIHAppException: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:322)
at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:170)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2472)
at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:47)
at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:1)
at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2913)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2625)
at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:577)
at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:468)
at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2128)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:366)
at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:300)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:292)
at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:855)
at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)
at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
at java.lang.Thread.run(Thread.java:662)
Caused by: Traceback (most recent call last):
File "<string>", line 3, in <module>
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:240)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
com.hyperion.odi.common.ODIHAppException: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at org.python.core.PyException.fillInStackTrace(PyException.java:70)
at java.lang.Throwable.<init>(Throwable.java:181)
at java.lang.Exception.<init>(Exception.java:29)
at java.lang.RuntimeException.<init>(RuntimeException.java:32)
at org.python.core.PyException.<init>(PyException.java:46)
at org.python.core.PyException.<init>(PyException.java:43)
at org.python.core.Py.JavaError(Py.java:455)
at org.python.core.Py.JavaError(Py.java:448)
at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java:177)
at org.python.core.PyObject.__call__(PyObject.java:355)
at org.python.core.PyMethod.__call__(PyMethod.java:215)
at org.python.core.PyMethod.instancemethod___call__(PyMethod.java:221)
at org.python.core.PyMethod.__call__(PyMethod.java:206)
at org.python.core.PyObject.__call__(PyObject.java:397)
at org.python.core.PyObject.__call__(PyObject.java:401)
at org.python.pycode._pyx15.f$0(<string>:6)
at org.python.pycode._pyx15.call_function(<string>)
at org.python.core.PyTableCode.call(PyTableCode.java:165)
at org.python.core.PyCode.call(PyCode.java:18)
at org.python.core.Py.runCode(Py.java:1204)
at org.python.core.Py.exec(Py.java:1248)
at org.python.util.PythonInterpreter.exec(PythonInterpreter.java:172)
at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
... 19 more
Caused by: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:240)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java:175)
... 33 more
Caused by: com.hyperion.odi.common.ODIHAppException: Properties [CLEAR_ALL_METADATA_BEFORE_LOAD, REPLACE_MODE] are mandatory.
at com.hyperion.odi.hfm.ODIHFMUtil.validateRequiredProperties(ODIHFMUtil.java:87)
at com.hyperion.odi.hfm.ODIHFMMetadataLoader$OptionsMetadataLoad.validate(ODIHFMMetadataLoader.java:66)
at com.hyperion.odi.hfm.ODIHFMMetadataLoader.validateOptions(ODIHFMMetadataLoader.java:188)
at com.hyperion.odi.hfm.ODIHFMAppStatement.validateLoadOptions(ODIHFMAppStatement.java:168)
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:195)
... 38 moreHi Ujjwal,
I remember Pratik had same issues and he resolved it.
Re: ODI Error while loading data to Classic HFM Application
Hope that fix your issue.
Bhabani -
Error while loading data on 0LIV_DS01 ODS through DTP
Hi Expert,
I am trying to load data on 0LIV_DS01 ODS through 2LIS_06_INV datasource data is comming correctly in PSA
while i'm trying to load data on ODS through DTP is runing properly till 9th package but at 10th package it showing me error
like "Exception condition "CTT_NOT_FOUND" raised"
Error Analysis
"A RAISE statement in the program "SAPLRSW0" raised the exception
condition "CTT_NOT_FOUND".
Since the exception was not intercepted by a superior
program, processing was terminated."
please tell me how i rectify the error....
Thanks and Regards
Lalit KumarHi,
This error has to do something with Currency Conversion.
While loading from the DSO to the cube, do you have any currency conversions?
Please look at the below threads:
Loading error in 0LIV_DS01
Raise Exception error: CTT_NOT_FOUND
Hope this helps.
Regards
Vishal -
Error While Loading data from HFM to Essbase
Hi,
I am trying to load data from a HFM application to Essbase application. The LKM i am using is "HFM Data to SQL" and the IKM is "SQL to Hyperion Essbase(DATA). I am also using couple of mapping expression like {CASEWHEN(HFMDATA."Account" = '3110', 'PL10100', HFMDATA."Account") } , {CASEWHEN(HFMDATA."Period" = 'January', 'Jan', HFMDATA."Period") } etc.
The error i am getting looks like this -
org.apache.bsf.BSFException: exception from Jython:
Traceback (innermost last):
File "<string>", line 23, in ?
com.hyperion.odi.essbase.ODIEssbaseException: Invalid column type specified for Data column [Account]
at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
at org.python.core.PyMethod.__call__(PyMethod.java)
at org.python.core.PyObject.__call__(PyObject.java)
at org.python.core.PyInstance.invoke(PyInstance.java)
I am mapping account dimension of HFM to to account dimension of essbase. Both source and target column type is "string" with length 80. Instead of keeping the target "Essbase application", if i keep an Oracle table, columns of which are exactly like essbase application, ODI is loading data perfectly with proper coversion.
Can anyody please help me out.
Many thanks.
N. Laha.Hi John,
Thank you very much for your response. Pasting here the SQl generated at step "8. Integration - Load data into Essbase" in the operator in the description tab, for your perusal -
from com.hyperion.odi.common import ODIConstants
from com.hyperion.odi.connection import HypAppConnectionFactory
from java.lang import Class
from java.lang import Boolean
from java.sql import *
from java.util import HashMap
# Get the select statement on the staging area:
sql= """select CASEWHEN(C3_ACCOUNT = '3110', 'PL10100', C3_ACCOUNT) "Account",CASEWHEN(C4_PERIOD = 'January', 'Jan', C4_PERIOD) "TimePeriod",'HSP_Inputvalue' "HSP_Rates",C8_SCENARIO "Scenario",CASEWHEN(C2_YEAR = '2007', 'FY09', C2_YEAR) "Years",CASEWHEN(C1_CUSTOM1 = '2012', 'Atacand', C1_CUSTOM1) "Product",CASEWHEN(C7_VIEW = 'YTD', 'Input', C7_VIEW) "Version",CASEWHEN(C6_VALUE = 'USD', 'Local', C6_VALUE) "Currency",C9_ENTITY "Entity",C5_DATAVALUE "Data" from "C$_0AZGRP_PLData" where (1=1) """
srcCx = odiRef.getJDBCConnection("SRC")
stmt = srcCx.createStatement()
srcFetchSize=30
stmt.setFetchSize(srcFetchSize)
rs = stmt.executeQuery(sql)
#load the data
stats = pWriter.loadData(rs)
#close the database result set, connection
rs.close()
stmt.close()
From the SQL, you may find the CASEWHEN expressions. Yes these i have added in the interface mapping section, with the help of ODI expression editor. Following are the expressions i have used, one statement for one mapping each. This is just to make the data of HFM, compatible to Essbase. e.g. - account number in HFM for which i am estracting the data is 3110, but there is no account '3110' in the essbase application, so essbase application wont be able to load the data. To make the HFM data compatible with Essbase, these casewhen statements have been used. I hope its fine to use such statements in mapping.
CASEWHEN(HFMDATA."Year" = '2007', 'FY09', HFMDATA."Year")
CASEWHEN(HFMDATA."View" = 'YTD', 'Input', HFMDATA."View")
CASEWHEN(HFMDATA."Entity" = '100106_LC', '1030GB', HFMDATA."Entity")
CASEWHEN(HFMDATA."Value" = 'USD', 'Local', HFMDATA."Value")
CASEWHEN(HFMDATA."Account" = '3110', 'PL10100', HFMDATA."Account")
CASEWHEN(HFMDATA."Custom1" = '2012', 'Atacand', HFMDATA."Custom1")
CASEWHEN(HFMDATA."Period" = 'January', 'Jan', HFMDATA."Period")
If you can get some idea from the SQL and help me out!
Many thanks.
N. Laha
Edited by: user8732338 on Sep 28, 2009 9:45 PM -
Error while loading metada into HFM Application
Can any one help me out from this stage in hfm.I am getting these errors frequently.Please send your comments on this .
Metadata Scan started at 21:41:36
PASS I - Analyzing Section Headers
PASS II - Validating member names and hierarchies
Scenario
Error: Member defined but not referenced in the hierarchy
Scenario
ActualEntity
Account
Custom1
Custom2
Custom3
Custom4
Currency
Scan ended at: 21:41:36
Elapsed time: 00:00:00You have a scenario which was defined, but is not attached to anything in the hierarchy section. Using the Win32 metadata manager, open the scenario dimension, and in the list tab check the box that shows orphans. Once you see which member isn't attached to the hierarchy, drag it where you want it, save and load the file.
--Chris -
Error while loading data from a file on application server
Hi all,
Facing an error while loading data from a flat file.
Error 'The argument '##yyyymmdd;@##' cannot be interpreted as a number ' while assigning character.
I changed the format of date fields (tried with number,general,date(International))in the xls. But i still get the same error.Did check all the data types in Data source all the fields are dats.
Can you please tell me what could be the problem?
Thank you all,
PraveenHi all,
As far as my first question i got through it but i had one more field in my flat file while actually is a time stamp, but in my flat file i have a data in this format
10/21/2006 5:11:48 AM which i need to change to 10/21/2006
one more note is i have some of the fields as NULL in this field
Last Updated Date
10/21/2006 5:11:48 AM
10/21/2006 5:11:48 AM
NULL
NULL
10/21/2006 5:11:48 AM
NULL
I want to display the values as 10/21/2006 and NULL as it is.
Please let me know if we have a conversion routine in datasource which can solve my problem.
Regards,
Praveen -
Error While loading data for LIS InfoSources.
Hi All,
I am repeatedly receiving load failure errors while loading data using 2lis_01_s001 (This is the case with all the InfoSources).
The error message is:
An error occurred in the source system.
Note
If the source system is a Client Workstation, then it is possible that the file that you wanted to load was being edited at the time of the data request. Make sure that the file is in the specified directory, that it is not being processed at the moment, and restart the request.
in our Quality system, we disabled the LIS Updation to No Update(R3) and loaded data and then again changed the Update Mode for No Updating to Asynchronous Update(R3). But now we are doing dataloading in Production. How to proceed. Should we have to disable the LIS Updating whenever we have to load the loads from R3 to BW.
Regards
JayHi Jayanthy,
Pls. check the order of the fields in the two set up tables for the S001 structure. The order of fields in both the tables should be the same.
You can see the structure in the TCode - SE11.
If the order is different, then you bneed to ask the BASIS person to change the order so that the order of fields in both the setup tables is same. This should fix the issue.
Thanks,
Raj -
Error while loading data from DS into cube
Hello All
I am getting the below error while loading data from Data source to the cube.
Record 1 :Time conversion from 0CALDAY to 0FISCPER (fiscal year ) failed with value 20070331
Wht am i suppose to do?
Need ur input in this regard.
Regards
RohitHi
Simply map 0calday to 0fiscper(you already done it)..You might have forgotten to map fiscal year variant in update rules....if it comes from souce,just map it..if it won't comes from source...make it as constant in update rules and give value..
If your business is April to March make it as 'V3'
If your business is January to Decemeber make it as 'K4'..
activate your update rules delete old data and upload again....
Hope it helps
Thanks
Teja -
Error while loading data from 0FI_TX_4 to ODS
Dear Friends,
One of our proccess chain getting error since a week, every day with same error while loading data from R3 data source 0FI_TX_4 to ODS.
Short dump is raised in R3 system.
"DBIF_RSQL_SQL_ERROR" CX_SY_OPEN_SQL_DBC
"SAPLBWFIR" or "LBWFIRU13"
"BWFIR_READ_BSET_CPUDT_DATA"
ORA-01555
The number of records are below 10000 only.
All other chains are OK.
we have already searched SDN and Notes...and applied below solution, but still no use
Increase the size of all the roll back segments
Increase number of REDO logs
Decrease the number of records per data packet
Any more suggessions please ??
Thanks
TonyHow are you telling that no. of records are below 10000 .Have you checked the no.of records in PSA???
Now ,I tell you something which I had done for this error .I increased the Package size such that whole data gets transferred in single data package.Please make the package size as large as possible so that all the data can be processed in single package.
I know you have read so many threads on the same but what I am suggesting is what I have done and successfully rectified this issue.
regards,
rocks -
Error while loading data to ODS
Hello BW folks ,
I am getting an error while loading data to ODS from R/3 data source.
The error message is " Error in PSA".
We have deleted yesterday data from PSA for the related infosource/data source successfully .
Since then this error is occuring. Previously data upload was successful.
In Detail tab of monitor we are getting error message in Request option ( red Mark )
and the same request is not going to R/3 .
<b>The related Job itself is not present in R/3 .</b>
This looks very strange and new to me .
Could anyone help me in rectifying this issue ?
Thanks in Advance.
Best Regards ,
Amol.Hello All,
Thanks for reply.
I tried to replicate the data source once again and activate the transfer rules in BW Production by running program in SE38 = RS_TRANS* activate all tranfer sturctures , then I got an error message that one infoobject is added to transfer rules in BW Production .
But interestingly the same Transfer rules do not have this infoobject in BW Development server . So I create once again a request for Transfer rules in Dev and transported it to BW production.
and This solved the problem
The error message in Monitor was " Error in PSA " and then request itself was not going to R/3 . But the real issue was with Transfer rules. So after rectifying it we have solved the problem of data upload.
Best Regards ,
Amol Kulkarni. -
10)example of errors while loading data and how do u resolve them?
)example of errors while loading data and how do u resolve them?
hai ram reddy this is ramprasad reddy....u can get this type of most commen errors....like below
As the frequent failures and errors , there is no fixed reason for the load to be fail , if you want it for the interview perspective I would answer it in this way.
a) Loads can be failed due to the invalid characters
b) Can be because of the deadlock in the system
c) Can be becuase of previuos load failure , if the load is dependant on other loads
d) Can be because of erreneous records
e) Can be because of RFC connections.....(sol: raise the ticket to BASIS team to provide the connection)
f)can be coz of missing master data.
g)can be coz of no data found in the source system.
h)Time stamp error........(sol: activate the data source and replicate the data source and load)
i)Short dump error.........(sol: delete the request and load once again)
j)Data locking when loading in parallel.
k)ODS activation failures.
l)No SID found........(sol: should load masterdata before transaction data) -
Getting error While loading data from ERP integrator to HFM
Hello,
We are getting the following error while loading the data from ERPI to HFM.
2013-12-31 22:44:54,133 INFO [AIF]: ERPI Process Start, Process ID: 300
2013-12-31 22:44:54,137 INFO [AIF]: ERPI Logging Level: DEBUG (5)
2013-12-31 22:44:54,139 INFO [AIF]: ERPI Log File: C:\Windows\TEMP\/aif_501_300.log
2013-12-31 22:44:54,141 INFO [AIF]: Jython Version: 2.5.1 (Release_2_5_1:6813, Sep 26 2009, 13:47:54)
[Java HotSpot(TM) 64-Bit Server VM (Oracle Corporation)]
2013-12-31 22:44:54,143 INFO [AIF]: Java Platform: java1.7.0_25
2013-12-31 22:44:56,821 INFO [AIF]: COMM Process Periods - Insert Periods - START
2013-12-31 22:44:56,828 DEBUG [AIF]:
SELECT l.SOURCE_LEDGER_ID
,l.SOURCE_LEDGER_NAME
,l.SOURCE_COA_ID
,l.CALENDAR_ID
,'0' SETID
,l.PERIOD_TYPE
,NULL LEDGER_TABLE_NAME
FROM AIF_BALANCE_RULES br
,AIF_COA_LEDGERS l
WHERE br.RULE_ID = 27
AND l.SOURCE_SYSTEM_ID = br.SOURCE_SYSTEM_ID
AND l.SOURCE_LEDGER_ID = br.SOURCE_LEDGER_ID
2013-12-31 22:44:56,834 DEBUG [AIF]:
INSERT INTO AIF_PROCESS_PERIODS (
PROCESS_ID
,PERIODKEY
,PERIOD_ID
,ADJUSTMENT_PERIOD_FLAG
,GL_PERIOD_YEAR
,GL_PERIOD_NUM
,GL_PERIOD_NAME
,GL_PERIOD_CODE
,GL_EFFECTIVE_PERIOD_NUM
,YEARTARGET
,PERIODTARGET
,IMP_ENTITY_TYPE
,IMP_ENTITY_ID
,IMP_ENTITY_NAME
,TRANS_ENTITY_TYPE
,TRANS_ENTITY_ID
,TRANS_ENTITY_NAME
,PRIOR_PERIOD_FLAG
,SOURCE_LEDGER_ID
SELECT DISTINCT brl.LOADID PROCESS_ID
,pp.PERIODKEY PERIODKEY
,prd.PERIOD_ID
,COALESCE(prd.ADJUSTMENT_PERIOD_FLAG, 'N') ADJUSTMENT_PERIOD_FLAG
,COALESCE(prd.YEAR,0) GL_PERIOD_YEAR
,COALESCE(prd.PERIOD_NUM,0) GL_PERIOD_NUM
,prd.PERIOD_NAME GL_PERIOD_NAME
,COALESCE(prd.PERIOD_CODE, CAST(COALESCE(prd.PERIOD_NUM,0) AS VARCHAR(38))) GL_PERIOD_CODE
,(COALESCE(prd.YEAR,0) * 10000 + COALESCE(prd.PERIOD_NUM,0)) GL_EFFECTIVE_PERIOD_NUM
,COALESCE(ppa.YEARTARGET, pp.YEARTARGET) YEARTARGET
,COALESCE(ppa.PERIODTARGET, pp.PERIODTARGET) PERIODTARGET
,'PROCESS_BAL_IMP' IMP_ENTITY_TYPE
,(COALESCE(prd.YEAR,0) * 10000 + COALESCE(prd.PERIOD_NUM,0)) IMP_ENTITY_ID
,prd.PERIOD_NAME IMP_ENTITY_NAME
,'PROCESS_BAL_TRANS' TRANS_ENTITY_TYPE
,(COALESCE(prd.YEAR,0) * 10000 + COALESCE(prd.PERIOD_NUM,0)) TRANS_ENTITY_ID
,pp.PERIODDESC TRANS_ENTITY_NAME
,'N' PRIOR_PERIOD_FLAG
,1 SOURCE_LEDGER_ID
FROM (
AIF_BAL_RULE_LOADS brl
INNER JOIN TPOVCATEGORY pc
ON pc.CATKEY = brl.CATKEY
INNER JOIN TPOVPERIOD_FLAT_V pp
ON pp.PERIODFREQ = pc.CATFREQ
AND pp.PERIODKEY >= brl.START_PERIODKEY
AND pp.PERIODKEY <= brl.END_PERIODKEY
LEFT OUTER JOIN TPOVPERIODADAPTOR_FLAT_V ppa
ON ppa.PERIODKEY = pp.PERIODKEY
AND ppa.PERIODFREQ = pp.PERIODFREQ
AND ppa.INTSYSTEMKEY = 'FMTEST2'
INNER JOIN AIF_GL_PERIODS_STG prd
ON prd.SOURCE_SYSTEM_ID = 3
AND prd.CALENDAR_ID IN ('10000')
AND prd.SETID = '0'
AND prd.PERIOD_TYPE = 'Month'
AND prd.ADJUSTMENT_PERIOD_FLAG = 'N'
AND prd.START_DATE > pp.PRIORPERIODKEY
AND prd.START_DATE <= pp.PERIODKEY
WHERE brl.LOADID = 300
ORDER BY pp.PERIODKEY
,GL_EFFECTIVE_PERIOD_NUM
2013-12-31 22:44:56,915 INFO [AIF]: COMM Process Periods - Insert Periods - END
2013-12-31 22:44:56,945 INFO [AIF]: COMM Process Periods - Insert Process Details - START
2013-12-31 22:44:56,952 DEBUG [AIF]:
INSERT INTO AIF_PROCESS_DETAILS (
PROCESS_ID
,ENTITY_TYPE
,ENTITY_ID
,ENTITY_NAME
,ENTITY_NAME_ORDER
,TARGET_TABLE_NAME
,EXECUTION_START_TIME
,EXECUTION_END_TIME
,RECORDS_PROCESSED
,STATUS
,LAST_UPDATED_BY
,LAST_UPDATE_DATE
SELECT PROCESS_ID
,ENTITY_TYPE
,ENTITY_ID
,ENTITY_NAME
,ENTITY_NAME_ORDER
,'AIF_EBS_GL_BALANCES_STG' TARGET_TABLE_NAME
,CURRENT_TIMESTAMP EXECUTION_START_TIME
,NULL EXECUTION_END_TIME
,0 RECORDS_PROCESSED
,'PENDING' STATUS
,'native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER' LAST_UPDATED_BY
,CURRENT_TIMESTAMP LAST_UPDATE_DATE
FROM (
SELECT DISTINCT PROCESS_ID
,IMP_ENTITY_TYPE ENTITY_TYPE
,IMP_ENTITY_ID ENTITY_ID
,IMP_ENTITY_NAME ENTITY_NAME
,(COALESCE(SOURCE_LEDGER_ID,0) * 100000000 + GL_EFFECTIVE_PERIOD_NUM) ENTITY_NAME_ORDER
FROM AIF_PROCESS_PERIODS
WHERE PROCESS_ID = 300
) q
ORDER BY ENTITY_NAME_ORDER
2013-12-31 22:44:56,963 DEBUG [AIF]:
INSERT INTO AIF_PROCESS_DETAILS (
PROCESS_ID
,ENTITY_TYPE
,ENTITY_ID
,ENTITY_NAME
,ENTITY_NAME_ORDER
,TARGET_TABLE_NAME
,EXECUTION_START_TIME
,EXECUTION_END_TIME
,RECORDS_PROCESSED
,STATUS
,LAST_UPDATED_BY
,LAST_UPDATE_DATE
SELECT PROCESS_ID
,ENTITY_TYPE
,ENTITY_ID
,ENTITY_NAME
,ENTITY_NAME_ORDER
,'TDATASEG' TARGET_TABLE_NAME
,CURRENT_TIMESTAMP EXECUTION_START_TIME
,NULL EXECUTION_END_TIME
,0 RECORDS_PROCESSED
,'PENDING' STATUS
,'native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER' LAST_UPDATED_BY
,CURRENT_TIMESTAMP LAST_UPDATE_DATE
FROM (
SELECT PROCESS_ID
,TRANS_ENTITY_TYPE ENTITY_TYPE
,MIN(TRANS_ENTITY_ID) ENTITY_ID
,TRANS_ENTITY_NAME ENTITY_NAME
,MIN(COALESCE(SOURCE_LEDGER_ID,0) * 100000000 + GL_EFFECTIVE_PERIOD_NUM) ENTITY_NAME_ORDER
FROM AIF_PROCESS_PERIODS
WHERE PROCESS_ID = 300
AND PRIOR_PERIOD_FLAG = 'N'
GROUP BY PROCESS_ID
,TRANS_ENTITY_TYPE
,TRANS_ENTITY_NAME
) q
ORDER BY ENTITY_NAME_ORDER
2013-12-31 22:44:56,970 INFO [AIF]: COMM Process Periods - Insert Process Details - END
2013-12-31 22:44:57,407 DEBUG [AIF]: EBS/FS GL Balances Print Variables - Printing Variables - START
2013-12-31 22:44:57,408 DEBUG [AIF]:
p_process_id: 300
p_sql_db_type: ORACLE
p_partitionkey: 12
p_rule_id: 27
p_source_system_id: 3
p_application_id: 26
p_target_application_type: HFM
p_is_multi_currency: true
p_data_load_method: CLASSIC_VIA_EPMI
p_bal_balance_method_code: STANDARD
p_bal_ledger_group_code: SINGLE
p_bal_amount_type: MONETARY
p_prd_entity_name: JAN-13
p_prd_period_id: 135
p_prd_gl_period_name: JAN-13
p_prd_source_ledger_id: 1
p_prd_source_coa_id: 101
p_source_ledger_id: 1
p_source_coa_id: 101
p_bal_actual_flag: A
p_bal_seg_column_name: SEGMENT1
p_max_ccid_loaded_to_stg: 148137
2013-12-31 22:44:57,408 DEBUG [AIF]: EBS/FS GL Balances Print Variables - Printing Variables - END
2013-12-31 22:44:57,806 INFO [AIF]: LKM EBS/FS Extract Type - Load Audit AND Full Refresh - START
2013-12-31 22:44:57,817 DEBUG [AIF]:
SELECT p.PROCESS_ID
,br.RULE_NAME
,l.SOURCE_LEDGER_NAME
FROM AIF_GL_LOAD_AUDIT aud
,AIF_PROCESSES p
,AIF_BALANCE_RULES br
,AIF_COA_LEDGERS l
WHERE aud.SOURCE_SYSTEM_ID = 3
AND aud.SOURCE_LEDGER_ID = 1
AND aud.GL_PERIOD_ID = 135
AND aud.BALANCE_TYPE = 'A'
AND p.PROCESS_ID = aud.LAST_LOADID
AND p.STATUS = 'RUNNING'
AND p.PROCESS_ID <> 300
AND br.RULE_ID = p.RULE_ID
AND l.SOURCE_SYSTEM_ID = aud.SOURCE_SYSTEM_ID
AND l.SOURCE_LEDGER_ID = aud.SOURCE_LEDGER_ID
2013-12-31 22:44:57,826 DEBUG [AIF]:
SELECT 'Y' VALID_FLAG
FROM GL_PERIOD_STATUSES
WHERE APPLICATION_ID = 101
AND SET_OF_BOOKS_ID = 1
AND PERIOD_NAME = 'JAN-13'
AND CLOSING_STATUS IN ( 'O','C','P' )
2013-12-31 22:44:57,847 DEBUG [AIF]:
SELECT 'Y' EXISTS_FLAG
FROM GL_TRACK_DELTA_BALANCES
WHERE SET_OF_BOOKS_ID = 1
AND PROGRAM_CODE = 'FEM'
AND PERIOD_NAME = 'JAN-13'
AND ACTUAL_FLAG = 'A'
AND EXTRACT_LEVEL_CODE = 'DTL'
AND CURRENCY_TYPE_CODE = 'B'
AND ENABLED_FLAG = 'Y'
2013-12-31 22:44:57,883 DEBUG [AIF]:
SELECT MAX(DELTA_RUN_ID) MAX_DELTA_RUN_ID
FROM GL_BALANCES_DELTA
WHERE SET_OF_BOOKS_ID = 1
AND PERIOD_NAME = 'JAN-13'
AND ACTUAL_FLAG = 'A'
2013-12-31 22:44:57,898 DEBUG [AIF]:
SELECT brl.EXECUTION_MODE
,( SELECT CASE COUNT(aud.DELTA_RUN_ID) WHEN 0 THEN 'N' ELSE 'Y' END
FROM AIF_GL_LOAD_AUDIT aud
WHERE aud.SOURCE_SYSTEM_ID = 3
AND aud.SOURCE_LEDGER_ID = 1
AND aud.GL_PERIOD_ID = 135
AND aud.BALANCE_TYPE = 'A'
AND aud.LAST_LOADID <> brl.LOADID
AND COALESCE( aud.STATUS, 'SUCCESS' ) = 'SUCCESS'
) GL_LOAD_AUDIT_SUCCESS_FLAG
,( SELECT CASE COUNT(aud.DELTA_RUN_ID) WHEN 0 THEN 'N' ELSE 'Y' END
FROM AIF_GL_LOAD_AUDIT aud
WHERE aud.SOURCE_SYSTEM_ID = 3
AND aud.SOURCE_LEDGER_ID = 1
AND aud.GL_PERIOD_ID = 135
AND aud.BALANCE_TYPE = 'A'
AND aud.LAST_LOADID <> brl.LOADID
AND aud.STATUS = 'RUNNING'
) GL_LOAD_AUDIT_RUNNING_FLAG
FROM AIF_BAL_RULE_LOADS brl
WHERE brl.LOADID = 300
2013-12-31 22:44:57,904 DEBUG [AIF]:
INSERT INTO AIF_GL_LOAD_AUDIT ( LAST_LOADID
,DELTA_RUN_ID
,SOURCE_SYSTEM_ID
,SOURCE_LEDGER_ID
,GL_PERIOD_ID
,BALANCE_TYPE
,GL_EXTRACT_TYPE
,STATUS
) VALUES ( 300
,0
,3
,1
,135
,'A'
,'FULLREFRESH'
,'RUNNING'
2013-12-31 22:44:57,907 DEBUG [AIF]:
DELETE FROM AIF_EBS_GL_BALANCES_STG
WHERE SOURCE_SYSTEM_ID = 3
AND SOURCE_COA_ID = 101
AND SOURCE_LEDGER_ID = 1
AND ACTUAL_FLAG = 'A'
AND PERIOD_NAME = 'JAN-13'
2013-12-31 22:44:59,283 INFO [AIF]: LKM EBS/FS Extract Type - Load Audit AND Full Refresh - END
2013-12-31 22:45:06,507 INFO [AIF]: COMM End Process Detail - Update Process Detail - START
2013-12-31 22:45:06,514 DEBUG [AIF]:
UPDATE AIF_PROCESS_DETAILS
SET STATUS = 'SUCCESS'
,RECORDS_PROCESSED = CASE
WHEN RECORDS_PROCESSED IS NULL THEN 0
ELSE RECORDS_PROCESSED
END + 57408
,EXECUTION_END_TIME = CURRENT_TIMESTAMP
,LAST_UPDATED_BY = CASE
WHEN ('native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER' IS NULL) THEN LAST_UPDATED_BY
ELSE 'native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER'
END
,LAST_UPDATE_DATE = CURRENT_TIMESTAMP
WHERE PROCESS_ID = 300
AND ENTITY_TYPE = 'PROCESS_BAL_IMP'
AND ENTITY_NAME = 'JAN-13'
2013-12-31 22:45:06,519 INFO [AIF]: COMM End Process Detail - Update Process Detail - END
2013-12-31 22:45:07,106 INFO [AIF]: EBS/FS Load Data - Load TDATASEG_T - START
2013-12-31 22:45:07,112 INFO [AIF]:
Import Data from Source for Period 'January 2013'
2013-12-31 22:45:07,115 DEBUG [AIF]:
SELECT brl.PARTITIONKEY
,brl.CATKEY
,brl.PLAN_TYPE
,CASE brl.PLAN_TYPE
WHEN 'PLAN1' THEN 1
WHEN 'PLAN2' THEN 2
WHEN 'PLAN3' THEN 3
WHEN 'PLAN4' THEN 4
WHEN 'PLAN5' THEN 5
ELSE 0
END PLAN_NUMBER
,brl.EXECUTION_MODE
,br.AMOUNT_TYPE
,br.BALANCE_SELECTION
,br.CURRENCY_CODE
,br.INCL_ZERO_BALANCE_FLAG
,br.BAL_SEG_VALUE_OPTION_CODE
,COALESCE(br.BALANCE_AMOUNT_BS, 'YTD') BALANCE_AMOUNT_BS
,COALESCE(br.BALANCE_AMOUNT_IS, 'PERIODIC') BALANCE_AMOUNT_IS
FROM AIF_BAL_RULE_LOADS brl
,AIF_TARGET_APPLICATIONS app
,AIF_BALANCE_RULES br
WHERE brl.LOADID = 300
AND app.APPLICATION_ID = brl.APPLICATION_ID
AND br.RULE_ID = brl.RULE_ID
2013-12-31 22:45:07,120 DEBUG [AIF]:
SELECT PERIODKEY
FROM TPOVPERIOD
WHERE PERIODDESC = 'January 2013'
2013-12-31 22:45:07,122 INFO [AIF]:
Import Data from Source for Ledger 'JWR Books'
2013-12-31 22:45:07,125 DEBUG [AIF]:
SELECT COA_SEGMENT_NAME
,ACCOUNT_TYPE_FLAG
,BALANCE_TYPE_FLAG
FROM AIF_COA_SEGMENTS
WHERE SOURCE_SYSTEM_ID = 3
AND SOURCE_COA_ID = '101'
AND (
ACCOUNT_TYPE_FLAG = 'Y'
OR BALANCE_TYPE_FLAG = 'Y'
2013-12-31 22:45:07,127 DEBUG [AIF]:
SELECT adim.BALANCE_COLUMN_NAME DIMNAME
,adim.DIMENSION_ID
,dim.TARGET_DIMENSION_CLASS_NAME
,(SELECT COA_SEGMENT_NAME
FROM AIF_COA_SEGMENTS cs
WHERE cs.COA_LINE_ID = tiie.IMPSOURCECOALINEID1
) COA_SEGMENT_NAME1
,(SELECT COA_SEGMENT_NAME
FROM AIF_COA_SEGMENTS cs
WHERE cs.COA_LINE_ID = tiie.IMPSOURCECOALINEID2
) COA_SEGMENT_NAME2
,(SELECT COA_SEGMENT_NAME
FROM AIF_COA_SEGMENTS cs
WHERE cs.COA_LINE_ID = tiie.IMPSOURCECOALINEID3
) COA_SEGMENT_NAME3
,(SELECT COA_SEGMENT_NAME
FROM AIF_COA_SEGMENTS cs
WHERE cs.COA_LINE_ID = tiie.IMPSOURCECOALINEID4
) COA_SEGMENT_NAME4
,(SELECT COA_SEGMENT_NAME
FROM AIF_COA_SEGMENTS cs
WHERE cs.COA_LINE_ID = tiie.IMPSOURCECOALINEID5
) COA_SEGMENT_NAME5
,(SELECT CASE mdd.ORPHAN_OPTION_CODE
WHEN 'CHILD' THEN 'N'
WHEN 'ROOT' THEN 'N'
ELSE 'Y'
END DIMENSION_FILTER_FLAG
FROM AIF_MAP_DIM_DETAILS_V mdd
,AIF_MAPPING_RULES mr
WHERE mr.PARTITIONKEY = tpp.PARTITIONKEY
AND mdd.RULE_ID = mr.RULE_ID
AND mdd.DIMENSION_ID = adim.DIMENSION_ID
) DIMENSION_FILTER_FLAG
,tiie.IMPCONCATCHAR
FROM TPOVPARTITION tpp
INNER JOIN AIF_TARGET_APPL_DIMENSIONS adim
ON adim.APPLICATION_ID = 26
INNER JOIN AIF_DIMENSIONS dim
ON dim.DIMENSION_ID = adim.DIMENSION_ID
LEFT OUTER JOIN TBHVIMPITEMERPI tiie
ON tiie.IMPGROUPKEY = tpp.PARTIMPGROUP
AND tiie.IMPFLDFIELDNAME = adim.BALANCE_COLUMN_NAME
AND tiie.IMPMAPTYPE = 'ERP'
WHERE tpp.PARTITIONKEY = 12
AND adim.BALANCE_COLUMN_NAME IS NOT NULL
ORDER BY adim.BALANCE_COLUMN_NAME
2013-12-31 22:45:07,154 DEBUG [AIF]:
INSERT INTO TDATASEG_T (
LOADID
,PARTITIONKEY
,CATKEY
,RULE_ID
,PERIODKEY
,VALID_FLAG
,CHANGESIGN
,CODE_COMBINATION_ID
,SOURCE_LEDGER_ID
,GL_PERIOD_YEAR
,GL_PERIOD_NUM
,YEAR
,PERIOD
,ATTR1
,ATTR2
,ATTR3
,ATTR4
,ATTR5
,ATTR6
,ATTR7
,ATTR8
,ATTR9
,ATTR10
,ATTR11
,ATTR12
,ATTR13
,ATTR14
,ACCOUNT
,ACCOUNTX
,ENTITY
,ENTITYX
,ICP
,ICPX
,UD1
,UD1X
,UD2
,UD2X
,UD3
,UD3X
,UD4
,UD4X
,DATAVIEW
,DATAKEY
,STAT_BALANCE_FLAG
,CURKEY
,AMOUNT_PTD
,AMOUNT_YTD
,AMOUNT
,AMOUNTX
SELECT pprd.PROCESS_ID LOADID
,12 PARTITIONKEY
,4 CATKEY
,27 RULE_ID
,pprd.PERIODKEY
,'Y' VALID_FLAG
,0 CHANGESIGN
,ccid.CODE_COMBINATION_ID
,bal.SOURCE_LEDGER_ID
,pprd.GL_PERIOD_YEAR
,pprd.GL_PERIOD_NUM
,pprd.YEARTARGET YEAR
,pprd.PERIODTARGET PERIOD
,pprd.PROCESS_ID ATTR1
,bal.SOURCE_SYSTEM_ID ATTR2
,bal.SOURCE_LEDGER_ID ATTR3
,pprd.GL_PERIOD_YEAR ATTR4
,pprd.GL_PERIOD_NAME ATTR5
,bal.ACTUAL_FLAG ATTR6
,bal.BUDGET_VERSION_ID ATTR7
,bal.ENCUMBRANCE_TYPE_ID ATTR8
,ccid.ACCOUNT_TYPE ATTR9
,NULL ATTR10
,NULL ATTR11
,NULL ATTR12
,NULL ATTR13
,NULL ATTR14
,ccid.SEGMENT4 ACCOUNT
,NULL ACCOUNTX
,ccid.SEGMENT1 ENTITY
,NULL ENTITYX
,NULL ICP
,NULL ICPX
,ccid.SEGMENT5 UD1
,NULL UD1X
,ccid.SEGMENT3 UD2
,NULL UD2X
,ccid.SEGMENT6 UD3
,NULL UD3X
,ccid.SEGMENT2 UD4
,NULL UD4X
,( CASE WHEN ccid.ACCOUNT_TYPE IN ('R','E','D','C') THEN 'Periodic' ELSE 'YTD' END ) DATAVIEW
,TDATASEG_DATAKEY_S.NEXTVAL
,'N' STAT_BALANCE_FLAG
,bal.CURRENCY_CODE CURKEY
,( CASE WHEN ccid.ACCOUNT_TYPE IN ('A','E','D') THEN 1 ELSE -1 END ) * ( bal.PERIOD_NET_DR - bal.PERIOD_NET_CR ) AMOUNT_PTD
,( CASE WHEN ccid.ACCOUNT_TYPE IN ('A','E','D') THEN 1 ELSE -1 END ) * ( bal.BEGIN_BALANCE_DR - bal.BEGIN_BALANCE_CR + bal.PERIOD_NET_DR - bal.PERIOD_NET_CR ) AMOUNT_YTD
,( CASE WHEN ccid.ACCOUNT_TYPE IN ('A','E','D') THEN 1 ELSE -1 END ) *
( CASE
WHEN ccid.ACCOUNT_TYPE IN ('R','E','D','C') THEN ( bal.PERIOD_NET_DR - bal.PERIOD_NET_CR )
ELSE ( bal.BEGIN_BALANCE_DR - bal.BEGIN_BALANCE_CR + bal.PERIOD_NET_DR - bal.PERIOD_NET_CR )
END
AMOUNT
,( CASE WHEN ccid.ACCOUNT_TYPE IN ('A','E','D') THEN 1 ELSE -1 END ) *
( CASE
WHEN ccid.ACCOUNT_TYPE IN ('R','E','D','C') THEN ( bal.PERIOD_NET_DR - bal.PERIOD_NET_CR )
ELSE ( bal.BEGIN_BALANCE_DR - bal.BEGIN_BALANCE_CR + bal.PERIOD_NET_DR - bal.PERIOD_NET_CR )
END
AMOUNTX
FROM AIF_EBS_GL_BALANCES_STG_V bal
,AIF_EBS_GL_CCID_STG ccid
,AIF_PROCESS_PERIODS pprd
WHERE bal.SOURCE_SYSTEM_ID = 3
AND bal.SOURCE_LEDGER_ID = 1
AND bal.ACTUAL_FLAG = 'A'
AND ccid.SOURCE_SYSTEM_ID = bal.SOURCE_SYSTEM_ID
AND ccid.SOURCE_COA_ID = bal.SOURCE_COA_ID
AND ccid.CODE_COMBINATION_ID = bal.CODE_COMBINATION_ID
AND pprd.PROCESS_ID = 300
AND pprd.PERIODKEY = '2013-01-01'
AND pprd.SOURCE_LEDGER_ID = bal.SOURCE_LEDGER_ID
AND pprd.GL_PERIOD_NAME = bal.PERIOD_NAME
AND (
bal.BEGIN_BALANCE_DR <> 0
OR bal.BEGIN_BALANCE_CR <> 0
OR bal.PERIOD_NET_DR <> 0
OR bal.PERIOD_NET_CR <> 0
AND bal.CURRENCY_CODE <> 'STAT'
AND bal.TRANSLATED_FLAG IS NULL
2013-12-31 22:45:09,269 INFO [AIF]: Monetary Data Rows Imported from Source: 12590
2013-12-31 22:45:09,293 DEBUG [AIF]:
INSERT INTO AIF_APPL_LOAD_AUDIT (
LOADID
,TARGET_APPLICATION_TYPE
,TARGET_APPLICATION_NAME
,PLAN_TYPE
,SOURCE_LEDGER_ID
,EPM_YEAR
,EPM_PERIOD
,SNAPSHOT_FLAG
,PARTITIONKEY
,CATKEY
,RULE_ID
,PERIODKEY
,EXPORT_TO_TARGET_FLAG
SELECT DISTINCT PROCESS_ID LOADID
,'HFM' TARGET_APPLICATION_TYPE
,'FMTEST2' TARGET_APPLICATION_NAME
,NULL PLAN_TYPE
,SOURCE_LEDGER_ID
,YEARTARGET EPM_YEAR
,PERIODTARGET EPM_PERIOD
,'Y' SNAPSHOT_FLAG
,12 PARTITIONKEY
,4 CATKEY
,27 RULE_ID
,PERIODKEY
,'N' EXPORT_TO_TARGET_FLAG
FROM AIF_PROCESS_PERIODS
WHERE PROCESS_ID = 300
AND PERIODKEY = '2013-01-01'
AND SOURCE_LEDGER_ID = 1
2013-12-31 22:45:09,297 DEBUG [AIF]:
INSERT INTO AIF_APPL_LOAD_PRD_AUDIT (
LOADID
,SOURCE_LEDGER_ID
,GL_PERIOD_ID
,DELTA_RUN_ID
,PARTITIONKEY
,CATKEY
,RULE_ID
,PERIODKEY
SELECT DISTINCT pprd.PROCESS_ID LOADID
,pprd.SOURCE_LEDGER_ID
,pprd.PERIOD_ID GL_PERIOD_ID
,(SELECT MAX(gl.DELTA_RUN_ID)
FROM AIF_GL_LOAD_AUDIT gl
WHERE gl.SOURCE_SYSTEM_ID = 3
AND gl.SOURCE_LEDGER_ID = pprd.SOURCE_LEDGER_ID
AND gl.BALANCE_TYPE = 'A'
AND gl.GL_PERIOD_ID = pprd.PERIOD_ID
) DELTA_RUN_ID
,12 PARTITIONKEY
,4 CATKEY
,27 RULE_ID
,pprd.PERIODKEY
FROM AIF_PROCESS_PERIODS pprd
WHERE pprd.PROCESS_ID = 300
AND pprd.PERIODKEY = '2013-01-01'
AND pprd.SOURCE_LEDGER_ID = 1
2013-12-31 22:45:09,302 INFO [AIF]:
Total Data Rows Imported from Source: 12590
2013-12-31 22:45:09,305 INFO [AIF]: EBS/FS Load Data - Load TDATASEG_T - END
2013-12-31 22:45:09,381 INFO [AIF]: COMM Update Data - Init DataLoadUtil - START
2013-12-31 22:45:09,385 INFO [AIF]: COMM Update Data - Init DataLoadUtil - END
2013-12-31 22:45:09,485 INFO [AIF]: COMM Update Data - Update TDATASEG_T/TDATASEGW - START
2013-12-31 22:45:09,491 DEBUG [AIF]:
DELETE FROM TDATASEG_T
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND AMOUNT = 0
2013-12-31 22:45:09,559 WARN [AIF]:
Warning: Data rows with zero balances exist
2013-12-31 22:45:09,636 INFO [AIF]: Zero Balance Data Rows Deleted: 1879
2013-12-31 22:45:09,654 DEBUG [AIF]:
SELECT DIMNAME
,CASE WHEN RULE_ID IS NULL THEN 'N' ELSE 'Y' END RULE_MAP_FLAG
,SRCKEY
,TARGKEY
,WHERECLAUSETYPE
,WHERECLAUSEVALUE
,CHANGESIGN
,SEQUENCE
,DATAKEY
,MAPPING_TYPE
FROM (
SELECT DISTINCT tdm.DIMNAME
,tdm.RULE_ID
,NULL SRCKEY
,NULL TARGKEY
,tdm.WHERECLAUSETYPE
,tdm.WHERECLAUSEVALUE
,NULL CHANGESIGN
,1 SEQUENCE
,COALESCE(tdm.SYSTEM_GENERATED_FLAG,'N') SYSTEM_GENERATED_FLAG
,NULL DATAKEY
,CASE
WHEN tdm.WHERECLAUSETYPE IS NULL THEN 1
ELSE 3
END MAPPING_TYPE
FROM TDATAMAP_T tdm
WHERE tdm.LOADID = 300
AND tdm.PARTITIONKEY = 12
AND tdm.TDATAMAPTYPE = 'ERP'
AND (tdm.RULE_ID IS NULL OR tdm.RULE_ID = 27)
AND tdm.WHERECLAUSETYPE IS NULL
UNION ALL
SELECT tdm.DIMNAME
,tdm.RULE_ID
,tdm.SRCKEY
,tdm.TARGKEY
,tdm.WHERECLAUSETYPE
,tdm.WHERECLAUSEVALUE
,tdm.CHANGESIGN
,CASE tpp.PARTSEQMAP
WHEN 0 THEN CASE
WHEN (tdm.WHERECLAUSETYPE = 'MULTIDIM') THEN 2
WHEN (tdm.WHERECLAUSETYPE = 'BETWEEN') THEN 3
WHEN (tdm.WHERECLAUSETYPE = 'LIKE') THEN 4
ELSE 0
END
ELSE tdm.SEQUENCE
END SEQUENCE
,COALESCE(tdm.SYSTEM_GENERATED_FLAG,'N') SYSTEM_GENERATED_FLAG
,tdm.DATAKEY
,CASE
WHEN tdm.WHERECLAUSETYPE IS NULL THEN 1
ELSE 3
END MAPPING_TYPE
FROM TDATAMAP_T tdm
INNER JOIN TPOVPARTITION tpp
ON tpp.PARTITIONKEY = tdm.PARTITIONKEY
WHERE tdm.LOADID = 300
AND tdm.PARTITIONKEY = 12
AND tdm.TDATAMAPTYPE = 'ERP'
AND (tdm.RULE_ID IS NULL OR tdm.RULE_ID = 27)
AND tdm.WHERECLAUSETYPE IN ('MULTIDIM','BETWEEN','LIKE')
) q
ORDER BY DIMNAME
,RULE_ID
,SEQUENCE
,SYSTEM_GENERATED_FLAG
,SRCKEY
2013-12-31 22:45:09,672 INFO [AIF]:
Processing Mappings for Column 'ACCOUNT'
2013-12-31 22:45:09,677 DEBUG [AIF]:
UPDATE TDATASEG_T
SET ACCOUNTX = ACCOUNT
,ACCOUNTR = 121
,ACCOUNTF = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '*' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND ACCOUNTX IS NULL
AND (1=1)
2013-12-31 22:45:10,044 INFO [AIF]: Data Rows Updated by Location Mapping 'Like' (LIKE): 10711
2013-12-31 22:45:10,053 INFO [AIF]:
Processing Mappings for Column 'ENTITY'
2013-12-31 22:45:10,057 DEBUG [AIF]:
UPDATE TDATASEG_T
SET ENTITYX = ENTITY
,ENTITYR = 122
,ENTITYF = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '*' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND ENTITYX IS NULL
AND (1=1)
2013-12-31 22:45:10,426 INFO [AIF]: Data Rows Updated by Location Mapping 'Like' (LIKE): 10711
2013-12-31 22:45:10,435 DEBUG [AIF]:
UPDATE TDATASEG_T
SET ENTITYX = ENTITY
,ENTITYR = 132
,ENTITYF = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '*' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND ENTITYX IS NULL
AND (1=1)
2013-12-31 22:45:10,446 INFO [AIF]: Data Rows Updated by Location Mapping 'DEFAULT' (LIKE): 0
2013-12-31 22:45:10,448 INFO [AIF]:
Processing Mappings for Column 'ICP'
2013-12-31 22:45:10,452 DEBUG [AIF]:
UPDATE TDATASEG_T
SET ICPX = '[ICP None]'
,ICPR = 130
,ICPF = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '[ICP None]' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND ICPX IS NULL
AND (1=1)
2013-12-31 22:45:10,784 INFO [AIF]: Data Rows Updated by Location Mapping 'Like' (LIKE): 10711
2013-12-31 22:45:10,798 INFO [AIF]:
Processing Mappings for Column 'UD1'
2013-12-31 22:45:10,802 DEBUG [AIF]:
UPDATE TDATASEG_T
SET UD1X = '[None]'
,UD1R = 124
,UD1F = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '[None]' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND UD1X IS NULL
AND (1=1)
2013-12-31 22:45:11,134 INFO [AIF]: Data Rows Updated by Location Mapping 'Like' (LIKE): 10711
2013-12-31 22:45:11,156 INFO [AIF]:
Processing Mappings for Column 'UD2'
2013-12-31 22:45:11,160 DEBUG [AIF]:
UPDATE TDATASEG_T
SET UD2X = '[None]'
,UD2R = 123
,UD2F = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '[None]' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND UD2X IS NULL
AND (1=1)
2013-12-31 22:45:11,517 INFO [AIF]: Data Rows Updated by Location Mapping 'Like' (LIKE): 10711
2013-12-31 22:45:11,531 INFO [AIF]:
Processing Mappings for Column 'UD3'
2013-12-31 22:45:11,535 DEBUG [AIF]:
UPDATE TDATASEG_T
SET UD3X = '[None]'
,UD3R = 125
,UD3F = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '[None]' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND UD3X IS NULL
AND (1=1)
2013-12-31 22:45:11,870 INFO [AIF]: Data Rows Updated by Location Mapping 'Like' (LIKE): 10711
2013-12-31 22:45:11,883 INFO [AIF]:
Processing Mappings for Column 'UD4'
2013-12-31 22:45:11,887 DEBUG [AIF]:
UPDATE TDATASEG_T
SET UD4X = '[None]'
,UD4R = 128
,UD4F = 3
,AMOUNTX = AMOUNTX * 1
,CHANGESIGN = CASE 1
WHEN -1 THEN CASE CHANGESIGN
WHEN 1 THEN 0
WHEN 0 THEN 1
ELSE CHANGESIGN
END
ELSE CHANGESIGN
END
,VALID_FLAG = CASE '[None]' WHEN 'IGNORE' THEN 'I' ELSE VALID_FLAG END
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND UD4X IS NULL
AND (1=1)
2013-12-31 22:45:12,192 INFO [AIF]: Data Rows Updated by Location Mapping 'Like' (LIKE): 10711
2013-12-31 22:45:12,204 DEBUG [AIF]:
UPDATE TDATASEG_T
SET ATTR14 = DATAKEY
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
2013-12-31 22:45:12,739 DEBUG [AIF]:
UPDATE TDATASEG_T
SET VALID_FLAG = 'N'
WHERE 1=1
AND (
(1=0)
OR TDATASEG_T.ACCOUNTX IS NULL
OR TDATASEG_T.ENTITYX IS NULL
OR TDATASEG_T.ICPX IS NULL
OR TDATASEG_T.UD1X IS NULL
OR TDATASEG_T.UD2X IS NULL
OR TDATASEG_T.UD3X IS NULL
OR TDATASEG_T.UD4X IS NULL
AND LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND VALID_FLAG = 'Y'
2013-12-31 22:45:12,754 INFO [AIF]:
Total Data Rows available for Export to Target: 10711
2013-12-31 22:45:12,773 INFO [AIF]: COMM Update Data - Update TDATASEG_T/TDATASEGW - END
2013-12-31 22:45:12,808 INFO [AIF]: COMM End Process Detail - Update Process Detail - START
2013-12-31 22:45:12,823 DEBUG [AIF]:
UPDATE AIF_PROCESS_DETAILS
SET STATUS = 'SUCCESS'
,RECORDS_PROCESSED = CASE
WHEN RECORDS_PROCESSED IS NULL THEN 0
ELSE RECORDS_PROCESSED
END + 10711
,EXECUTION_END_TIME = CURRENT_TIMESTAMP
,LAST_UPDATED_BY = CASE
WHEN ('native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER' IS NULL) THEN LAST_UPDATED_BY
ELSE 'native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER'
END
,LAST_UPDATE_DATE = CURRENT_TIMESTAMP
WHERE PROCESS_ID = 300
AND ENTITY_TYPE = 'PROCESS_BAL_TRANS'
AND ENTITY_NAME = 'January 2013'
2013-12-31 22:45:12,829 INFO [AIF]: COMM End Process Detail - Update Process Detail - END
2013-12-31 22:45:12,987 INFO [AIF]: COMM Update YTD Amounts - Update YTD Amounts - START
2013-12-31 22:45:12,993 DEBUG [AIF]:
SELECT brl.PARTITIONKEY
,brl.CATKEY
,pprd.YEARTARGET
,pprd.PERIODTARGET
,pprd.SOURCE_LEDGER_ID
FROM AIF_BAL_RULE_LOADS brl
,AIF_PROCESS_PERIODS pprd
WHERE brl.LOADID = 300
AND pprd.PROCESS_ID = brl.LOADID
GROUP BY brl.PARTITIONKEY
,brl.CATKEY
,pprd.YEARTARGET
,pprd.PERIODTARGET
,pprd.SOURCE_LEDGER_ID
HAVING COUNT(*) > 1
2013-12-31 22:45:12,995 INFO [AIF]: COMM Update YTD Amounts - Update YTD Amounts - END
2013-12-31 22:45:13,052 INFO [AIF]: COMM Load TDATAMAPSEG/TDATASEG - Load TDATAMAPSEG/TDATASEG - START
2013-12-31 22:45:13,057 DEBUG [AIF]:
SELECT brl.PARTITIONKEY
,brl.CATKEY
,brl.EXECUTION_MODE
FROM AIF_BAL_RULE_LOADS brl
WHERE brl.LOADID = 300
2013-12-31 22:45:13,059 DEBUG [AIF]:
SELECT PERIODKEY
FROM AIF_APPL_LOAD_AUDIT
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND RULE_ID = 27
ORDER BY PERIODKEY
2013-12-31 22:45:13,061 INFO [AIF]:
Processing Data for PeriodKey '2013-01-01'
2013-12-31 22:45:13,065 DEBUG [AIF]:
DELETE FROM TDATAMAPSEG
WHERE PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
AND (
TDATAMAPTYPE = 'ERP'
OR (
TDATAMAPTYPE = 'MULTIDIM'
AND EXISTS (
SELECT 1
FROM TDATAMAPSEG parent
WHERE parent.PARTITIONKEY = TDATAMAPSEG.PARTITIONKEY
AND parent.DATAKEY = TDATAMAPSEG.TARGKEY
AND parent.CATKEY = TDATAMAPSEG.CATKEY
AND parent.PERIODKEY = TDATAMAPSEG.PERIODKEY
AND parent.TDATAMAPTYPE = 'ERP'
2013-12-31 22:45:13,074 INFO [AIF]: Number of Rows deleted from TDATAMAPSEG: 8
2013-12-31 22:45:13,077 DEBUG [AIF]:
INSERT INTO TDATAMAPSEG (
DATAKEY
,PARTITIONKEY
,CATKEY
,PERIODKEY
,DIMNAME
,SRCKEY
,SRCDESC
,TARGKEY
,WHERECLAUSETYPE
,WHERECLAUSEVALUE
,CHANGESIGN
,SEQUENCE
,VBSCRIPT
,TDATAMAPTYPE
,SYSTEM_GENERATED_FLAG
SELECT DATAKEY
,PARTITIONKEY
,4
,'2013-01-01'
,DIMNAME
,SRCKEY
,SRCDESC
,TARGKEY
,WHERECLAUSETYPE
,WHERECLAUSEVALUE
,CHANGESIGN
,SEQUENCE
,VBSCRIPT
,TDATAMAPTYPE
,SYSTEM_GENERATED_FLAG
FROM TDATAMAP_T
WHERE LOADID = 300
2013-12-31 22:45:13,081 INFO [AIF]: Number of Rows inserted into TDATAMAPSEG: 8
2013-12-31 22:45:13,083 DEBUG [AIF]:
DELETE FROM TDATASEG
WHERE LOADID < 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND RULE_ID = 27
AND PERIODKEY = '2013-01-01'
2013-12-31 22:45:15,659 INFO [AIF]: Number of Rows deleted from TDATASEG: 10711
2013-12-31 22:45:15,728 DEBUG [AIF]:
INSERT INTO TDATASEG (
DATAKEY
,PARTITIONKEY
,CATKEY
,PERIODKEY
,CURKEY
,DATAVIEW
,CALCACCTTYPE
,CHANGESIGN
,JOURNALID
,AMOUNT
,AMOUNTX
,AMOUNT_PTD
,AMOUNT_YTD
,DESC1
,DESC2
,ACCOUNT
,ACCOUNTX
,ACCOUNTR
,ACCOUNTF
,ENTITY
,ENTITYX
,ENTITYR
,ENTITYF
,ICP
,ICPX
,ICPR
,ICPF
,UD1
,UD1X
,UD1R
,UD1F
,UD2
,UD2X
,UD2R
,UD2F
,UD3
,UD3X
,UD3R
,UD3F
,UD4
,UD4X
,UD4R
,UD4F
,UD5
,UD5X
,UD5R
,UD5F
,UD6
,UD6X
,UD6R
,UD6F
,UD7
,UD7X
,UD7R
,UD7F
,UD8
,UD8X
,UD8R
,UD8F
,UD9
,UD9X
,UD9R
,UD9F
,UD10
,UD10X
,UD10R
,UD10F
,UD11
,UD11X
,UD11R
,UD11F
,UD12
,UD12X
,UD12R
,UD12F
,UD13
,UD13X
,UD13R
,UD13F
,UD14
,UD14X
,UD14R
,UD14F
,UD15
,UD15X
,UD15R
,UD15F
,UD16
,UD16X
,UD16R
,UD16F
,UD17
,UD17X
,UD17R
,UD17F
,UD18
,UD18X
,UD18R
,UD18F
,UD19
,UD19X
,UD19R
,UD19F
,UD20
,UD20X
,UD20R
,UD20F
,ATTR1
,ATTR2
,ATTR3
,ATTR4
,ATTR5
,ATTR6
,ATTR7
,ATTR8
,ATTR9
,ATTR10
,ATTR11
,ATTR12
,ATTR13
,ATTR14
,ARCHIVEID
,HASMEMOITEM
,STATICDATAKEY
,LOADID
,RULE_ID,
CODE_COMBINATION_ID
,STAT_BALANCE_FLAG
,VALID_FLAG
SELECT
DATAKEY
,PARTITIONKEY
,CATKEY
,PERIODKEY
,CURKEY
,DATAVIEW
,CALCACCTTYPE
,CHANGESIGN
,JOURNALID
,AMOUNT
,AMOUNTX
,AMOUNT_PTD
,AMOUNT_YTD
,DESC1
,DESC2
,ACCOUNT
,ACCOUNTX
,ACCOUNTR
,ACCOUNTF
,ENTITY
,ENTITYX
,ENTITYR
,ENTITYF
,ICP
,ICPX
,ICPR
,ICPF
,UD1
,UD1X
,UD1R
,UD1F
,UD2
,UD2X
,UD2R
,UD2F
,UD3
,UD3X
,UD3R
,UD3F
,UD4
,UD4X
,UD4R
,UD4F
,UD5
,UD5X
,UD5R
,UD5F
,UD6
,UD6X
,UD6R
,UD6F
,UD7
,UD7X
,UD7R
,UD7F
,UD8
,UD8X
,UD8R
,UD8F
,UD9
,UD9X
,UD9R
,UD9F
,UD10
,UD10X
,UD10R
,UD10F
,UD11
,UD11X
,UD11R
,UD11F
,UD12
,UD12X
,UD12R
,UD12F
,UD13
,UD13X
,UD13R
,UD13F
,UD14
,UD14X
,UD14R
,UD14F
,UD15
,UD15X
,UD15R
,UD15F
,UD16
,UD16X
,UD16R
,UD16F
,UD17
,UD17X
,UD17R
,UD17F
,UD18
,UD18X
,UD18R
,UD18F
,UD19
,UD19X
,UD19R
,UD19F
,UD20
,UD20X
,UD20R
,UD20F
,ATTR1
,ATTR2
,ATTR3
,ATTR4
,ATTR5
,ATTR6
,ATTR7
,ATTR8
,ATTR9
,ATTR10
,ATTR11
,ATTR12
,ATTR13
,ATTR14
,ARCHIVEID
,HASMEMOITEM
,STATICDATAKEY
,LOADID
,RULE_ID,
CODE_COMBINATION_ID
,STAT_BALANCE_FLAG
,VALID_FLAG
FROM TDATASEG_T
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
2013-12-31 22:45:16,838 INFO [AIF]: Number of Rows inserted into TDATASEG: 10711
2013-12-31 22:45:16,858 DEBUG [AIF]:
DELETE FROM TDATASEG_T
WHERE LOADID = 300
AND PARTITIONKEY = 12
AND CATKEY = 4
AND PERIODKEY = '2013-01-01'
2013-12-31 22:45:17,123 INFO [AIF]: Number of Rows deleted from TDATASEG_T: 10711
2013-12-31 22:45:17,153 DEBUG [AIF]:
DELETE FROM TDATAMAP_T
WHERE LOADID = 300
2013-12-31 22:45:17,156 INFO [AIF]: Number of Rows deleted from TDATAMAP_T: 8
2013-12-31 22:45:17,161 INFO [AIF]: COMM Load TDATAMAPSEG/TDATASEG - Load TDATAMAPSEG/TDATASEG - END
2013-12-31 22:45:17,993 DEBUG [AIF]:
SELECT CASE app.METADATA_LOAD_METHOD
WHEN 'EPMA' THEN CASE dim.TARGET_DIMENSION_CLASS_NAME
WHEN 'Generic' THEN dim.TARGET_DIMENSION_NAME
ELSE dim.TARGET_DIMENSION_CLASS_NAME
END
ELSE dim.TARGET_DIMENSION_NAME
END TARGET_DIMENSION_NAME
,adim.BALANCE_COLUMN_NAME
FROM AIF_TARGET_APPLICATIONS app
,AIF_TARGET_APPL_DIMENSIONS adim
,AIF_DIMENSIONS dim
WHERE app.APPLICATION_ID = 26
AND adim.APPLICATION_ID = app.APPLICATION_ID
AND dim.DIMENSION_ID = adim.DIMENSION_ID
AND dim.TARGET_DIMENSION_CLASS_NAME IN ('Custom1','Custom2','Custom3','Custom4','Generic')
2013-12-31 22:45:17,997 DEBUG [AIF]:
SELECT SCENARIO "Scenario"
,YEAR "Year"
,PERIOD "Period"
,DATAVIEW "View"
,DATAVALUE "Value"
,ACCOUNT "Account"
,ENTITY "Entity"
,ICP "ICP"
,UD2 "Area"
,UD1 "Tail"
,UD3 "Special"
,UD4 "Facility"
,AMOUNT "DataValue"
FROM AIF_HS_BALANCES
WHERE LOADID = 300
2013-12-31 22:45:18,000 INFO [SimpleAsyncTaskExecutor-9]: ODI Hyperion Financial Management Adapter
2013-12-31 22:45:18,002 INFO [SimpleAsyncTaskExecutor-9]: Load task initialized.
2013-12-31 22:45:18,028 INFO [AIF]: LKM COMM Load Data into HFM - Load Data to HFM - START
2013-12-31 22:45:18,031 DEBUG [AIF]:
INSERT INTO AIF_PROCESS_DETAILS (
PROCESS_ID
,ENTITY_TYPE
,ENTITY_ID
,ENTITY_NAME
,ENTITY_NAME_ORDER
,TARGET_TABLE_NAME
,EXECUTION_START_TIME
,EXECUTION_END_TIME
,RECORDS_PROCESSED
,STATUS
,LAST_UPDATED_BY
,LAST_UPDATE_DATE
) VALUES (
300
,'PROCESS_BAL_EXP_HFM'
,NULL
,'FMTEST2'
,NULL
,NULL
,CURRENT_TIMESTAMP
,NULL
,NULL
,'RUNNING'
,'native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER'
,CURRENT_TIMESTAMP
2013-12-31 22:45:18,034 INFO [SimpleAsyncTaskExecutor-9]: Connecting to Financial Management application [FMTEST2] on [10.150.20.40] using user-name [admin].
2013-12-31 22:45:18,155 INFO [SimpleAsyncTaskExecutor-9]: Connected to Financial Management application.
2013-12-31 22:45:18,157 INFO [SimpleAsyncTaskExecutor-9]: HFM Version: 11.1.2.2.300.
2013-12-31 22:45:18,160 INFO [SimpleAsyncTaskExecutor-9]: Options for the Financial Management load task are:
<Options>
<Option name=LOG_FILE_NAME value=C:\Windows\TEMP\/aif_501_300.log/>
<Option name=IMPORT_MODE value=Replace/>
<Option name=CONSOLIDATE_ONLY value=false/>
<Option name=CONSOLIDATE_PARAMETERS value=""/>
<Option name=LOG_ENABLED value=true/>
<Option name=ACCUMULATE_WITHIN_FILE value=false/>
<Option name=DEBUG_ENABLED value=true/>
<Option name=CONSOLIDATE_AFTER_LOAD value=false/>
<Option name=FILE_CONTAINS_SHARE_DATA value=false/>
</Options>
2013-12-31 22:45:18,168 INFO [SimpleAsyncTaskExecutor-9]: Load Options validated.
2013-12-31 22:45:18,176 ERROR [SimpleAsyncTaskExecutor-9]: Error occurred during load process ORA-00904: "DATAVALUE": invalid identifier
com.hyperion.odi.common.ODIHAppException: ORA-00904: "DATAVALUE": invalid identifier
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:216)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java:175)
at org.python.core.PyObject.__call__(PyObject.java:355)
at org.python.core.PyMethod.__call__(PyMethod.java:215)
at org.python.core.PyMethod.instancemethod___call__(PyMethod.java:221)
at org.python.core.PyMethod.__call__(PyMethod.java:206)
at org.python.core.PyObject.__call__(PyObject.java:397)
at org.python.core.PyObject.__call__(PyObject.java:401)
at org.python.pycode._pyx161.f$0(<string>:98)
at org.python.pycode._pyx161.call_function(<string>)
at org.python.core.PyTableCode.call(PyTableCode.java:165)
at org.python.core.PyCode.call(PyCode.java:18)
at org.python.core.Py.runCode(Py.java:1204)
at org.python.core.Py.exec(Py.java:1248)
at org.python.util.PythonInterpreter.exec(PythonInterpreter.java:172)
at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:322)
at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:170)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2472)
at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:47)
at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:1)
at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2913)
at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2625)
at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:558)
at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:464)
at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:2093)
at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1889)
at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor$2.doAction(StartScenRequestProcessor.java:580)
at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:216)
at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor.doProcessStartScenTask(StartScenRequestProcessor.java:513)
at oracle.odi.runtime.agent.processor.impl.StartScenRequestProcessor$StartScenTask.doExecute(StartScenRequestProcessor.java:1066)
at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:126)
at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:82)
at java.lang.Thread.run(Thread.java:724)
Caused by: java.sql.SQLSyntaxErrorException: ORA-00904: "DATAVALUE": invalid identifier
at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:462)
at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:405)
at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:931)
at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:481)
at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:205)
at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:548)
at oracle.jdbc.driver.T4CStatement.doOall8(T4CStatement.java:202)
at oracle.jdbc.driver.T4CStatement.executeForDescribe(T4CStatement.java:942)
at oracle.jdbc.driver.OracleStatement.executeMaybeDescribe(OracleStatement.java:1283)
at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1441)
at oracle.jdbc.driver.OracleStatement.executeQuery(OracleStatement.java:1690)
at oracle.jdbc.driver.OracleStatementWrapper.executeQuery(OracleStatementWrapper.java:446)
at com.hyperion.odi.hfm.ODIHFMAppWriter.loadData(ODIHFMAppWriter.java:212)
... 38 more
2013-12-31 22:45:18,208 DEBUG [AIF]:
UPDATE AIF_PROCESS_DETAILS
SET STATUS = 'FAILED'
,RECORDS_PROCESSED = CASE
WHEN RECORDS_PROCESSED IS NULL THEN 0
ELSE RECORDS_PROCESSED
END + 0
,EXECUTION_END_TIME = CURRENT_TIMESTAMP
,LAST_UPDATED_BY = CASE
WHEN ('native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER' IS NULL) THEN LAST_UPDATED_BY
ELSE 'native://DN=cn=911,ou=People,dc=css,dc=hyperion,dc=com?USER'
END
,LAST_UPDATE_DATE = CURRENT_TIMESTAMP
WHERE PROCESS_ID = 300
AND ENTITY_TYPE = 'PROCESS_BAL_EXP_HFM'
AND ENTITY_NAME = 'FMTEST2'
2013-12-31 22:45:18,210 FATAL [AIF]: Error in LKM CThe issue is that you are mapping "Data Value" to amount in the Target Application import format:
SELECT SCENARIO "Scenario"
,YEAR "Year"
,PERIOD "Period"
,DATAVIEW "View"
,DATAVALUE "Value"
,ACCOUNT "Account"
,ENTITY "Entity"
,ICP "ICP"
,UD2 "Area"
,UD1 "Tail"
,UD3 "Special"
,UD4 "Facility"
,AMOUNT "DataValue"
FROM AIF_HS_BALANCES
WHERE LOADID = 300
You need to map AMOUNT to "AMOUNT" in the HFM Application. Check that the dimension mapping is correct for the class in the target application and that your import format is going to the proper target dimension(Amount). -
Error while loading data into application
I have created new application successfully.
Loaded LCM backup successfully.
Imported all the rules successfully.
While loading data in plan1 through Essbase Administrator Services. i got this error:
"Parallel dataload enabled: [1] block prepare threads, [1] block write threads.
There were errors, look in D:\Oracle\Middleware\EPMSystem11R1\products\Essbase\eas\client\dataload.err
Database import completed ['hyplive6'.'Plan1']
Output columns prepared: [0]"
This file shows "D:\Oracle\Middleware\EPMSystem11R1\products\Essbase\eas\client\dataload.err"
\\ Member No Period Not Found In Database
"BegBalance" "Jan" "Feb" "Mar" "Apr" "May" "Jun" "Jul" "Aug" "Sep" "Oct" "Nov" "Dec" "No Period" "MTD AVERAGE" "JAN MTD" "FEB MTD" "MAR MTD" "APR MTD" "MAY MTD" "JUN MTD" "JUL MTD" "AUG MTD" "SEP MTD" "OCT MTD" "NOV MTD" "DEC MTD"
I have little clue and new at this hyperion and stuff.
Any help will be appreciated.
RegardsMember No Period Not Found In Database
"BegBalance" "Jan" "Feb" "Mar" "Apr" "May" "Jun" "Jul" "Aug" "Sep" "Oct" "Nov" "Dec" "No Period" "MTD AVERAGE" "JAN MTD" "FEB MTD" "MAR MTD" "APR MTD" "MAY MTD" "JUN MTD" "JUL MTD" "AUG MTD" "SEP MTD" "OCT MTD" "NOV MTD" "DEC MTD"
check your period dimension -
Error while Loading data through .csv file
Hi,
I am getting below date error when loading data through into Olap tables through .csv file.
Data stored in .csv is 20071113121100.
TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
TRANSF_1_1_1> TE_7007 Transformation Evaluation Error [<<Expression Error>> [TO_DATE]: invalid string for converting to Date
... t:TO_DATE(u:'2.00711E+13',u:'YYYYMMDDHH24MISS')]
TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
TRANSF_1_1_1> TT_11132 Transformation [Exp_FILE_CHNL_TYPE] had an error evaluating output column [CREATED_ON_DT_OUT]. Error message is [<<Expression Error>> [TO_DATE]: invalid string for converting to Date
... t:TO_DATE(u:'2.00711E+13',u:'YYYYMMDDHH24MISS')].
TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
TRANSF_1_1_1> TT_11019 There is an error in the port [CREATED_ON_DT_OUT]: The default value for the port is set to: ERROR(<<Expression Error>> [ERROR]: transformation error
... nl:ERROR(u:'transformation error')).
TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
TRANSF_1_1_1> TT_11021 An error occurred moving data from the transformation Exp_FILE_CHNL_TYPE: to the transformation W_CHNL_TYPE_DS.
TRANSF_1_1_1> CMN_1761 Timestamp Event: [Mon Mar 29 15:06:17 2010]
TRANSF_1_1_1> CMN_1086 Exp_FILE_CHNL_TYPE: Number of errors exceeded threshold [1].
Any help is greatly appreciated.
Thanks,
Poojak1) Wrong format, you wont get much support loading OLAP cubes in here I think.
2) Has your CSV file been anywhere near excel by chance the conversion of 20071113121100 to 2.00711E+13 looks very much like what I see when Excel has an invalid number mask / precision specified for a cell.
*** We are using EBS. File was already generated by consultants through SQL query. Getting error while loading though Informatica. Target table is setup as date datatype and source is String(19).
Expression in Informatica is setup as below.
IIF(ISNULL(CREATED_ON_DT), NULL, TO_DATE(CREATED_ON_DT, 'YYYYMMDDHH24MISS'))
Thanks,
Poojak
Maybe you are looking for
-
What does this message mean and how can I avoid it?
This computer's local hostname "joseph-santos-computer-4.local" is already in use on this network. The name has been changed to "joseph-santos-computer-5.local".
-
How Clear the Session After we complete the Execution.. session.invalidate(); is not working ... rs2=st.executeQuery("select deptname from department where collid='"+coll1+"' "); while(rs2.next()) //deptname= rs2.getString(1); data1.add(rs2.getString
-
Function may not be use in SQL - PLS-00231
Here is my code, -- Database 10g Enterprise Edition Release 10.2.0.2.0 declare v1 varchar2(2000); function test_func ( v2 varchar2) return varchar2 is begin return v2 || 'x' || v2; end; begin -- ok v1 := test_func('Hello'); --- Error PLS-00231: Funct
-
Business Content query cant able to see in Bex.
Hi BW Experts, I have installed the query 0PS_C04_Q1000 from the business Content.After installing, I couldnt find the query in Bex analyzer. But in the business Content the query is showing that it is already installed. Can u pls help on this issue
-
Adjusting space around "textwrap' after opening PM7 file in InDesign
Love PageMaker as it suits my basic needs, but printer needs Indesign file. In PageMaker 7, I have numerous photos with captions (67). The photo and caption have been textwrapped together. When I open the file in InDesign 4, both the caption and phot