Import Data (Aggregate Storage) -Load data parallel
I am using the following maxl:
import database 'app'.'db' data connect as 'xxx' identified by 'xxxac' using multiple rules_file 'L_Lo_1','L_Lo_2','L_Lo_3','L_Lo_All' to
load_buffer_block starting with buffer_id 100 on error to 'C:\dataload.err';
I am getting the following error/warning:
Execution Message: 'L_Lo_1','L_Lo_2','L_Lo_3','L_Lo_All' doesn't exist.
Does anyone know how I can fix this? I would appreciate your help.
Thank you.
I'd probably try a few different things...
First, if that is actually your statement copy-pasted, as you seem to be missing the word 'write' in the 'on error' clause? If it's not, please paste the actual syntax you are using. Obviously, I appreciate you are hiding the SQL credentials. I don't suppose your password has any special characters (like a single or double quote) in it that might be creating problems?
Second, I'd try losing the quotes around your rules files. According to the TechRef '_' is an alphabetic character, so there should be no need to quote them (although the TechRef is wrong about MaxL quoting sometimes).
Third, if that doesn't work, try using double quotes (and double-backslash in the error path) everywhere.
I am assuming those load rules all actually exist on the Essbase server and in the database you are loading. I also wonder if you're doing this on a supported platform, because the error path suggests a personal laptop install. :)
Similar Messages
-
Derived Cells in Aggregate storage
<BR>The aggregate storage loads obviously ignore the derived cells. Is there a way to get these ignored records diverted to a log or error file to view and correct the data at the source system !?<BR><BR>Has anybody tried any methods for this !? Any help would be much appreciated.<BR><BR>-Jnt
did yuo ever get this resolved? we are running into the same problem.We have an ASO db which requires YTD calcs and TB Last. We've tried using two separate options (CASE and IF) statements on the YTD, Year and Qtr members (ie MarYTD). Both worked and now concerned about performance. Any suggestions?
-
Loading data using send function in Excel to aggregate storage cube
Hi there
just got version 9.3.1. installed. Can finally load to aggregate storage database using excel essbase send. however, very slow, especially when loading many lines of data. Block storage much much faster. Is there any way you can speed up loading to aggreagate storage data base? Or is this an architectural issue and therefore not much can be done?As far as I know, it is an architectural issue.. Further, I would expect it to slow down even further if you have numerous people writing back simultaneously because, as I understand it, they are throttling the update process on the server side so a single user is actually 'writing' at a time. At least this is better than earlier versions where other users couldn't even do a read when the database was being loaded; I believe that restriction has been lifted as part of the 'trickle-feed' support (although I haven't tested it)..
Tim Tow
Applied OLAP, Inc -
Aggregate Storage Backup level 0 data
<p>When exporting level 0 data from aggregate storage through abatch job you can use a maxL script with "export database[dbs-name] using server report_file [file_name] to data_file[file_name]". But how do I build a report script that exportsall level 0 data so that I can read it back with a load rule?</p><p> </p><p>Can anyone give me an example of such a report script, thatwould be very helpful.</p><p> </p><p>If there is a better way to approach this matter, please let meknow.</p><p> </p><p>Thanks</p><p>/Fredrik</p>
<p>An example from the Sample:Basic database:</p><p> </p><p>// This Report Script was generated by the Essbase QueryDesigner</p><p><SETUP { TabDelimit } { decimal 13 } { IndentGen -5 }<ACCON <SYM <QUOTE <END</p><p><COLUMN("Year")</p><p><ROW("Measures","Product","Market","Scenario")</p><p>// Page Members</p><p>// Column Members</p><p>// Selection rules and output options for dimension: Year</p><p>{OUTMBRNAMES} <Link ((<LEV("Year","Lev0,Year")) AND ( <IDESC("Year")))</p><p>// Row Members</p><p>// Selection rules and output options for dimension:Measures</p><p>{OUTMBRNAMES} <Link ((<LEV("Measures","Lev0,Measures")) AND (<IDESC("Measures")))</p><p>// Selection rules and output options for dimension: Product</p><p>{OUTMBRNAMES} <Link ((<LEV("Product","SKU")) AND ( <IDESC("Product")))</p><p>// Selection rules and output options for dimension: Market</p><p>{OUTMBRNAMES} <Link ((<LEV("Market","Lev0,Market")) AND ( <IDESC("Market")))</p><p>// Selection rules and output options for dimension:Scenario</p><p>{OUTMBRNAMES} <Link ((<LEV("Scenario","Lev0,Scenario")) AND (<IDESC("Scenario")))</p><p>!</p><p>// End of Report</p><p> </p><p>Note that no attempt was made here to eliminate shared membervalues.</p>
-
Clear Partial Data in an Essbase Aggregate storage database
Can anyone let me know how to clear partial data from an Aggregate storage database in Essbase v 11.1.13? We are trying to clear some data in our dbase and don’t want to clear out all the data. I am aware that in Version 11 Essbase it will allow for a partial clear if we write using mdx commands.
Can you please help me on the same by giving us some examples n the same?
Thanks!John, I clearly get the difference between two. What I am asking is in the EAS tool itself for v 11.1.1.3 we have option - right clicking on the DB and getting option "Clear" and in turn sub options like "All Data", "All aggregations" and "Partial Data".
I want to know more on this option part. How will this option know which partial data to be removed or will this option ask us to write some MAXL query for the same"? -
How to filter some illegal rows when SQL Loader import data
I want to import data in a csv file by SQL Loader.
but , I don't want to import some illegal rows
when the column 'name' is null
how can I modify the SQL Loader ctrl file?Hi,
refer this blogpost:
http://gennick.com/allnull.html
thanks,
X A H E E R -
How to export&import data using sql *loader
Hi all,
How to export&import data from sql*loader. Give me the clear steps..
Thanks in AdvanceHi did you already exported data from SQL SERVER? if not using SQL*LOADER you cannot export data. SQL*LOADER is only mean for importing data from flat files(usually text files) into ORACLE tables.
for importing data into oracle tables using sql*loader use below steps
1) create a sql*loader control file.
it looks like as follows
LOAD DATA
INFILE 'sample.dat'
BADFILE 'sample.bad'
DISCARDFILE 'sample.dsc'
APPEND
INTO TABLE emp
TRAILING NULLCOLS
or for sample script of control file search google.
2) at command prompt issue following
$ sqlldr test/test
enter control file=<give control file name which you create earlier>
debug any errors (if occured) -
How to Import data via SQL Loader with characterset UTF16 little endian?
Hello,
I'm importing data from text file into one of my table which contains blob column.
I've specified following in my control file.
-----Control file-------
LOAD DATA
CHARACTERSET UTF16
BYTEORDER LITTLE
INFILE './DataFiles/Data.txt'
BADFILE './Logs/Data.bad'
INTO TABLE temp_blob truncate
FIELDS TERMINATED BY " "
TRAILING NULLCOLS
(GROUP_BLOB,CODE)
Problem:
SQL Loader always importing data via big endian. Is there any method available using which we can convert these data to little endian?
ThanksA new preference has been added to customize the import delimiter in main code line. This should be available as part of future release.
-
Aggregate storage data export failed - Ver 9.3.1
Hi everyone,
We have two production server; Server1 (App/DB/Shared Services Server), Server2 (Anaytics). I am trying to automate couple of our cubes using Win Batch Scripting and MaxL. I can export the data within EAS successfully but when I use the following command in a MaxL Editor, it gives the following error.
Here's the MaxL I used, which I am pretty sure that it is correct.
Failed to open file [S:\Hyperion\AdminServices\deployments\Tomcat\5.0.28\temp\eas62248.tmp]: a system file error occurred. Please see application log for details
[Tue Aug 19 15:47:34 2008]Local/MyAPP/Finance/admin/Error(1270083)
A system error occurred with error number [3]: [The system cannot find the path specified.]
[Tue Aug 19 15:47:34 2008]Local/MyAPP/Finance/admin/Error(1270042)
Aggregate storage data export failed
Does any one have any clue that why am I getting this error.
Thnx in advance!
Regards
FGThis error was due to incorrect SSL settings for our shared services.
-
I have macbook without a firewire port, I have usb 2.0 port, now my os is not working I can not get through apple logo loading , I can not enter safe mode, I can only enter one user mode, how can I backup my data, I have very important data in my hdd
Here is what worked for me:
My usb hub, being usb2, was too fast. I moved the wire to a usb port directory on my pc. That is a usb1 port which is slow enough to run your snyc. -
Segmentation fault error during data load in parallel with multiple rules
Hi,
I'm trying to do sql data load in parallel with multiple rules (4 or 5 rules, maybe), i'm getting a "segmentation fault" error. I tested 3 rules file and it worked fine. we're using Essbase system 9.3.2., with UDB (v8) as the sql data source. ODBC driver is DataDirect 5.2 DB2 Wire Protocol Driver (ARdb222). Please let me know if you have any information on this.
thx.
YHi Thad,
I was wondering, if system is unicode or non unicode that should not matter the amount and currency field . As currencies are defined by SAP and it is in pure English at least a currency code part of 3 Chars.
Could this because of some incosistency of data ??
I would like to know for Which currency had some special characters it it in that particular record ??
Hope that helps.
Regards
Mr Kapadia -
Materialized View with "error in exporting/importing data"
My system is a 10g R2 on AIX (dev). When I impdp a dmp from other box, also 10g R2, in the dump log file, there is an error about the materialized view:ORA-31693: Table data object "BAANDB"."MV_EMPHORA" failed to load/unload and is being skipped due to error:
ORA-02354: error in exporting/importing data
Desc mv_emphora
Name Null? Type
C_RID ROWID
P_RID ROWID
T$CWOC NOT NULL CHAR(6)
T$EMNO NOT NULL CHAR(6)
T$NAMA NOT NULL CHAR(35)
T$EDTE NOT NULL DATE
T$PERI NUMBER
T$QUAN NUMBER
T$YEAR NUMBER
T$RGDT DATEAs i ckecked here and Metalink, I found the info is less to do with the MV? what was the cause?The total lines are 25074. So I used the GREP from the OS to get the lines involved with the MV. Here are:
grep -n -i "TTPPPC235201" impBaanFull.log
5220:ORA-39153: Table "BAANDB"."TTPPPC235201" exists and has been truncated. Data will be loaded but all dependent metadata will be skipped due to table_exists_action of truncate
5845:ORA-39153: Table "BAANDB"."MLOG$_TTPPPC235201" exists and has been truncated. Data will be loaded but all dependent meta data will be skipped due to table_exists_action of truncate
8503:. . imported "BAANDB"."TTPPPC235201" 36.22 MB 107912 rows
8910:. . imported "BAANDB"."MLOG$_TTPPPC235201" 413.0 KB 6848 rows
grep -n -i "TTCCOM001201" impBaanFull.log
4018:ORA-39153: Table "BAANDB"."TTCCOM001201" exists and has been truncated. Data will be loaded but all dependent metadata will be skipped due to table_exists_action of truncate
5844:ORA-39153: Table "BAANDB"."MLOG$_TTCCOM001201" exists and has been truncated. Data will be loaded but all dependent metadata will be skipped due to table_exists_action of truncate
9129:. . imported "BAANDB"."MLOG$_TTCCOM001201" 9.718 KB 38 rows
9136:. . imported "BAANDB"."TTCCOM001201" 85.91 KB 239 rows
grep -n -i "MV_EMPHORA" impBaanFull.log
8469:ORA-39153: Table "BAANDB"."MV_EMPHORA" exists and has been truncated. Data will be loaded but all dependent metadata will be skipped due to table_exists_action of truncate
8558:ORA-31693: Table data object "BAANDB"."MV_EMPHORA" failed to load/unload and is being skipped due to error:
8560:ORA-12081: update operation not allowed on table "BAANDB"."MV_EMPHORA"
25066:ORA-31684: Object type MATERIALIZED_VIEW:"BAANDB"."MV_EMPHORA" already exists
25072: BEGIN dbms_refresh.make('"BAANDB"."MV_EMPHORA"',list=>null,next_date=>null,interval=>null,implicit_destroy=>TRUE,lax=>
FALSE,job=>44,rollback_seg=>NULL,push_deferred_rpc=>TRUE,refresh_after_errors=>FALSE,purge_option => 1,parallelism => 0,heap_size => 0);
25073:dbms_refresh.add(name=>'"BAANDB"."MV_EMPHORA"',list=>'"BAANDB"."MV_EMPHORA"',siteid=>0,export_db=>'BAAN'); END;the number in front of each line is the line number of the import log.
Here is my syntax of import pmup:impdp user/pw SCHEMAS=baandb DIRECTORY=baanbk_data_pump DUMPFILE=impBaanAll.dmp LOGFILE=impBaanAll.log TABLE_EXISTS_ACTION=TRUNCATEYes I can create the MV manually and I have no problem to refresh manually it after the inmport. -
Error importing data in 10.2.0.4
We have a production 10.2.0.4 database.
The database had crashed and I needed to export data and import data back in using data pump.
However, I am receiveing the follwing error while importing:
RA-31693: Table data object "R"."RIC_TRANSACTION" failed to load/unload and is being skipped due to error:
ORA-02354: error in exporting/importing data
ORA-39776: fatal Direct Path API error loading table "R"."RIC_TRANSACTION"
ORA-00600: internal error code, arguments: [klaprs_12], [], [], [], [], [], [], []
Has anyone faced this?Khushboo Srivastava wrote:
expdp [email protected] dumpfile=TABLES_032910.dmp logfile=FULL_032910.log
directory=data1 job_name=FILLRIC_EXPORT schemas=RUSER,RUSER2 parallel=4
impdp [email protected] directory=data1 dumpfile=TABLES_032910.dmp logfile=IC_IMPORT.log job_name=IMPORT3 schemas=RUSER,RUSER2
import command generates the error:
RA-31693: Table data object "RUSER"."RIC_TRANSACTION" failed to load/unload and is being skipped due to error:
ORA-02354: error in exporting/importing data
ORA-39776: fatal Direct Path API error loading table "RUSER"."RIC_TRANSACTION"
ORA-00600: internal error code, arguments: [klaprs_12], [], [], [], [], [], [], []ora-0600 is an otherwise undefined error. You are probably going to have to open an SR with Oracle Support. At the very least, go to MetaLink and use the ora-0600 lookup tool. -
Incremental Load in Aggregate Storage
<p>Hi,</p><p> </p><p>From what I understand, Aggregate Storage (ASO) clears all dataif a new member gets added to the outline.</p><p>This is unlike Block Storage (BSO) where we can restructure thecube if new member is added to the outline.</p><p> </p><p>We need to load data daily into an ASO cube and the cubecontains 5 yrs of data. We may get a new member in the customerdimension daily. Is there a way we can retain (restructure)existing data when updating the customer dimension and then add thenew data? Otherwise, we will have to rebuild the cube daily andtherefore reload 5 yrs of data (about 600 million recs) on a dailybasis.</p><p> </p><p>Is there a better way of doing this in ASO?</p><p> </p><p>Any help would be appreciated.</p><p> </p><p>Thanks</p><p>--- suren_v</p>
Good information Steve. Is the System 9 Essbase DB Admin Guide available online? I could not find it here: <a target=_blank class=ftalternatingbarlinklarge href="http://dev.hyperion.com/resource_library/technical_documentation">http://dev.hyperion.com/resour...echnical_documentation</a><BR><BR>(I recently attended the v7 class in Dallas and it was excellent!)<BR><BR><BR><blockquote>quote:<br><hr><i>Originally posted by: <b>scran4d</b></i><BR>Suren:<BR><BR><BR><BR>In the version 7 releases of Essbase ASO, there is not a way to hold on to the data if a member is added to the outline; data must be reloaded each time.<BR><BR><BR><BR>This is changed in Hyperion's latest System 9 release, however.<hr></blockquote><BR><BR>
-
Essbase data error importing data back to cube.
Hi
We have some text measures in cube.We loaded data using smart view.Once the data is loaded.we did export data from cube and when i import data back to the same cube i am getting the following error message below.
Parallel data load enabled: [1] block prepare threads, [1] block write threads.
Member [Sample text Measure] is a text measure. Only text values can be loaded to the Smart List in a text measure. [1004] Records Completed
Unexpected Essbase error 1003073.
I have done some analysis on this error message but could not find any solution.
please suggest me where it is going wrong.
Thanks in advance
Kris
Edited by: user9363364 on Apr 19, 2010 5:19 PM
Edited by: user9363364 on Apr 20, 2010 3:38 PM
Edited by: user9363364 on Apr 21, 2010 1:55 PMHi,
I don't know if I'm directing you correctly as I only loaded data, but never imported the exported text data. However, from the 11.1.1.3 dbag, below is found on P#193:
*"100-10" "New York" "Cust Index" #Txt:"Highly Satisfied"*
This shows that- Perhaps, we may've to prefix *#Txt:* before the double-quoted Text value.
Wanna give a try?
Maybe you are looking for
-
Problem in oracle net services
Dear People. we are trying to create a physical standby database using DGMGRL CLI .After the configuration is made and while enabling the configuration is made i get this error ORA-16607: one or more databases have failed.When i checked the alert log
-
Question about different file name send to different receivers?
Hi, Gurus, my senario: in ftp server there are many files, which are either .doc or .xls. All .doc must be sent to e-mail address departmentA(a)companyname.com. All .xls must be sent to departmentB(a)companyname.com. I use Mail Package and keep atta
-
As above. This is really starting to **** me off.
-
I'm trying to learn to play Texas Holdem and have tried several different free sites but have not been able to complete a query. I type in fulltiltpoker.com or pokerstars.net and nothing happens. Does the Mac not support these sites?
-
Interactive movies within interactive movies
I have been creating a gallery for a client that changes preview images in response to thumbnail images being clicked, I have the basics working but now my client wants further interaction in some of the previews that include using buttons on the pre