Query about TRAILING NULLCOLS usage
Hi,
I'm using SQL Loader for loading flat files to Tables.
Here is my datafile
empid|name|dept|salary|grade|
1|shekhar|projects|1000|C|
2|ravi|projects|800|D|
3|rob|projects|1200||
4|Mathew|projects|1400||
my ctl file
LOAD DATA
INFILE emp.txt "str '|\r\n'"
APPEND INTO TABLE EMP
FIELDS TERMINATED BY '|'
EMP_ID
, NAME
, DEPT
, SALARY
, GRADE
Its giving error message
Column not found before end of logical record (use TRAILING NULLCOLS)
I think TRAILING NULLCOLS is used to put null for the columns which are not found. But in my case the grade column value is empty for the empid 3 & 4
If I'm wrong please explain me the usage of TRAILING NULLCOLS.
Thanks
the record delimiter specified in the INFILE clause is not considered part of the record. for record 3 and 4, SQL Loader is parsing:
3|rob|projects|1200|
4|Mathew|projects|1400|
There are only 4 field values for these records because the last '|' is stripped off. Try changing the INFILE line to
INFILE emp.txt "str '\r\n'"
This should preserve the ending '|'
Of course. the other option is to us TRAILING NULLCOLS in the control file.
Similar Messages
-
Hi,
I have a Oracle 9.2 DB with 2 cpu's,when i took a statspack report i found that CPU Time is very hight i,e more that 81%....
one of the query is taking near about 51% CPU Usage...
i.e
SELECT /*+ INDEX (OM_COURSE_FEE OCF_CFC_CODE) */DISTINCT CF_COURSE_CODE FROM OM_COURSE_FEE,OT_STUDENT_FEE_COL_HEAD,OT_STUDENT_FEE_COL_DETL WHERE SFCH_SYS_ID = SFCD_SFCH_SYS_ID AND
SFCD_FEE_TYPE_CODE = CF_TYPE_CODE AND CF_COURSE_CODE IN ( 'PE1','PE2','CCT','CPT' ) AND SFCH_TXN_CODE = :b1 AND SFCD_SUBSCRBE_JOURNAL_YN IN ( 'T','R','1','C' ) AND SFCH_APPR_UID IS NOT NULL
AND SFCH_APPR_DT IS NOT NULL AND SFCH_NO = :b2 AND NOT EXISTS (SELECT 'X' FROM OM_STUDENT_HEAD WHERE STUD_SRN = SFCH_STUD_SRN_TEMP_NO AND NVL(STUD_CLO_STATUS,0) != 1 AND
NVL(STUD_REG_STATUS,0) != 23 AND STUD_COURSE_CODE != 'CCT' AND CF_COURSE_CODE= STUD_COURSE_CODE ) ORDER BY 1 DESCExplain Plan for......
SQL> select * from table(dbms_xplan.display());
PLAN_TABLE_OUTPUT
| Id | Operation | Name
| Rows | Bytes | Cost | Pstart| Pstop |
| 0 | SELECT STATEMENT |
PLAN_TABLE_OUTPUT
| 1 | 69 | 45 | | |
| 1 | SORT UNIQUE |
| 1 | 69 | 27 | | |
| 2 | TABLE ACCESS BY GLOBAL INDEX ROWID | OT_STUDENT_FEE_COL_DETL
| 1 | 12 | 2 | ROWID | ROW L |
| 3 | NESTED LOOPS |
| 1 | 69 | 9 | | |
PLAN_TABLE_OUTPUT
| 4 | NESTED LOOPS |
| 1 | 57 | 7 | | |
| 5 | TABLE ACCESS BY GLOBAL INDEX ROWID | OT_STUDENT_FEE_COL_HEAD
| 1 | 48 | 5 | ROWID | ROW L |
| 6 | INDEX SKIP SCAN | OT_STUDENT_FEE_COL_HEAD_UK0
1 | 1 | | 4 | | |
| 7 | INLIST ITERATOR |
| | | | | |
PLAN_TABLE_OUTPUT
| 8 | TABLE ACCESS BY INDEX ROWID | OM_COURSE_FEE
| 1 | 9 | 2 | | |
| 9 | INDEX RANGE SCAN | OCF_CFC_CODE
| 1 | | 1 | | |
| 10 | FILTER |
| | | | | |
| 11 | TABLE ACCESS BY GLOBAL INDEX ROWID| OM_STUDENT_HEAD
PLAN_TABLE_OUTPUT
| 1 | 21 | 4 | ROWID | ROW L |
| 12 | INDEX RANGE SCAN | IDM_STUD_SRN_COURSE
| 1 | | 3 | | |
| 13 | INDEX RANGE SCAN | IDM_SFCD_FEE_TYPE_CODE
| 34600 | | 1 | | |
PLAN_TABLE_OUTPUT
Note: cpu costing is off, PLAN_TABLE' is old version
21 rows selected.
SQL>Statspack report
DB Name DB Id Instance Inst Num Release Cluster Host
ai 1372079993 ai11 1 9.2.0.6.0 YES ai1
Snap Id Snap Time Sessions Curs/Sess Comment
Begin Snap: 175 12-Dec-08 13:21:33 ####### .0
End Snap: 176 12-Dec-08 13:56:09 ####### .0
Elapsed: 34.60 (mins)
Cache Sizes (end)
~~~~~~~~~~~~~~~~~
Buffer Cache: 3,264M Std Block Size: 8K
Shared Pool Size: 608M Log Buffer: 977K
Load Profile
~~~~~~~~~~~~ Per Second Per Transaction
Redo size: 5,727.62 21,658.54
Logical reads: 16,484.89 62,336.32
Block changes: 32.49 122.88
Physical reads: 200.46 758.03
Physical writes: 5.08 19.23
User calls: 97.43 368.44
Parses: 11.66 44.11
Hard parses: 0.39 1.48
Sorts: 3.22 12.19
Logons: 0.02 0.06
Executes: 36.70 138.77
Transactions: 0.26
% Blocks changed per Read: 0.20 Recursive Call %: 28.65
Rollback per transaction %: 20.95 Rows per Sort: 131.16
Instance Efficiency Percentages (Target 100%)
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Buffer Nowait %: 100.00 Redo NoWait %: 100.00
Buffer Hit %: 98.79 In-memory Sort %: 99.99
Library Hit %: 98.92 Soft Parse %: 96.65
Execute to Parse %: 68.21 Latch Hit %: 99.98
Parse CPU to Parse Elapsd %: 60.50 % Non-Parse CPU: 99.48
Shared Pool Statistics Begin End
Memory Usage %: 90.06 89.79
% SQL with executions>1: 72.46 72.46
% Memory for SQL w/exec>1: 69.42 69.51
Top 5 Timed Events
~~~~~~~~~~~~~~~~~~ % Total
Event Waits Time (s) Ela Time
CPU time 3,337 81.43
db file sequential read 60,550 300 7.32
global cache cr request 130,852 177 4.33
db file scattered read 72,915 101 2.46
db file parallel read 3,384 75 1.84
Cluster Statistics for DB: ai Instance: ai11 Snaps: 175 -176
Global Cache Service - Workload Characteristics
Ave global cache get time (ms): 1.3
Ave global cache convert time (ms): 2.1
Ave build time for CR block (ms): 0.1
Ave flush time for CR block (ms): 0.3
Ave send time for CR block (ms): 0.3
Ave time to process CR block request (ms): 0.7
Ave receive time for CR block (ms): 4.9
Ave pin time for current block (ms): 0.0
Ave flush time for current block (ms): 0.0
Ave send time for current block (ms): 0.3
Ave time to process current block request (ms): 0.3
Ave receive time for current block (ms): 2.8
Global cache hit ratio: 1.5
Ratio of current block defers: 0.0
% of messages sent for buffer gets: 1.4
% of remote buffer gets: 0.1
Ratio of I/O for coherence: 1.1
Ratio of local vs remote work: 9.7
Ratio of fusion vs physical writes: 0.1
Global Enqueue Service Statistics
Ave global lock get time (ms): 0.8
Ave global lock convert time (ms): 0.0
Ratio of global lock gets vs global lock releases: 1.1
GCS and GES Messaging statistics
Ave message sent queue time (ms): 0.4
Ave message sent queue time on ksxp (ms): 2.7
Ave message received queue time (ms): 0.2
Ave GCS message process time (ms): 0.1
Ave GES message process time (ms): 0.1
% of direct sent messages: 19.4
% of indirect sent messages: 43.5
% of flow controlled messages: 37.1
GES Statistics for DB: ai Instance: ai11 Snaps: 175 -176
Statistic Total per Second per Trans
dynamically allocated gcs resourc 0 0.0 0.0
dynamically allocated gcs shadows 0 0.0 0.0
flow control messages received 0 0.0 0.0
flow control messages sent 0 0.0 0.0
gcs ast xid 0 0.0 0.0
gcs blocked converts 1,231 0.6 2.2
gcs blocked cr converts 2,432 1.2 4.4
gcs compatible basts 0 0.0 0.0
gcs compatible cr basts (global) 658 0.3 1.2
gcs compatible cr basts (local) 57,822 27.9 105.3
gcs cr basts to PIs 0 0.0 0.0
gcs cr serve without current lock 0 0.0 0.0
gcs error msgs 0 0.0 0.0
gcs flush pi msgs 821 0.4 1.5
gcs forward cr to pinged instance 0 0.0 0.0
gcs immediate (compatible) conver 448 0.2 0.8
gcs immediate (null) converts 1,114 0.5 2.0
gcs immediate cr (compatible) con 42,094 20.3 76.7
gcs immediate cr (null) converts 396,284 190.9 721.8
gcs msgs process time(ms) 42,220 20.3 76.9
gcs msgs received 545,133 262.6 993.0
gcs out-of-order msgs 5 0.0 0.0
gcs pings refused 1 0.0 0.0
gcs queued converts 0 0.0 0.0
gcs recovery claim msgs 0 0.0 0.0
gcs refuse xid 0 0.0 0.0
gcs retry convert request 0 0.0 0.0
gcs side channel msgs actual 2,397 1.2 4.4
gcs side channel msgs logical 232,024 111.8 422.6
gcs write notification msgs 15 0.0 0.0
gcs write request msgs 278 0.1 0.5
gcs writes refused 1 0.0 0.0
ges msgs process time(ms) 4,873 2.3 8.9
ges msgs received 39,769 19.2 72.4
global posts dropped 0 0.0 0.0
global posts queue time 0 0.0 0.0
global posts queued 0 0.0 0.0
global posts requested 0 0.0 0.0
global posts sent 0 0.0 0.0
implicit batch messages received 39,098 18.8 71.2
implicit batch messages sent 33,386 16.1 60.8
lmd msg send time(ms) 635 0.3 1.2
lms(s) msg send time(ms) 2 0.0 0.0
messages flow controlled 196,546 94.7 358.0
messages received actual 182,783 88.0 332.9
messages received logical 584,848 281.7 1,065.3
messages sent directly 102,657 49.4 187.0
messages sent indirectly 230,329 110.9 419.5
msgs causing lmd to send msgs 9,169 4.4 16.7
msgs causing lms(s) to send msgs 3,347 1.6 6.1
msgs received queue time (ms) 142,759 68.8 260.0
msgs received queued 584,818 281.7 1,065.2
msgs sent queue time (ms) 99,300 47.8 180.9
msgs sent queue time on ksxp (ms) 608,239 293.0 1,107.9
msgs sent queued 230,391 111.0 419.7
msgs sent queued on ksxp 229,013 110.3 417.1
GES Statistics for DB: ai Instance: ai11 Snaps: 175 -176
Statistic Total per Second per Trans
process batch messages received 65,059 31.3 118.5
process batch messages sent 50,959 24.5 92.8
Wait Events for DB: ai Instance: ai11 Snaps: 175 -176
-> s - second
-> cs - centisecond - 100th of a second
-> ms - millisecond - 1000th of a second
-> us - microsecond - 1000000th of a second
-> ordered by wait time desc, waits desc (idle events last)
Avg
Total Wait wait Waits
Event Waits Timeouts Time (s) (ms) /txn
db file sequential read 60,550 0 300 5 110.3
global cache cr request 130,852 106 177 1 238.3
db file scattered read 72,915 0 101 1 132.8
db file parallel read 3,384 0 75 22 6.2
latch free 7,253 1,587 52 7 13.2
enqueue 44,947 0 16 0 81.9
log file parallel write 2,140 0 6 3 3.9
db file parallel write 341 0 5 14 0.6
global cache open x 1,134 3 4 4 2.1
CGS wait for IPC msg 166,993 164,390 4 0 304.2
library cache lock 3,169 0 3 1 5.8
log file sync 494 0 3 5 0.9
row cache lock 702 0 3 4 1.3
DFS lock handle 6,900 0 2 0 12.6
control file parallel write 689 0 2 3 1.3
control file sequential read 2,785 0 2 1 5.1
wait for master scn 687 0 2 2 1.3
global cache null to x 699 0 2 2 1.3
global cache s to x 778 5 1 2 1.4
direct path write 148 0 0 3 0.3
SQL*Net more data to client 3,621 0 0 0 6.6
global cache open s 149 0 0 2 0.3
library cache pin 78 0 0 2 0.1
ksxr poll remote instances 3,536 2,422 0 0 6.4
LGWR wait for redo copy 12 6 0 9 0.0
buffer busy waits 23 0 0 5 0.0
direct path read 9 0 0 10 0.0
buffer busy global CR 5 0 0 17 0.0
SQL*Net break/reset to clien 172 0 0 0 0.3
global cache quiesce wait 4 1 0 7 0.0
KJC: Wait for msg sends to c 86 0 0 0 0.2
BFILE get length 67 0 0 0 0.1
global cache null to s 9 0 0 1 0.0
BFILE open 6 0 0 0 0.0
BFILE read 60 0 0 0 0.1
BFILE internal seek 60 0 0 0 0.1
BFILE closure 6 0 0 0 0.0
cr request retry 4 4 0 0 0.0
SQL*Net message from client 201,907 0 180,583 894 367.8
gcs remote message 171,236 43,749 3,975 23 311.9
ges remote message 79,324 40,722 2,017 25 144.5
SQL*Net more data from clien 447 0 9 21 0.8
SQL*Net message to client 201,901 0 0 0 367.8
Background Wait Events for DB: ai Instance: ai11 Snaps: 175 -176
-> ordered by wait time desc, waits desc (idle events last)
Avg
Total Wait wait Waits
Event Waits Timeouts Time (s) (ms) /txn
enqueue 44,666 0 16 0 81.4
latch free 4,895 108 6 1 8.9
log file parallel write 2,140 0 6 3 3.9
db file parallel write 341 0 5 14 0.6
CGS wait for IPC msg 166,969 164,366 4 0 304.1
DFS lock handle 6,900 0 2 0 12.6
control file parallel write 689 0 2 3 1.3
control file sequential read 2,733 0 2 1 5.0
wait for master scn 687 0 2 2 1.3
ksxr poll remote instances 3,528 2,419 0 0 6.4
LGWR wait for redo copy 12 6 0 9 0.0
db file sequential read 7 0 0 9 0.0
global cache null to x 26 0 0 2 0.0
global cache open x 16 0 0 1 0.0
global cache cr request 1 0 0 0 0.0
rdbms ipc message 28,636 20,600 16,937 591 52.2
gcs remote message 171,254 43,746 3,974 23 311.9
pmon timer 708 686 2,022 2856 1.3
ges remote message 79,305 40,719 2,017 25 144.5
smon timer 125 0 1,972 15776 0.2
SQL ordered by Gets for DB: ai Instance: ai11 Snaps: 175 -176
-> End Buffer Gets Threshold: 10000
-> Note that resources reported for PL/SQL includes the resources used by
all SQL statements called within the PL/SQL code. As individual SQL
statements are also reported, it is possible and valid for the summed
total % to exceed 100
CPU Elapsd
Buffer Gets Executions Gets per Exec %Total Time (s) Time (s) Hash Value
17,503,305 84 208,372.7 51.1 676.25 1218.80 17574787
SELECT /*+ INDEX (OM_COURSE_FEE OCF_CFC_CODE) */DISTINCT CF_COU
RSE_CODE FROM OM_COURSE_FEE,OT_STUDENT_FEE_COL_HEAD,OT_STUDENT
FEECOL_DETL WHERE SFCH_SYS_ID = SFCD_SFCH_SYS_ID AND SFCD_FE
E_TYPE_CODE = CF_TYPE_CODE AND CF_COURSE_CODE IN ( 'PE1','PE2',
'CCT','CPT' ) AND SFCH_TXN_CODE = :b1 AND SFCD_SUBSCRBE_JOURNAuser00726 wrote:
somw of the changes that have been done....
cai11_lmd0_15771.trc.
Thu Oct 2 13:35:48 2008
CKPT: Begin resize of buffer pool 3 (DEFAULT for block size 8192)
CKPT: Current size = 2512 MB, Target size = 3072 MB
CKPT: Resize completed for buffer pool DEFAULT for blocksize 8192
Thu Oct 2 13:35:50 2008
ALTER SYSTEM SET db_cache_size='3072M' SCOPE=BOTH SID='icai11';
Thu Oct 2 14:04:34 2008
ALTER SYSTEM SET sga_max_size='5244772679680' SCOPE=SPFILE SID='*';
ALTER SYSTEM SET sga_max_size='5244772679680' SCOPE=SPFILE SID='*';
Thu Oct 2 15:24:14 2008
CKPT: Begin resize of buffer pool 3 (DEFAULT for block size 8192)
CKPT: Current size = 3072 MB, Target size = 2512 MB
CKPT: Resize completed for buffer pool DEFAULT for blocksize 8192
Thu Oct 2 15:24:20 2008
ALTER SYSTEM SET db_cache_size='2512M' SCOPE=BOTH SID='icai11';
Thu Oct 2 15:32:33 2008
CKPT: Begin resize of buffer pool 3 (DEFAULT for block size 8192)
CKPT: Current size = 2512 MB, Target size = 3072 MB
CKPT: Resize completed for buffer pool DEFAULT for blocksize 8192
Thu Oct 2 15:32:34 2008
ALTER SYSTEM SET db_cache_size='3072M' SCOPE=BOTH SID='icai11';
Thu Oct 2 15:36:46 2008
ALTER SYSTEM SET shared_pool_size='640M' SCOPE=BOTH SID='icai11';
Thu Oct 2 16:33:52 2008
CKPT: Begin resize of buffer pool 3 (DEFAULT for block size 8192)
CKPT: Current size = 3072 MB, Target size = 2512 MB
CKPT: Resize completed for buffer pool DEFAULT for blocksize 8192
Thu Oct 2 16:33:56 2008
ALTER SYSTEM SET db_cache_size='2512M' SCOPE=BOTH SID='icai11';
Thu Oct 2 16:39:30 2008
ALTER SYSTEM SET pga_aggregate_target='750M' SCOPE=BOTH SID='icai11';Just to make certain that I am not missing anything, if the above you set (all scaled to GB just for the sake of comparison):
sga_max_size=4885GB
db_cache_size=3GB
shared_pool_size=0.625GB
pga_aggregate_target=0.733GB
The SQL statement is forcing the use of a specific index, is this the best index?:
SELECT /*+ INDEX (OM_COURSE_FEE OCF_CFC_CODE) */DISTINCT CF_COURSE_CODE FROM OM_COURSE_FEE,OT_STUDENT_FEE_COL_HEAD,OT_STUDENT_FEE_COL_DETL WHERE SFCH_SYS_ID = SFCD_SFCH_SYS_ID AND
SFCD_FEE_TYPE_CODE = CF_TYPE_CODE AND CF_COURSE_CODE IN ( 'PE1','PE2','CCT','CPT' ) AND SFCH_TXN_CODE = :b1 AND SFCD_SUBSCRBE_JOURNAL_YN IN ( 'T','R','1','C' ) AND SFCH_APPR_UID IS NOT NULL
AND SFCH_APPR_DT IS NOT NULL AND SFCH_NO = :b2 AND NOT EXISTS (SELECT 'X' FROM OM_STUDENT_HEAD WHERE STUD_SRN = SFCH_STUD_SRN_TEMP_NO AND NVL(STUD_CLO_STATUS,0) != 1 AND
NVL(STUD_REG_STATUS,0) != 23 AND STUD_COURSE_CODE != 'CCT' AND CF_COURSE_CODE= STUD_COURSE_CODE ) ORDER BY 1 DESC
Unfortunately, explain plans, even with dbms_xplan.display, may not show the actual execution plan - this is more of a problem when the SQL statement includes bind variables (capturing a 10046 trace at level 8 or 12 will help). With the information provided, it looks like the problem is with the number of logical reads performed: 17,503,305 in 84 executions = 208,373 logical reads per execution. Something is causing the SQL statement to execute inefficiently, possibly a join problem between tables, possibly the forced use of an index.
From one of my previous posts related to this same SQL statement:
SELECT /*+ INDEX (OM_COURSE_FEE OCF_CFC_CODE) */ DISTINCT
CF_COURSE_CODE
FROM
OM_COURSE_FEE,
OT_STUDENT_FEE_COL_HEAD,
OT_STUDENT_FEE_COL_DETL
WHERE
SFCH_SYS_ID = SFCD_SFCH_SYS_ID
AND SFCD_FEE_TYPE_CODE = CF_TYPE_CODE
AND CF_COURSE_CODE IN ( 'PE1','PE2','CCT','CPT' )
AND SFCH_TXN_CODE = :b1
AND SFCD_SUBSCRBE_JOURNAL_YN IN ( 'T','R','1','C' )
AND SFCH_APPR_UID IS NOT NULL
AND SFCH_APPR_DT IS NOT NULL
AND SFCH_NO = :b2
AND NOT EXISTS (
SELECT
'X'
FROM
OM_STUDENT_HEAD
WHERE
STUD_SRN = SFCH_STUD_SRN_TEMP_NO
AND NVL(STUD_CLO_STATUS,0) != 1
AND NVL(STUD_REG_STATUS,0) != 23
AND STUD_COURSE_CODE != 'CCT'
AND CF_COURSE_CODE = STUD_COURSE_CODE)
ORDER BY
1 DESC
| Id | Operation | Name | Rows | Bytes | Cost | Pstart| Pstop |
| 0 | SELECT STATEMENT 1 69 34
| 1 | SORT UNIQUE 1 69 22
| 2 | TABLE ACCESS BY GLOBAL INDEX ROWID | OT_STUDENT_FEE_COL_DETL | 1 | 12 | 2 | ROWID | ROW L |
| 3 | NESTED LOOPS 1 69 9
| 4 | NESTED LOOPS 1 57 7
| 5 | TABLE ACCESS BY GLOBAL INDEX ROWID | OT_STUDENT_FEE_COL_HEAD | 1 | 48 | 5 | ROWID | ROW L |
| 6 | INDEX SKIP SCAN OT_STUDENT_FEE_COL_HEAD_UK0| 1 | 1 | 4 |
| 7 | INLIST ITERATOR |
| 8 | TABLE ACCESS BY INDEX ROWID | OM_COURSE_FEE | 1 | 9 | 2 | | |
| 9 | INDEX RANGE SCAN | OCF_CFC_CODE | 1 | | 1 | | |
| 10 | FILTER
| 11 | TABLE ACCESS BY GLOBAL INDEX ROWID| OM_STUDENT_HEAD | 1 | 21 | 4 | ROWID | ROW L |
| 12 | INDEX RANGE SCAN | IDM_STUD_SRN_COURSE | 1 | | 3 | | |
| 13 | INDEX RANGE SCAN | IDM_SFCD_FEE_TYPE_CODE | 34600 | | 1 | | |It appears, based just on the SQL statement, that there is no direct relationship between OM_COURSE_FEE and OT_STUDENT_FEE_COL_HEAD, yet the plan above indicates that the two tables are being joined together, likely as a result of the index hint. There is the possibility of additional predicates being generated by Oracle which will make this possible without introducing a Cartesian join, but those predicates are not displayed with an explain plan (they will appear with a DBMS_XPLAN). I may not be remembering correctly, but if the optimizer goal is set to first rows, a Cartesian join might appear in a plan as a nested loops join - Jonathan will know for certain if that is the case.
Have you investigated the above as a possible cause? If you know that there is a problem with this one SQL statement, a 10046 trace at level 8 or 12 is much more helpful than a Statspack report.
Charles Hooper
IT Manager/Oracle DBA
K&M Machine-Fabricating, Inc. -
Hi,
My 20G Zen Sleek Photo has just recently transformed itself into a useless brick.
No matter what I try and do, it always ends up freezing after about 5 minutes usage. When you force a reset it does unlock but when it comes back on it sticks on the Creative "copyright" start up screen.
I've reformatted and re-installed the firmware several times but with the same result.
font face="comic sans ms,sans-serif">
The only time it totally unlocks is when you either:
() let the battery fully drain and then recharge it or
(2) you disconnect the battery lead and then press the On/Off switch for several minutes.
I've looked everywhere for a solution/reason but nothing. You find plenty of people with the same problem but no answers. Surely someone knows how to fix the problem
Is it a faulty battery (unlikely because mine charges up fine and holds it charge well)
Is it the hard dri've (very unlikely since the player can start up correctly [given the above recovery procedure] and play tracks within the 5-0 minutes you have available)
Is it the firmware (I doubt it, I can't see what the firmware's got to do with a delayed lock up, having worked fine up to that point)
Is it the motherboard (maybe....., to my mind it points to a component failure that allows an electric charge to build up and not dissipate, the only way to discharge it is to short the battery pins [with, I hasten to add, the battery removed!])
font face="comic sans ms,sans-serif">
Has anyone within the Creative community successfully solved this problem Has anyone received a decent answer from Creative
I find it staggering that the guys at Creative have washed their hands of this problem. They, after all, are the most likely people to identify the problem - only they have a schematic of the motherboard that would allow them to investigate and locate the problem.
For the record the firmware I'm using is v .0.0.
Any guidance would be much appreciated,
VictorIs there really no one with an answer to this problem or have I overlooked a previous post that explained the fault.
BR
Victor -
How to ensure File Read Adapter handles like SQLLOADER's TRAILING NULLCOLS
I have to read a CSV file which has a specific format but sometimes the trailing columns values can be missing and i would like to handle it in such way that they values are treated as null. This is similar to the SQLLoader's TRAILING NULLCOLS clause.
Currently if my data is as shown below:
As-Of Date,As-Of-Time,Bank ID,Bank Name,State,Acct No,Acct Type,Acct Name,Currency,BAI Type Code,Tran Desc,Debit Amt,Credit Amt,0 Day Flt Amt,1 Day Flt Amt,2+ Day Flt Amt,Customer Ref No,Value Date,Location,Bank Reference,Tran Status,Descriptive Text,Descriptive Text1,Descriptive Text2,Descriptive Text3,Descriptive Text4,Descriptive Text5,Descriptive Text6
20061031,23:59:00,121000248,"WELLS FARGO BANK, N.A.",CA,4121235097,COMMERCIAL DDA,Silicon Image - A/P,USD,475,CHECK PAID,55.86,0,0,0,0,51689,10/31/2006,,IA000313659233,POSTED,,,,,,,
20061031,23:59:00,121000248,"WELLS FARGO BANK, N.A.",CA,4121235097,COMMERCIAL DDA,Silicon Image - A/P,USD,475,CHECK PAID,1377.57,0,0,0,0,51685,10/31/2006,,IA000210166161,POSTED
20061031,23:59:00,121000248,"WELLS FARGO BANK, N.A.",CA,4121235097,COMMERCIAL DDA,Silicon Image - A/P,USD,475,CHECK PAID,1435,0,0,0,0,51621,10/31/2006,,IA000627084628,POSTED
It reads the first row properly , but when in encounters the second row POSTED column , it tries to find the terminated column and since it is not present it reads data from the next line until it finds a comma.
This is not the way i would like to handle it.
We get around 1000-2000 records in that CSV from the Bank and they cant change the way they are sending the file. So , please help in resolving this issue.
Thanks
SridharHi thanks for the reply
well i've been having a play around but haven't got it to work yet,
Here's what i have so far
private void jButton2MouseClicked(java.awt.event.MouseEvent evt) {
BufferedReader sb = new BufferedReader();
sb.append(""+jTextField1.setText()+",");
sb.append(""+jFormattedTextField1.setText()+"\n");
sb.append(""+jTextField2.setText()+",");
sb.append(""+jFormattedTextField2.setText()+"\n");
sb.append(""+jTextField3.setText()+",");
sb.append(""+jFormattedTextField3.setText()+"\n");
sb.append(""+jTextField4.setText()+",");
sb.append(""+jFormattedTextField4.setText()+"\n");
sb.append(""+jTextField5.setText()+",");
sb.append(""+jFormattedTextField5.setText()+"\n");
ReadFromFile(sb.toString());
private void ReadFromFile(String s) {
try {
BufferedReader in = new BufferedReader(new FileReader("/users/data.txt"));
String str;
while ((str = in.readLine()) != null) {
Process(str);
in.close();
} catch (IOException e) {
}// TODO add your handling code here:
}im thinking the problem is the parts that say sb.append but i don't know what to in it's place? -
I DO NOT want to send anonymous reports to Mozilla about my computer usage and so on. How do I get rid of the question that comes up every time I open Fiorfox up?
See:
* https://support.mozilla.org/en-US/kb/Options%20window%20-%20Advanced%20panel
* go to Firefox/Tools-Option-Advance and uncheck submit performance data -
I have a recently acquired an iPhone 4. Is it true if I am using PhotoStream to upload photos to my PC via iCloud, photos are only pushed when wifi is available? This means I don't have to worry about extra data usage, right?
Thank you.Correct.
On an iOS device, new photos you take will be automatically uploaded to your photo stream when you leave the Camera app and are connected to Wi-Fi. Note: My Photo Stream does not push photos over cellular connections. -
Hi,
i had a query about local storage.
I've a machine that hosts weblogic and tangosol. i've an ejb that accesses a distributed cache i.e NamedCache cache = CacheFactory.get("MyCache")
i modified tangosol-coherence.xml and set local-storage to false ( for distributed cache) and replaced the file in coherence.jar.
i'm using an overflow scheme and the back map uses a disk scheme.
i also start a separate standalone instance of tangosol and i set the system property of local storage to true for the standalone instance.
i start the standalone instance first and then weblogic.
The idea is ensure that the tangosol instance in weblogic or the weblogic JVM should not participate in storing data (hence local storage false).
only the JVM for the standalone instance should store data (hence local storage true -system property).
i wanted to know whether the property "local-storage" is pertinent to a member(machine) or to a JVM?
the reason for this doubt: as i'm using a disk scheme, tangosol creates a file for an overflow (e.g lh014402~.tp). i can see two such files when ideally i would have wanted only one for the tangosol instance.
-rw-r--r-- 1 zephyr users 8364032 2005-06-23 17:02 lh014402~.tp
-rw-r--r-- 1 zephyr users 8364032 2005-06-23 17:02 lh014403~.tp.
can you please let me know if we can configure tangosol in such a way that we can two separate instances running with local stroage false for one and true for the other?
Awaiting your reply
Thanks
VinayI would suggest leaving the default 'local-storage' value set to 'true' in the tangosol-coherence.xml and just use the JVM argument to control the local storage of each individual node. Then start the stand alone instance normally (I assume you are using the com.tangosol.net.DefaultCacheServer) and start the WebLogic instance with the following:
java [...] -Dtangosol.coherence.distributed.localstorage=false [...]
Hope this helps.
Later,
Rob Misek
Tangosol, Inc. -
Big Troubles on designing Query about special customers' counting
Hello buddies:
I meet a problem on designing Query about special customers' counting. Let me describe the requirment first. I want to create a query with BEX , and there is a key figure with very special logic.
That is: to list the counts of the customers which has more than one sales records in a time period from sales data.
For example :
when the user excute the query , he or she must input a time period ( 2007.01~2007.03 e.g)
then the query output as follow:
District Cust-sount
North-Zone 100
South-Zone 120
The Main trouble are :
1. Threr are no document number in the detail of sales data document records. so I could not count the sales times with document number.
2. The time period is not fixed value, it depends on the user's input, so I can not define the counting logic in the update rule or in the query with fixed time period.
Anybody who met similar requirement pls show me your hand and give your solutions, thanks very much.
JasonHi,
Your solution sounds a good way to count the distinct customers. but in my case, one salse line item must not be recognize as one sales record, instead, one customer's all sales line items occurs in one day must be recognize as one sales record ( or we say that one sales behavior).
for example:
customer product quantity date
cust001 prod001 10 2007.06.06
cust001 prod002 20 2007.06.06
the two line items above means one sales record for the customer "cust001".
so I could not simply use the CKF : (( Counter ) *FV2 ) > 1 .
Best Regards,
Jason -
Query about licensing Jdeveloper
Dear Friends,
I have a query about licensing of Jdeveloper development tool. I understand that Jdeveloper is Free tool.That is we do not require license to use Jdeveloper for development as well as production.
Recent I heard that Jdeveloper is free only if we purchase Oracle Application Server. Is it correct ? Does one need to purchase Jdeveloper license if it is being deployed on any other App. server eg. Jboss etc ?
Can anyone throw light on the same ?
Many thanks,
VaijHi,
JDeveloper is free! Oracle ADF - the binding layer - ADF BC, and ADF Faces need an OracleAs licence
Frank -
New issue in R 12.1.3 in AP while query about inovice was recorded in AP
i record new invoice in ap on release R 12.1.3 and when query about it on invoice form the error appeare was it
forms
FRM-40735:POST-QUERY trigger raised unhandled exception ORA-04063
how can someone help usHi,
Please see these docs.
R12:Getting FRM-40735 Post-Query Trigger On Quering Invoice [ID 1209736.1]
After Applying Patch APXINWKB.fmb Is Not Working [ID 1159124.1]
R12.1.1 APXINWKB Invoice Workbench Form Comes Up With 'ORA-01403' [ID 949942.1]
12.1.1: FRM-40735: Post-Query Trigger Raised Unhandled Exception ORA-04063 [ID 1077613.1]
Query on Invoices, Getting "FRM-40735: POST-QUERY trigger raised unhandled exception ORA-4063" [ID 1076609.1]
Thanks,
Hussein -
When Clause with FIELDS TERMINATED BY and TRAILING NULLCOLS option?
Dear all
If i have the following datafields:
E;15;
D;14;16;
E;18;
to load into table tab1 (Type_ligne VARCHAR2(1),
Nombre1 NUMBER(2),
Nombre2 NUMBER(2))
Is it possible to have the following control file :
LOAD DATA
TRUNCATE
INTO TABLE Tab1
when (1) ='E'
FIELDS TERMINATED BY ';' TRAILING NULLCOLS
(Type_Ligne CHAR,
Nombre1 INTEGER EXTERNAL)
INTO TABLE Tab1
when (1) ='D'
FIELDS TERMINATED BY ';' TRAILING NULLCOLS
(Type_Ligne CHAR,
Nombre1 INTEGER EXTERNAL,
Nombre2 INTEGER EXTERNAL)
If no! please show me the correct syntax.Yes, you can, but for each WHEN clause after the first one, you have to specify the starting position for the first field, as shown in bold below.
LOAD DATA
TRUNCATE
INTO TABLE Tab1
when (1) ='E'
FIELDS TERMINATED BY ';' TRAILING NULLCOLS
(Type_Ligne CHAR,
Nombre1 INTEGER EXTERNAL)
INTO TABLE Tab1
when (1) ='D'
FIELDS TERMINATED BY ';' TRAILING NULLCOLS
(Type_Ligne POSITION(1) CHAR,
Nombre1 INTEGER EXTERNAL,
Nombre2 INTEGER EXTERNAL) -
I have one query about table entries.
i have one query about table entries.
suppose for particular table we maintained 5 entries in dev server. actaully in the dev we have only these 5 entries.
In production we have 200 entries actually.
If we move the cts from dev to production ,we will get 205 entries right
please help me in this.If i understood correctly, It is a Z table and you have done some changes in DEV system. If you move this to Production, it wont effect the production entries.
There are 2 different ways you can trnasport a Table. 1. with Table entries 2. Without table entries.
If you transport with Table entries then the DEV entries in this case 5 will be moved to PRODUCTION totaling 205.
If you transport without table entries, then in Production you will find only 200 entries. -
Hi Support,
I have a query about the DR -Standby:
1) Is FAL_CLIENT and FAL_SERVER parameter mandatory in both Primary and standy node?
2) How many standy by we can create for primary node?
ThanksDear Prashanth,
Quantity of 3 is rejected so you blocked it? why you want to bring it again to Quality? Use of putting in block is as follows:
Since quantity of 3 is rejected and need to be kept aside you are blocking the stock, the purchase person will send back this material to vendor directly from block stock (MBRL or MIGO). This is best practice.
There can be other situation that your vendor came to your place and done rework om that 3 blocked items and you are satisfied with the quality then you can use respective movement for transferring from block to unrestricted, if you want lot to be generated then assign 08 insp type and do inspection.
Hope this will help you better.
Best Regards,
Shekhar -
Query about the "User Decision
Hi,
I have a query about the u201CUser Decisionu201D, Suppose I have received the 10 quantity and lot get created for 10. At time of UD (QA11) I put 7 quantities in unrestricted and 3 in blocked stock .what we generally use to do with that block stock? Whether we again transfer it to quality stock (Movement 349)? Or move directly to unrestricted (movement: 343).
I tried to move the blocked stock in u201Cquality stocku201D in transaction MB1B with movement 349 but it had given error u201CChange the inspection stock of material 9112223334A in QM onlyu201D. In QM how to transfer the blocked stock in quality stock?
Regards,
PKDear Prashanth,
Quantity of 3 is rejected so you blocked it? why you want to bring it again to Quality? Use of putting in block is as follows:
Since quantity of 3 is rejected and need to be kept aside you are blocking the stock, the purchase person will send back this material to vendor directly from block stock (MBRL or MIGO). This is best practice.
There can be other situation that your vendor came to your place and done rework om that 3 blocked items and you are satisfied with the quality then you can use respective movement for transferring from block to unrestricted, if you want lot to be generated then assign 08 insp type and do inspection.
Hope this will help you better.
Best Regards,
Shekhar -
Query about multiple connection pools under one database
Hi,
I have s query about connection pool, now we have a requirement,
we have two schemas in one db, some data store in one schema, some in another schema,
all tables are the same, but data is different, we want to retrive all data under two schemas,
so we need two connection pools under one database,
I have set two system DSN, and each connection pool was mapping to one DSN,
but after I importing tables into RPD, when I view data, there is a dialog let me select connection pool. so If this, when we drag columns in answer, it will definitely get wrong.
so how to realize this function about multiple connection pools under one database and we can get data normally.Hi,
Try this step
1)Better to create two different DSN for the same database with different user id and password
2)now create multiple connection pool in the same database in u r RPD physical layer .
also refer this link : for imporving performance
http://obiee101.blogspot.com/2009/01/obiee-multiple-connection-pools.html
http://gerardnico.com/wiki/dat/obiee/connection_pool
Thanks
Deva
Maybe you are looking for
-
Access 2007 Runtime & Windows 8.1
Has anyone had success installing Access 2007 Runtime on Windows 8.1? (This particular computer only has Office Home installed). When I attempted this, the Runtime program seemed to install okay (opened a copy of AccessRuntime.exe as Administrator) -
-
Import from Word- all links linking too low down so that bread crumbs don't display
When I import from Word, all the links being created are linking to the topic title, instead of just to the topic: e.g. href="CreateLineOfSight.htm #CreateLineOfSight">CreateLineOfSight</a></p></td> As a result, whenever the topic is longish, opening
-
A thumbnail button option to hide or show thumbnails on slideshows
A great feature would be the option to hide or show thumbnails when using the full screen slideshow widget, I would suggest an option to turn this feature on in the flyout box so when viewing a full screeen slide show a grid icon appears thats brings
-
POP and IMAP aren't starting?
I used the "Configure Mail Service" option, setup both POP and IMAP access for a couple of my users in the Workgroup Manager, and the Mail Service status is "Running" (also running are Outgoing/Incoming SMTP, Junk Mail Filtering, and Virus Scanning),
-
Substitution Variables just don't work for me
I have a database that I want to copy data from one version of existing data to another version using substitution variables.<BR><BR>I have a version of (target) data named "Flash Sep Test" that I want to copy to, and a version (source) "HQ1" with th