Query of some time related fields
Hi experts
In pm order operation "Dates" view, there are some time fields in the screen include Earliest scheduled start time,
Earliest scheduled finish time,Latest scheduled start time,Latest scheduled finish time.
and actual start time,actual finish time.
I found the earliest and latest time were calculated automatically and they have display status,can not be changed.I want to know how system calculate and get the time?
after technical close the pm order. I found the actual time is still blank. what should I do to get the actual time and make it display in the "Dates" screen?
thanks in advance!
Mark,
Useful reading: [Understanding Maintenance Order Scheduling|http://www.sapfans.com/forums/viewtopic.php?p=711224]
Maybe you need to activate the "forward scheduling with time" in the IMG???
PeteA
Similar Messages
-
Definitions of time-related fields of usage tracking data
Hi all,
After reading the related documentations on usage tracking, I'm still confused about the true meaning of the following fields:
COMPILE_TIME_SEC The time in seconds required to compile the query.
C UM_DB_TIME_SEC The total amount of time in seconds that the Oracle BI Server waited for back-end physical databases on behalf of a logical query.
TOTAL_TIME_SEC The time in seconds that the Oracle BI Server spent working on the query while the client waited for responses to its query requests.
START_TS The date and time the logical query was submitted.
END_TS The date and time the logical query finished. The start and end timestamps also reflect any time the query spent waiting for resources to become available.
My questions are:
1. Is TOTAL_TIME_SEC = C U M_DB_TIME_SEC + COMPILE_TIME_SEC? Why are there cases where C U M_DB_TIME_SEC > TOTAL_TIME_SEC ?
2. I've logged cases where (END_TS - START_TS) > TOTAL_TIME_SEC. Why would this happen?
Could someone clarify the meaning of these fields for me please? Thanks very much!!
=================
I have logged the following data with usage tracking, posting here as examples:
1.
compile_time_sec = 0
c um_db_time_sec = 23
total_time_sec = 12
start_ts = 18-AUG-09 (14:15:47)
end_ts = 18-AUG-09 (14:16:00)
end_ts - start_ts (in sec) = 13
cache_ind_flg = N
num_cache_hits = 0
num_cache_inserted = 1
2.
compile_time_sec = 0
c um_db_time_sec = 125
total_time_sec = 110
start_ts = 21-AUG-09 (15:45:37)
end_ts = 21-AUG-09 (16:47:35)
end_ts - start_ts (in sec) = 3718
cache_ind_flg = N
num_cache_hits = 0
num_cache_inserted = 0
3.
compile_time_sec = 0
c um_db_time_sec = 1252
total_time_sec = 932
start_ts = 24-AUG-09 (17:23:40)
end_ts = 24-AUG-09 (17:49:27)
end_ts - start_ts (in sec) = 1547
cache_ind_flg = N
num_cache_hits = 0
num_cache_inserted = 0Check out DOC ID 973090.1 in metalink.
-
Same query takes different time to fetch result from Database
Hi all,
I am having a scenario in which a query is taking different time keeping other environmental variables constant.
When that query runs for 1 user it takes just 2 minutes to fetch result and the DB connection becomes inactive after fetching the result.
But if I run the same query after some time in similar environment it takes 25 minutes. sometimes 40 minutes to execute and give the result to the app server from the same database.
I am not able to understand this behavior from DB. Can anybody try to explain this behavior?
The details of the DB are,
Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
PL/SQL Release 11.2.0.1.0 - Production
CORE 11.2.0.1.0 Production
TNS for 32-bit Windows: Version 11.2.0.1.0 - Production
NLSRTL Version 11.2.0.1.0 - Production
OS Details:
Windows 2008 server Enterprise Edition.
I tried analyzing that query in oracle, it recommended there are lots of hard parsing in that query.
Regards,user10915512 wrote:
Hi all,
I am having a scenario in which a query is taking different time keeping other environmental variables constant.
When that query runs for 1 user it takes just 2 minutes to fetch result and the DB connection becomes inactive after fetching the result.
But if I run the same query after some time in similar environmentBut not exactly the same environment. So what is different that it is only "similar"?
it takes 25 minutes. sometimes 40 minutes to execute and give the result to the app server from the same database.
I am not able to understand this behavior from DB. Can anybody try to explain this behavior?Run a statspack on the 'well behaved' query and on the 'not well behaved' query. Compare and contrast the results.
>
The details of the DB are,
Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
PL/SQL Release 11.2.0.1.0 - Production
CORE 11.2.0.1.0 Production
TNS for 32-bit Windows: Version 11.2.0.1.0 - Production
NLSRTL Version 11.2.0.1.0 - Production
OS Details:
Windows 2008 server Enterprise Edition. To paraphrase Forest Gump, "My momma always said Windows was like a box of chocolates ...."
>
>
I tried analyzing that query in oracle, it recommended there are lots of hard parsing in that query.
Regards, -
Query problem: Some fields missing for org unit
I'm currently trying to fix a Minimum Compliance Report with some missing fields.
For a specific organisational unit no data is being generated for distribution and logistics.
How can i make sure organisational units generate a compliance figure correctly. The formula in the query is fine however when i generate the report there are X's in some fields.
How can i find the error?
any advice will be appreciated
Cheers
NeilHi Neil,
Have you tried using the function NDIV0 or NOERR in your formulas? The X in some of the fields may be due to undefined values (e.g. division by zero).
Hope this helps,
Juice -
Query Designer slows down after working some time with it
Hi all,
the new BEx Query Designer slows down when working some time with it. The longer it remains open, the slower it gets. Especially formula editing slows down extremely.
Did anyone of you encounter the same problem? Do you have an idea, how to fix this. To me it seems as if the Designer allocates more and more RAM and does not free that up.
My version: BI AddOn 7.X, Support Package 13, Revision 467
Kind regards,
PhilippI have seen a similar problem on one of my devices, the 'Samsung A-920'. Every time the system would pop up the 'Will you allow Network Access' screen , the imput from all keypresses from then on would be strangely delayed. It looked like the problem was connected with the switching from my app and the system dialog form. I tried for many many long hours / days to fix this, but just ended up hacking my phone to remove the security questions. After removing the security questions my problem went away.
I don't know if it's an option in your application, but is it possible to do everything using just one Canvas, and not switch between displayables? You may want to do an experiment using a single displayable Canvas, and just change how it draws. I know this will make user input much more complicated, but you may be able to avoid the input delays.
In my case, I think the device wasn't properly releasing / un-registering the input handling from the previous dialogs, so all keypresses still went through the non-current network-security dialog before reaching my app. -
Blank exif date time original field in some files (scanned images)
I have had some issues with scanned files or psd files where the date time original field does not exist or is blank. Mostly with scanned files. I would like to edit the field and fix the date; but the EXIF doesn't exist on scanned images or is a shorter data set excluding the date time original field.
Lightroom won't update or insert missing EXIF date fields into master files. However, if you export a file with a date/time edited in Lightroom, it will have valid EXIF date/time fields.
-
Occasionally my caps lock key blinks some time later my keyboard locks up what is causing the blinking and is it related to my keyboard locking.
This blinking occurs when no keyboarding is going on.
Back-up is not running.
I am on-line.Anthony:
The normal practice is to use the Shift key when typing a single Capital letter, as in a name or the beginning of a sentence, and the Caps Lock key to type all caps. The easiest thing would be for you to retrain your little finger to use the Shift instead of Caps Lock. I know of no way to display an icon at the bottom of the monitor to display that the caps lock key is engaged.
cornelius -
How to copy data from a Input query to real time Infocube
Hi All, I need some help on IP related stuff. The details are as below:
I am using two real time infocubes and one Multiprovider on them.
First Infocube u2018PRODUCTu2019 with the following Objects:
Product Group ,Product,Region ,Fiscal Year , Fiscal Year Varient ,0currency
Gross Sales
Second Infocube u2018C_PRODUCT u2018 with the following Objects:
Product Group ,Product,Region ,Fiscal Year , Fiscal Year Varient ,0currency
Gross Sales and Forecast Sales
( u2018Forecast Salesu2019 in the extra object in the second Infocube. Remaining objects are same in the both the cubes)
First Infocube u2018Productu2019 has data but the second cube does not have any data.
I have created a Multiprovider ZMulti and in that I have included these two cubes. On this Multi Provider, I have created a Aggrigation Level(The details of the aggregation level is provided in the next page).
On this Aggrigation level I have built a Input Query by which I run the report.
(In the Rows area, I have taken Product Group, Product, Region and Fical year.
In the Key Figure section, I have taken Gross Sales and Forecast Sales( which is part of the second Infocube C_Product). These two Key Figures are made Input as Input ready. And the Query is Input ready query. In the Filter, I have taken Fiscal Year Variant. )
Data available in the first cube and no data currently in the second cube. Now executed the report and report displayed information from the first cube.
EX:
Product Product Region Year Gross Sales Forecast Sales
Group
PG01 Beer North 2008 10000 u2026u2026u2026
PG01 Beer South 2008 20000 u2026u2026u2026
PG01 Beer East 2008 15000 u2026u2026u2026
The second column is empty and there is no data for the Forecast sales Key Figure in the column.
Things look OK till the above.
Now based on the Gross Sales, I am entering forecast Sales values and I need to save the entire record to the second cube.
For this in the report, I have given the Forecast sales values in the second column and right clicked the mouse and form the menu which I got, selected the u2018SAVEu2019 Option to save the data. For some time, I can see u2018Processingu2019 message and later values at u2018Forecast Salesu2019 are disappearing.
Data is not saving to the cube second cube.
====================================================
Am I missing any thing?
In the Query, Do I need to provide InfoProvider and restrict it to specific value?
In the query, Do I need to include, 0currency object?
Could you please guide me on this?
=============================================
Details of the Aggrigation level.
I have created an Aggrigation Level ZAL_Prod on MultiProvider.
Selected all the objects and activated.
In the Filter, I have taken 0Fiscal Year Variant and saved it.Restrict Forecast sales to the Plan cube in the Query.
that should fix it. -
Caltyp24_Call Type Real Time All Fields Service Level Today discrepancy
Hi All,
During our initial phase of deployment to CUIC reporting tool we want to offer similar features to our team to ensure smoother transition. However, I'm seeing discrepancy in data pulled by two different reporting engines - Webview 7.5 and CUIC 8.x. I have attached the screen shot of both highlighted in red.
Details:
CUIC/WebView Report: caltyp24_Call Type Real Time All Fields;
Version of CUIC: 8.5(4) build 1 (8_5_4_10000_99);
ICM Version: Release 7.5.1.0, Build 23684;
Column with Discrepancy: Service level today.
Any help much appreciated.
Thanks in advance,
RamchandThanks very much for responding so quickly Kartik. The data is pulled from the ServiceLevelToday column in Call_Type_Real_Time table. Here is the query; I have truncated some columns for brevity:
SELECT Call_Type.EnterpriseName,
Call_Type_Real_Time.CallTypeID,
Call_Type_Real_Time.ServiceLevelCallsToday,
Call_Type_Real_Time.ServiceLevelToday,
Call_Type_Real_Time.TalkTimeToday,
Call_Type_Real_Time.ReturnReleaseHalf,
RouterCallsAbandToAgentTo5 = Call_Type_Real_Time.RouterCallsAbandToAgentTo5,
RouterCallsAbandToAgentHalf = Call_Type_Real_Time.RouterCallsAbandToAgentHalf,
CallsAtVRUNow = Call_Type_Real_Time.CallsAtVRUNow,
CallsAtAgentNow = Call_Type_Real_Time.CallsAtAgentNow
FROM Call_Type, Call_Type_Real_Time
where Call_Type.CallTypeID = Call_Type_Real_Time.CallTypeID
Order BY Call_Type.EnterpriseName, Call_Type_Real_Time.DateTime
We used the templates through CDN that mimic the real time and historic reports from Web View. This is the only real time report where we noticed the discrepancy in service level today column. We have the Standard CUIC license so I am unable to verify the report definition through CUIC. However, I should assume that both reports are referring to the same columns and tables. Do we have any alternatives? Please suggest…thanks -
Write a query for which has its field name as input parameter
Hi All,
I need to write a query for SSRS report. StartDate, EndDate and ChangedField are the input parameters.
Let's consider that we are taking employee details. The table contains lot of fields like emp name,add1,add2,city,state,country,designation,rank etc. We have almost 30 fields are there. These field name are coming as input parameter "ChangedField"
. I need to show the records which belong to the start date and end date also data change for the field which is given in the input parameter "ChangedField". That meand If I give ChangedField="city", we need to check for data which
has been updated with new City name in the given time period.
Please suggest.
Regards,
JulieHi Julie,
Per my understanding that you have many fields in the table to be displayed in the report, some of the fields will always show and others should be hide and show based on the selection of the parameters which include the columns names, right?
By default, we can't display columns dynamically in the report automatically without insert this column in the report before. So, in your scenario, I will suggest you to add all the columns in the reports first and then add show/hide expression on
the column visibility based on the selection of the parameter "ChangedField", when some of the fields selectted they will be display, otherwise they will be hidden. For the fields which you want to be always displayed will not need to add the expression
in the column visibility.
Details steps about how to acheive this, please referencet to the article below:
Displaying Dynamic Columns in SSRS Report
If you still have any problem, please feel free to ask.
Regards,
Vicky Liu
Vicky Liu
TechNet Community Support -
Improve Query Suggestions Load time
How can I improve load time for Pre Query Suggestions on search home page when user start typing ??
I notice it was slow in loading when I hit first time in the morning so I try to warm up by adding
""http://SiteName/_api/search/suggest?querytext='sharepoint' ""
in to warm up script but even during the day time after hitting few times it is slower some times . Any reason ?
Do you think moving Query Component to WFE will do any help here?
Pleas let me know - Thanks .Hi,
Query Suggestions should work at a high level overview is:
• You issue a query within a Search Center site and get results..
• When you hover over or click a result.. this gets stored as a “RecordPageClick” method and will be stored in the Web Applications W3WP process…
• Every five minutes ( I believe ) this W3WP will flush these Recorded Page Clicks and pass them over to the Search Proxy…
• This will then store them in a table in the SSA ( Search Service Application ) Admin DB
• By default, once a day, there is a timer job, Prepare Query Suggestions, that will run
• It does a check to see if the same term has been queried for and clicked at least 6 times and then will move them to another table in the same DB..
• If there are successful queries\clicks > 6 times for a term, and they move to the appropriate table, then when you start to type a word, like “share”
• This will fire a “method” over to the Search proxy servers called “GetQuerySuggestions” and will check that Admin DB to see if they have any matches..
• If so, then it will show up in your Search Box as like “SharePoint” for a suggestion…
Other components involved with the Query Suggestions:
Timer Jobs
Prepare query suggestions Daily
Query Classification Dictionary Update for Search Application Search Service Application Minutes
Query Logging Minutes
Database
MSSQLogQuerySuggestion (SearchDB) This gets cleaned up when we run the timer job
MSSQLogQueryString (SearchDB) Info on the Query String
MSSQLogSearchCounts (SearchDB) Info on the click counts
MSSQLogQuerySuggestion Looks like this may be where the hits for suggestions are stored
So the issue might related to timer job, database, connection between SharePoint server and SQL server. There is a similar case caused by DistributedCache.
If you move query component on to another sever, this may improve the process related to Search service, however, it may affect the performance due to networking.
Please collect verbose ULS log per steps below:
Enable Verbose logging via Central Admin> Monitoring> Reporting>Configure diagnostic logging(You can simply set all the categories with Verbose level and I can filter myself).
Reproduce the issue(Try to remove SSRS).
Record the time and get a capture of the error message(including the correlation ID if there is). Collect the log file which is by default located in the folder: <C:\Program files\common files\Microsoft Shared\web server extensions\15\LOGS>.
Stop verbose logging.
Regards,
Rebecca Tu
TechNet Community Support
Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
[email protected]. -
Approach to tune a query in short time
Hi All,
Oracle 10g I know this question is asked number of times and there are many good replies to them.
But I just want to know how to approach a completely new query ( like the task given to me to fine tume a query in 1 day when I dont have even the slightest idea about how to proceed) if the timeline is very stringent and by just looking at the explain plan, you have to take the decision.
I am just posting my query here and what I am looking for is some lead on how to identify the congetion point which is where this query takes long time ( in my case some 15 mins as reported to me)
select
"LEGAL ENTITY",
"Legal Entity Description",
"Cluster",
"Sub_Cluster",
"Account",
rownum,
"Moody_Rating",
"Process_Date",
"Merge_Description",
rownum,
"Merge_Description",
"is_id_ic",
"is_n",
"cusip",
"isin",
"credit_spread_PV01",
"amount",
"Market_Value",
"Currency",
"Sensitivity_Type",
"maturity_Date",
"Exception_Flag",
"Base_Security_Id",
DECODE(sign("Market_Value"),-1,DeCode(SigN("Recovery"),-1,"Recovery",('-'||"Recovery")), ABS("Recovery")) as "Recovery"
from
select
le.name "LEGAL ENTITY",
le.display_name "Legal Entity Description",
mn4.display_name "Cluster",
mn3.display_name "Sub_Cluster",
bookname.display_name "Account",
(SELECT RATING_NAME
FROM moody_rating
where moody_rating_id = i.moody_rating_id) "Moody_Rating",
to_char(to_date(:v_cob_date,'DD-MM-YY'),'YYYYMMDD') "Process_Date",
ss.issuer "Merge_Description",
PART.MARS_ISSUER "is_id_ic",
PART.PARTICIPANT_NAME "is_n",
NULL "cusip",
NULL "isin",
NULL "credit_spread_PV01",
NULL "amount",
sum(mtmsens.sensitivity_value) "Market_Value",
(SELECT distinct cc.CCY
FROM legacy_country CC
INNER JOIN MARSNODE MN ON CC.countryisocode = MN.NAME
and mn.close_date is null
INNER JOIN MARSNODETYPE MNT ON MN.TYPE_ID =
MNT.NODE_TYPE_ID
AND MNT.NAME = 'COUNTRY'
and mnt.close_date is null
where MN.NODE_ID = part.country_domicile_id
and cc.begin_cob_date <= :v_cob_date
and cc.end_cob_date > :v_cob_date
and rownum < 2) "Currency",
'CREDITSPREADMARKETVALUE' "Sensitivity_Type",
NULL "maturity_Date",
NULL "Exception_Flag",
NULL "Base_Security_Id",
sum(ss.sensitivity_value) "Recovery"
from staging_position sp
left JOIN position p on (
p.feed_instance_id = sp.feed_instance_id
AND p.feed_row_id = sp.feed_row_id)
left JOIN staging_instrument si on (si.feed_instance_id =
sp.feed_instance_id AND
si.position_key =
sp.position_key)
left join book b on (b.book_id = p.book_id and
b.begin_cob_date <= :v_cob_date and
b.end_cob_date > :v_cob_date)
left join marsnode bk on (b.book_id = bk.node_id and
bk.close_date is null)
left join marsnode le on (b.leg_ent_id = le.node_id and
le.close_date is null)
left join marsnode bookname on (bookname.node_id = p.book_id and
bookname.close_date is null)
left join marsnodelink mnl on p.book_id = mnl.node_id
and :v_bus_org_hier_id =
mnl.hierarchy_id
and mnl.close_date is null
and :v_cob_date >= mnl.begin_cob_date
and :v_cob_date < mnl.end_cob_date
left join marsnode mn on mn.node_id = mnl.parent_id
and mn.close_date is null
left join marsnodelink mnl2 on mn.node_id = mnl2.node_id
and :v_bus_org_hier_id =
mnl2.hierarchy_id
and mnl2.close_date is null
and :v_cob_date >= mnl2.begin_cob_date
and :v_cob_date < mnl2.end_cob_date
left join marsnode mn2 on mn2.node_id = mnl2.parent_id
and mn2.close_date is null
left join marsnodelink mnl3 on mn2.node_id = mnl3.node_id
and :v_bus_org_hier_id =
mnl3.hierarchy_id
and mnl3.close_date is null
and :v_cob_date >= mnl3.begin_cob_date
and :v_cob_date < mnl3.end_cob_date
left join marsnode mn3 on mn3.node_id = mnl3.parent_id
and mn3.close_date is null
left join marsnodelink mnl4 on mn3.node_id = mnl4.node_id
and :v_bus_org_hier_id =
mnl4.hierarchy_id
and mnl4.close_date is null
and :v_cob_date >= mnl4.begin_cob_date
and :v_cob_date < mnl4.end_cob_date
left join marsnode mn4 on mn4.node_id = mnl4.parent_id
and mn4.close_date is null
--sensitivity data
left JOIN STAGING_SENSITIVITY ss ON (ss.FEED_INSTANCE_ID =
sp.FEED_INSTANCE_ID AND
ss.FEED_ROW_ID =
sp.FEED_ROW_ID)
--sensitivity data
left JOIN STAGING_SENSITIVITY mtmsens ON (mtmsens.FEED_INSTANCE_ID =
sp.FEED_INSTANCE_ID AND
mtmsens.FEED_ROW_ID =
sp.FEED_ROW_ID)
LEFT join xref_domain_value_map XREF on (XREF.Src_Value =
ss.issuer and
XREF.close_action_id is null and
XREF.Begin_Cob_Date <=
:v_cob_date and
XREF.End_Cob_Date >
:v_cob_date AND
xref.domain_map_id = 601 AND
xref.source_system_id = 307 AND xref.ISSUE_ID is not null)
Left join ISSUE i on (i.issue_id = xref.issue_id)
LEFT join participant PART ON (PART.PARTICIPANT_ID =
XREF.TGT_VALUE and
PART.Close_Action_Id is null and
PART.Begin_Cob_Date <= :v_cob_date and
PART.End_Cob_Date > :v_cob_date)
left join moody_rating RATING on (rating.moody_rating_id =
i.MOODY_RATING_ID)
where sp.feed_instance_id in
(select fbi.feed_instance_id
from feed_book_status fbi ,
feed_instance fi
where fbi.cob_date = :v_cob_date
and fbi.feed_instance_id = fi.feed_instance_id
and fi.feed_id in (
select feed_id from feed_group_xref where feed_group_id in (
select feed_group_id from feed_group where description like 'CDO Feeds')
and close_action_id is null
and sp.Feed_Row_Status_Id = 1
and ss.sensitivity_type = 'CREDITSPREADDEFAULT'
and mtmsens.sensitivity_type = 'MTMVALUE'
and le.name='161'
group by le.name,
le.display_name,
mn3.display_name,
mn4.display_name,
mn.display_name,
i.moody_rating_id,
ss.issuer,
PART.MARS_ISSUER,
PART.PARTICIPANT_NAME,
sp.feed_instance_id,
part.country_domicile_id,
bookname.display_name) And the explain plan
SELECT STATEMENT, GOAL = CHOOSE Cost=19365 Cardinality=1 Bytes=731
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MOODY_RATING Cost=1 Cardinality=1 Bytes=9
INDEX UNIQUE SCAN Object owner=MARS Object name=PK_MOODY_RATING Cost=0 Cardinality=1
HASH UNIQUE Cost=77 Cardinality=1 Bytes=488
COUNT STOPKEY
HASH JOIN Cost=76 Cardinality=1 Bytes=488
NESTED LOOPS Cost=68 Cardinality=1 Bytes=460
HASH JOIN Cost=66 Cardinality=1 Bytes=450
HASH JOIN Cost=59 Cardinality=1 Bytes=412
NESTED LOOPS Cost=51 Cardinality=1 Bytes=402
HASH JOIN Cost=49 Cardinality=1 Bytes=392
NESTED LOOPS Cost=42 Cardinality=1 Bytes=359
NESTED LOOPS Cost=40 Cardinality=1 Bytes=349
NESTED LOOPS Cost=37 Cardinality=1 Bytes=300
NESTED LOOPS Cost=34 Cardinality=1 Bytes=251
HASH JOIN Cost=32 Cardinality=1 Bytes=241
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=27
NESTED LOOPS Cost=24 Cardinality=1 Bytes=231
NESTED LOOPS Cost=21 Cardinality=1 Bytes=204
NESTED LOOPS Cost=18 Cardinality=1 Bytes=171
NESTED LOOPS Cost=16 Cardinality=1 Bytes=136
NESTED LOOPS Cost=13 Cardinality=1 Bytes=86
NESTED LOOPS Cost=10 Cardinality=1 Bytes=37
VIEW Object owner=MARS Cost=7 Cardinality=1 Bytes=10
FILTER
CONNECT BY WITH FILTERING
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_PARENT_ID Cost=3 Cardinality=250 Bytes=2500
HASH JOIN Cost=5 Cardinality=1 Bytes=62
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHYROOT Cost=2 Cardinality=5 Bytes=175
NESTED LOOPS
CONNECT BY PUMP
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=7 Cardinality=1 Bytes=39
INDEX RANGE SCAN Object owner=MARS Object name=IDX_MNL_HI_PI_NI Cost=3 Cardinality=4
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=27
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=49
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=50
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODETYPE Cost=2 Cardinality=1 Bytes=35
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODETYPE Cost=1 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=NODE_ASSOC Cost=3 Cardinality=1 Bytes=33
INDEX RANGE SCAN Object owner=MARS Object name=PK_NODE_ASSOC Cost=1 Cardinality=3
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
VIEW Object owner=MARS Cost=7 Cardinality=1 Bytes=10
FILTER
CONNECT BY WITH FILTERING
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_PARENT_ID Cost=3 Cardinality=250 Bytes=2500
HASH JOIN Cost=5 Cardinality=1 Bytes=62
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHYROOT Cost=2 Cardinality=5 Bytes=175
NESTED LOOPS
CONNECT BY PUMP
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=7 Cardinality=1 Bytes=39
INDEX RANGE SCAN Object owner=MARS Object name=IDX_MNL_HI_PI_NI Cost=3 Cardinality=4
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1 Bytes=10
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=NODE_ASSOC Cost=3 Cardinality=1 Bytes=49
INDEX RANGE SCAN Object owner=MARS Object name=PK_NODE_ASSOC Cost=1 Cardinality=3
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=49
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1 Bytes=10
VIEW Object owner=MARS Cost=7 Cardinality=1 Bytes=33
FILTER
CONNECT BY WITH FILTERING
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_PARENT_ID Cost=3 Cardinality=250 Bytes=2500
HASH JOIN Cost=5 Cardinality=1 Bytes=62
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHYROOT Cost=2 Cardinality=5 Bytes=175
NESTED LOOPS
CONNECT BY PUMP
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=7 Cardinality=1 Bytes=39
INDEX RANGE SCAN Object owner=MARS Object name=IDX_MNL_HI_PI_NI Cost=3 Cardinality=4
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1 Bytes=10
VIEW Object owner=MARS Cost=7 Cardinality=1 Bytes=10
FILTER
CONNECT BY WITH FILTERING
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_PARENT_ID Cost=3 Cardinality=250 Bytes=2500
HASH JOIN Cost=5 Cardinality=1 Bytes=62
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHYROOT Cost=2 Cardinality=5 Bytes=175
NESTED LOOPS
CONNECT BY PUMP
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=7 Cardinality=1 Bytes=39
INDEX RANGE SCAN Object owner=MARS Object name=IDX_MNL_HI_PI_NI Cost=3 Cardinality=4
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
VIEW Object owner=MARS Cost=7 Cardinality=1 Bytes=38
FILTER
CONNECT BY WITH FILTERING
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_PARENT_ID Cost=3 Cardinality=250 Bytes=2500
HASH JOIN Cost=5 Cardinality=1 Bytes=62
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHYROOT Cost=2 Cardinality=5 Bytes=175
NESTED LOOPS
CONNECT BY PUMP
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=7 Cardinality=1 Bytes=57
INDEX RANGE SCAN Object owner=MARS Object name=IDX_MNL_HI_PI_NI Cost=3 Cardinality=4
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=36
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1 Bytes=10
VIEW Object owner=MARS Cost=7 Cardinality=1 Bytes=28
FILTER
CONNECT BY WITH FILTERING
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_PARENT_ID Cost=3 Cardinality=250 Bytes=2500
HASH JOIN Cost=5 Cardinality=1 Bytes=62
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHYROOT Cost=2 Cardinality=5 Bytes=175
NESTED LOOPS
CONNECT BY PUMP
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=7 Cardinality=1 Bytes=57
INDEX RANGE SCAN Object owner=MARS Object name=IDX_MNL_HI_PI_NI Cost=3 Cardinality=4
TABLE ACCESS FULL Object owner=MARS Object name=MARSHIERARCHY Cost=2 Cardinality=1 Bytes=27
COUNT
VIEW Object owner=MARS Cost=19365 Cardinality=1 Bytes=731
HASH GROUP BY Cost=19365 Cardinality=1 Bytes=1112
NESTED LOOPS OUTER Cost=19364 Cardinality=1 Bytes=1112
NESTED LOOPS OUTER Cost=19361 Cardinality=1 Bytes=1040
NESTED LOOPS OUTER Cost=19361 Cardinality=1 Bytes=1037
NESTED LOOPS OUTER Cost=19360 Cardinality=1 Bytes=1019
NESTED LOOPS OUTER Cost=19357 Cardinality=1 Bytes=951
NESTED LOOPS OUTER Cost=19354 Cardinality=1 Bytes=914
NESTED LOOPS OUTER Cost=19351 Cardinality=1 Bytes=877
NESTED LOOPS OUTER Cost=19337 Cardinality=1 Bytes=820
NESTED LOOPS OUTER Cost=19334 Cardinality=1 Bytes=783
NESTED LOOPS OUTER Cost=19320 Cardinality=1 Bytes=726
NESTED LOOPS OUTER Cost=19317 Cardinality=1 Bytes=707
NESTED LOOPS OUTER Cost=19303 Cardinality=1 Bytes=650
NESTED LOOPS OUTER Cost=19300 Cardinality=1 Bytes=613
NESTED LOOPS Cost=19285 Cardinality=1 Bytes=556
NESTED LOOPS Cost=19280 Cardinality=1 Bytes=443
NESTED LOOPS OUTER Cost=19275 Cardinality=1 Bytes=330
HASH JOIN RIGHT SEMI Cost=17457 Cardinality=1 Bytes=248
VIEW Object owner=SYS Object name=VW_NSO_1 Cost=1119 Cardinality=30 Bytes=150
HASH JOIN Cost=1119 Cardinality=30 Bytes=2040
TABLE ACCESS FULL Object owner=MARS Object name=FEED_GROUP Cost=2 Cardinality=5 Bytes=120
HASH JOIN Cost=1116 Cardinality=1607 Bytes=70708
TABLE ACCESS FULL Object owner=MARS Object name=FEED_GROUP_XREF Cost=13 Cardinality=701 Bytes=14721
HASH JOIN Cost=1102 Cardinality=3602 Bytes=82846
INDEX RANGE SCAN Object owner=MARS Object name=IDX_FBS_CD_FII_BI Cost=22 Cardinality=3602 Bytes=46826
TABLE ACCESS FULL Object owner=MARS Object name=FEED_INSTANCE Cost=1024 Cardinality=670264 Bytes=6702640
NESTED LOOPS Cost=16337 Cardinality=324 Bytes=78732
HASH JOIN Cost=14324 Cardinality=1977 Bytes=302481
NESTED LOOPS OUTER Cost=11 Cardinality=1 Bytes=114
NESTED LOOPS Cost=8 Cardinality=1 Bytes=95
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=5 Cardinality=1 Bytes=59
INDEX RANGE SCAN Object owner=MARS Object name=IDX_NODE1 Cost=3 Cardinality=2
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=BOOK Cost=3 Cardinality=2 Bytes=72
INDEX RANGE SCAN Object owner=MARS Object name=IDX_BOOK_LEI_BCD Cost=2 Cardinality=4
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=19
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
PARTITION RANGE ALL Cost=13995 Cardinality=3854299 Bytes=150317661
TABLE ACCESS FULL Object owner=MARS Object name=POSITION Cost=13995 Cardinality=3854299 Bytes=150317661
PARTITION RANGE ITERATOR Cost=2 Cardinality=1 Bytes=90
PARTITION HASH ITERATOR Cost=2 Cardinality=1 Bytes=90
TABLE ACCESS BY LOCAL INDEX ROWID Object owner=MARS Object name=STAGING_POSITION Cost=2 Cardinality=1 Bytes=90
INDEX UNIQUE SCAN Object owner=MARS Object name=PK_STAGINGPOSITON Cost=1 Cardinality=1
PARTITION HASH ITERATOR Cost=1819 Cardinality=1 Bytes=82
TABLE ACCESS BY LOCAL INDEX ROWID Object owner=MARS Object name=STAGING_INSTRUMENT Cost=1819 Cardinality=1 Bytes=82
INDEX RANGE SCAN Object owner=MARS Object name=PK_STAGINGINSTRUMENT Cost=9 Cardinality=2551
PARTITION RANGE ITERATOR Cost=5 Cardinality=1 Bytes=113
PARTITION HASH ITERATOR Cost=5 Cardinality=1 Bytes=113
TABLE ACCESS BY LOCAL INDEX ROWID Object owner=MARS Object name=STAGING_SENSITIVITY Cost=5 Cardinality=1 Bytes=113
INDEX RANGE SCAN Object owner=MARS Object name=IDX_SENSITIVITY_FEED_ROW_ID Cost=3 Cardinality=8
PARTITION RANGE ITERATOR Cost=5 Cardinality=1 Bytes=113
PARTITION HASH ITERATOR Cost=5 Cardinality=1 Bytes=113
TABLE ACCESS BY LOCAL INDEX ROWID Object owner=MARS Object name=STAGING_SENSITIVITY Cost=5 Cardinality=1 Bytes=113
INDEX RANGE SCAN Object owner=MARS Object name=IDX_SENSITIVITY_FEED_ROW_ID Cost=3 Cardinality=8
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=14 Cardinality=1 Bytes=57
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_NODE_ID Cost=2 Cardinality=14
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=37
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=14 Cardinality=1 Bytes=57
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_NODE_ID Cost=2 Cardinality=14
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=19
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=14 Cardinality=1 Bytes=57
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_NODE_ID Cost=2 Cardinality=14
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=37
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODELINK Cost=14 Cardinality=1 Bytes=57
INDEX RANGE SCAN Object owner=MARS Object name=FKI_15632_NODE_ID Cost=2 Cardinality=14
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=37
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=MARSNODE Cost=3 Cardinality=1 Bytes=37
INDEX RANGE SCAN Object owner=MARS Object name=PK_MARSNODE Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=XREF_DOMAIN_VALUE_MAP Cost=3 Cardinality=1 Bytes=68
INDEX RANGE SCAN Object owner=MARS Object name=IDX_XDVM_DMI_SV_BCD Cost=2 Cardinality=1
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=ISSUE Cost=1 Cardinality=1 Bytes=18
INDEX UNIQUE SCAN Object owner=MARS Object name=PK_ISSUE Cost=0 Cardinality=1
INDEX UNIQUE SCAN Object owner=MARS Object name=PK_MOODY_RATING Cost=0 Cardinality=1 Bytes=3
TABLE ACCESS BY INDEX ROWID Object owner=MARS Object name=PARTICIPANT Cost=3 Cardinality=1 Bytes=72
INDEX RANGE SCAN Object owner=MARS Object name=PK_PARTICIPANT Cost=2 Cardinality=1Hi,
in your explain plan:
HASH JOIN RIGHT SEMI Cost=17457 Cardinality=1 Bytes=248
VIEW Object owner=SYS Object name=VW_NSO_1 Cost=1119 Cardinality=30 Bytes=150
HASH JOIN Cost=1119 Cardinality=30 Bytes=2040
TABLE ACCESS FULL Object owner=MARS Object name=FEED_GROUP Cost=2 Cardinality=5 Bytes=120
HASH JOIN Cost=1116 Cardinality=1607 Bytes=70708
TABLE ACCESS FULL Object owner=MARS Object name=FEED_GROUP_XREF Cost=13 Cardinality=701 Bytes=14721
HASH JOIN Cost=1102 Cardinality=3602 Bytes=82846
INDEX RANGE SCAN Object owner=MARS Object name=IDX_FBS_CD_FII_BI Cost=22 Cardinality=3602 Bytes=46826
TABLE ACCESS FULL Object owner=MARS Object name=FEED_INSTANCEThis part has the highest costs (this doesn't always mean it is slow). So this leads me to the WHERE clause where feed_group, feed_group_xref and feed_instance full are used. Maybe this can be improved, although the cardinality is not that high, so a full table can be the best. So the question is can indexes help here?
Furthermore there is the full table scan on POSITION:
TABLE ACCESS FULL Object owner=MARS Object name=POSITION Cost=13995 Cardinality=3854299 Bytes=150317661This looks also a large tabel (3 million + records), so is it possible to get this part smaller?
Herald ten Dam
http://htendam.wordpress.com -
How to add related fields from one transaction to other transacction?
Hi all,
I have a problem. I need to add an OPPORTUNITY field (SALESCYCLE) to the ACTIVITY transaction because of an upgrade and I'm now in a new world... The CRM 2007 is an unexplored place to me.
I added some custom fields (Z fields from the CUSTOMER_H) by adding the context node BTCUSTOMERH to the BT126H_APPT component, view ApptDetails, with transaction BSP_WD_CMPWB (thanks Stephen Johannes for your blogs) and all works correctly.
I tried doing the same with the SALESCYCLE and it doesn't worked... :_(
I added the context node BTOPPORTH, but I'm sure I failed at the wizard step with the "Higher level context node" or with the "BOL Relation", but I can't find the correct answer. I found the field in the "Configuration tab", I put it in my view but the GET method doesn't work... The field remains always void.
I don't have problems adding activity related fields, but when I try to add some flow related (opportunity one in my case) fields it's different.
I have been looking for any solution but I have only seen that I can add this attribute directly to one context node and modify some BAdI or do any thing similar, but I don't know how to do that either. I will accept this solution if someone tells me how to do that... But I think the first solutions could be the most appropiate (I can be wrong... of course!)
Thanks in advance....
SergioThanks for your answer, Subhasis!
Your solution is correct if I want to add to the Activity a field related to the Activity or some field on the DocFlow structure (the case you described), but it's not my case. I did that with customer fields before and all worked correctly.
My scene is like this. I have an Opportunity and there is a follow-up Activity created. When I want to see the Activity, I need to show on the Activity transaction the preceding Opportunity field SALESCYCLE, which is at the CRMD_OPPORT_H table (BTOPPORTH context node).
The problem comes when I add this context node to the Activity. I think the application is trying to get the field from the CRMD_OPPORT_H with ACTIVITY data and it returns void to the field. I think it forgets to get the opportunity data first and after that try to get the SALESCYCLE value from the CRMD_OPPORT_H (with the Opportunity GUID).
By the moment, I have added the attribute SALESCYCLE to the BTDOCFLOW context node and, at the moment he do the GET, I change the field it gets by the GUID of the opportunity:
dref = current->get_property( 'SALESCYCLE' ). "#EC NOTEXT
dref = current->get_property( 'OBJKEY_A' ). "#EC NOTEXT
Then I go to the CRMD_OPPORT_H directly to get the SALESCYCLE value...
I know it's a bad solution... but it works and I don't know any other solution by now. If anyone can help me to do it correctly I'll be greatly appreciated!
Regards,
Sergio -
How to Convert a 'Time Stamp' field to Date......?
Hello Gurus:
I am using a Generic Extractor to pull some stats data. One of the field is a Time Stamp field, which tells me the date and time of Transaction run. I only need "Date". I tried mapping the stamp field to Date info-object hoping for a automatic determination of the Date.... (like the Week, Month happens). Well, I was wrong!
I only need the Date part. How can I do this in BI? is teher a specific SAP info-object for this that could be used..? If not, what would be the routine. The Date field that comes in is a SAP format... "u201C20,070,505.0000000u201D meaning the Transaction was run
on 05/05/2007. How can I get this convertion in BI transaformation?? is there a special Routine...?
Also, we are trying to use "Time Stamp" based DELTA from a generic extraction. The field used is a System date with time stamp. Value looks like u201C20100,505,170,113.2918470u201D, u201C20100,505,170,113.4748820u201D Every record pulled has a different value... probably giving me the time at which the record was created or generated. Now, with DELTA specifid on this type of field, what will be the 'Current Status" value in Delta queue? Is it based on the last record created and will that be the highest value? For next Delta extraction, does it compare the extraction time stamp to this value and bring in those records that are greater than the 'Current Status" value? I guess with that we can have as many deltas as we want..... right?? is there any particular setting I should be doing or worried/ careful about setting etc....?
Thanks for the response in advance....
Best..... SMaaThanks Akshay,
Your suggestion worked right away and was probably the easiest of all. I tried the Function Module that Pravender suggested,
but it did not return me the date back.... may be because it does not need the Micro Seconds in it....? Routine as suggested by
Satyam might work too, but did not go there....!
Thanks to all of you.
Best..... SMaa -
While running the query how much time it will taken, I want to see the time
Hi Folks
I would like to know ... While running the query how much time it will be taken, I want to see the time? in WEBI XI R2.....
Plz let me know the answer.......Hi Ravi,
The time a report runs is estimated based on the last time it was run. So you need to run the report once before you can see how long it will take. Also it depends on several factors... the database server could cache some queries so running it a second time immediately after the first time could be quicker. And there is the chance of changing filters to bring back different sets of data.
You could also schedule a report and then check the scheduled instance's status properties and view how long a report actually ran.
Good luck
Maybe you are looking for
-
How to use SequenceRunner with FlexUnit 4
Hi, I just read the Wiki page about Fluint Sequences at http://docs.flexunit.org/index.php?title=Sequences#Sequences_from_Fluint. I really would like to use sequences as it seems a great way to test flows in UI components. So I created a first test b
-
Convert .fla or .swf to a print file?
Our reviewers want to see online courses in a print format. Is there any way to export a Flash course (all frames) into a PDF or other printable file? The only way I know is to screen capture each frame (ugh!), which is inefficient at best. I downloa
-
JMS : Using SUN MQUEUE as a JMS Provider
Hi guys, I'm having quite some trouble in configuring a JMS scenario... I'm using SUN MQueue as a JMS Provider. I've installed the respective drivers, imq.jar, fscontext.jar and jms.jar. I had no trouble in deploying this drivers on the SDM. Now, aft
-
Hi, I've created a book using iPhoto. However, the ribbon at the top of the screen (the one showing what the pages look like in a smaller version), shows a few pages that are blank. But when I made the pages (below the ribbon in the main part of the
-
Drop Down loading problem - Java Applet
Hi, We are on the process of migrating our applet based application to SUN JVM. We have a applet drop down box in the screen which fetches the data from database when the focus is gained. The code is working fine in Microsoft JVM. But when I covert t