Using GoldenGate SQLEXEC For Lookup Queries

My target table T is the rsult of joining two source tables - S1 and S2. For each new row in S1 I need to look up S2 attributes and apply combined result to T.
I am not able to find any extamples of this functionality inplemented on the source. All examples I have seen are implemented in the MAP statement on the target.
I was able to use SQLEXEC on the target to call lookup proc on the source via dblink. That seems backwards.
Is there a better way to do this?

Looks like the solution is to resolve mapping on the source.
After I configured mapping on the source the data is no longer replicating.
The extract is running without complaints and correctly reports DMLs and successful EXECSQL calls.
The replicat is also running without complaints, it just sits there as if it has nothing to do.
What am I doing wrong?
Attaching reports.
GGSCI (pkdb5) 1> view report ext1
Oracle GoldenGate Capture for Oracle
Version 11.1.1.0.0 Build 078
HP/UX, IA64, 64bit (optimized), Oracle 10 on Jul 28 2010 15:49:30
Copyright (C) 1995, 2010, Oracle and/or its affiliates. All rights reserved.
Starting at 2011-03-30 14:39:47
Operating System Version:
HP-UX
Version U, Release B.11.31
Node: pkdb5
Machine: ia64
soft limit hard limit
Address Space Size : unlimited unlimited
Heap Size : 4294967296 4294967296
File Size : unlimited unlimited
CPU Time : unlimited unlimited
Process id: 1412
Description:
** Running with the following parameters **
EXTRACT ext1
TARGETDEFS /u02/GG/source/dirsql/testdef.sql
USERID ggs_owner, PASSWORD *********
RMTHOST pkdb6, MGRPORT 7809
RMTTRAIL /u01/GG/source/dirdat/rt
TRANLOGOPTIONS ASMUSER goldgatesys@PK5_ASM, ASMPASSWORD ***********
TABLE atrifonov.test1, SQLEXEC (SPNAME get_lkp, PARAMS (p_col1 = col1)), COLMAP (col1 = col1, col2 = col2, col3 = get_lkp.p_col2), TARG
ET dwh.test1;
2011-03-30 14:39:47 INFO OGG-01635 BOUNDED RECOVERY: reset to initial or altered checkpoint.
Bounded Recovery Parameter:
Options = BRRESET
BRINTERVAL = 4HOURS
BRDIR = /u02/GG/source
CACHEMGR virtual memory values (may have been adjusted)
CACHEBUFFERSIZE: 64K
CACHESIZE: 8G
CACHEBUFFERSIZE (soft max): 4M
CACHEPAGEOUTSIZE (normal): 4M
PROCESS VM AVAIL FROM OS (min): 16G
CACHESIZEMAX (strict force to disk): 13.99G
Database Version:
Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
PL/SQL Release 10.2.0.4.0 - Production
CORE 10.2.0.4.0 Production
TNS for HPUX: Version 10.2.0.4.0 - Production
NLSRTL Version 10.2.0.4.0 - Production
Database Language and Character Set:
NLS_LANG environment variable specified has invalid format, default value will be used.
NLS_LANG environment variable not set, using default value AMERICAN_AMERICA.US7ASCII.
NLS_LANGUAGE = "AMERICAN"
NLS_TERRITORY = "AMERICA"
NLS_CHARACTERSET = "UTF8"
Warning: your NLS_LANG setting does not match database server language setting.
Please refer to user manual for more information.
Maximum supported ASM read buffer size is 28 KB
Maximum supported ASM read buffer size is 28 KB
2011-03-30 14:39:48 INFO OGG-00546 Default thread stack size: 262144.
2011-03-30 14:39:48 INFO OGG-00547 Increasing thread stack size from 262144 to 1048576.
2011-03-30 14:39:48 INFO OGG-01515 Positioning to begin time Mar 30, 2011 2:39:38 PM.
2011-03-30 14:40:10 INFO OGG-01516 Positioned to (Thread 1) Sequence 11498, RBA 326680592, SCN 0.0, Mar 30, 2011 2:39:38 PM.
2011-03-30 14:40:10 INFO OGG-01515 Positioning to begin time Mar 30, 2011 2:39:38 PM.
2011-03-30 14:40:32 INFO OGG-01516 Positioned to (Thread 2) Sequence 10747, RBA 343953424, SCN 0.0, Mar 30, 2011 2:39:38 PM.
2011-03-30 14:40:32 INFO OGG-01517 Position of first record processed for Thread 2, Sequence 10747, RBA 343953424, SCN 4.269826912
1, Mar 30, 2011 2:39:40 PM.
2011-03-30 14:40:32 INFO OGG-01517 Position of first record processed for Thread 1, Sequence 11498, RBA 326680592, SCN 4.269826912
2, Mar 30, 2011 2:39:41 PM.
2011-03-30 14:40:37 INFO OGG-01226 Socket buffer size set to 27985 (flush size 27985).
2011-03-30 14:40:37 INFO OGG-01055 Recovery initialization completed for target file /u01/GG/source/dirdat/rt000022, at RBA 1233.
2011-03-30 14:40:37 INFO OGG-01478 Output file /u01/GG/source/dirdat/rt is using format RELEASE 10.4/11.1.
2011-03-30 14:40:37 INFO OGG-01026 Rolling over remote file /u01/GG/source/dirdat/rt000023.
2011-03-30 14:40:37 INFO OGG-01053 Recovery completed for target file /u01/GG/source/dirdat/rt000023, at RBA 872.
2011-03-30 14:40:37 INFO OGG-01057 Recovery completed for all targets.
** Run Time Messages **
TABLE resolved (entry ATRIFONOV.TEST1):
TABLE ATRIFONOV.TEST1, SQLEXEC (SPNAME get_lkp, PARAMS (p_col1 = col1)), COLMAP (col1 = col1, col2 = col2, col3 = get_lkp.p_col2), TA
RGET dwh.test1;
Using the following key columns for source table ATRIFONOV.TEST1: COL1.
2011-03-30 14:43:02 INFO OGG-01021 Command received from GGSCI: STOP.
* ** Run Time Statistics ** *
Report at 2011-03-30 14:43:04 (activity since 2011-03-30 14:41:53)
Output to /u01/GG/source/dirdat/rt:
From Table ATRIFONOV.TEST1 to DWH.TEST1:
# inserts: 1
# updates: 0
# deletes: 0
# discards: 0
Stored procedure get_lkp:
attempts: 1
successful: 1
REDO Queue Statistics
-- Write Operations ------------------ -- Read Operations -------------------
Queue Name Size Count Waited Signaled Count Waited Signaled
+++consumer 1 Control 128 0 0 0 0 0 0
coordinator Control 128 1 0 0 1 0 0
coordinator Records 2048 1 0 0 1 122 1
Redo Thread 1 Control 128 1 0 0 1 0 0
Redo Thread 1 Records 2048 1 0 0 1 44 1
Redo Thread 2 Control 128 1 0 0 1 0 0
Redo Thread 2 Records 2048 0 0 0 0 149 0
CACHE OBJECT MANAGER statistics
CACHE MANAGER VM USAGE
vm current = 64K vm anon queues = 64K
vm anon in use = 0 vm file = 0
vm used max = 64K ==> CACHE BALANCED
CACHE CONFIGURATION
cache size = 8G cache force paging = 13.99G
buffer min = 64K buffer highwater = 4M
pageout eligible size = 4M
CACHE Transaction Stats
trans active = 0 max concurrent = 5
non-zero total = 2.89K trans total = 2.89K
CACHE File Caching
disk current = 0 disk total = 0
disk caching = 0 file cached = 0
file retrieves = 0
CACHE MANAGEMENT
buffer links = 0 anon gets = 0
forced unmaps = 0 cnnbl try = 0
cached out = 0 force out = 0
Allocation Request Distribution
< 128B: 0
128B: 1 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
Cached Transaction Size Distribution
0: 2.89K
< 4K: 1
4K: 0 0 | 16K: 0 0
64K: 0 0 | 256K: 0 0
1M: 0 0 | 4M: 0 0
16M: 0 0 | 64M: 0 0
256M: 0 0 | 1G: 0 0
4G: 0 0 | 16G: 0 0
64G: 0 0 | 256G: 0 0
1T: 0 0 | 4T: 0 0
16T: 0 0 | 64T: 0 0
256T: 0 0 |1024T: 0 0
QUEUE Statistics:
num queues = 15 default index = 0
cur len = 0 max len = 0
q vm current = 0 vm max = 0
q hits = 0 q misses = 1
queue size q hits curlen maxlen cannibalized
0 64K 0 1 1 0
1 128K 0 0 0 0
2 256K 0 0 0 0
3 512K 0 0 0 0
4 1M 0 0 0 0
5 2M 0 0 0 0
6 4M 0 0 0 0
7 8M 0 0 0 0
8 16M 0 0 0 0
9 32M 0 0 0 0
10 64M 0 0 0 0
11 128M 0 0 0 0
12 256M 0 0 0 0
13 512M 0 0 0 0
14 1G 0 0 0 0
================================================================================
CACHE POOL #0
POOL INFO group: ext1 id: p1412_Redo Thread 2 instance: 2 tid: 0000000000000007
trans active = 0 trans concurrent (max) = 3
trans total = 2.18K (2230 )
flag = 0x00000005
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
================================================================================
CACHE POOL #1
POOL INFO group: ext1 id: p1412_Redo Thread 1 instance: 1 tid: 0000000000000006
trans active = 0 trans concurrent (max) = 3
trans total = 724 (724 )
flag = 0x00000005
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 1 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
================================================================================
CACHE POOL #2
POOL INFO group: ext1 id: p1412_ORA-LOB-MEMPOOL instance: 0 tid: 0000000000000000
trans active = 0 trans concurrent (max) = 0
trans total = 0 (0 )
flag = 0x00000009
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
================================================================================
CACHE POOL #3
POOL INFO group: ext1 id: p1412_BLOB instance: 0 tid: 0000000000000000
trans active = 0 trans concurrent (max) = 0
trans total = 0 (0 )
flag = 0x00000000
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
QUEUE Statistics:
num queues = 15 default index = 0
cur len = 0 max len = 0
q vm current = 0 vm max = 0
q hits = 0 q misses = 1
queue size q hits curlen maxlen cannibalized
0 64K 0 1 1 0
1 128K 0 0 0 0
2 256K 0 0 0 0
3 512K 0 0 0 0
4 1M 0 0 0 0
5 2M 0 0 0 0
6 4M 0 0 0 0
7 8M 0 0 0 0
8 16M 0 0 0 0
9 32M 0 0 0 0
10 64M 0 0 0 0
11 128M 0 0 0 0
12 256M 0 0 0 0
13 512M 0 0 0 0
14 1G 0 0 0 0
================================================================================
CACHE POOL #0
POOL INFO group: ext1 id: p1412_Redo Thread 2 instance: 2 tid: 0000000000000007
trans active = 0 trans concurrent (max) = 3
trans total = 2.18K (2230 )
flag = 0x00000005
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
================================================================================
CACHE POOL #1
POOL INFO group: ext1 id: p1412_Redo Thread 1 instance: 1 tid: 0000000000000006
trans active = 0 trans concurrent (max) = 3
trans total = 724 (724 )
flag = 0x00000005
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 1 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
================================================================================
CACHE POOL #2
POOL INFO group: ext1 id: p1412_ORA-LOB-MEMPOOL instance: 0 tid: 0000000000000000
trans active = 0 trans concurrent (max) = 0
trans total = 0 (0 )
flag = 0x00000009
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
================================================================================
CACHE POOL #3
POOL INFO group: ext1 id: p1412_BLOB instance: 0 tid: 0000000000000000
trans active = 0 trans concurrent (max) = 0
trans total = 0 (0 )
flag = 0x00000000
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
EXT1.rpt: END
GGSCI (pkdb6) 1> view report rep2
Oracle GoldenGate Delivery for Oracle
Version 11.1.1.0.0 Build 078
HP/UX, IA64, 64bit (optimized), Oracle 11 on Jul 28 2010 16:09:58
Copyright (C) 1995, 2010, Oracle and/or its affiliates. All rights reserved.
Starting at 2011-03-30 14:40:39
Operating System Version:
HP-UX
Version U, Release B.11.31
Node: pkdb6
Machine: ia64
soft limit hard limit
Address Space Size : unlimited unlimited
Heap Size : 4294967296 4294967296
File Size : unlimited unlimited
CPU Time : unlimited unlimited
...skipping...
Oracle GoldenGate Delivery for Oracle
Version 11.1.1.0.0 Build 078
HP/UX, IA64, 64bit (optimized), Oracle 11 on Jul 28 2010 16:09:58
Copyright (C) 1995, 2010, Oracle and/or its affiliates. All rights reserved.
Starting at 2011-03-30 14:40:39
Operating System Version:
HP-UX
Version U, Release B.11.31
Node: pkdb6
Machine: ia64
soft limit hard limit
Address Space Size : unlimited unlimited
Heap Size : 4294967296 4294967296
File Size : unlimited unlimited
CPU Time : unlimited unlimited
Process id: 29967
Description:
** Running with the following parameters **
REPLICAT rep2
ASSUMETARGETDEFS
USERID ggs_owner, PASSWORD *********
MAP atrifonov.test1, TARGET dwh.test1, KEYCOLS (col1);
CACHEMGR virtual memory values (may have been adjusted)
CACHEBUFFERSIZE: 64K
CACHESIZE: 512M
CACHEBUFFERSIZE (soft max): 4M
CACHEPAGEOUTSIZE (normal): 4M
PROCESS VM AVAIL FROM OS (min): 1G
CACHESIZEMAX (strict force to disk): 881M
Database Version:
Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production
PL/SQL Release 11.2.0.2.0 - Production
CORE 11.2.0.2.0 Production
TNS for HPUX: Version 11.2.0.2.0 - Production
NLSRTL Version 11.2.0.2.0 - Production
Database Language and Character Set:
NLS_LANG environment variable specified has invalid format, default value will be used.
NLS_LANG environment variable not set, using default value AMERICAN_AMERICA.US7ASCII.
NLS_LANGUAGE = "AMERICAN"
NLS_TERRITORY = "AMERICA"
NLS_CHARACTERSET = "AL32UTF8"
Warning: your NLS_LANG setting does not match database server language setting.
Please refer to user manual for more information.
Opened trail file /u01/GG/source/dirdat/rt000022 at 2011-03-30 14:40:39
Switching to next trail file /u01/GG/source/dirdat/rt000023 at 2011-03-30 14:40:39 due to EOF, with current RBA 1233
Opened trail file /u01/GG/source/dirdat/rt000023 at 2011-03-30 14:40:39
2011-03-30 14:40:39 INFO OGG-01014 Positioning with begin time: Mar 30, 2011 2:40:23 PM, starting record time: Mar 3
0, 2011 2:40:37 PM at extseqno 23, extrba 872.
** Run Time Messages **
Opened trail file /u01/GG/source/dirdat/rt000023 at 2011-03-30 14:40:39
Processed extract process graceful restart record at seq 23, rba 872.
2011-03-30 14:43:10 INFO OGG-01021 Command received from GGSCI: STOP.
* ** Run Time Statistics ** *
Last record for the last committed transaction is the following:
Trail name : /u01/GG/source/dirdat/rt000023
Hdr-Ind : E (x45) Partition : . (x04)
UndoFlag : . (x00) BeforeAfter: A (x41)
RecLength : 30 (x001e) IO Time : 2011-03-30 14:41:50.000095
IOType : 5 (x05) OrigNode : 255 (xff)
TransInd : . (x03) FormatType : R (x52)
SyskeyLen : 0 (x00) Incomplete : . (x00)
AuditRBA : 11498 AuditPos : 327552200
Continued : N (x00) RecCount : 1 (x01)
2011-03-30 14:41:50.000095 Insert Len 30 RBA 931
Name: DWH.TEST1
Reading /u01/GG/source/dirdat/rt000023, current RBA 1080, 0 records
Report at 2011-03-30 14:43:10 (activity since 2011-03-30 14:40:39)
No records were replicated.
Last log location read:
FILE: /u01/GG/source/dirdat/rt000023
SEQNO: 23
RBA: 1080
TIMESTAMP: Not Available
EOF: YES
READERR: 400
CACHE OBJECT MANAGER statistics
CACHE MANAGER VM USAGE
vm current = 0 vm anon queues = 0
vm anon in use = 0 vm file = 0
vm used max = 0 ==> CACHE BALANCED
CACHE CONFIGURATION
cache size = 512M cache force paging = 881M
buffer min = 64K buffer highwater = 4M
pageout eligible size = 4M
CACHE Transaction Stats
trans active = 0 max concurrent = 0
non-zero total = 0 trans total = 0
CACHE File Caching
disk current = 0 disk total = 0
disk caching = 0 file cached = 0
file retrieves = 0
CACHE MANAGEMENT
buffer links = 0 anon gets = 0
forced unmaps = 0 cnnbl try = 0
cached out = 0 force out = 0
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0
Cached Transaction Size Distribution
0: 0
< 4K: 0
4K: 0 0 | 16K: 0 0
64K: 0 0 | 256K: 0 0
1M: 0 0 | 4M: 0 0
16M: 0 0 | 64M: 0 0
256M: 0 0 | 1G: 0 0
4G: 0 0 | 16G: 0 0
64G: 0 0 | 256G: 0 0
1T: 0 0 | 4T: 0 0
16T: 0 0 | 64T: 0 0
256T: 0 0 |1024T: 0 0
QUEUE Statistics:
num queues = 15 default index = 0
cur len = 0 max len = 0
q vm current = 0 vm max = 0
q hits = 0 q misses = 0
queue size q hits curlen maxlen cannibalized
0 64K 0 0 0 0
1 128K 0 0 0 0
2 256K 0 0 0 0
3 512K 0 0 0 0
4 1M 0 0 0 0
5 2M 0 0 0 0
6 4M 0 0 0 0
7 8M 0 0 0 0
8 16M 0 0 0 0
9 32M 0 0 0 0
10 64M 0 0 0 0
11 128M 0 0 0 0
12 256M 0 0 0 0
13 512M 0 0 0 0
14 1G 0 0 0 0
================================================================================
CACHE POOL #0
POOL INFO group: rep2 id: p29967_BLOB
trans active = 0 trans concurrent (max) = 0
trans total = 0 (0 )
flag = 0x00000000
last error = (0=<none>)
Allocation Request Distribution
< 128B: 0
128B: 0 0 | 512B: 0 0
2K: 0 0 | 8K: 0 0
32K: 0 0 | 128K: 0 0
512K: 0 0 | 2M: 0 0
8M: 0 0 | 32M: 0 0
128M: 0 0 | 512M: 0 0
2G: 0 0 | 8G: 0

Similar Messages

  • Using a treemap for lookup in a jsp grid

    Hi
              I am trying to use a treemap to lookup the names for id´s in a table.
              for example.
              Lookup table:
              ID.....Name
              1......carrot
              2......bananna
              the main table looks like this:
              row....id
              1.......2
              2.......1
              3.......1
              4.......2
              in my jsp page I want it to look like this:
              row....id
              1......bananna
              2......carrot
              3......carrot
              4......bananna
              I have the lookup table loaded into a treemap in the begin part of my pageflow.
              is there a special way that I can refer to the treemap in a grid view, for example, with a netui-data:expressionColumn ?
              thanks for any help.
              /Nigel
              Message was edited by NigelGuest at Feb 3, 2005 2:34 AM

    hi,
              important here to aknowledge that the {} databinding are compile time. This implies a whole lot of restrictions. The scenario below works for me:
              - getLookupMap in the pageflow returns the lookup map
              - getList in the pageflow returns the list
              The following code then works for me
              <!-- in the heading of my JSP -->
              // get the lookup map into pagecontext
              <netui-data:getData resultId="lookup"
              value="{pageFlow.lookupMap}"/>
              // pull down as a new page attribute
              <%
              Map lookup = (Map)pageContext.getAttribute("lookup");
              %>
              Now, in my repeater ( of gridview ) I implement the column showing the actual mapped value like this:
              // to get the item key
              <netui-data:getData resultId="key"
              value="{container.item}"/>
              // to set the mapped attribute value
              <%
              pageContext.setAttribute(
              "lookupValue",
              lookup.get(pageContext.getAttribute("key")));
              %>
              // to render the mapped attribute value
              <netui:label value="{pageContext.lookupValue}"
              defaultValue=" " />
              Probably not exactly what you asked for, but hopefully close enough to guide you in the right direction.
              - Anders M.

  • Use of Infoview for BI Queries

    We are in the early stages of deciding how to position BOBJ in relation to our existing BI 7.0 installation. We have developed a number of queries through Business Explorer Query Designer. Is it possible to access these through Infoview? Or do we need to develop a Crystal Report, for example, to run against the query?

    Hi Peter,
    you can use the BW query as a source for the BusinessObjects client tools. here some examples:
    BusinessObjects and SAP - Overview
    /people/ingo.hilgefort/blog/2008/02/07/businessobjects-and-sap-part-i
    Crystal Reports and BI Queries
    /people/ingo.hilgefort/blog/2008/02/19/businessobjects-and-sap-part-2
    Crystal Reports and BI Hierarchies
    /people/ingo.hilgefort/blog/2008/02/27/businessobjects-and-sap-part-3
    Crystal Reports and SAP R/3`
    /people/ingo.hilgefort/blog/2008/03/23/businessobjects-and-sap-part-4
    Web Intelligence and SAP BI
    /people/ingo.hilgefort/blog/2008/03/24/businessobjects-and-sap-part-5
    In case the idea was to use BusinessExplorer web reporting launched from InfoView, then this would lead to launching the URL but you still would get asked for the SAP credentials as InfoView is not generating a SSO token to be consumed by the BEx web reporting runtime.
    Ingo

  • Create key mapping using import manager for lookup table FROM EXCEL file

    hello,
    i would like create key mapping while importing the values via excel file.
    the source file containing the key, but how do i map it to the lookup table?
    the properties of the table has enable the creation of mapping key. but during the mapping in import manager, i cant find any way to map the key mapping..
    eg
    lookup table contains:
    Material Group
    Code
    excel file contain
    MatGroup1  Code   System
    Thanks!
    Shanti

    Hi Shanti,
    Assuming you have already defined below listed points
    1)  Key Mapping "Yes" to your lookup table in MDM Console
    2) Created a New Remote System in MDM console
    3) proper rights for your account for updating the remote key values in to data manager through import manager.
    Your sample file can have Material Group and Code alone which can be exported from Data Manager by File-> Export To -> Excel, if you have  data already in Data Manager.
    Open your sample file through Import Manager by selecting  the remote system for which you want to import the Key mapping.
    (Do Not select MDM as Remote System, which do not allows you to maintain key mapping values) and also the file type as Excel
    Now select your Soruce and Destination tables, under the destination fields you will be seeing a new field called [Remote Key]
    Map you source and destination fields correspondingly and Clone your source field code by right clicking on code in the source hierarchy and map it to Remote Key if you want the code to be in the remote key values.
    And in the matching criteria select destination field code as a Matching field and change the default import action to Update NULL fields or UPDATED MAPPED FIELDS as required,
    After sucessfull import you can check the Remote Key values in Data Manager.
    Hope this helps
    Thanks
    Sowseel

  • How to use different iViews for variables and queries?

    Hi all!
    I need to use separtate iViews for storing variables and extracting queries.
    For example, I have one variable "Currency" for two different queries,  and I need that all of these queries (in separate iViews) would use the value of currency was put by user in the first iView with only one input.
    How can I do this task? Is it possible at all?
    Thanks for replies.

    Hello Alex,
    You have to use EP client framework (EPCF) in your development.
    I hope this idea will help you.
    Kind regards, Pino.

  • Capture @@ROWCOUNT for all QUERIES using SQL AUDIT

    I have a requirement where the customer wants to audit all SELECT queries made to a specific table  and also capture
    "No of rows returned" for these queries made to the table. I was able to capture various SELECT queries happening in the database /Table using SQL AUDIT, Wondering if anybody can suggest how to capture no of rows affected along
    with it, Since we have numerous stored procedures in the system we wont be able to modify existing stored procedures.

    Good day Vish_SQL,
    There are several options that you can use, which fit different cases. for example:
    1. Using extended events (My prefered solution for most cases like this)
    2. Using profiler (older option)
    3. Using view (with the same name as the table, and rename the tables related to the issue) instead of the original table, adding to the view a simple function or SP (CLR for example). the function return the
    column value, but behind the  screen write information to another table. It is a very rare case that you need this option, and I can't recommend
    it.
    4. using applications like GREENSQL which give another level between the application and the SQL Server. The users connect to the server but the server do not listen to remote connection but the external app dose.
    In this case the app get the query and send it t the server (after security or monitoring if you need)
    * it was much simpler if you want to monitor delete, update, or insert for example, since in those cases you could work with AFTER trigger.
    ** I highly recommend you to monitor the application that connect to the SQL Server rather than the SQL Server. if you can.
    Check this:
    http://solutioncenter.apexsql.com/auditing-select-statements-on-sql-server/
      Ronen Ariely
     [Personal Site]    [Blog]    [Facebook]

  • Read data from MDM For Lookup and Flat table using MDM ABAP API

    Hi,
    I have requriment to read data from MDM from FLAT and Lookup table using MDM ABAP API. My design  is like this ,
    I have one ITEMS (Main table in MDM) and inside that i have one Lookup flat table ITEM_TYPE , my requriment is to read Item number and its related Item type.
    From ABAP.
    Please help if any body has any idea.
    Regards,
    Shyam

    HI Guys,
    I found my solution by myself. Below is the solution , hope this will help others:-
    Retrieve data from MDM  using MDM ABAP API.
    Step- 1. Create structure in SAP with the same name as that of MDM field code for MDM Main table.
    Step-2. Create another structure in SAP having all  lookup fields of MDM , fieldname in ECC must be same as that of MDM field
    code.
    Step-3.Create structure in SAP for  individual lookup field(Single Field only)   with the same name as MDM Field code.
    Step-4.
    DATA: IT_QUERY            TYPE STANDARD TABLE OF MDM_QUERY,  "MDM_QUERY_TABLE,
          WA_QUERY            TYPE  MDM_QUERY,
          WA_CDT_TEXT         TYPE  MDM_CDT_TEXT,
          IT_RESULT_SET_KEY   TYPE  MDM_SEARCH_RESULT_TABLE,
          WA_RESULT_SET_KEY   TYPE  MDM_SEARCH_RESULT,
          WA_STRING           TYPE  STRING.
    DATA:<Internal table> TYPE STANDARD TABLE OF <SAP Str Having all LOOKup Fields>    
    DATA: :<Internal table>TYPE STANDARD TABLE OF <SAP Str one LOOKup field>,
         <Workarea> LIKE LINE OF :<Internal table>.
    *PASS LOGICAL OBJECT NAME.
    V_LOG_OBJECT_NAME = 'Logical object name defined in Customization'.
    Define logon language, country & region for server
    WA_LANGUAGE-LANGUAGE = 'eng'.
    WA_LANGUAGE-COUNTRY = 'US'.
    WA_LANGUAGE-REGION = 'USA'.
    TRY.
        CREATE OBJECT LR_API
          EXPORTING
            IV_LOG_OBJECT_NAME = V_LOG_OBJECT_NAME.
    ENDTRY.
    CONNECT to repository. Apply particular logon language info
    CALL METHOD LR_API->MO_ACCESSOR->CONNECT
      EXPORTING
        IS_REPOSITORY_LANGUAGE = WA_LANGUAGE.
    *NOW PASS ITEM NO AND GET KEY FROM MDM.
    CLEAR WA_QUERY.
    WA_QUERY-PARAMETER_CODE  = <MDM FIELD CODE>. "Field code
    WA_QUERY-OPERATOR        = 'EQ'. "Contains
    WA_QUERY-DIMENSION_TYPE  = 1. "Field search
    WA_QUERY-CONSTRAINT_TYPE = 8. "Text search
    WA_STRING                = <Field Value>.
    GET REFERENCE OF WA_STRING INTO WA_QUERY-VALUE_LOW.
    APPEND WA_QUERY TO IT_QUERY.
    CLEAR WA_QUERY.
    *PASS ITEM NUMBER AND GET RELATED KEY FROM MDM.
    TRY.
        CALL METHOD LR_API->MO_CORE_SERVICE->QUERY
          EXPORTING
            IV_OBJECT_TYPE_CODE = <MDM Main Table>
            IT_QUERY            = IT_QUERY
          IMPORTING
            ET_RESULT_SET       = IT_RESULT_SET_KEY.
      CATCH CX_MDM_COMMUNICATION_FAILURE .
      CATCH CX_MDM_KERNEL .
      CATCH CX_MDM_NOT_SUPPORTED .
      CATCH CX_MDM_USAGE_ERROR .
      CATCH CX_MDM_PROVIDER .
      CATCH CX_MDM_SERVER_RC_CODE .
    ENDTRY.
    Pass record id into keys.
    LOOP AT IT_RESULT_SET_KEY INTO WA_RESULT_SET_KEY.
      WA_KEYS = WA_RESULT_SET_KEY-RECORD_IDS.
    ENDLOOP.
    WA_RESULT_SET_DEFINITION-FIELD_NAME = <Look field name>.
    APPEND WA_RESULT_SET_DEFINITION TO IT_RESULT_SET_DEFINITION.
    CALL METHOD LR_API->MO_CORE_SERVICE->RETRIEVE
      EXPORTING
        IV_OBJECT_TYPE_CODE      = <MDM Main Table>
        IT_RESULT_SET_DEFINITION = IT_RESULT_SET_DEFINITION
        IT_KEYS                  = WA_KEYS
      IMPORTING
        ET_RESULT_SET            = IT_RESULT_SET.
    LOOP AT IT_RESULT_SET INTO
            WA_RESULT_SET.
    *PASS KEYS INTO MAIN TABLE TO GET Structure for FALT or Look up Table
      TRY.
          CALL METHOD LR_API->MO_CORE_SERVICE->RETRIEVE_SIMPLE
            EXPORTING
              IV_OBJECT_TYPE_CODE = <MDM Main Table>
              IT_KEYS             = WA_KEYS
            IMPORTING
              ET_DDIC_STRUCTURE =<SAP Strct having all Look up fileds of MDM>         
      ENDTRY.
      LOOP AT <SAP Strct having all Look up fileds of MDM> INTO <Work area>.
        CLEAR WA_KEYS.
        APPEND <Work area>-field name TO WA_KEYS.
        CALL METHOD LR_API->MO_CORE_SERVICE->RETRIEVE_SIMPLE
          EXPORTING
            IV_OBJECT_TYPE_CODE = <MDM Lookup table name>
            IT_KEYS             = WA_KEYS
          IMPORTING
            ET_DDIC_STRUCTURE   = <Single Structure in SAP For Lookup field>.
        READ TABLE <Single Structure in SAP For Lookup field>. INTO <Work Area> INDEX 1.
    Here you can get the value of realted lookup fields associated with main table data.
      ENDLOOP.
    ENDLOOP.
    LR_API->MO_ACCESSOR->DISCONNECT( ).
    Edited by: Shyam Babu Sah on Nov 24, 2009 4:52 AM

  • Can I use same result set for two queries?

    Hi,
    Can I use the same result set for two queries?
    For example:
    ResultSet rs = null;
    rs = ps.executeQuery(query1);
    while (rs.next()) {
    rs = ps.executeQuery(query2);
    while (rs.next()) {
    Is it OK to use? or Is there any preformance issues?
    Appreciate your help.
    Thanks in advance.
    Prasad Vagolu

    Sure. You have a variable of ResultSet type. First you create a ResultSet and assign it to that variable. Then later you create another ResultSet and assign it to that same variable. Assigning an object reference to a variable takes essentially no time, and at any rate it isn't any faster to assign it to a different variable. Also, no matter whether you use one variable or two, you are still creating two ResultSets.
    However, you really should be closing your ResultSets after you finish using them. That's nothing to do with the assigning-to-a-variable question.

  • OLE DB provider 'MSOLAP' cannot be used for distributed queries because the provider is configured to run in single-threaded

    Hopefully this will save somebody some trouble.
    Running 64bit Enterprise SQL and SSAS with Service pack 2 installed.
    Also running Proclarity so 32bit mode Reporting Services is running.
    When trying to create a linked server to my OLAP database I was continually getting the following Error:
    OLE DB provider 'MSOLAP' cannot be used for distributed queries because the provider is configured to run in single-threaded apartment mode. (Microsoft SQL Server, Error: 7308)
    Many posts suggested I select the "in Proc" check box under the olap provider, but this did not help.
    Finally, instead of using the IDE to create the linked server I used a script to call sp_addlinkedserver and used @provider='MSOLAP.3'.  This fixed the problem.
    If you have a more clear idea of why I was having the issue in the first place, feel free to let me know what you think.

    Try this thread:
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/f02a3921-7d0b-4038-97bb-9f17d381b345/linked-ssas-server-fails?forum=sqlanalysisservices
    Talk to me now on

  • Performace Issue using Crystal Report For enterprise and BEx Queries

    Hi all;
        We are generating the following error stack when trying to build a report on top of a BEX query using Crystal Report for Enterprise :
        |7C4F8ECE44034DB897AD88D6F98B028B3|2011 12 12 17:24:21.277|+0100|>>|E| |crj|20380|  56|ModalContext    | |2|0|0|0|BIPSDK.InfoStore:query|CHVXRIL0047:20380:56.174:1|-|-|BIPSDK.InfoStore:query|CHVXRIL0047:20380:56.174:1|Cut2PbOe3UdzgckPBHn8spEab|||||||||com.crystaldecisions.sdk.occa.infostore.internal.InfoObjects||Assertion failed: Java plugin for CommonConnection is not loaded.
    java.lang.AssertionError
         at com.businessobjects.foundation.logging.log4j.Log4jLogger.assertTrue(Log4jLogger.java:52)
         at com.crystaldecisions.sdk.occa.infostore.internal.InfoObjects.newInfoObject(InfoObjects.java:576)
         at com.crystaldecisions.sdk.occa.infostore.internal.InfoObjects.continueUnpackHelper(InfoObjects.java:548)
         at com.crystaldecisions.sdk.occa.infostore.internal.InfoObjects.continueUnpack(InfoObjects.java:489)
         at com.crystaldecisions.sdk.occa.infostore.internal.InfoObjects.startUnpack(InfoObjects.java:464)
         at com.crystaldecisions.sdk.occa.infostore.internal.InternalInfoStore$XRL3WireStrategy.startUnpackTo(InternalInfoStore.java:1484)
         at com.crystaldecisions.sdk.occa.infostore.internal.InternalInfoStore$XRL3WireStrategy.startUnpackTo(InternalInfoStore.java:1464)
         at com.crystaldecisions.sdk.occa.infostore.internal.InternalInfoStore.unpackAll(InternalInfoStore.java:910)
         at com.crystaldecisions.sdk.occa.infostore.internal.InternalInfoStore.queryHelper(InternalInfoStore.java:944)
         at com.crystaldecisions.sdk.occa.infostore.internal.InternalInfoStore.queryHelper(InternalInfoStore.java:929)
         at com.crystaldecisions.sdk.occa.infostore.internal.InternalInfoStore.query_aroundBody24(InternalInfoStore.java:798)
         at com.crystaldecisions.sdk.occa.infostore.internal.InternalInfoStore.query(InternalInfoStore.java:1)
         at com.crystaldecisions.sdk.occa.infostore.internal.InfoStore.query_aroundBody20(InfoStore.java:175)
         at com.crystaldecisions.sdk.occa.infostore.internal.InfoStore.query_aroundBody21$advice(InfoStore.java:42)
         at com.crystaldecisions.sdk.occa.infostore.internal.InfoStore.query(InfoStore.java:1)
         at com.businessobjects.mds.securedconnection.cms.services.olap.OlapCmsSecuredConnectionService.getConnectionObject(OlapCmsSecuredConnectionService.java:125)
         at com.businessobjects.mds.securedconnection.cms.services.olap.OlapCmsSecuredConnectionService.getOlapSecuredConnection(OlapCmsSecuredConnectionService.java:191)
         at com.businessobjects.mds.securedconnection.loader.internal.SecuredConnectionLoaderImpl.getOlapConnectionFromSecuredConnection(SecuredConnectionLoaderImpl.java:83)
         at com.businessobjects.mds.securedconnection.loader.internal.SecuredConnectionLoaderImpl.getConnectionFromSecuredConnection(SecuredConnectionLoaderImpl.java:60)
         at com.businessobjects.dsl.services.workspace.impl.DirectOlapAccessDataProviderBuilder.loadSecuredConnection(DirectOlapAccessDataProviderBuilder.java:193)
         at com.businessobjects.dsl.services.workspace.impl.DirectOlapAccessDataProviderBuilder.loadSecuredConnection(DirectOlapAccessDataProviderBuilder.java:176)
         at com.businessobjects.dsl.services.workspace.impl.DirectOlapAccessDataProviderBuilder.provideUniverseFromCms(DirectOlapAccessDataProviderBuilder.java:63)
         at com.businessobjects.dsl.services.datasource.impl.AbstractUniverseProvider.provideUniverse(AbstractUniverseProvider.java:41)
         at com.businessobjects.dsl.services.workspace.impl.AbstractDataProviderBuilder.updateQuerySpecDataProvider(AbstractDataProviderBuilder.java:119)
         at com.businessobjects.dsl.services.workspace.impl.AbstractDataProviderBuilder.updateDataProvider(AbstractDataProviderBuilder.java:106)
         at com.businessobjects.dsl.services.workspace.impl.AbstractDataProviderBuilder.addDataProvider(AbstractDataProviderBuilder.java:49)
         at com.businessobjects.dsl.services.workspace.impl.WorkspaceServiceImpl.addDataProvider(WorkspaceServiceImpl.java:56)
         at com.businessobjects.dsl.services.workspace.impl.WorkspaceServiceImpl.addDataProvider(WorkspaceServiceImpl.java:45)
         at com.crystaldecisions.reports.dsl.shared.DSLTransientUniverseServiceProvider.createSessionServicesHelper(DSLTransientUniverseServiceProvider.java:72)
         at com.crystaldecisions.reports.dsl.shared.DSLServiceProvider.createSessionServices(DSLServiceProvider.java:428)
         at com.businessobjects.crystalreports.designer.qpintegration.DSLUtilities.getServiceProvider(DSLUtilities.java:279)
         at com.businessobjects.crystalreports.designer.qpintegration.InitializeDSLRunnable.run(InitializeDSLRunnable.java:82)
         at org.eclipse.jface.operation.ModalContext$ModalContextThread.run(ModalContext.java:121)
    Here seems to be that a plugin is not loaded : com.crystaldecisions.sdk.occa.infostore.internal.InfoObjects||Assertion failed: Java plugin for CommonConnection is not loaded.
    could this affect the performance of Crystal Reports for Enterprise and how could I fix this ?
    Best Regards
    Anis

    Venkat,
    Thanks for your response. Please note, however, the transaction RAD1 does not exist. Let me provide more details about the current settings of the InfoObject.
    The Characteristic is 'Item' (0CS_ITEM) and upon going to RSA1 >  Modeling > InfoObjects > Item (0CS_ITEM) > Right Click > Display > Business Explorer (tab) > Text Type is set to 'Long Text' and BEx description is set to 'Long description' already.
    When I run/execute the query with this Item characteristic, the results in BEx Analyzer is showing appropriate long text, however, Crystal Report for Enterprise shows short text only
    K
    Edited by: Kumar Pathak on Feb 3, 2012 6:18 PM

  • GoldenGate SQLEXEC

    Hi,
    If possible delete this thread mistakenly created twice
    I am not getting how to delete this entry
    Hi,
    I want to do transformations with GoldenGate + many to one - one to many replication.
    For this there are following methods as per my understanding :-
    1. table to table to mapping like which source_table.column to which Target_Table.column
    2. use the sql for transformation/ join query as it is in GG parameter file.
    To implement the second method as per my understanding we have SQLEXEC
    Is there any other way if no then could anybody please explain how to use sqlexec (if there is some example step by step with both configuration files it will be great.)
    I have tried myself pasting the contents below, but with this I am not able to get the updated value same time. I mean when I do second DML then the earlier updated value is reflected in target side.
    extract_
    extract srce110
    setenv (ORACLE_SID=GGDB)
    USERID ggsrc, PASSWORD oracle1234
    RMTHOST localhost, mgrport 7810
    RMTTRAIL ./dirdat/sb
    DDL INCLUDE ALL
    DDLOPTIONS ADDTRANDATA
    sequence src1.*;
    TABLE src1.stocksrc;
    replicat_
    REPLICAT trgr110
    setenv (ORACLE_SID=GGDB)
    ASSUMETARGETDEFS
    USERID ggsrc, PASSWORD oracle1234
    discardfile ./dirrpt/trgr110.dsc, append, megabytes 100
    -- MAP src1.seq1, TARGET src2.seq1;
    -- SQLEXEC " UPDATE SRC2.STOCKSRC SET STK_NAME = 'TEST' ";
    TABLE src1.stocksrc, SQLEXEC (ID MYid, QUERY " UPDATE SRC2.STOCKSRC SET STK_NAME = 'TEST' " , NOPARAMS ), TARGET src2.stocksrc;
    MAP src1.stocksrc, TARGET src2.stocksrc;
    TIA,
    VIKAS
    Edited by: Vikas Panwar on Oct 8, 2012 7:30 PM

    Hi Artem,
    Thanks for the reply but I don't think that defgen approach will work in case I want to execute big queries in GG config file.
    Let me explain little further, suppose I want to use joins and there are lot of such queries which I can obtain from current DB setup
    SO there are two options as per my understanding either I go DEFINITIONS approach and go for individual mapping or I use the available queries from DB as it is in GG config. So to start with second approach i.e. to understand SQLEXEC parameter I was trying Update query.
    HOPE u understood my problem
    TIA,
    Vikas

  • Need Different Selection screen for different Queries in a Workbook

    Hi,
    I have created a workbook with Multiple tabs in BI 7.0.  Each Tab has different Queries and each query has different Selection screens (Variable Selections).
    When i open the workbook and refresh it, the selection screen is appearing only for one query.  All the queries are refreshed by this single selection screen, though each query has different Variable selections.  What i need is a seperate selection screen i.e seperate Variable selection appearing for each queries, when i refresh each one of them.
    Is it possible to do this?  If anybody has tried this, help me in solving this issue.  Thanks for ur time.
    Regards,
    Murali

    Murali,
    If you un-check the 'Display Duplicate Variables Only Once' this WILL solve your problem.
    When you Refresh, you should be presented with a single variable selection dialog box, but it should contain an area for each Query (DataProvider) that is embedded in the Workbook.
    This is the case if the queries are all on the same tab, or on different tabs.
    However, if you have multiple tabs each with a query on it, each query must have it's own DataProvider. If all queries are based on the same DataProvider, it will not work as the Workbook only 'sees' one Query for which it needs variable input.
    If you REALLY want multiple variable selection dialog boxes, then maybe the best way to do this is to have the queries in separate Workbooks.
    If you don't want the User to have to open 5 queries manually, you could use a Macro in each Workbook that runs on opening, to open the next Workbook in the sequence.
    I hope this makes sense!
    Regards
    Steve

  • HT3354 how can i use one table for reference to another

    how can i use a table for a referance to another eg when i type a word in a cell, i will like it to match the word with another table then return the information in the cell i am using

    you can use vlookup() (or any of the lookup family of functions) to locate an item based on a key value:
    Here is an example of something you can do with two tables:
    The table on the right is title "Data" and stores a list of names with age and favorite color.
    The table on the left uses the value in the first column to lookup up information in the table Data
    in the table on the left:
    B2=IFERROR(A2&" is " & VLOOKUP(A2, Data :: A:D, 2, 0)&" years old and likes the color "& VLOOKUP(A2, Data :: A:D, 3, 0), "NOT FOUND")
    I know this look complicated.  so I'll break it up into smalled pieces:
    first the "&" is called the concatenate operator and joins two strings.  like this:
    a string is a set of characters between double quotes.
    so "string 1" & "string 2" becomes "string 1string2"  or "Sam " & "Jones" becomes "Sam Jones"
    you can use cell references instead of strings directly in which case the concatenation is performed on the contents of the cells.
    so if cell A1 contains "Hi " and the cell A2 contains "There"  then A1 & A2 will result in "Hi There"
    so you could add the formula
    A3=A1 & A2
    this is short hand for select cell A3 then type everything including the A3 so that A3 contains "=A1 & A2" (omit the double quote)
    OK.  So the formula I provided concatenates several items together:
    it concatenates A2, then the string " is " then a formula, then the string " years old and likes the color " then a formula
    the two formulas (highlighted in blue) perform a lookup of the value in cell A2 in columns A thru D of the table named "Data".  If if finds the value in cell A2 in the first column of the lookup range in the table Data (column A) then it returns the value from the same row but in the second or third column.
    all that is in a function calld iserror() to trap the condition where the calue you enter in A2 does not exist in the table Data:
    You will find the Numbers users guide and function reference helpful.  You can download then from Apple here:
    http://support.apple.com/manuals/#productivitysoftware

  • How do I use "Search Query" for nontrivial searches?

    After a few years in the community, this is my _first_ question to the boards ;-)
    In speaking with a colleague today, I learned about an NI forum feature that would be useful for me. When I tried to configure it, I then learned that I didn't know how to use it. When I tried to learn how to use it, I further learned that it wasn't fully documented. So, I'd like some clarification :-)
    The feature in question is "Search Query" which will send you an email when a forum search matches your query. What I don't know is:
    What are the rules (syntax) for search queries? For example, I would like to search for single words, all the words that I specify, as well as exact phrases. How do I differentiate between multiword queries and exact phrase queries?
    Is it possible to have more than one Search Query?
    If so, how do you delineate them? Would I use commas, a new line, something else?
    I'm asking because the available (and discoverable) documentation doesn't mention what to do when you want more than one Search Query. The help text on the subscription page [1] makes me think that only one query is possible, and says a Search Query:
    Sends you mail every time a message matching your query is submitted. Search queries take up to 15 minutes to take effect. Saving a blank query will disable this feature and will stop any mail from being sent.
    The other place I tried was the NI Forum FAQ [2], which makes me think that more than one query is possible:
    If you add a board, thread, message, or search term to your subscriptions, the system will send you an e-mail every time someone posts to the board, or replies to the message or thread. If you prefer not to receive an e-mail for your subscriptions, you can subscribe to the RSS feed of a board, thread, user, or search term.
    The FAQ goes on about what to do with board, thread, and message subscriptions, but doesn't say anything more about search subscriptions.
    The last place I looked was on the search results page itself. In the blue header that precedes the results, there's a link on the far left that says "Search Options" with a single option that says "Subscribe to this search's RSS Feed". While I could subscribe and use an RSS reader, I really would prefer email. I suppose if only one Search Query is possible, then this will be my workaround.
    As a comment, why does the Search Options menu not also have a "Subscribe to this search" option, which would add another entry in the Search Query box? This is a noticeably different interface than the Thread Options menu.
    [1] Subscriptions & Bookmarks
    http://forums.ni.com/ni/user_subscriptions
    [2] Frequently Asked Questions
    http://forums.ni.com/ni/help_faq
    Joe Friedchicken
    NI VirtualBench Application Software
    Get with your fellow hardware users :: [ NI's VirtualBench User Group ]
    Get with your fellow OS users :: [ NI's Linux User Group ] [ NI's OS X User Group ]
    Get with your fellow developers :: [ NI's DAQmx Base User Group ] [ NI's DDK User Group ]
    Senior Software Engineer :: Multifunction Instruments Applications Group
    Software Engineer :: Measurements RLP Group (until Mar 2014)
    Applications Engineer :: High Speed Product Group (until Sep 2008)
    Solved!
    Go to Solution.

    Hi Joe,
    Thanks for all of the questions and congrats on posting your first question!
    Unfortunately I think all of your questions come from the valid assumption that the search query functionality is more robust than it actually is.  The search query functionality really only serves a single purpose and that is to email you when your phrase has been posted to the forums.  For me, I use it to email me when my name is mentioned in the forums.  It serves this purpose well since I have it set up to work off just a single search term.  However I find that multiple word phrases are combined with an OR which is not as useful.  The search query does not adhere to any common search syntax that you would expect (I believe we are the only community that uses this feature so it has not been improved upon).
    The best way to set up both multiple word queries and multiple different queries is to use the RSS feed as you mentioned.  This way you can tailor your search very specifically using different advanced search options and subscribe to them all in one feed reader without constantly getting emails.  The search query emails do not adhere to the digest subscription settings.  I find that the RSS feeds work very well for searches and I am fond of using them, but I do use an RSS Reader for many things so it is not out of my normal workflow.
    Thank you for reading the Forum FAQ.  I've done some work to improve it but I think it has a long way to go and it's good to know that people read it
    Regards,
    Laura
    Web Support & Operations
    National Instruments

  • To use one SessionInfo for multiple frames

    how to use one SessionInfo for master frame and other frame the detail
    Thanks for your reply.

    You can instantiate the SessionInfo and other dataproducer objects in frame1 and bind them to controls in the first and second frames.
    If you plan to change/execute queries pass the sessionInfo instance to the second frame.
    Hope this helps.
    Sathish
    null

Maybe you are looking for

  • How can I collapse all my sub-folders when bookmarking a page?

    In Safari when I go to bookmark a page all my sub-folders and their own sub-folders are shown, Is there a way to collapse them all into just the main folders and then being able to choose the correct sub-folder from them to bookmark the page. I have

  • How to post a table/set of records without losing formatting while posting

    how to post a table/set of records without losing formatting while posting in this message forum. Thanks

  • Payment term

    Hi GURU I want to get the report for purchasing payment terms for all of our creditors. can any one tell me the table are tcode kind regards sunitha

  • Splitting MIDI?

    Is it possible to split a MIDI file, to 2 arrays: one with the notes, and one with the duration of the notes? We have a project on school, we need to load in a MIDI file, then get the note that the player is playing, and match it with an image. Anyon

  • Setting in for slate / clapboard

    i am new to premiere and want to set the cursor to the exact point where the clapboard peaks on the audio but when I go to click it goes to the frame before or after, in Final Cut I had to hold shift to get the exact point. how do I do this in premie