R&R issue with extracts

Hi All,
We have an issue with the R&R demon when creating a new extract. The queue AC_Extract holds with a DDIC_WRITE error. After deleting it and releasing the queue, the extract will eventually end succesfully. We have the same problem with all MSA sites. What does the error mean and what can we do to prevent it?
Thanks!

Hi,
isn't it SPE_DDIC_WRITE which causes the problem?!
Then in case you are running 5.0 please check note 765953.
Regards,
Wolfhard

Similar Messages

  • Issue with extracting node values of an XMLElement with multiple namespaces

    I have created a Table with XMLType Column
    CREATE TABLE xmlNode
    ( ID numeric(10) primary key,
    DATA XMLType);
    Then inserted one row
    INSERT INTO xmlNode(ID, DATA) VALUES(1, '<lineItemElement orderLineItemID="12323" transactionType="New" xmlns="http://rcss.bell.ca/schema/lineitemmessages">
    <rcss:catalogeID xmlns:rcss="http://rcss.bell.ca/schema/PartnerBatch">3234</rcss:catalogeID>
    <rcss:venueID xmlns:rcss="http://rcss.bell.ca/schema/PartnerBatch">345345</rcss:venueID>
    <rcss:startDate xmlns:rcss="http://rcss.bell.ca/schema/PartnerBatch">2007-09-10T00:00:00Z</rcss:startDate>
    <rcss:endDate xmlns:rcss="http://rcss.bell.ca/schema/PartnerBatch">2007-09-10T00:00:00Z</rcss:endDate>
    <rcss:contact contactID="234234" xmlns:rcss="http://rcss.bell.ca/schema/PartnerBatch">
    <rcss:address>
    <rcss:streetAddress1>?asdsd</rcss:streetAddress1>
    <rcss:streetAddress2>?sdss</rcss:streetAddress2>
    <rcss:city>?ddf</rcss:city>
    <rcss:province>?sdcfsdcf</rcss:province>
    <rcss:country>sds</rcss:country>
    <rcss:postalCode>1121332</rcss:postalCode>
    </rcss:address>
    </rcss:contact>
    </lineItemElement>')
    Then i ran:
    SELECT
    extractValue(data,'/lineItemElement/rcss:catalogeID','xmlns=http://rcss.bell.ca/schema/lineitemmessages, xmlns:rcss=http://rcss.bell.ca/schema/PartnerBatch')
    from xml where id=1
    There is nothing returned.
    Please advise how could I get the node value when there is namespace?
    I know how to extract values if there is one namespace but i'm not able to extract values if there are two namespaces.
    Any help is appreciable!!
    Thanks in Advance.

    I fixed that problem too. What I thought was a singleton was actually a repeating element. I needed to be able to "multiply" the repeating node for the main node, but I couldn't find a concise way of doing this in SQL. So instead I did it in with two loops.
    FOR rec IN (
    SELECT
         extractvalue(value(e), 'surchargeEntry/company',p_namespace) company
         ,extractvalue(value(e), 'surchargeEntry/surchargeType',p_namespace) surcharge_type
         ,extractvalue(value(e), 'surchargeEntry/shortDescription',p_namespace) short_description
         ,extractvalue(value(e), 'surchargeEntry/longDescription',p_namespace) long_description
         ,e.column_value surcharge_entry_xml
         FROM TABLE(xmlsequence(extract(p_xml, 'surchargeInquiryResponse' ||
         '/surchargeAttributes/surchargeEntry'
                                                                                    ,p_namespace))) e
    ) LOOP
    FOR rec_country IN (
    SELECT
         extractvalue(value(c), 'country/originationCountry',p_namespace) origination_country
         ,extractvalue(value(c), 'country/destinationCountry',p_namespace) destination_country
         FROM TABLE(xmlsequence(extract(rec.surcharge_entry_xml, 'surchargeEntry/country',p_namespace))) c
    LOOP
    nafta_fee_obj.append_surcharge_entry_data(
              v_arr
              ,r_surcharge_entry(
              rec.company
              ,rec.surcharge_type
              ,rec.short_description
              ,rec.long_description
              ,rec_country.origination_country
              ,rec_country.destination_country
    END LOOP;
    END LOOP;
    I populate a collection, which is easy to pass around between routines. I can easily convert the collection to a ref cursor, and index-by array, comma-delimited string, etc.
    Apparently, the namespace is required in SQL, but not in staright PL/SQL.
    I've also noticed that the docs say xmltype() has all these member functions, but when I try to use some of them, I get an error that it's undefined. Like getNumVal(). There are some functions I can use in SQL, or PL/SQL, but not vice-versa.
    - Dan Clamage

  • Issue with extraction of data to Delta Queue

    Hi All,
    Background:- One LO Data source each from Application 8, 11 and 12 was active in LBWE. Update method was specified as Queued Delta. However this Data source was not utilized to pull any data to any of the Global BI Systems. These data sources did not exist in RSA7. There was a business requirement that these Data sources were to be enhanced, along with this other Data sources under these applications were also required to be enhanced.  There were entries in LBWQ which used to get dumped out at regular interval by running the SAP Standard Collector jobs. Since no Delta was initialized for these extractors, no entries were passed on to RSA7.
    Problem:- We did not clear the entries or took a downtime before sending transports and now the collector job is dumping saying the structures have changed, which is correct.  I believe we wonu2019t require any data sitting in LBWQ as none of the extractors had any entry in RSA7. This problem is happening only in our Production system and we did not face this problem in any other system even though there were entries existing in LBWQ.
    What is the solution to fix this? Is the only option now is to delete the Queue from LBWQ and then run the collector job or is there any other way and will that fix the problem?

    Hi Victor,
    As u said that u were not using the DS for Application 8,11 and 12.. in this case I believe that u dont have any data in BW as well(Historical data). As u have enhanced these data sources with the new requirement, I hope u must have changed the structures in BW side as well to load the data to the newly added fields.
    So in ur case u need to replicate the DS in BW and then activate the Transfer structure for all of them.
    And then flushout the LBWQ for all those 3 applications and then take the setup tables fill for the same (but this requires the down time in prod to make sure that u wont loose any records).
    Before setup check SM13 whether u have any blocked entries for all these entries if so then u can delete them as well before taking the setup.
    Once the setup is done then u can take Init without data transfer and then take repair full loads in BW for the targets.(Activate ur V3 once the init- without data transfer took place for the regular deltas) .
    Thanks
    Assign points if this helps

  • Issue with AQ notification callback.

    Hi,
    I'm using OCCI API to subscribe for event notification.
    I was able to open subscription and the callback is triggered when the messages are enqueued from DB.
    I'm having issues with extracting the data in the callback.
    From the oracle doc, I see that only RAW payload is supported if the queue is non-persistent. Different payloads are supported if the queue is persistent and that the user is supposed to explicitly dequeue messages since the messages are not provided as part of NotifyResult argument of the callback.
    My question is how do I dequeue messages when the callback is triggered.
    I tried to use the consumer with in the callback as below:
    Consumer consumer(_connection);
    consumer.setConsumerName("fbe_ac1_usr");
    consumer.setDequeueMode(oracle::occi::aq::Consumer::DEQ_REMOVE);
    consumer.setQueueName("xrfdbo.fbe_ac1_queue");
    Message mes = consumer.receive(Message::OBJECT, "CONSOLIDATED_UPDT_MSG_T", "XRFDBO");
    PObject* obj = mes.getObject();
    I'm getting the exception "21301ORA-21301: not initialized in object mode"
    I believe this is due to Environment being created in EVENTS mode to register for subscription.
    If I change the mode to OBJECT when I create Environment, I'm getting exception "ORA-01010: invalid OCI operation" while registering for subscription.
    Is there any other way to dequeue DB type messages when callback is triggered ?
    In order to create a consumer within the callback, I changed my Connection object pointer as static. Is this the right way ?
    Thanks

    Environment modes are cumulative. You need to define both object and events mode at the same time, which you can do with a bitwise OR of EVENTS and OBJECT.
    You are spot on about non-persistent queues. They are deprecated (and slightly buggy, you get the raw payload as an XML document if you dereference the payload pointer in the notification callback function, but I find things go missing, like timestamps for instance), it's replaced by a new kind of persistent queue that behaves similarly to NP queues. I forget the marketing name now, let me go look it up and update the thread.
    EDIT: buffered queues.
    http://docs.oracle.com/cd/B19306_01/server.102/b14257/aq_intro.htm#sthref76
    It's a persistent queuing but the messages live in shared memory until the streams_pool_size is reached. I could be wrong, but I think Oracle built on this to provide continuous query notification.

  • Issue with generic extraction (Generic Delta)

    I have one more issue with Generic Delta Extraction. Here I selected the field related to my requirement but when I am trying to save I am getting error Still OLTP have errors.
    I need some help when we select generic delta wht are the settings. Like now I am working on Inventory Management.
    Please Some one help me as soon as possible.
    Thanks In advance.....
    Regds
    SDR.

    Hi ,
    I think there are some of the Keyfigure fields which needs 0Unit or 0Currency field to be present in the Extract Structure but they are not present . Check All unit of measure and Currency unit fields are not hidden .
    Regards,
    Vijay.

  • Issue with Data Extraction from OAGIS 9 User Area Tags ...

    We are working on displaying the data coming from OAGIS9 into Excel sheet. We created a custom Schema for the Excel sheet and are able to display all the data in OAGIS9 except for the ones in UserArea. We are facing issue in extracting data from tags present in UserArea of OAGIS 9 schema and transforming it into our Custom built schema.
    Can you please suggest us some way in which we extract data from userArea?
    Thanks in advance ~

    Here is the example with instructions, we have just done it and working
    1.You need to have the custom schema that is coming with in the <UserArea>/<any>
    For example this custom schema is like this one shown below and you are using this CutomerPartyMaster of OAGIS
    <?xml version="1.0" encoding="utf-8"?>
    <xs:schema xmlns="http://www.userarea.com/mdm/oagisextensions" elementFormDefault="qualified" targetNamespace="http://www.userarea.com/mdm/oagisextensions" xmlns:xs="http://www.w3.org/2001/XMLSchema">
    <xs:element name="Customer" type="CustomerType" />
    <xs:complexType name="CustomerType">
    <xs:sequence>
    <xs:element minOccurs="0" name="Gender" type="xs:string" />
    <xs:element minOccurs="0" name="DOB" type="xs:date" />
    <xs:element minOccurs="0" name="SSN" type="xs:string" />
    <xs:element minOccurs="0" name="PAN" type="xs:string" />
    </xs:sequence>
    </xs:complexType>
    </xs:schema>
    and the Input data snippet will be like this inside CustomerPartyMaster
    <tnsa:UserArea>
         <tnso:Customer>
              <tnso:Gender>string</tnso:Gender>
              <tnso:DOB>1980-10-13</tnso:DOB>
              <tnso:SSN>string</tnso:SSN>
              <tnso:PAN>string</tnso:PAN>
         </tnso:Customer>
    </tnsa:UserArea>
    where xmlns:tnsa="http://www.openapplications.org/oagis/9" and xmlns:tnso="http://www.userarea.com/mdm/oagisextensions"
    Lets assume your Target Schema is like this where only DOB is coming from userarea
    <CreateRequest>
    <record>
    <firstname></firstname>
    <lastname></lastname>
    <dob></dob>
    </record>
    </CreateRequest>
    2.declare the name space of incoming userarea custom schema in your xsl file like this
    xmlns:nsext="http://www.userarea.com/mdm/oagisextensions"
    3.here is xsl code that copies element from CustomerPartyMaster/UserArea
    <xsl:template match="/">
    <imp1:CreateRequest>
    <xsl:for-each select="/inp1:ESBProcessCustomerRequest/body/ns4:ProcessCustomerPartyMaster/ns4:DataArea/ns4:CustomerPartyMaster">
         <imp1:record>
    <xsl:for-each select="ns4:UserArea/nsext:Customer/nsext:DOB">
              <imp1:dob>
              <xsl:value-of select="."/>
              </imp1:dob>
         <imp1:record>
    </xsl:for-each>     
    where imp1 is target schema namespace
         ns4 is OAGIS standard namespace.
    *****************

  • Multiple issues with Creative Cloud File Syncing

    I'm have issues with syncing files on my mac to Adobe Creative Cloud. I don't seem to have issues syncing via drag and drop to my web browser. I keep getting this error while syncing via finder:
    I also have a fairly complex PSD with folders that i'm toggling on and off using Layer Comps. I don't see any difference when switching Layer Comps within Extract in the browser. I do see the changes when turning on and off the folders though.. What's the deal with that?
    Why does it take so long to render the PSD after I turn on a folder or switch layer comps? It takes about 30 seconds to re-render, which is crippling for our workflow.
    This would be an amazing tool for our developers if these 3 issues were resolved.

    Thanks for the information.
    Could you now send me some log files please..
    The log files can be found here:
    Mac: /Users/<yourusername>/Library/Application Support/Adobe/CoreSync/
    Windows: C:\Users\<yourusername>\AppData\Roaming\Adobe\CoreSync\
    The logs have the date in the filename, like "CoreSync-2014-03-25.log". Please compress (zip) all the CoreSync-2014-MM-DD.log files and email them to me directly at [email protected]
    Thanks
    Warner

  • Issues with using the output redirection character with newer NXOS versions?

    Has anyone seen any issues with using the output redirection character with newer NXOS versions?
    Am receiving "Error 0x40870004 while copying."
    Simply copying a file from bootflash to tftp is ok.
    This occurs for both 3CDaemon and Tftpd32 softwares.
    Have tried it on multiple switches - same issue.
    Any known bugs?
    thanks!
    The following is an example of bad (NXOS4.1.1b) and good (SANOS3.2.1a)
    MDS2# sho ver | inc system
      system:    version 4.1(1b)
      system image file is:    bootflash:///m9200-s2ek9-mz.4.1.1b.bin
      system compile time:     10/7/2008 13:00:00 [10/11/2008 09:52:55]
    MDS2# sh int br > tftp://10.73.54.194
    Trying to connect to tftp server......
    Connection to server Established. Copying Started.....
    TFTP put operation failed:Access violation
    Error 0x40870004 while copying tftp://10.73.54.194/
    MDS2# copy bootflash:cpu_logfile tftp://10.73.54.194
    Trying to connect to tftp server......
    Connection to server Established. Copying Started.....
    |
    TFTP put operation was successful
    MDS2#
    ck-ci9216-001# sho ver | inc system
      system:    version 3.2(1a)
      system image file is:    bootflash:/m9200-ek9-mz.3.2.1a.bin
      system compile time:     9/25/2007 18:00:00 [10/06/2007 06:46:51]
    ck-ci9216-001# sh int br > tftp://10.73.54.194
    Trying to connect to tftp server......
    |
    TFTP put operation was successful

    Please check with new version of TFTPD 32 server. The error may be due to older version of TFPT server, the new version available solved this error. Files are getting uploaded with no issues.
    1. Download tftpd32b.zip from:
    http://tftpd32.jounin.net/tftpd32_download.html
    2. Copy the tftpd32b.zip file into an empty directory and extract it.
    3. Copy the file you want to transver into the directory containing tftpd32.exe.
    4. Run tftpd32.exe from that directory. The "Base Directory" field should show the path to the directory containing the file you want to transfer.
    At this point, the tftpserver is ready to begin serving files. As devices request files, the main tftpd32 window will log the requests.
    Best Regards...

  • Performance Issues with large XML (1-1.5MB) files

    Hi,
    I'm using an XML Schema based Object relational storage for my XML documents which are typically 1-1.5 MB in size and having serious performance issues with XPath Query.
    When I do XPath query against an element of SQLType varchar2, I get a good performance. But when I do a similar XPath query against an element of SQLType Collection (Varray of varchar2), I get a very ordinary performance.
    I have also created indexes on extract() and analyzed my XMLType table and indexes, but I have no performance gain. Also, I have tried all sorts of storage options available for Collections ie. Varray's, Nested Tables, IOT's, LOB's, Inline, etc... and all these gave me same bad performance.
    I even tried creating XMLType views based on XPath queries but the performance didn't improve much.
    I guess I'm running out of options and patience as well.;)
    I would appreciate any ideas/suggestions, please help.....
    Thanks;
    Ramakrishna Chinta

    Are you having similar symptoms as I am? http://discussions.apple.com/thread.jspa?threadID=2234792&tstart=0

  • Performance issues with 0CO_OM_WBS_1

    We use BW3.5 & R/3 4.7 and encounter huge performance issues with 0CO_OM_WBS_1? Always having to do a full load involving approx 15M records even though there are on the average 100k new records since previous load. This takes a longtime.
    Is there a way to delta-enable this datasource?

    Hi,
    This DS is not delta enabled and you can only do a full load.  For a delta enabled one, you need to use 0CO_OM_WBS_6.  This works as other Financials extractors, as it has a safety delta (configurable, default 2 hours, in table BWOM_SETTINGS).
    What you should do is maybe, use the WBS_6 as a delta and only extract full loads for WBS_1 for shorter durations.
    As you must have an ODS for WBS_1 at the first stage, I would suggest do a full load only for posting periods that are open.  This will reduce the data load.
    You may also look at creating your own generic data source with delta; if you are clear on the tables and logic used.
    cheers...

  • Issue with 0hrposition master data

    We are extracting data from SAP using 0HRPOSITION_ATTR datasource. I noticed that data is not maintained correctly in master data tables in BW and giving us incorrect results in reporting. Consider below mentioned scenario:
    Position A created as vacant on 04/01/2006 with start date (BEGDA/ Valid from) as 04/01/2006 and End date (ENDDA/ Valid to) as 12/31/9999 and is vacant. Below mentioned entry is shown under maintain master data for 0HRPOSITION in BW:
    Position Valid To Valid From Position Vacant
    A 03/31/2006 01/01/1000
    A 12/31/9999 04/01/2006 X
    Position A is now delimited on 09/15/2006 as it’s no more required. In SAP, position has record only from 04/01/2006 till 09/15/2006 as vacant. When record is extracted in BW, it creates below mentioned entry in master data table.
    Position Valid To Valid From Position Vacant
    A 03/31/2006 01/01/1000
    A 09/15/2006 04/01/2006 X
    <b>A 12/31/9999 09/16/2006 X</b>
    Entry 09/16- 12/31 is incorrect as position doesn’t exist for this duration. If we report on 0HRPOSTION with key date as 09/30/2006, it shows position A as vacant though position no longer exists.
    Has anyone come across this situation. any help is greatly appreciated.
    Kamal
    P.S: Milind Rane...I was searching through the forums and came across your post.I would appreciate if you could let me know how you solved this issue...
    Message was edited by:
            Kamal K

    HI KK
    I have a similar issue. Can you please let me know how this Issue with 0HRPOSITION_ATTR extractor was resolved. In my case i have incorrect data in BW when reporting is done. THrough the extractor untill the PSA i have the correct records coming  but in the master data 0HRPOSITION i have incorrect records.
    Please help.
    Thanks
    Hari
    Message was edited by:
            SAPCOOL

  • Issue with 0HRPOSITION_ATTR extractor

    We are live with BW HR application and having issue with 0HRPOSITION master data. We are extracting data from SAP using 0HRPOSITION_ATTR datasource. I noticed that data is not maintained correctly in master data tables in BW and giving us incorrect results in reporting. Consider below mentioned scenario:
    Position A created as vacant on 04/01/2006 with start date (BEGDA/ Valid from) as 04/01/2006 and End date (ENDDA/ Valid to) as 12/31/9999 and is vacant. Below mentioned entry is shown under maintain master data for 0HRPOSITION in BW:
    Position Valid To   Valid From  Position Vacant
    A        03/31/2006 01/01/1000
    A        12/31/9999 04/01/2006        X
    Position A is now delimited on 09/15/2006 as it’s no more required. In SAP, position has record only from 04/01/2006 till 09/15/2006 as vacant. When record is extracted in BW, it creates below mentioned entry in master data table.
    Position  Valid To    Valid From  Position Vacant
    A         03/31/2006  01/01/1000
    A         09/15/2006  04/01/2006        X
    A         12/31/9999  09/16/2006        X
    Entry 09/16- 12/31 is incorrect as position doesn’t exist for this duration. If we report on 0HRPOSTION with key date as 09/30/2006, it shows position A as vacant though position no longer exists.
    Has anyone come across this situation. any help is greatly appreciated.
    Thanks,
    Milind

    Hi
    I have nt worked with this data source.
    1)Is your data source delta enabled?
    2)Try running RSA3 and check for record which is not coming fine. If in RSA3 record is fine then check calculations(if any ) in update/transfer rule.
    Let me know if it helps..
    Sorabh
    Assign point if it helps**
    Message was edited by: Sorabh Arora

  • Dell Laptops E7440 & E7240 - Issues with connecting to MDT 2012 deployment share

    Hi,
      Good afternoon.Need help on this please
    -Test laptops Iam using are 2 * E7440 and 1*E7240. One laptop of model E7440 has our image installed and other two (1*E7440 & 1*E7240) are shipped from Dell with default Win 7 64 bit OS.
    -Test desktops Iam using are Dell Optiplex 3020 & Dell 3010 . Both models have our own images of Win 7 64 Bit.
    -Laptops and desktops are on same network .
    -For OS deployments we have MDT 2012.
    -While there are no issues with desktops, all the laptops mentioned above don’t connect to MDT deployments share . The detailed error message I get when I run a litetouch ISO is ‘A connection to the deployment share could not be made. Connection OJ.Possible
    cause:invalid credentials. Retry : Try again to connect to the deployment share.’
    -Have injected all possible network & storage drivers mentioned by Dell &  Intel into MDT boot image but the problem still persists. Still WinPE doesn’t detect network adapter wherein IPaddresses are not picked in WinPE mode.
    -Have been testing each and every driver by manually loading them into Winpe mode to see if they pick up the IP addresses and I finally found one driver by name ‘e1d63x64.inf ‘ which I downloaded directly from Intel website. This one is picking IP Address
    in WinPE mode when loaded manually. Also have tested the connectivity to MDT server in WinPE mode with net use command and it was successful.
    Have injected this driver into MDT Laptop deployment share, regenerated the ISO and tried again however Iam still unable to connect to the deployment share with litetouch ISO.
    The above driver is very much part of set of network drivers on Dell website too.But it doesn’t work.
    -Have also tried combination of network & storage drivers from Dell website  (along with the above one mentioned ) injecting them into the MDT laptop boot image but no luck.
    -Every time I make changes to drivers in boot image,I update deployment share & recreate boot image .
    -This very litetouch iso connects to  MDT deployments share for Dell desktop models 3010 & 3020 but have issues with laptop models E7440 & E7240. This means any connectivity or access issues can be ruled out. Also no problem with our image cause
    this happens even with brand new shipped laptops with your inbuild win 7 OS.
    -Have tried all options suggested on internet,tested all the drivers suggested but no luck.
    -For example I tried latest E7440 netwrok & storage drivers ,Dell Win PE drivers,drivers from Intel website,etc  individually as well as in combinations but the issue persists.
    -If you google many express similar issues with these laptop models.
    Though I used the drivers suggested by various bloggers,nothing worked for me.
    Any help on this is appreciated.Thanks in advance.

    Hi,
    I have several E6420, O3010 and since a week also some E7440 and E7240 on site.
    I never had issues using the current Dell PE package for MDT.
    At the moment my PE contains only the drivers from the Dell PE 3.0 package (WinPE3.0-Drivers-A13-D67JC)
    and all my systems mentioned abobe have no issues connecting to the deployment share.
    (btw: I have also some older E6400, E6410 and O390 I do no longer support, they
    are supported by the pack Dell-WinPE-Drivers-A05.CAB according
    to the Dell site, but I just started a deplymenst on a E6410 and is is also workign with just the latest PE pack)
    I'd
    suggest you start over with your deplyoment share using at least MDT2012 Update 1.
    Create a folder "PE" under "Out-of-Box Drivers" and extract latest Dell PE pack and import he drivers there
    Create a selection profile e.g. "PE Drivers" and include just the "PE" folder created above
    When configuring the Windows PE  settings of your deployment share select "PE
    Drivers" selection profile for the driver injection in the "Drivers and Patches" to keep the PE image free of not needed drives.

  • Issue with creating a new line in Glossary widget

    I'm using Captivate 6 and have an issue with trying to display text on a new line in the description field of the Glossary widget. In playback, a carriage return displays as \n
    I read in another discussion (from early last year), that this issue had been resolved and that the fix would be shared soon on the captivate blog for non-subscription users. Does anyone know if the fix has been shared yet - I can't find it anywhere?
    Thanks for any help you can provide.
    Tenneil

    Hi Tenneil,
    Greetings from Adobe.
    Please click on the below link and download the updated Glosarry Interaction.
    https://files.acrobat.com/a/preview/0a17bb28-049e-4ac7-b8f7-dbdacef5e91c
    Extract the zip file to your desktop.
    Copy "08_Glossary.wdgt" from the extracted zip file.
    Go to : C:\Program Files (x86)\Adobe\Adobe Captivate 6 (32 Bit)\Gallery\Interactions (For 32 bit Captivate) or C:\Program Files\Adobe\Adobe Captivate 6 (64 Bit)\Gallery\Interactions (For 64 bit Captivate).
    Replace the glossary widget in the interaction folder. Restart Captivate 6. (Make sure that you take the backup of the original Glosarry Widget)
    Apply the widget again and check for the issue.
    Let me know if it works for you.
    Regards,
    Mayank

  • Issue with stock qty of component in Structured Articles in SAP BI

    Dear Experts,
    We are extracting stock in BI through SAP standard data source. We have used 2lis_03_bx and 2lis_03_bf data sources. We are facing issue with stock data validation due to structure articles (pre-pack) available in SAP source system.
    We have stock of structure articles available at DC in source system.  When we extract this information in BI, this article gets split into its components automatically. Due to which stock of components articles gets increased with huge amounts.
    To analysis it in further detail, we have created below scenario:
    Article Document Number:  4900000827
    Article Number (Pre-pack):  300000082901
    Article Number (Component): 300000082002
    Stock of pre- pack article at this article document number: 174 pc
    Stock of component article at this article document number: 348 pc (due to splitting of pre-pack article) + 151 (its own stock movement). Total stock for this article is 499. But actually total stock for this article should be 151 only.
    As per our Business requirement, pre-pack article should not be split into its components. And stock in BI should be totally in sync with SAP stock.
    I kindly request you to suggest the solution to rectify the same.
    Thanks in advance.
    --Vikas

    Hi Vikas,
    Check the OSS note L 1320386 - Frequently asked questions regarding BW Trade Foundation too.
    Regards,
    Rajesh

Maybe you are looking for

  • Why is the trigger triggering on all fields in a table?

    I have a trigger and the text is as follows. The trigger should just write the emplid to the table PS_DM_ADDR_CHG_EMP whenever there is an update to the fields STREET1,STREET2 ,STREET3 ,CITY ,STATE ,ZIP ,COUNTRY in PS_PERSONAL_DATA. But some how it i

  • Does the sound return after doing a system restore ?

    operating system windows 7 problem with sound

  • Calling stored procedure

    Hello Techies I want to call a stored procedure which does not return anything but it executes the query inside itself. How should I take the output from it �I want 3 values from that procedure. The stored procedure is::::::: CREATE Procedure NewPFAt

  • SQL Plus in Oracle 10g XE -

    How can i works in oracle 10g XE using the SQL Plus. When I try to connect, show-me and error.

  • Java on OS390 (MVS)

    Anybody here solved any of the issues of using Java on an MVS mainframe (running zOS in fact)? I'm specifically referring to: * Editing (square brackets) * Uploading .java and/or .class files from NT * Running within CICS - is there a 3270 display op