Issue when inserting Child Entity (Notes) with Contact

Hi,
I am getting the following error message when inserting Contacts with a child entity i.e. Notes which is quite misleading
SOAP fault from Item: OnDemandServicesObjMgr_eng_44477.log(SBL-DAT-00542)Method 'SetFieldValue' of business component 'Contact Note' (integration component 'Contact_Contact Note') returned the following error: "Access denied.(SBL-DAT-00542)"(SBL-EAI-04376
As suggested in earlier post I have tried providing values for fields NotesID and ContactID as well and even though it gives this error message.
Thanks
Arif

Hi, From the error message it looks like the access profile settings for the WS user doesn't have required privileges to create a note for contact. You make have to check access profile settings first and then try it
-- Venky CRMIT

Similar Messages

  • Update/Insert child entity record in siebel when a session is saved in WD

    Hi,
    We have a requirement where we need to create/update a child entity's record in siebel when user clicks on save session link in WD window after the rule execution. We are using IO mapping. Account is the parent entity and let's say B is the child entity.We are storing row_id of the new created B record in a field of Account BC(before the WD call we are creating a new B record).
    So, in the rule we need to update one field of B entity such that we can update the new created B record using EAI Siebel Adapter BS in the Policy Automation Save Session IO Post WF. We wrote like below.
    reason(B's field) = RejectionReason (Account's attribute) if
    ButtonOption = "Reject"
    and
    B's id = Account's field where B's row_id is stored for the new record
    but we found that we had to use rule table for this.
    We used the rule table like below(could not paste the table):
    rule_loop
    Reason
    RejectionReason     ButtonOption = “Reject”
    And
    B’s row_id = B’s row_id in Account
    Reason     otherwise
    It is setting B's Reason field for the particular record of B, but for rest of the B's records it is setting Reason to null although we do not want to change the Reason for rest of the B records.
    Could you please help us on this? Also how to navigate to the question screen directly if any attribute value is null? We added the screen flow but now it is directly showing the summary screen even if the attribute value is null.
    Thanks in advance!
    Regards,
    Subhradeep

    Hi Frank,
    Yes, the parent entity is "Account" and the child entity is a new entity(B).One Account can have multiple B records(1:M relationship).Before calling WD one B record is created for the Account and we are storing the row id of that record in Account BC.
    In WD screen we have dropdown where user can select "Reject" option for the product offered to him.
    user selects "Reject": in the next screen he will select the Rejection Reason and when he saves the session the Reason field of the new B record will be updated.
    We have created an IO to send data to OPA where Account is the primary BC.So if we can update the Reason of B in the rule(for the new B record) then we can directly use the IntegrationObject returned from OPA and use EAI Siebel Adapter BS to update that B record in Policy Automation Save Session IO Post WF.
    So we need:
    Update Reason for the new B record, do not update the Reason for rest of the B records.We had written in a rule table like below:
    rule_loop
    Reason
    RejectionReason     ButtonOption = “Reject” and B’s row_id = B’s row_id in Account
    Reason     otherwise
    (B’s row_id in Account, RejectionReason are global attributes for Account, Reason is the B's attribute)
    when we used the IntegrationObject in Siebel(after the save session) it is updating Reason of the correct B record(new one)...but for rest of the B records Reason becomes null(the otherwise condition). We do not need to update the Reason for the remaining B records.
    Please could you suggest us how to stop updating the Reason of the remaing B records in the rule itself?
    Thanks & Regards,
    Subhradeep

  • Performance issue when moves child component

    Hi
         I am using flex 3.5. When i moved a HBox(with 150 labels and textinputs), from container A to B, it takes lot of time. Any ideas Or tricks to improve the performance.
    Thanks,
    Sathyamoorthi.

    Hi Alex,
         After a long try on that big source code, i found something makes it bad app. There is a code like this.
        1.   canvas.addEventListener("updateComplete", UpdateCompleteHandler);
         private function UpdateCompleteHandler(event:Event):void
              (event.currentTarget as Canvas).height = ....
            2.
            override protected function updateDisplayList(unscaledWidth:Number, unscaledHeight:Number):void
                   AddEventListener();
            here numchildren would be around 250 fields from 7 panels.
            private function AddEventListener():void
                   for(var count:int = 0; count < this.numChildren; count ++)
                        (this.getChildAt(count) as compoent).setStyle('backgroundColor', 'red');
    The above two scenorios eat CPU time alot. when i fixed them everything comes back good.
    I don't say add child is working faster. but now it is good.
    My question is how can i find which component eats CPU time alot? There is only 7 instance in this case. But they ate CPU time always.
    Is there any ability to track CPU usage via profiler.

  • Multi-track duration issue when inserting clips

    I apologize if I missed this anywhere.  Tried to search first.
    I'm using Audition cs6 and I have a clip in my waveform viewer.  I have already created a new multi-track session that matches the specs of the source files.
    HOWEVER, when I make a selection on the source file and select "Insert into multitrack" and then select the session, it inserts it 14-hours and 22-minutes into the blank 30-second session.  This throws off the whole viewing slider and it becomes very difficult to manipulate the timeline.  If I simply drag the file from the browser and then re-trim it, it's fine, but not using the insert command.
    Why is this and is there a way to correct this?
    Any help would be wonderful
    -Todd
    Windows 7 64bit
    12GB RAM
    Production Premium CS6

    It sounds like the file has BWF Time Reference metadata embedded.  You can verify this by opening Window > Metadata and under the BWF tab, check the Time Reference field.  To disable this behavior, uncheck Preferences > Multitrack > Use embedded timecode when inserting clips into Multitrack.

  • Start background Job when another is finished (NOT with the JobSteps)

    Hi guys,
    i need your help.
    I've already searched here in forum, but i wasn't able to find a good solution.
    I have this problem.
    I have a program that create a background job with the FM FM JOB_OPEN.. SUBMIT report with parameters .. JOB_CLOSE.
    I want that if i run this report again (20 secs after for example) it does:
    - Check if there is an already running Job with the same name (means with state 'R') (This task is simply, with a select on TBTO table)
    - If it's found, it have to create a new job with the same name that starts automatically AFTER the first running job is finished (don't care about the end-state of the first job).
    I've already tried with pred_jobcount & PRED_JOBNAME parameters of the JOB_CLOSE FM but it doesen't works!
    The JOB_CLOSE, creates a job in Planned state. But when the first Job is finished, the second job(Planned) doesen't start automatically.
    In this scenario, i CANNOT use endless loops (wait until the first job is finished and then submit the second) Job-steps (one job that contains multiple steps) events (i have to start only one job after the predecessor is finished), because this report could be run many times and each job should be collected like a "stack" (only when the first job is finshed the second "registered" should be started and so on, until the aren't more planned jobs).
    <REMOVED BY MODERATOR>
    Thx a lot for your help.
    Andrea
    Edited by: Alvaro Tejada Galindo on Jun 12, 2008 12:19 PM

    Hi Veda
    i can tell u but ... some reward points are very appreciated ....
    I'm joking (of course)
    Here the question:
    I have a program (called A) that submit a new program (called B) with the JOB_OPEN .. submit JOB_CLOSE. The program B should start only if another program B (called before for example) is finished.
    Here the solution.
    I add a parameter (with no-display clausole) to the program B. In this parameter i pass to the program, the job number returned by the JOB_OPEN function. 
    When i create the JOB with the function, the "jobname" parameter is set with value 'G_DELIVERY' (Here u can change the name of the job as u want : this is the jobname that u see in SM37 transaction).
    In the start-of-selection of program B i put a "waiting" procedure like this :
    First i save a timestamp of system-date and system-time (called for example r_date and r_time)
    select from table TBTCO all the jobs called "G_DELIVERY" with jobnumber <> from the jobnumber parameter (that means exclude itselfs)  with status running ('R')  -> that have startdate /starttime less than the r_date and r_time <- (this is the key of the selection that solve the problem).
    if it is found (means there is another running job started before this one).
    wait up to 60 seconds. "for example
    repeat the selection.
    endif.
    When the job called before ends, this one programs exits from the loop and continues. If u submit more programs "B", they'll works like a stack.
    I should say to u just one thing.... I solved my problem in another way (because i've changed the logic so this problem was no more) so i didn't implemented that logic, but it should works very good.
    Try it and tell me!
    Bye
    Andrea

  • Strange PDC issue when inserting compressor on a group with side chain

    Hi Forum People!
    Seem to have hit a brick wall trouble shooting this issue, and hope someone here can help.
    Whenever I insert a compressor on a bus, and send a sidechain trigger to the compressor, PDC stops working for all multi-output synths. I can sidechain stereo tracks without an issue, but not group tracks with mutliple tracks routed through the groups.
    For the record, as soon as I remove the send to the sidechain, PDC works normally.
    Is this a limitation of Logic's PDC, or is there a workaround that I haven't discovered?
    Many thanks for your time
    Steve

    LOL! Sorry, didn't have time to respond yesterday. The problem was the way in which I'd set-up the group. Here's how I had the group set up, which broke PDC:
    I created an auxiliary track (the group), inserted a compressor, and routed my drums direct to this group.
    Here's how I fixed the problem:
    instead of creating an aux channel for the group, I routed my drum outputs through an unused bus, which in turn created the group channel automatically.
    I've only been using Logic for 1 month, having used Cubase for nearly 20 years, so realizing some things are done very differently.
    Thanks chaps!
    Steve

  • Performance issue when inserting into spatial indexed table with JDBC

    We have a table named 'feature' which has a "sdo_geometry" column, and we created spatial index on that column,
    CREATE TABLE feature ( id number, desc varchar, oshape sdo_gemotry)
    CREATE INDEX feature_sp_idx ON feature(oshape) INDEXTYPE IS MDSYS.SPATIAL_INDEX;
    Then we executed the following SQL to insert about 800 records into that table(We tried this by using DB visualizer and
    our Java application, both of them were using JDBC driver to connect to the oracle 11gR2 database) .
    insert into feature(id,desc,oshape) values (1001,xxx,xxxxx);
    insert into feature (id,desc,oshape) values (1002,xxx,xxxxx);
    insert into feature (id,desc,oshape) values (1800,xxx,xxxxx);
    We encoutered the same problem as this topic
    Performance of insert with spatial index
    It takes nearly 1 secs for inserting one record,compare to 50 records inserted per sec without spatial index,
    which is 50x drop in peformance when doing insertion with spatial index.
    However, when we copy and paste those insertion scripts into Oracle Client(same test and same table with spatial index), we got a totally different performance result:
    more than 50 records inserted in 1 secs, just as fast as the insertion without building spatial index.
    Is it because Oracle Client is not using JDBC? Perhaps JDBC was got something wrong when updating those spatial indexed tables.
    Edited by: 860605 on 19/09/2011 18:57
    Edited by: 860605 on 19/09/2011 18:58
    Edited by: 860605 on 19/09/2011 19:00

    Normally JDBC use auto-commit. So each insert can causes a commit.
    I don't know about Oracle Client. In sqlplus, insert is just a insert,
    and you execute "commit" to explicitly commit your changes.
    So maybe this is the reason.

  • JDBC issue when accessing an Entity Bean

    Hi,
    WLCS 3.1
    WLS 5.1 sp6
    Windows NT
    Oracle 8.1.6
    I am having problems with an Entity bean, I have originaly deployed and had it
    running on cloudscape. I decided I need to deploy it to Oracle 8.1.6, on startup
    of WLCS the bean was deployed to the connection pool correctly using Weblogic
    JDriver.
    The bit I cannot understand is when I try and do a FindByPrimaryKey on the Entity
    Bean the following exception is thrown by oracle, which leads me to believe this
    must be an Oracle/Jdbc problem.
    Here you can see that the bean has been deployed on a valid connection pool.
    Tue Sep 18 15:47:39 GMT+03:00 2001:<I> <EJB JAR deployment /export/home/mccann/tester2.jar>
    EJB home interface: 'com.testerHome' deployed bound to the JNDI name
    : 'tester'
    Heres the exception raised when a findByPrimaryKey is invoked on the bean....
    Tue Sep 18 15:47:57 GMT+03:00 2001:<I> <EJB JAR deployment /export/home/mccann/t
    ester2.jar> Exception in non-transactional EJB invoke:
    java.sql.SQLException: ORA-00942: table or view does not exist
    at weblogic.db.oci.OciCursor.getCDAException(OciCursor.java:228)
    at weblogic.jdbcbase.oci.Statement.executeUpdate(Statement.java:869)
    at weblogic.jdbc.pool.PreparedStatement.executeUpdate(PreparedStatement.
    java:65)
    at com.testerEJBPSWebLogic_CMP_RDBMS.create(testerEJBPSWebLogic_CMP_RDBM
    S.java:199)
    at com.testerEJBPSWebLogic_CMP_RDBMS.create(testerEJBPSWebLogic_CMP_RDBM
    S.java:162)
    at weblogic.ejb.internal.EntityEJBContext.create(EntityEJBContext.java:1
    18)
    at weblogic.ejb.internal.StatefulEJBObject.postCreate(StatefulEJBObject.
    java:268)
    at com.testerEJBEOImpl.create(testerEJBEOImpl.java:61)
    at com.testerEJBHomeImpl.create(testerEJBHomeImpl.java:32)
    at com.testerEJBHomeImpl_WLSkel.invoke(testerEJBHomeImpl_WLSkel.java:68)
    at weblogic.rmi.extensions.BasicServerObjectAdapter.invoke(BasicServerOb
    jectAdapter.java:347)
    at weblogic.rmi.extensions.BasicRequestHandler.handleRequest(BasicReques
    tHandler.java:69)
    at weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest
    .java:15)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:135)
    Any help here would be great, this is annoying problem because the Entity bean
    deploys fine and no issues are raised at startup time, so the table must of existed
    and the structure must of been correct. But minutes after the server started I
    ran my client which tried to invoke this bean with the above exception.
    Thanks
    Wayne.

    Hi Wayne,
    I'd recommend you installing of latest SP for WL plus installing of Oracle
    8.1.7 thing driver.
    Regards,
    Slava Imeshev
    "Wayne Highland" <[email protected]> wrote in message
    news:[email protected]...
    >
    Hi,
    WLCS 3.1
    WLS 5.1 sp6
    Windows NT
    Oracle 8.1.6
    I am having problems with an Entity bean, I have originaly deployed andhad it
    running on cloudscape. I decided I need to deploy it to Oracle 8.1.6, onstartup
    of WLCS the bean was deployed to the connection pool correctly usingWeblogic
    JDriver.
    The bit I cannot understand is when I try and do a FindByPrimaryKey on theEntity
    Bean the following exception is thrown by oracle, which leads me tobelieve this
    must be an Oracle/Jdbc problem.
    Here you can see that the bean has been deployed on a valid connectionpool.
    >
    Tue Sep 18 15:47:39 GMT+03:00 2001:<I> <EJB JAR deployment/export/home/mccann/tester2.jar>
    EJB home interface: 'com.testerHome' deployed bound to the JNDI name
    : 'tester'
    Heres the exception raised when a findByPrimaryKey is invoked on thebean....
    >
    Tue Sep 18 15:47:57 GMT+03:00 2001:<I> <EJB JAR deployment/export/home/mccann/t
    ester2.jar> Exception in non-transactional EJB invoke:
    java.sql.SQLException: ORA-00942: table or view does not exist
    at weblogic.db.oci.OciCursor.getCDAException(OciCursor.java:228)
    atweblogic.jdbcbase.oci.Statement.executeUpdate(Statement.java:869)
    atweblogic.jdbc.pool.PreparedStatement.executeUpdate(PreparedStatement.
    java:65)
    atcom.testerEJBPSWebLogic_CMP_RDBMS.create(testerEJBPSWebLogic_CMP_RDBM
    S.java:199)
    atcom.testerEJBPSWebLogic_CMP_RDBMS.create(testerEJBPSWebLogic_CMP_RDBM
    S.java:162)
    atweblogic.ejb.internal.EntityEJBContext.create(EntityEJBContext.java:1
    18)
    atweblogic.ejb.internal.StatefulEJBObject.postCreate(StatefulEJBObject.
    java:268)
    at com.testerEJBEOImpl.create(testerEJBEOImpl.java:61)
    at com.testerEJBHomeImpl.create(testerEJBHomeImpl.java:32)
    atcom.testerEJBHomeImpl_WLSkel.invoke(testerEJBHomeImpl_WLSkel.java:68)
    atweblogic.rmi.extensions.BasicServerObjectAdapter.invoke(BasicServerOb
    jectAdapter.java:347)
    atweblogic.rmi.extensions.BasicRequestHandler.handleRequest(BasicReques
    tHandler.java:69)
    atweblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest
    java:15)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:135)
    Any help here would be great, this is annoying problem because the Entitybean
    deploys fine and no issues are raised at startup time, so the table mustof existed
    and the structure must of been correct. But minutes after the serverstarted I
    ran my client which tried to invoke this bean with the above exception.
    Thanks
    Wayne.

  • Error when inserting in a table with an identity column

    Hi,
    I am new to Oracle SOA suite and ESB.
    I have been through the Oracle training and have worked for about 2 months with the tooling.
    We have a Database adabter that inserts data in 5 Tables with relations to each other.
    Each table has his own not NULL Identity column.
    When running/ testing the ESB service we get the error at the end of this post.
    From this we learned that the Database adapter inserts the value NULL in the identity column.
    We cannot find in the documentation how to get the database adabter to skip this first column and ignore it.
    Is this possible within the wizard? Our impression is no
    Is this possible somwhere else/
    And if so How can we do this?
    If anyone can help it would be greatly appreciated
    Pepijn
    Generic error.
    oracle.tip.esb.server.common.exceptions.BusinessEventRejectionException: An unhandled exception has been thrown in the ESB system. The exception reported is: "org.collaxa.thirdparty.apache.wsif.WSIFException: esb:///ESB_Projects/GVB_PDI_PDI_Wegschrijven_Medewerkergegevens/testurv.wsdl [ testurv_ptt::insert(VastAdresCollection) ] - WSIF JCA Execute of operation 'insert' failed due to: DBWriteInteractionSpec Execute Failed Exception.
    insert failed. Descriptor name: [testurv.VastAdres]. [Caused by: Cannot insert explicit value for identity column in table 'VastAdres' when IDENTITY_INSERT is set to OFF.]
    ; nested exception is:
         ORABPEL-11616
    DBWriteInteractionSpec Execute Failed Exception.
    insert failed. Descriptor name: [testurv.VastAdres]. [Caused by: Cannot insert explicit value for identity column in table 'VastAdres' when IDENTITY_INSERT is set to OFF.]
    Caused by Uitzondering [TOPLINK-4002] (Oracle TopLink - 10g Release 3 (10.1.3.3.0) (Build 070608)): oracle.toplink.exceptions.DatabaseException
    Interne uitzondering: com.microsoft.sqlserver.jdbc.SQLServerException: Cannot insert explicit value for identity column in table 'VastAdres' when IDENTITY_INSERT is set to OFF.Foutcode: 544
    Call:INSERT INTO dbo.VastAdres (ID, BeginDatum, Einddatum, Land, Plaats, Postcode, VolAdres) VALUES (?, ?, ?, ?, ?, ?, ?)
         bind => [null, 1894-06-24 00:00:00.0, 1872-09-04 00:00:00.0, Nederland, Wijdewormer, 1456 NR, Oosterdwarsweg 8]
    Query:InsertObjectQuery(<VastAdres null />).

    Hi,
    Click on the resources tab in the ESB system/ Project to see the ESB system design and all the components in it.
    Click on the Database adapter that you want to edit..and make the necesary changes..
    Check this link.
    http://download-uk.oracle.com/docs/cd/B31017_01/core.1013/b28764/esb008.htm for "6.8.2 How to Modify Adapter Services" section.
    If you are calling a database procedure which inturn makes the insert, you will have to make changes in the database and you job would be much simpler. It seems there are limitations on what you can change in the Database adapter once it is created. Please check the link for further details.
    Thanks,
    Rajesh

  • Issues when using a Pixelink camera with LabView 9.0

    In my lab, we have a program for video acquisition using LabView and the low level camera functions for Pixelink cameras.  We run into some problems trying to run this program and other demo gui programs that came with the Pixelink SDK.  In our data acquisition program, we have problems with our image coming out all scrambled.  We use the low level functions to obtain a frame from the camera in the from of a vector.  We then reshape the vector into an array and then send the array to a vi that turns it into an image.  When we try to view this image using the standard Labview image viewer, the image comes out all scrambled if the zoom factor is not set to .5 on the image viewer.  The full resolution of the image is supposed to be a 1024x1280 image which when we use the Capture OEM software for the camera provided by Pixelink we can see.  The image that we can see through LabView when we have the correct zoom factor seems to be only the top left quarter of what the entire image is supposed to be.  We also encounter strange problems when we attempt to use the demo programs for labview 8.6 that are provided with the SDK.  When trying to run the simpleGUI.vi, for example, the program connects to the camera just fine.  After you hit play to view an image, the problems start.  The program will play the video for a few seconds until it just freezes up and quits responding.  Labview must be exited and restarted in order for the camera to connect again.  I am not sure if it makes a difference that we are running labview 9.0 on our computers or not.  Any help you have on these issues would be of great help.  Attached are pictures of the scrambled/unscrambled image and of the part of our block diagram for camera control.
    Attachments:
    Incorrectimage.jpg ‏126 KB
    Errorexplanation.jpg ‏543 KB
    correct image.jpg ‏22 KB

    IMAQdx is the driver that National Instruments provides to interface Firewire, GigE and USB cameras (Directshow compliant USB).  This is a licensed driver, so if you don't currently have it, then testing in MAX is not an option. 
    So if I understand correctly, you have the Pixelink software which works fine, but you have Pixelink VI's that are used in LabVIEW, but those do not work correctly?  That to me seems like a driver issue, which I'm not sure how much I can help with since it is not our driver. 
    Do you know what dll's the Pixel link VIs are calling into?  Are they Pixelink dlls?  If so, I would suggest speaking with the manufacturer or perhaps on of our other users has worked with Pixelink before.
    Sorry I can't be much more helpful.
    Tejinder Gill
    National Instruments
    Applications Engineer
    Visit ni.com/gettingstarted for step-by-step help in setting up your system.

  • Performance issue when opening DOCX Word Doc with Excel Links

    Is this a bug or is there a config setting that can prevent Word from needlessly opening Excel repeatedly (upon initial open of a DOCX Word doc) when pre-existing links should NOT be updated. The behavior is good using 2003 formats (DOC & XLS),
    but varies significantly (~100:1 range in performance) depending on the test scenario (see below) using DOCX with XLSX.
    We can't move our environment to the new XML formats until we can find a fix. We were led to believe Microsoft knew this to be a bug in Office 2007, but it has NOT been corrected in Office 2010. Omitting a detailed explanation for why we need to
    do this, here's our test scenarios:
    -Currently testing Office 2010 Pro Trial version on WIN XP SP3. (We have found similar results using Office 2007 with Win7 or XP).
    -Word Option Deselected/DISABLED: "Update Automatic Links at Open"
    -The Word Doc contains 100 linked Excel tables and is 130K in size (DOCX).
    -Each link points to the same 10x10 cell range in one Excel Workbook located in same directory as the Word Doc, although other tests using separate directories produced similar results.
    -The Excel Links were inserted into Word via Paste Special/Paste Link/MS Excel Worksheet Object.
    -The workbook (XLSX) is 15K in size.
    -All Links are set for "Manual" update, our testing showed no difference if links are set to "Auto".
    -Again, Word is set to NOT update Automatic links when the document is opened.
    -Results shown below are for local hard drive test.
    -Network response times are slightly greater on 100mb LAN
    -Word is completely restarted prior to each test to make sure there's no application caching involved.
    LOCAL DRIVE RESPONSE TIMES TO OPEN THE WORD DOC in Docx format:
    1. Open the Word Doc from Word with Excel closed - 175 seconds (Excel appears to be opened and closed once per link)
    2. Open the Word Doc from Word with Excel open & workbook closed - 33 seconds (Excel appears to be accessed once per link)
    3. Open the Word Doc from Word with the linked workbook already open - 7 seconds (Excel appears to be accessed once per link)
    4. Rename Excel XLSX workbook so that Word can't find it. Then open Doc the from Word. 1-2 seconds (Excel does not appear to be invoked)
    5. Repeat ALL of the above scenarios 1-4 using DOC and XLS files. These files are created with "Save AS" option based on original test files and then re-pasting the links.  1-2 seconds to open the Word Doc (Excel does not appear to be invoked)
    Summary
    Using DOCX/XLSX, if Word can't find the linked spreadsheet it gives up and performs nicely during file-open. Otherwise, Word performs needless, time-consuming access to Excel, and in the worst case, opens and closes Excel repeatedly in the background.
    Testing with DOC/XLS formats gives good performance across all scenarios.

    It is a known bug and one for which AFAIK, there has been no fix.  The only way that I could get around it in a particular application was to do something like save the linked information as document variables and then on opening the document recreate
    all the information.  My memory on exactly what I did is a bit hazy, but I can go back and check if you are interested.
    -- Hope this helps.
    Doug Robbins - Word MVP,
    dkr[atsymbol]mvps[dot]org
    Posted via the Community Bridge
    "galmcrantz" wrote in message news:[email protected]...
    Is this a bug or is there a config setting that can prevent Word from needlessly opening Excel repeatedly (upon initial open of a DOCX Word doc) when pre-existing links should NOT be updated. The behavior is good using 2003 formats (DOC & XLS), but varies
    significantly (~100:1 range in performance) depending on the test scenario (see below) using DOCX with XLSX.
    We can't move our environment to the new XML formats until we can find a fix. We were led to believe Microsoft knew this to be a bug in Office 2007, but it has NOT been corrected in Office 2010. Omitting a detailed explanation for why we need to do this,
    here's our test scenarios:
    -Currently testing Office 2010 Pro Trial version on WIN XP SP3. (We have found similar results using Office 2007 with Win7 or XP).
    -Word Option Deselected/DISABLED: "Update Automatic Links at Open"
    -The Word Doc contains 100 linked Excel tables and is 130K in size (DOCX).
    -Each link points to the same 10x10 cell range in one Excel Workbook located in same directory as the Word Doc, although other tests using separate directories produced similar results.
    -The Excel Links were inserted into Word via Paste Special/Paste Link/MS Excel Worksheet Object.
    -The workbook (XLSX) is 15K in size.
    -All Links are set for "Manual" update, our testing showed no difference if links are set to "Auto".
    -Again, Word is set to NOT update Automatic links when the document is opened.
    -Results shown below are for local hard drive test.
    -Network response times are slightly greater on 100mb LAN
    -Word is completely restarted prior to each test to make sure there's no application caching involved.
    LOCAL DRIVE RESPONSE TIMES TO OPEN THE WORD DOC in Docx format:
    1. Open the Word Doc from Word with Excel closed - 175 seconds (Excel appears to be opened and closed once per link)
    2. Open the Word Doc from Word with Excel open & workbook closed - 33 seconds (Excel appears to be accessed once per link)
    3. Open the Word Doc from Word with the linked workbook already open - 7 seconds (Excel appears to be accessed once per link)
    4. Rename Excel XLSX workbook so that Word can't find it. Then open Doc the from Word. 1-2 seconds (Excel does not appear to be invoked)
    5. Repeat ALL of the above scenarios 1-4 using DOC and XLS files. These files are created with "Save AS" option based on original test files and then re-pasting the links.  1-2 seconds to open the Word Doc (Excel does not appear to be invoked)
    Summary
    Using DOCX/XLSX, if Word can't find the linked spreadsheet it gives up and performs nicely during file-open. Otherwise, Word performs needless, time-consuming access to Excel, and in the worst case, opens and closes Excel repeatedly in the background.
    Testing with DOC/XLS formats gives good performance across all scenarios.
    Doug Robbins - Word MVP dkr[atsymbol]mvps[dot]org

  • Handling EDT issues when listening to model events with ModelViewPresenter

    I'm planning on using model-view-presenter (passive screen/humble dialog variant) http://martinfowler.com/eaaDev/PassiveScreen.html as the basis for my Swing application.
    view <-> presenter <-> model
    The view is very thin and dumb and only shows what presenter tells it to. Presenter listens to UI events from the view, updates the model based on them, listens to model update events, and reads view state and updates view state.
    I'm now searching for a good practise for handling EDT-threading problem. The model is inherently multithreaded in the sense that we have several threads running concurrently. They for example fetch data from different sources and update the model accordingly. An event is created, presenter listens to it and updates the view.
    There's also operations that the user initiates. Those start in EDT, presenter does some actions on the model and possibly updates the view state as a result. If the action is fast, the whole thing can be done in EDT. If the operation takes a long time, the processing needs to jump out of EDT at some point to not block the UI.
    When thinking about the events that are created from model due to some background thread, where should we jump back to EDT? There's options:
    a) jump to EDT when presenter updates the view
    view (EDT) <- presenter (either EDT or other thread) <- model (either EDT or other thread)
    b) jump to EDT when moving to presenter layer
    view (EDT) <- presenter (EDT) <- model (either EDT or other thread)
    a)
    - could be implemented by creating an AOP aspect for all View interfaces(?) The aspect would jump to EDT if not there yet always when any method of the interface is called
    - if one ModelEvent is received in non-EDT thread in Presenter, and 15 different View setters need to be called to update the view state, causes 15 separate SwingUtilities.invokeLater() calls...problem?
    - model events and eventlisteners are straight forward to implement. Presenter and model do not need to be aware of Swing (existence of event dispatch thread)
    - Presenter is multithreaded and presenter objects need to be implemented so that threading problems do not occur
    b)
    - could be implemented for example by creating special model EventListener implementations that jump back to EDT when event is received
    - model events and eventlisteners are not as straightforward to do as with a). Somewhere the events need to jump to EDT (event source or listener) and you cannot code just "simple plain" listeners and event sources
    - Presenter needs to be aware of Swing (existence of event dispatch thread and need to jump back to it when events come from model) at least in the sense that it uses the correct EventListener implementations
    - View interface could still check that thread = EDT when methods are called (using similar AOP) to ensure threading rules are not broken
    - Presenter can be single threaded (always in EDT) and presenter implementer does not need to worry as much about threading
    - note that above point requires that user-initiated long running operations, that can't be run in EDT, need to jump from EDT the model layer. If view initiates new non-EDT threads (for example using SwingWorker), Presenter still does have multiple threads running on it (although only "one way", from view to model) and needs to be made threadsafe. This is starting to sound bad, but even if Presenter is single threaded only "one way", it may still be easier to implement than presenter that is completely multithreaded?
    I think major point in deciding is the difficulty of creating thread safe Presenters. Thread safe code is notoriously difficult to implement and even things that seem trivial at first sight can cause big problems. I can't help thinking that since the application's core is already multithreaded by nature, how much more difficult can it be to make the Presenters thread safe as well? Presenters should be quite simple in any case, can receiving events from model and calling a few setters on the view cause threading problems?
    Note that this problem is not limited to programs where background threads update the data....any program where some long running operations are run outside of EDT, and may cause events that are listened to in some presenter, has to solve this issue.
    Any views or ideas?

    RickyT wrote:
    Hello Helpful Nokia Users,
    When I am using my BH 103 blue tooth stereo earphones to listen to music on my N900, the sound drops out for about half a second (or less) then the music sounds slower (like a record player that has been slowed down) for about 4 seconds, then goes back to normal. This sometimes happens on each song, one after the other, sometimes it is just @ random times.
    This NEVER happens when using plug in head phones or when I hook it up to my stereo.
    The BH 103 are about 2 years old.... would the battery be failing? Causing them to run out of power super quickly? My N95 8GB used to beep when the battery in the earphones was low, would the N900 be "pausing" instead?
    Are the BH 103 not compatiable with the N900.... even though Nokia store lists them as accessories. Would I be better off with a newer model of ear phones?
    Cheers for your help!
    i think its the battery issue .
    Reality is wrong....dreams are for real... 2pac .
    don't forget to hit that green kudos

  • Issues when installing Hyper-V Platform with 8 network filters running

    Hi. This isn't so much a question as a solution in case anyone else has this problem.
    I tried to install Hyper-V Platform on Windows 8.1 Pro but it kept undoing changes after 93%.
    I had it previously installed on the same hardware so knew there weren't any limitations there. Some forums suggested running "Dism /Online /Cleanup-Image /ScanHealth" followed by "Dism /Online /Cleanup-Image /RestoreHealth" but this
    made no difference.
    Eventually, I looked up a forums that suggested "HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\Network\MaxNumFilters" had to be set if a maximum of 8 filters were in use. So I created a DWORD entry with a value
    of 14 (maximum on previous Windows versions) but it still didn't work!
    I checked how many networks filter were active with this powershell script:
    get-itemproperty hklm:\SYSTEM\CurrentControlSet\Control\Network | format-list -property MaxNumFilters
    $children = get-childitem -path hklm:\system\currentcontrolset\control\network\"{4d36e974-e325-11ce-bfc1-08002be10318}" | get-childitem | where-object {$_.PSChildName -eq "Ndi"} | get-itemproperty | where-object {$_.FilterClass} |get-itemproperty -name FilterClass
    "Filter Count: " + $children.count
    "Filter List: "
    $children | foreach-object -process { get-itemproperty -path $_.PSParentPath } | format-list -property Description,PSPath
    Still 8 were active! I then read that Windows 8.1 should scale MaxNumFilters automatically so I removed MaxNumFilters from the registry. Finally, I uninstalled my Anti-Virus (Kaspersky in my case) then tried installing Hyper-V platform again and it worked!
    After that, I ran the powershell script again and it detected that 9 filters were running. It looks like a bug whereby Windows is not scaling up the MaxNumFilters properly when Hyper-V platform is installed.
    This might only be an issue with Kaspersky and/or if exactly 8 filters are running but I imagine there are many more conditions under which this might occur.

    Hello ZXCallum,
    Thanks for your sharing.
    Best regards,
    Fangzhou CHEN
    Fangzhou CHEN
    TechNet Community Support

  • Issues when Insert and Analyze at the same time

    Hi All,
    We have a weekend job that used to take 12-15 hours usually and insert around 8 millions of records to a table from external table. Last weekend the same load took 60 hours and it move to Monday and there are some business impact.
    While we investigate the same issue we identified that there is one more weekend analyze job for all the tables and indexes and took around almost same time and we noticed that the table we inserted the 8 million records alone it took around 15 hours. This table now has around 91 Core records.
    We need to identify why this jobs taken long time as almost 4 times compare to the normal scenario.
    I would like to know what exactly happen when Analyze and insert happened to the table at the same time. And What all issues can occur because this.
    Best Regards,
    Shijo

    Ok,
    It is unclear what version you are using as you forgot to post that.
    Furthermore it is unclear what 'Analyze' is, as analyze statistics was already deprecated in 9i.
    Secondly, most people is this forum speak English, and they are accustomed international units only.
    Sybrand Bakker
    Senior Oracle DBA

  • I accidentaly clicked "remove" when using find my iphone...how do I log in when my iphone is not with me to find it?

    i was using "find my iphone" and when given the location it was an old location. A post read that the iphone was either turned off or the wifi was not being used. A remove or cancel  button appears with this message. I accidentally clicked cancel and now when trying to retrive information about the location of my device it says that I must refresh page or set up instructions. There must be a way to reinstate my account without my device physically present but I do not know how...can someone help me?

    What ever device you still have on had.  Go to settings - general - about - name - then the name.
    I have 4 devices 2 iphone and ipad - which I've name differently so I don't get confused when using find my iphone.

Maybe you are looking for

  • How do I recall an e-mail?

    How do I recall an e-mail that I have sent?

  • Macbook Vs. iMac G5

    I have a 2.1 20" iMac G5. Is the Macbook going to outperform my iMac? I'm not talking about the macbook pro, I'm asking about just the macbook. Thanks! iMac G5, 2.1GHZ, 20"   Mac OS X (10.4.6)   1.5GB RAM

  • SHA1 Signature not in PKCS#7 Format

    Hello, we got a Problem with a Signet XML request. We want to Communicate with a service Provider via XML request. The interface of the Service Provider want to have a SHA1 signature of the Data we send. As fare es I now the SSF Library is only suppo

  • Mail not talkinto to lineone account

    I have a lineone PAYG email account that I've had for years. Tiscali now look after lineone accounts and since they did some maintanence a couple of days ago, Mail will not talk to my lineone account. When i try to retrieve mail it keeps asking for m

  • My CD drive isn't working anymore

    Hi to all. I've had my Macbook pro for about...4 or 5 years now. I'm trying to copy some music from CD's i recently bought but for some reason my CD-Rom drive just keeps spitting CD's out. I put the CD in, and it spits it right back out without it ev