Issues with Parallels

First off, Parallels customer support stinks. I have called and emailed twice with no response.
My issue is, Parallels won't see any of my peripheral devices from within Windows XP, yet it "disconnects" them from the Mac interface so I can't use them there either. My biggest issue with this has to do with printing and scanning.
How can I get Parallels to see my printer???

"I have called and emailed twice with no response. "
Try their forums then. You'll stand a better chance of getting the help you need there, rather than an Apple forum:
http://forums.parallels.com

Similar Messages

  • Issue with parallel operation of SAP NW SSO 2.0 and SNC Client Encryption (Logon Groups)

    Hi!
    One of our customers is using the SNC Client Encryption solution to ensure encryption using SNC (based on Kerberos Technology) for their SAP GUI Dialog connections. They have lots of SAP backends DEV, QAS, PRD all with the SNC Client Encryption SNC Lib installed. The profile parameter snc/identity/as contains the following value: p:CN=SAP/<ServiceAccount>@<DOMAIN>.
    Example: p:CN=SAP/[email protected]
    The customer is using one AD Service Account "SNCServiceUser" with one registered SPN "SAP/SNCServiceUser" for all systems (yes, this is not recommended... but the case).
    Important: All users use group entries in the SAP Logon (saplogin.ini). Means, for SAP logon the SNC name can not be manually configured on the SAP Front End. With group logons, the application server's SNC name is dynamically requested by the message server each time a SAP GUI connection is started. The SNC Name is greyed out in this case as dynamically obtained from the applications servers profile parameter snc/identity/as.
    Now our customer implements SAP NetWeaver Single Sign-On 2.0 within his landscape. Based on the Secure Login Server 2.0 (SP3) he likes to use X.509 based authentication to his AS ABAP backends using SAP GUI SNC while others still use SNC Client Encryption.
    Replacing the SNC Library on the AS ABAP
    The Secure Login Library 2.0 (SP3) has been installed on one of the ABAP systems and the SNC Client Encryption SNC Library (which is based on SSO 1.0) is no longer used, thus we changed the parameter snc/gssapi_lib to point to the new SNC library. We removed the old PSE.ZIP containing the keytab and created the new SAPSNCSKERB.PSE incl. the keytab and proper credentials. To ensure parallel operation, we kept the snc/identity/as value as is =  p:CN=SAP/[email protected].
    After restarting the system with initialized Secure Login Library 2.0, still the SNC client encryption works fine for existing users.
    The problem
    We created on the Secure Login Server an SNC certificate for the AS ABAP which has the following X.509 Distinguised Name Fomat: CN=SAP/[email protected] This is to avoid having to change the snc/identity/as to an "real" X.509 DN which would lead to non-working SNC Client Encryption for all the other users using SAP GUI and logon groups.
    As soon as we install the PSE via STRUST on the system the SNC Client Encryption solution stops working with error „Server refuses kerberos key exchange“.
    As part of an pilot implementation we have installed Secure Login Client 2.0 (SP3) on some test PCs. The test PC with SLC is able to perform Single Sign-On with SNC based on X.509 (incl. Encryption) to the ABAP system.
    Seems the SAP System now only tries to do X.509 based authentication thus key exchange fails. The problem is, we cannot change the snc/identity/as value because of the logon groups. If we were able to do so, we would in any case set the server identity to X.509 DN and in addition create the SAPSNCSKERB.PSE incl. keytab. This should work, as confirmed by SAP see this post.  
    Any ideas how to solve this and have both solutions in parallel?
    Appreciate any help.
    Regards,
    Carsten

    Hi all,
    we was able to fix the issue. It was an issue with the customers cluster configuration and the  $SECUDIR variable. This tricky issue leads to non working or sporadic working SNC Client Encryption...
    This was how the configuration looks before:
    Environment variable $SECUDIR is defined:
    "/ABCDEF<SID>/usr/sap/<SID>/DVEBMGSxx/sec“
    sapgenpse seclogin -l -v
    running seclogin with USER="<SID>adm"
    Credentials for username '<SID>adm':
    0 (LPS:OFF):
             (LPS:OFF): /ABCDEF<SID>/usr/sap/<SID>/DVEBMGSxx/sec/SAPSNCSKERB.pse
    1 (LPS:OFF):
             (LPS:OFF): /usr/sap/<SID>/DVEBMGSxx/sec/SAPSNCS.pse
    After changing the $SECUDIR to "/usr/sap/<SID>/DVEBMGSxx/sec“ and re-creating the credentials, it worked like a charm.
    As a result of this we can confirm, this configuration and SNC Client Encryption works with CommonCryptoLib in parallel to the SSO configuration.
    And Valerie was right with 2. SLC starting from V. 1.0 SP2 PL3 was able to convert the CN= part of the SNC Name into an SPN, was my mistake. In addition SNC Client Encryption starting from Version 1 SP1 PL1 does this also.. just to make this clear
    Thread closed hope this helps someone
    Carsten

  • Issues with parallel processing in Logical Database PCH and PNP

    Has anyone encountered issues when executing programs in parallel  that utilizes the logical database PCH or PNP?
    Our scenario is the following:
    We having have 55 concurrent jobs that execute a program that use the logical database PCH at a given time.  We load the the PCHINDEX table with the code below.
          wa_pchindex-plvar = '01'.
          wa_pchindex-otype = 'S'.
          wa_pchindex-objid_low = index_objid.
          APPEND wa_pchindex TO pchindex.
    We have seen instances where when the program is executed in parallel, with each process having its own range of positions id's, that some positions are dropped or some are added that is outside the range of the given process.
    For example:
    process 1 has a range of positions ID's 1-10
    process 2 has a range of positions ID's 11-20
    process 3 has a range of positions ID's 21-30
    Process 3 drops position 25 and adds position 46.
    Has anyone faced a similar issue?
    Thanks for your help.
    Best Regards,
    Duke

    Hi,
    first of all, you should read [Using Parallel Execution|http://download.oracle.com/docs/cd/B19306_01/server.102/b14223/usingpe.htm#DWHSG024] in documentation for your version - almost all of these topics are covered there.
    1. According to my server specification how much DOP i can specify.It depends not only on number of CPU. More important factors are settings of PARALLEL_MAX_SERVERS and PARALLEL_ADAPTIVE_MULTI_USER.
    2. Which option for Setting Parallel is good - Using the 'alter table A parallel 4' or passing the parallel hints in the sql statementsIt depends on your application. When setting PARALLEL on a table, all SQL dealing with that table would be considered for parallel execution. So if it is normal for your app to use parallel access to that table, it's OK. If you want to use PX on a limited set of SQL, then hints or session settings are more appropriate.
    3. We have a batch processing jobs which are loading data into the tables from flat files (24*7) using sql loader. is it possible to parallel this operation and any negative effect if enabled parallel.Yes, refer to documentation.
    4. Query or DML - which one will be perform best with parallel option.Both may take advantages of using PX (with some restrictions to Parallel DML) and both may run slower than non-PX versions.
    5. What are the negative issue if parallel option is enabled.1) Object checkpoint happens before starting parallel FTS (true for >=10gR2, before that version tablespace checkpoint was used)
    2) More CPU and memory resources are used with PX - it may be both benefit and an issue, especially with concurrent PX.
    6. what are the things to be taken care while enabling the parallel option.Read the documentation - it contains almost all you need to know. Since you are using RAC, you sould not forget about method of PX slaves load balancing between nodes. If you are on 10g, refer to INSTANSE_GROUPS/PARALLEL_INSTANCE_GROUPS parameters, if you are using 11g then properly configure services.

  • OIM 11g R1. Issue with parallel participant type approval workflow.

    Hi All,
    I have a request level approval with the human task assignment participant type set to parallel. There will be two participants in the workflow with approve and reject having outcome value as 50 each. When i submit a request, i can see the both the request is properly allocated to both the users. But when i log in with the individual users and search the approval tasks in the self service page, i cant see it properly assigned.
    In the approval tasks table, i cant see the values for the following columns request ID, request type, beneficiary, request target. Whereas i can see the value for the requester as "Anonymour user" and it is assigned to a proper user. When i try to approve, i get the following error in the console.
    ========================================================================================================================
    <25 Nov, 2012 1:49:09 PM IST> <Warning> <oracle.adfinternal.view.faces.renderkit
    .rich.NavigationPaneRenderer> <BEA-000000> <ILLEGAL_COMPONENT_HIERARCHY : A chil
    d component that is not a commandNavigationItem (or a separator when hint=choice
    ) was found.>
    <25 Nov, 2012 2:22:57 PM IST> <Error> <oracle.iam.request.impl> <IAM-2050076> <N
    o request found with the request ID .>
    <25 Nov, 2012 2:22:57 PM IST> <Error> <oracle.iam.tasklist.agentry.task> <IAM-20
    60009> <Error occurred while approving task from BPEL>
    <25 Nov, 2012 2:22:57 PM IST> <Error> <oracle.iam.tasklist.agentry.task> <IAM-20
    60013> <Exception thrown:
    oracle.iam.request.exception.RequestServiceException: IAM-2050076:No request fou
    nd with the request ID .
    at oracle.iam.request.repository.RequestRepository.getRequestClone(Reque
    stRepository.java:891)
    at oracle.iam.request.impl.RequestEngine.getBasicRequestDetails(RequestE
    ngine.java:3877)
    at oracle.iam.request.impl.RequestServiceImpl.getBasicRequestData(Reques
    tServiceImpl.java:94)
    at oracle.iam.request.api.RequestServiceEJB.getBasicRequestDatax(Unknown
    Source)
    at sun.reflect.GeneratedMethodAccessor1000.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
    sorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at com.bea.core.repackaged.springframework.aop.support.AopUtils.invokeJo
    inpointUsingReflection(AopUtils.java:310)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMetho
    dInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:182)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMetho
    dInvocation.proceed(ReflectiveMethodInvocation.java:149)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntrodu
    ctionInterceptor.doProceed(DelegatingIntroductionInterceptor.java:131)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntrodu
    ctionInterceptor.invoke(DelegatingIntroductionInterceptor.java:119)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMetho
    dInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.jee.spi.MethodInvocationVisit
    orImpl.visit(MethodInvocationVisitorImpl.java:37)
    at weblogic.ejb.container.injection.EnvironmentInterceptorCallbackImpl.c
    allback(EnvironmentInterceptorCallbackImpl.java:54)
    at com.bea.core.repackaged.springframework.jee.spi.EnvironmentIntercepto
    r.invoke(EnvironmentInterceptor.java:50)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMetho
    dInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.interceptor.ExposeInvocat
    ionInterceptor.invoke(ExposeInvocationInterceptor.java:89)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMetho
    dInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntrodu
    ctionInterceptor.doProceed(DelegatingIntroductionInterceptor.java:131)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntrodu
    ctionInterceptor.invoke(DelegatingIntroductionInterceptor.java:119)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMetho
    dInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.framework.JdkDynamicAopPr
    oxy.invoke(JdkDynamicAopProxy.java:204)
    at $Proxy337.getBasicRequestDatax(Unknown Source)
    at oracle.iam.request.api.RequestService_dnwrzl_RequestServiceRemoteImpl
    .__WL_invoke(Unknown Source)
    at weblogic.ejb.container.internal.SessionRemoteMethodInvoker.invoke(Ses
    sionRemoteMethodInvoker.java:40)
    at oracle.iam.request.api.RequestService_dnwrzl_RequestServiceRemoteImpl
    .getBasicRequestDatax(Unknown Source)
    at sun.reflect.GeneratedMethodAccessor999.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
    sorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at weblogic.ejb.container.internal.RemoteBusinessIntfProxy.invoke(Remote
    BusinessIntfProxy.java:85)
    at $Proxy170.getBasicRequestDatax(Unknown Source)
    at sun.reflect.GeneratedMethodAccessor998.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
    sorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflecti
    on(AopUtils.java:307)
    at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynami
    cAopProxy.java:198)
    at $Proxy336.getBasicRequestDatax(Unknown Source)
    at oracle.iam.request.api.RequestServiceDelegate.getBasicRequestData(Unk
    nown Source)
    at oracle.iam.tasklist.agentry.task.ApproveActor.perform(ApproveActor.ja
    va:106)
    at oracle.iam.consoles.faces.mvc.canonic.Model.perform(Model.java:565)
    at oracle.iam.consoles.faces.mvc.self.Model.perform(Model.java:178)
    at oracle.iam.consoles.faces.mvc.canonic.Model.prepare(Model.java:460)
    at oracle.iam.consoles.faces.mvc.self.Model.prepare(Model.java:173)
    at oracle.iam.consoles.faces.mvc.canonic.Controller.doPrepare(Controller
    .java:225)
    at oracle.iam.consoles.faces.mvc.canonic.Controller.doSelectOperationNav
    igation(Controller.java:88)
    at oracle.iam.consoles.faces.render.canonic.UIEntrypod$CommandToolbarBut
    tonActionListener.processAction(UIEntrypod.java:677)
    at javax.faces.event.ActionEvent.processListener(ActionEvent.java:88)
    at org.apache.myfaces.trinidad.component.UIXComponentBase.broadcast(UIXC
    omponentBase.java:675)
    at org.apache.myfaces.trinidad.component.UIXCommand.broadcast(UIXCommand
    .java:179)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent$1.r
    un(ContextSwitchingComponent.java:92)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent._pr
    ocessPhase(ContextSwitchingComponent.java:361)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent.bro
    adcast(ContextSwitchingComponent.java:96)
    at oracle.adf.view.rich.component.fragment.UIXInclude.broadcast(UIXInclu
    de.java:102)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent$1.r
    un(ContextSwitchingComponent.java:92)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent._pr
    ocessPhase(ContextSwitchingComponent.java:361)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent.bro
    adcast(ContextSwitchingComponent.java:96)
    at oracle.adf.view.rich.component.fragment.UIXInclude.broadcast(UIXInclu
    de.java:96)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl.broadcastEvents
    (LifecycleImpl.java:902)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl._executePhase(L
    ifecycleImpl.java:313)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl.execute(Lifecyc
    leImpl.java:186)
    at javax.faces.webapp.FacesServlet.service(FacesServlet.java:265)
    at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run
    (StubSecurityHelper.java:227)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecuri
    tyHelper.java:125)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.jav
    a:300)
    at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.help.web.rich.OHWFilter.doFilter(Unknown Source)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.adf.model.servlet.ADFBindingFilter.doFilter(ADFBindingFilter.j
    ava:205)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.adfinternal.view.faces.webapp.rich.RegistrationFilter.doFilter
    (RegistrationFilter.java:106)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl$FilterL
    istChain.doFilter(TrinidadFilterImpl.java:446)
    at oracle.adfinternal.view.faces.activedata.AdsFilter.doFilter(AdsFilter
    .java:60)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl$FilterL
    istChain.doFilter(TrinidadFilterImpl.java:446)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl._doFilt
    erImpl(TrinidadFilterImpl.java:271)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl.doFilte
    r(TrinidadFilterImpl.java:177)
    at org.apache.myfaces.trinidad.webapp.TrinidadFilter.doFilter(TrinidadFi
    lter.java:92)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.iam.platform.auth.web.PwdMgmtNavigationFilter.doFilter(PwdMgmt
    NavigationFilter.java:121)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.iam.platform.auth.web.OIMAuthContextFilter.doFilter(OIMAuthCon
    textFilter.java:107)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.adf.library.webapp.LibraryFilter.doFilter(LibraryFilter.java:1
    75)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:111)
    at java.security.AccessController.doPrivileged(Native Method)
    at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:31
    3)
    at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUt
    il.java:413)
    at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.jav
    a:94)
    at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:1
    61)
    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:13
    6)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsF
    ilter.java:27)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.ja
    va:56)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationActio
    n.wrapRun(WebAppServletContext.java:3715)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationActio
    n.run(WebAppServletContext.java:3681)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(Authenticate
    dSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:
    120)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppS
    ervletContext.java:2277)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletC
    ontext.java:2183)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.j
    ava:1454)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:178)
    >
    ========================================================================================================================
    Please share any thoughts on this.
    Thanks,
    Srini

    Hi Kevin,
    Thanks a lot for the response. But, when i deployed the composite SAR in to the server from Jdeveloper, i checked the option to "Overwrite any existing composite with same revision ID". SO, i used the same revision ID (Say 1.0), will this also need to be disabled?
    Thanks,
    Srini

  • Help with parallel arrays of different data types

    Hello all, I am having an issue with parallel arrays. My program requires me to read information from a text file into 4 parallel arrays (2 String and 2 double). My text file needs to look something like this:
    John Johnson
    0000004738294961
    502.67
    1000.000
    Jane Smith
    0000005296847913
    284.51
    1000.000
    ...and so on
    Where the first thing is the name (obviously), an account number, the balance in the account, and the credit limit. I just cant figure out how to read everything into the arrays. We havent learned anything too heavy, and this seems a little too advanced for my class, but I guess we will manage. Any help will be appreciated. Thanks guys.
    Casey

    Man this is a dumb homework assignment. The requirements scream out for a class along the lines of
    public class Account{
      private String name, number;
      private double balance,creditlimit;
       // more code here
    }and then to use a List of Account objects.
    Anyway what's your actual problem. There's nothing very hard about it. A loop. So....
    You should consider posting (formatted) code showing what you have done and where exactly you are stuck.

  • Issue with "read by other session" and a parallel MERGE query

    Hi everyone,
    we have run into an issue with a batch process updating a large table (12 million rows / a few GB, so it's not that large). The process is quite simple - load the 'increment' from a file into a working table (INCREMENT_TABLE) and apply it to the main table using a MERGE. The increment is rather small (usually less than 10k rows), but the MERGE runs for hours (literally) although the execution plan seems quite reasonable (can post it tomorrow, if needed).
    The first thing we've checked is AWR report, and we've noticed this:
    Top 5 Timed Foreground Events
    Event     Waits     Time(s)     Avg wait (ms)     % DB time     Wait Class
    DB CPU           10,086           43.82     
    read by other session     3,968,673     9,179     2     39.88     User I/O
    db file scattered read     1,058,889     2,307     2     10.02     User I/O
    db file sequential read     408,499     600     1     2.61     User I/O
    direct path read     132,430     459     3     1.99     User I/OSo obviously most of the time was consumed by "read by other session" wait event. There were no other queries running at the server, so in this case "other session" actually means "parallel processes" used to execute the same query. The main table (the one that's updated by the batch process) has "PARALLEL DEGREE 4" so Oracle spawns 4 processes.
    I'm not sure how to fix this. I've read a lot of details about "read by other session" but I'm not sure it's the root cause - in the end, when two processes read the same block, it's quite natural that only one does the physical I/O while the other waits. What really seems suspicious is the number of waits - 4 million waits means 4 million blocks, 8kB each. That's about 32GB - the table has about 4GB, and there are less than 10k rows updated. So 32 GB is a bit overkill (OK, there are indexes etc. but still, that's 8x the size of the table).
    So I'm thinking that the buffer cache is too small - one process reads the data into cache, then it's removed and read again. And again ...
    One of the recommendations I've read was to increase the PCTFREE, to eliminate 'hot blocks' - but wouldn't that make the problem even worse (more blocks to read and keep in the cache)? Or am I completely wrong?
    The database is 11gR2, the buffer cache is about 4GB. The storage is a SAN (but I don't think this is the bottleneck - according to the iostat results it performs much better in case of other batch jobs).

    OK, so a bit more details - we've managed to significantly decrease the estimated cost and runtime. All we had to do was a small change in the SQL - instead of
    MERGE /*+ parallel(D DEFAULT)*/ INTO T_NOTUNIFIED_CLIENT D /*+ append */
      USING (SELECT
          FROM TMP_SODW_BB) S
      ON (D.NCLIENT_KEY = S.NCLIENT_KEY AND D.CURRENT_RECORD = 'Y' AND S.DIFF_FLAG IN ('U', 'D'))
      ...(which is the query listed above) we have done this
    MERGE /*+ parallel(D DEFAULT)*/ INTO T_NOTUNIFIED_CLIENT D /*+ append */
      USING (SELECT
          FROM TMP_SODW_BB AND DIFF_FLAG IN ('U', 'D')) S
      ON (D.NCLIENT_KEY = S.NCLIENT_KEY AND D.CURRENT_RECORD = 'Y')
      ...i.e. we have moved the condition from the MERGE ON clause to the SELECT. And suddenly, the execution plan is this
    OPERATION                           OBJECT_NAME             OPTIONS             COST
    MERGE STATEMENT                                                                 239
      MERGE                             T_NOTUNIFIED_CLIENT
        PX COORDINATOR
          PX SEND                       :TQ10000                QC (RANDOM)         239
            VIEW
              NESTED LOOPS                                      OUTER               239
                PX BLOCK                                        ITERATOR
                  TABLE ACCESS          TMP_SODW_BB             FULL                2
                    Filter Predicates
                      OR
                        DIFF_FLAG='D'
                        DIFF_FLAG='U'
                  TABLE ACCESS          T_NOTUNIFIED_CLIENT       BY INDEX ROWID    3
                    INDEX               AK_UQ_NOTUNIF_T_NOTUNI    RANGE SCAN        2
                      Access Predicates
                        AND
                          D.NCLIENT_KEY(+)=NCLIENT_KEY
                          D.CURRENT_RECORD(+)='Y'
                      Filter Predicates
                        D.CURRENT_RECORD(+)='Y' Yes, I know the queries are not exactly the same - but we can fix that. The point is that the TMP_SODW_BB table contains 1639 rows in total, and 284 of them match the moved 'IN' condition. Even if we remove the condition altogether (i.e. 1639 rows have to be merged), the execution plan does not change (the cost increases to about 1300, which is proportional to the number of rows).
    But with the original IN condition (that turns into an OR combination of predicates) in the MERGE ON clausule, the cost suddenly skyrockets to 990.000 and it's damn slow. It seems like a problem with cost estimation, because once we remove one of the values (so there's only one value in the IN clausule), it works fine again. So I guess it's a planner/estimator issue ...

  • Issue in Parallels with Vista

    Parallels installed with a full version of Windows Vista Ultimate. Flight Simulator X installed with the bulk going to a portable hard drive, flight sim says I have an issue with my video card drivers and that flight simulator will now exit. How, if possible, do I update the video card drivers for my Macbook pro.

    If you absolutely must play games on your MBP and the desired performance and/or requirements exceed what you get running virtualized under Parallels, you can always install the Microsoft O.S. and run natively with Boot Camp.
    You can also try with other virtualization products to see if they work better with your games. Besides Parallels Desktop, you have [VMWare Fusion|http://www.vmware.com/products/fusion> and you have [Sun VirtualBox|http://www.virtualbox.org>. This last one is free.
    Plus, you can try using Windows XP instead of the Vista bloatware.

  • Are there any known issues upgrading to 10.8.3 from 10.7.5 with Parallels Desktop already installed in 10.7.5?

    Are there any known issues upgrading to 10.8.3 from 10.7.5 with Parallels Desktop already installed in 10.7.5?

    Welcome to Apple Support Communities
    Most of the applications compatible with Lion are also compatible with Mountain Lion, so you shouldn't have any problem. Mountain Lion supports the latest update for Parallels 7 and 8, so if you haven't got one of these versions, you should upgrade it before upgrading Mac OS X. That application uses extensions, so it may cause kernel panics

  • Windows XP or Vista for use with Parallels on MacBook 2.16 GHz?

    As described in my earlier post, I'm currently switching from my G4 PowerBook to a new MacBook 13" 2.16 GHz.
    I'll have to use Windows parallel to OS X (plan to do that with Parallels Desktop) and would like to know what you recommend:
    Windows XP or Windows Vista?
    Vista should be better in security issues, however it should also consume much more ressources and thus run slower ... is that right?
    What's your experience?
    Martin
    MacBook 2.16 GHz, 2 GB RAM, 120 GB HD (white)   Mac OS X (10.4.9)  

    Hi Martin,
    personally I favour XP over Vista (at least for now and the near future), because as you mentioned Vista a much more hungrier than XP and also because there are still a lot of applications which do not work correctly or at all in Vista.
    Regards
    Stefan

  • After several repairs, still having heating issue with MacBook Pro (mid '09)

    EDIT: this post got removed the first time for not being constructive enough. I've now edited the post.
    This is going to be a lengthy post, so bear with me.
    2 months ago, I suddenly noticed my MacBook Pro (medio 2009, 15", 4GB RAM, GeForce 9400M) heating up a lot harder than usual. I'm familiar with the heating issues these laptops have, but this wasn't normal. Looking at iStat Pro, I saw my fans almost taking off like an airplane at 5700 rpm, while watching only a livestream on my browser. I know Shockwave Flash takes up quite a chunk of your CPU, but I had never encountered this problem beforehand.
    The biggest problem wasn't just the heat or the fans, but the fact that my computer also started reacting rather slowly. Sometimes I'd even get beachballed, or the whole browser stopped responding for half a minute...
    So I tried to see if gaming was still an option. Of course not. I had been playing League of Legends for several months without any issues. Whilst playing, I'd kick up my fans manually to about 4000 rpm, and they'd stay there throughout the whole game. Now, when I start up the game, it plays smoothly for the first 2 minutes. After the temperatures rise to a critical level, the fans start spinning up to 5700 rpm, and the whole game plays like a powerpoint presentation.
    On a sidenote: it sort of 'lags' every 1-2 seconds, and sometimes, after a long period, the game acts somewhat normal again (20 mins). This game uses wine to run, which isn't the most optimal way. So I tried playing Team Fortress 2, too. Probably even worse results! (but again noticeable 1-2 second interval lags)
    Before anyone starts screaming: "Don't use your Mac for gaming!" etc., I wish to point out that these two games had been running perfectly for the past few months/years. Something's definitely wrong.
    After looking for some solutions myself, I decided to go to an authorized Apple Service in my city. A few days later, they told me my logic board was broken, and I had to pay almost 700 euros ($ 900-1000) to replace it. So I did. I got it back, but it wasn't fixed. Brought it back in. Got it back a week later: "Your new logic board we got from Apple was already broken, but we hadn't noticed. We put a new one in it. It's fixed." Again, same problem.
    We're almost 2 months later and 700 euros poorer, still no result. I've brought it in about 6-7 times now, here's a list of what they've tried so far:
    - In total 3 new logic boards
    - New MagSafe (because they discovered it was giving too much power too my laptop)
    - Fresh re-install
    - Replaced RAM
    - Checked fans etc. (they do this by default)
    - A few tests which proved that it was repaired (but I have my doubts with these tests...)
    They told me now that it's fixed, and they basically won't look into it anymore. I must contact Apple myself, if I wasn't satisfied. The tests they performed showed no signs of heating issues. (One of the tests was playing 4 1080p movies at the same time. When I watch a 1080p movie on Youtube, my laptop is taking off like a plane) They ran a stress test all night long, showing no temperatures above 98°C. I got it up to 105°C in a few minutes according to iStat Pro.
    They advised me to install Mountain Lion, which would make the issues with "high-performance sites" go away. I haven't seen much improvement, though.
    This whole experience has been quite frustrating, especially now that I'm gonna need my laptop for University again. Isn't there anything I can do? I've paid almost 700 euros and missed my laptop for 2 months, without any results.
    I've got more information regarding the issues if needed, ask away. I'm getting desperate.
    I've been browsing around the support communites, and found that a lot of people are having issues since they upgraded to OS X Lion. Could this be the case? Should I try downgrading to Snow Leopard?
    I've also found a few topics on refreshing the cooling pasta on the heatsink. Would they have noticed any problems with this after replacing the Logic Board several times? Or should I check it myself? I'm not too keen on opening up my laptop myself...
    Nick.

    Thought of buying Parallels? Fusion?
    Might want to see if this gets moved to the MacBook Pro forum.
    I'd think Windows 8.1 is smart enough (what I use 90% time since the Dev Preview came out couple yrs ago it seems like) to find the driver.

  • New to Mac - Issues with imported Excel Docs

    Imported my Excel docs from my PC and they made the trip but when I open them, some don't work properly. I have some billing statements that I made with Excel templates. Whenever I open and try to save them, I get an error message:
    Users/JaneDoe/Documents/Private Practice/Billing Statements/Jones 2 16 05 MAC.numbers
    I'm having some issues with Word documents that have tables in them also.
    I love my new MAC but have many business forms that I need to use. Any help would be appreciated.

    Thanks for your answer, I suspected as much. I've been thinking about this and have decided that since those are historical files and not working files, I'll just leave them alone and create new billing statements using Mac. It's certainly easy enough to do.
    The Word documents, though, are a different story. I've scanned pages out of some manuals and many of the pages are questionnaires with tables to indicate answers. Some documents are multiple pages and every other page I view is blank. It's difficult and sometimes impossible to delete the blank page or move the table up. If I can't salvage these, I'll have to go back into the manuals and scan the pages all over again.
    And I don't want to partition my hard drive, run Windows parallel or invest in Office for Mac because that all defeats the purpose of my transition to a Mac in the first place - ease of use.
    Thanks again.

  • Performance Issues with SDV03V02

    Hi,
      The order rescheduling program(SDV03V02) runtime is too long. We have searched for notes in the OSS database but have not found any suitable notes. One of the oss notes suggested implementing indexes on VAPMA and VBUK but that has not helped either. We are thinking of writing a Z program which would submit the standard SAP program into multiple jobs-The Z program would schedule the program into multiple jobs using different sets of materials.
      Has anyone faced runtime issues with this program? Any other suggestions?
    Regards,
    Aravind

    Aravind,
    the program invokes availability check for all backourders based on material.
    I would check the following :
    - how many documents it's going to process ?
    - if the number of documents is too big and there are really old docs - you may have a problem with order status update (your order never goes to "complete" status even if all items are rejected or fully shipped - just an example)....
    - if you really need to reschedule many orders - you'll have to consider parallel rescheduling.... BUT it's not as simple as submit parallel jobs for different materials (material ranges)....as jobs will very likely fail due to a contention.
    - another option to speed up the process (may really help) is to change timing of all order outputs which are set to process immediately to '1' -> to  be processed via RSNAST00, then schedule a job after your rescheduling is complete to process all these outputs. It may really help as you separate output processing from avail.check/order save.... and since same order may be saved several times during rescheduling -> it may really help. You can do it by defining a "timing" routine for your output types.
    - always good to run SQL trace or to just monitor this job in SM50 to see what it's doing... OR run it under new USER ID and check statistics for that ID to see which programs/tables took the most time.
    Would be good to measure an average order processing time (check how many orders were processed in this job... I hope it has this info in the spool or you can calculate it based on messages in job log). It may help to determine if the problem is casued by too slow order processing or just by increased order volume.
    There are many options to look at regarding rescheduling...but I would start from order processing performance.

  • Performance issues with pipelined table functions

    I am testing pipelined table functions to be able to re-use the <font face="courier">base_query</font> function. Contrary to my understanding, the <font face="courier">with_pipeline</font> procedure runs 6 time slower than the legacy <font face="courier">no_pipeline</font> procedure. Am I missing something? The <font face="courier">processor</font> function is from [url http://www.oracle-developer.net/display.php?id=429]improving performance with pipelined table functions .
    Edit: The underlying query returns 500,000 rows in about 3 minutes. So there are are no performance issues with the query itself.
    Many thanks in advance.
    CREATE OR REPLACE PACKAGE pipeline_example
    IS
       TYPE resultset_typ IS REF CURSOR;
       TYPE row_typ IS RECORD (colC VARCHAR2(200), colD VARCHAR2(200), colE VARCHAR2(200));
       TYPE table_typ IS TABLE OF row_typ;
       FUNCTION base_query (argA IN VARCHAR2, argB IN VARCHAR2)
          RETURN resultset_typ;
       c_default_limit   CONSTANT PLS_INTEGER := 100;  
       FUNCTION processor (
          p_source_data   IN resultset_typ,
          p_limit_size    IN PLS_INTEGER DEFAULT c_default_limit)
          RETURN table_typ
          PIPELINED
          PARALLEL_ENABLE(PARTITION p_source_data BY ANY);
       PROCEDURE with_pipeline (argA          IN     VARCHAR2,
                                argB          IN     VARCHAR2,
                                o_resultset      OUT resultset_typ);
       PROCEDURE no_pipeline (argA          IN     VARCHAR2,
                              argB          IN     VARCHAR2,
                              o_resultset      OUT resultset_typ);
    END pipeline_example;
    CREATE OR REPLACE PACKAGE BODY pipeline_example
    IS
       FUNCTION base_query (argA IN VARCHAR2, argB IN VARCHAR2)
          RETURN resultset_typ
       IS
          o_resultset   resultset_typ;
       BEGIN
          OPEN o_resultset FOR
             SELECT colC, colD, colE
               FROM some_table
              WHERE colA = ArgA AND colB = argB;
          RETURN o_resultset;
       END base_query;
       FUNCTION processor (
          p_source_data   IN resultset_typ,
          p_limit_size    IN PLS_INTEGER DEFAULT c_default_limit)
          RETURN table_typ
          PIPELINED
          PARALLEL_ENABLE(PARTITION p_source_data BY ANY)
       IS
          aa_source_data   table_typ;-- := table_typ ();
       BEGIN
          LOOP
             FETCH p_source_data
             BULK COLLECT INTO aa_source_data
             LIMIT p_limit_size;
             EXIT WHEN aa_source_data.COUNT = 0;
             /* Process the batch of (p_limit_size) records... */
             FOR i IN 1 .. aa_source_data.COUNT
             LOOP
                PIPE ROW (aa_source_data (i));
             END LOOP;
          END LOOP;
          CLOSE p_source_data;
          RETURN;
       END processor;
       PROCEDURE with_pipeline (argA          IN     VARCHAR2,
                                argB          IN     VARCHAR2,
                                o_resultset      OUT resultset_typ)
       IS
       BEGIN
          OPEN o_resultset FOR
               SELECT /*+ PARALLEL(t, 5) */ colC,
                      SUM (CASE WHEN colD > colE AND colE != '0' THEN colD / ColE END)de,
                      SUM (CASE WHEN colE > colD AND colD != '0' THEN colE / ColD END)ed,
                      SUM (CASE WHEN colD = colE AND colD != '0' THEN '1' END) de_one,
                      SUM (CASE WHEN colD = '0' OR colE = '0' THEN '0' END) de_zero
                 FROM TABLE (processor (base_query (argA, argB),100)) t
             GROUP BY colC
             ORDER BY colC
       END with_pipeline;
       PROCEDURE no_pipeline (argA          IN     VARCHAR2,
                              argB          IN     VARCHAR2,
                              o_resultset      OUT resultset_typ)
       IS
       BEGIN
          OPEN o_resultset FOR
               SELECT colC,
                      SUM (CASE WHEN colD > colE AND colE  != '0' THEN colD / ColE END)de,
                      SUM (CASE WHEN colE > colD AND colD  != '0' THEN colE / ColD END)ed,
                      SUM (CASE WHEN colD = colE AND colD  != '0' THEN 1 END) de_one,
                      SUM (CASE WHEN colD = '0' OR colE = '0' THEN '0' END) de_zero
                 FROM (SELECT colC, colD, colE
                         FROM some_table
                        WHERE colA = ArgA AND colB = argB)
             GROUP BY colC
             ORDER BY colC;
       END no_pipeline;
    END pipeline_example;
    ALTER PACKAGE pipeline_example COMPILE;Edited by: Earthlink on Nov 14, 2010 9:47 AM
    Edited by: Earthlink on Nov 14, 2010 11:31 AM
    Edited by: Earthlink on Nov 14, 2010 11:32 AM
    Edited by: Earthlink on Nov 20, 2010 12:04 PM
    Edited by: Earthlink on Nov 20, 2010 12:54 PM

    Earthlink wrote:
    Contrary to my understanding, the <font face="courier">with_pipeline</font> procedure runs 6 time slower than the legacy <font face="courier">no_pipeline</font> procedure. Am I missing something? Well, we're missing a lot here.
    Like:
    - a database version
    - how did you test
    - what data do you have, how is it distributed, indexed
    and so on.
    If you want to find out what's going on then use a TRACE with wait events.
    All nessecary steps are explained in these threads:
    HOW TO: Post a SQL statement tuning request - template posting
    http://oracle-randolf.blogspot.com/2009/02/basic-sql-statement-performance.html
    Another nice one is RUNSTATS:
    http://asktom.oracle.com/pls/asktom/ASKTOM.download_file?p_file=6551378329289980701

  • Performance issues with Homesharing?

    I have a Time Capsule as the base station for my wireless network, then 2 Airport Express setup to extend the network around the house, an iMac i7 as the main iTunes Library and couple of iPads, and a couple of Apple TVs. Everything has the latest software, but I have several performance issues with Home sharing. I've done several tests making sure nothing is taking additional bandwidth, so here are the list of issues:
    1) With nothing else running, trying playing a movie via home sharing in an iPad 2 which is located on my iMac, it stops and I have to keep pressing the play button over and over again. I typically see that the iPad tries to download part of the movie first and then starts playing so that it deals with the bandwidth, but in many cases it doesn't.
    2) When trying to play any iTunes content (movies, music, photos, etc) from my Apple TV I can see my computer library, but when I go in on any of the menus, it says there's no content. I have to reboot the Apple TV and then problem fixed. I's just annoying that I have to reboot.
    3) When watching a Netflix movie on my iPad and with Airplay I send the sound to some speakers via Airplay through an Airport Express. At time I lose the connection to the speakers.
    I've complained about Wifi's instability, but here I tried to keep everything with Apples products to avoid any compatibility issues and stay within N wireless technology, which I understood it was much more stable.
    Has anyone some suggestions?

    Hi,
    you should analyze the db after you have loaded the tables.
    Do you use sequences to generate PKs? Do you have a lot of indexex and/or triggers on the tables?
    If yes:
    make sure your sequence caches (alter sequence s cache 10000)
    Drop all unneeded indexes while loading and disable trigger if possible.
    How big is your Redo Log Buffer? When loading a large amount of data it may be an option to enlarge this buffer.
    Do you have more then one DBWR Process? Writing parallel can speed up things when a checkpoint is needed.
    Is it possible using a direct load? Or do you already direct load?
    Dim

  • Performance issues with data warehouse loads

    We have performance issues with our data warehouse load ETL process. I have run
    analyze and dbms_stats and checked database environment. What other things can I do to optimize performance? I cannot use statspack since we are running Oracle 8i. Thanks
    Scott

    Hi,
    you should analyze the db after you have loaded the tables.
    Do you use sequences to generate PKs? Do you have a lot of indexex and/or triggers on the tables?
    If yes:
    make sure your sequence caches (alter sequence s cache 10000)
    Drop all unneeded indexes while loading and disable trigger if possible.
    How big is your Redo Log Buffer? When loading a large amount of data it may be an option to enlarge this buffer.
    Do you have more then one DBWR Process? Writing parallel can speed up things when a checkpoint is needed.
    Is it possible using a direct load? Or do you already direct load?
    Dim

Maybe you are looking for

  • How to send email notifications from a workflow?

    I am having problems trying to get notification emails to work from a workflow. I have done the following: 1.Checked notification to file is commented out in waveset.properties 2.Set SMTP server and port setting in Configure > Servers > email templat

  • How do I get iCloud to work on iPad2?

    I type in icloud in the address and it keeps going to set up this iPad 2.   All settings for iCloud are set up per instructions.  Am I missing a step here?  Comes up on my MacBook  pro just fine.  Want to be able to use beta iWorks on this device. Th

  • How Can I Control (or Understand) Genius Mix Lists?

    I am running iTunes 11. I decided to turn on Genius Mix (Menu--> Store--> Turn On Genius). This was fine. There was some information exchanged, some calculations done, etc. and an hour or so later I had a category under "Genius" of "Genius Mixes". Th

  • Switching to another supplier because...

    Constant re-boots No service information (programme guide0 displayed) Recording only manages half way through a programme I player streaming issues. Why do I pay £50+ for Infinity and phone when it doesnt work?

  • Best way to use live traced images?

    I've got a hand drawn logo I want to use on a photoshopped image.  If I live trace it in illustrator and then live paint it, what's the best way to import it into photoshop for use?  I'd like to do some advanced colouring and touching up of the logo