Should I specify the Parallel parameter for an non-RAC database?

The Oracle documatation state as the following:
"The Oracle Database 10g Release 2 database controls and balances all parallel operations, based upon available resources, request priorities and actual system load." It show that Oracle can optimize the Parallel level automaticly.
Should I specify the Parallel parameter for a non-RAC database? Most of the transactions are small OLTP.

What parallel parameter are you talking about?
Generally, you may benefit from parallelization in a very similar manner on RAC as on single instance system. And it is in both cases not sufficient to change the value of any initialization parameter to achieve parallelization of queries, DDL or DML.
Kind regards
Uwe
http://uhesse.wordpress.com

Similar Messages

  • How do I update/change the connect descriptor for a target RAC database in Cloud Control

    I have Cloud Control 12c Release 3 set up and monitoring many hosts and targets. However when monitoring my RAC (2 nodes), it cannot connect to the database instances (although it shows them as up). I get the message "IO Error: The Network Adapter could not establish the connection". I can see that the connect descriptior is using the vip, rather than the hostname.
    How can I update the connect descriptor within Cloud Control to the correct values?

    Modify the monitoring configuration following the below listed note
    How to Modify/Update Monitoring Configuration of a Target Database or Listener in 12C Cloud Control (Doc ID 1385769.1)

  • Using a variable for the "format" parameter for disk/directory name

    In the interest of maintainability, I'm trying to use an environment variable in the format parameter. I've found that RMAN is able to interpret the $ORACLE_HOME and $ORACLE_BASE variables, but others give the error:
    RMAN-03006: non-retryable error occurred during execution of command: backup
    RMAN-07004: unhandled exception during command execution on channel t1
    RMAN-10035: exception raised in RPC: ORA-07217: sltln: environment variable cannot be evaluated.
    RMAN-10031: ORA-19583 occurred during call to DBMS_BACKUP_RESTORE.BACKUPPIECECREATE
    Is there any more specific documentation about what variables are supported within RMAN? (I've found nothing in the manuals beyond the excerpt I included below.) Does anyone know if the "port-specific directory" mentioned in the excerpt below
    can be modified/customized?
    Thanks....
    format 'format_string'
    specifies the filename to use for the backup piece. Any name that is legal as a sequential filename on the platform is allowed, provided that each backup piece has a unique name. If backing up to disk, then any legal disk filename is allowed, provided it is unique. If you do not specify the format parameter, RMAN stores the backup pieces in a port-specific directory ($ORACLE_HOME/dbs on UNIX).

    i have used a different directory other then port-specific one and i am able to backup .
    format '/disk2/app/oracle/test/%d_%s_%t' , i have given full path names instead of substitution variables. i am not sure how
    they(variables) will work though in rman.

  • Used to convert all my videos into iPhone mp4 format using Any Video Converter - then add it to my iPhone 3GS. But last few days, I cant add video files to iTunes. and the iTunes doesnt even specify the reason except for that "files are not supported"

    Everyting was perfect earlier. Used to convert all my videos into iPhone mp4 format using Any Video Converter - then add it to my iPhone 3GS. But last few days, I cant add video files to iTunes. and the iTunes doesnt even specify the reason except for that "files are not supported"
    Everything is fine.
    -They are the right format
    -Quicktime is updated
    -I have tried simply dragging them into the 'movie' library
    -I have tried going to file<add file to library
    Nothing works, has anyone else had this problem and found a way around it?
    Any and all help appreciated.

    In addition to Mike's suggestions,  you only have 4GB of RAM and Mavericks does use more RAM than other versions. You may also want to look at the apps that startup on login and the possiblity of upgrading RAM.

  • Where do we specify the login details for an FTP adapter ?

    Hi,
    I am trying to do a simple ftp using an SOA-ESB service. I have created an FTP service using the FTP adapter. The FTP adapter asked abt the file name and directory. Where do I specify the host details like hostname, userid and Password?
    Thanks
    RV

    when you start the adapter wizard you specify a JNDI name it will be something like eis/Ftp/sample.
    You can make this anything but use the same notion, e.g. eis/Ftp/[hostname_user]
    You then login to Enterprise Manager
    oc4j_soa -> applications -> FTP Adapter (this may not be exact
    There you can specify the connection details for that server. Name the JNDI name eis/Ftp/[hostname_user]
    you can also modify the oc4j-ra.xml file directly for the FTP adapter
    this is found
    Oracle_Home\j2ee\[home or oc4j_soa]\application-deployments\default\FtpAdapter
    cheers
    James

  • How to pass a value to the export parameter for a method (public instance)?

    Hello,
      I am trying ABAP OO newly. How to pass a value to the export parameter for a method (public instance) called in a report? This *export parameter has a reference type of structure.
    Thanks in advance,
    Ranjini

    Hi,
    "class definition
    class lcl... definition.
      public section.
        method m1 imporitng par type ref to data.  "now you can pass any reference to the method, but this way you have to maintain this reference dynamically inside the method (you can't be sure what that reference is really "pointing" at)
    endclass.
    "in program
    data: r_lcl type ref to lcl...
    create object r_lcl.
    call method r_lcl
      exporting
         par    =  "pass any reference variable here
    Regards
    Marcin

  • The parallel process for mrp.

    hi exports
    we plan to do the scope of planning for the total planning as a background job.
    while doing that system ask for the parallal processing for mrp
    what is customize step and procedure to do the parallel process for mrp.

    Dear Raj,
    With the help of parallel processing procedures, you can significantly improve the runtime of the total planning run.
    To process in parallel, you can either select various sessions on the application server or various servers.
    Parallel processing runs according to packages using the low-level code logic:
    The work package, with a fixed number of materials that are internally defined in the program, is distributed over the individual servers/sessions. Once a server/session has finished processing a package, it starts processing the next package.
    If a low-level code is being planned, the servers/sessions that have finished must wait until the last server/session has finished its package to avoid inconsistencies. Then the next low-level code is processed per packages.
    The parallel processing procedure is switched on in the initial screen of total planning.
    Activities
    Define the application server with the number of sessions that can be used:
    If you want to define various servers for parallel processing, enter the server with the number of sessions.
    If you only want to use one server, but several sessions, enter the application server and the appropriate number of sessions.
    Further notes
    Parallel processing shortens the time required for calculation, however, it cannot shorten the database time as the system still only operates using one database.
    The Customizing Transaction is   OMIQ
    Regards
    PSV

  • How should i get the delta data for those new object's previous delta data?

    Hallo all,
    i have 2lis_11_vahdr and i am using a standard and loading delta! now i have appended some new fields in the datasource then in infosource then in ods and infocube..! now how should i get the delta data for those new object's previous delta data???
    Jimmy

    Hi,
    After appending new fields in the datasoure.It is not possible to delta load, u should have to load everything
    Thanks,
    Shreya

  • [Solved] if(Transaction specified for a non-transactional database) then

    I am getting started with BDBXML 2.4.14 transactions and XQuery update functionality and I am having some difficulty with 'node insert ...' and transactions failing with 'Transaction specified for a non-transactional database'
    Thanks for helping out.
    Setup:
    I have coded up a singleton manager for the XmlManger with a ThreadLocal holding the transaction and a query method to execute XQueries. The setup goes like this:
    environmentConfig = new EnvironmentConfig();
    environmentConfig.setRunRecovery(true);               environmentConfig.setTransactional(true);               environmentConfig.setAllowCreate(true);               environmentConfig.setRunRecovery(true);               environmentConfig.setInitializeCache(true);                environmentConfig.setTxnMaxActive(0);               environmentConfig.setInitializeLocking(true);               environmentConfig.setInitializeLogging(true);               environmentConfig.setErrorStream(System.err);
    environmentConfig.setLockDetectMode(LockDetectMode.MINWRITE);               environmentConfig.setJoinEnvironment(true);               environmentConfig.setThreaded(true);
    xmlManagerConfig = new XmlManagerConfig();               xmlManagerConfig.setAdoptEnvironment(true);               xmlManagerConfig.setAllowAutoOpen(true);               xmlManagerConfig.setAllowExternalAccess(true);
    xmlContainerConfig = new XmlContainerConfig();               xmlContainerConfig.setAllowValidation(false);               xmlContainerConfig.setIndexNodes(true);               xmlContainerConfig.setNodeContainer(true);
    // initialize
    instance.xmlManager = new XmlManager(instance.getEnvironment(),                    instance.getXmlManagerConfig());
    instance.xmlContainer = instance.xmlManager.openContainer(                              containerName, instance.getXmlContainerConfig());
    private ThreadLocal<XmlTransaction> transaction = new ThreadLocal<XmlTransaction>();
    public XmlTransaction getTransaction() throws Exception {
              if (transaction.get() == null) {
                   XmlTransaction t = xmlManager.createTransaction();
                   log.info("Transaction created, id: " + t.getTransaction().getId());
                   transaction.set(t);
              } else if (log.isDebugEnabled()) {
                   log.debug("Reusing transaction, id: "
                             + transaction.get().getTransaction().getId());
              return transaction.get();
         private XmlQueryContext createQueryContext(String docName) throws Exception {
              XmlQueryContext context = xmlManager.createQueryContext(
                        XmlQueryContext.LiveValues, XmlQueryContext.Lazy);
              List<NamespacePrefix> namespacePrefixs = documentPrefixes.get(docName);
              // declare ddi namespaces
              for (NamespacePrefix namespacePrefix : namespacePrefixs) {
                   context.setNamespace(namespacePrefix.getPrefix(), namespacePrefix
                             .getNamespace());
              return context;
         public XmlResults xQuery(String query) throws Exception {
              XmlQueryExpression xmlQueryExpression = null;
              XmlQueryContext xmlQueryContext = getQueryContext(docName);
              try {
                   xmlQueryExpression = xmlManager.prepare(getTransaction(), query,
                             xmlQueryContext);
                   log.info(query.toString());
              } catch (Exception e) {
                   if (xmlQueryContext != null) {
                        xmlQueryContext.delete();
                   throw new DDIFtpException("Error prepare query: " + query, e);
              XmlResults rs = null;
              try {
                   rs = xmlQueryExpression.execute(getTransaction(), xmlQueryContext);
              // catch deadlock and implement retry
              catch (Exception e) {
                   throw new DDIFtpException("Error on query execute of: " + query, e);
              } finally {
                   if (xmlQueryContext != null) {
                        xmlQueryContext.delete();
                   xmlQueryExpression.delete();
              return rs;
    <?xml version="1.0" encoding="UTF-8"?>
    <Test version="0.1">
    <Project id="test-project" agency="dda">
    <File id="large-doc.xml" type="ddi"/>
    <File id="complex-doc.xml" type="ddi"/>
    </Project>
    <Project id="2nd-project" agency="test.org"/>
    </Test>
    Problem:
    All the queries are run through the xQuery method and I do delete the XmlResults afterwards. How do I get around the 'Transaction specified for a non-transactional database' what is the transactions doing? How do I get state information out of a transaction? What am I doing wrong here?
    1 First I insert a node:
    Transaction created, id: -2147483647
    Adding document: large-doc.xml to xml container
    Reusing transaction, id: -2147483647
    Working doc: ddieditor.xml
    Root element: Test
    Reusing transaction, id: -2147483647
    insert nodes <Project id="JUnitTest" agency="test.org"></Project> into doc("dbxml:/ddieditor.dbxml/ddieditor.xml")/Test
    Reusing transaction, id: -2147483647
    2 Then do a query:
    Reusing transaction, id: -2147483647
    doc("dbxml:/ddieditor.dbxml/ddieditor.xml")/Test/Project/@id
    Reusing transaction, id: -2147483647
    3 The same query again:
    Reusing transaction, id: -2147483647
    doc("dbxml:/ddieditor.dbxml/ddieditor.xml")/Test/Project/@id
    Reusing transaction, id: -2147483647
    4 Delete a node:
    Reusing transaction, id: -2147483647
    delete node for $x in doc("dbxml:/ddieditor.dbxml/ddieditor.xml")/Test/Project where $x/@id = '2nd-project' return $x
    Reusing transaction, id: -2147483647
    5 Then an error on query:
    Reusing transaction, id: -2147483647
    doc("dbxml:/ddieditor.dbxml/ddieditor.xml")/Test/Project/@id
    Reusing transaction, id: -2147483647
    Transaction specified for a non-transactional database
    com.sleepycat.dbxml.XmlException: Error: Invalid argument, errcode = DATABASE_ERROR
         at com.sleepycat.dbxml.dbxml_javaJNI.XmlResults_hasNext(Native Method)
         at com.sleepycat.dbxml.XmlResults.hasNext(XmlResults.java:136)
    Message was edited by:
    jannikj

    Ok got it solved by increasing the locks lockers and mutex's I allso increased the the log buffer size:
    environmentConfig = new EnvironmentConfig();
                   // general environment
                   environmentConfig.setAllowCreate(true);
                   environmentConfig.setRunRecovery(true); // light recovery on startup
                   //environmentConfig.setRunFatalRecovery(true); // heavy recovery on startup
                   environmentConfig.setJoinEnvironment(true); // reuse of environment: ok
                   environmentConfig.setThreaded(true);
                   // log subsystem
                   environmentConfig.setInitializeLogging(true);
                   environmentConfig.setLogAutoRemove(true);
                   environmentConfig.setLogBufferSize(128 * 1024); // default 32KB
                   environmentConfig.setInitializeCache(true); // shared memory region
                   environmentConfig.setCacheSize(2500 * 1024 * 1024); // 250MB cache
                   // transaction
                   environmentConfig.setTransactional(true);
                   environmentConfig.setTxnMaxActive(0); // live forever, no timeout               
                   // locking subsystem
                   environmentConfig.setInitializeLocking(true);
    environmentConfig.setMutexIncrement(22);
    environmentConfig.setMaxMutexes(200000);
    environmentConfig.setMaxLockers(200000);
    environmentConfig.setMaxLockObjects(200000); // default 1000
    environmentConfig.setMaxLocks(200000);
    // deadlock detection
                   environmentConfig.setLockDetectMode(LockDetectMode.MINWRITE);
    In the docs by Oracle it is limited information given regarding the impact of these settings and their options. Can you guys point in a direction where I can find some written answers or it hands on?

  • Error -The trustAnchors parameter must be non-empty.

    Hi to all,
    i ve a send mail program, which i modified fm smtpsend.java,the sample come with javamail 1.4.3. this is working when i run from command prompt. but when i run this inside in jdeveloper 11g "the trustAnchors parameter must be non-empty" error is coming.
    i m using Weblogic Server 11g(10.3.2) & jdk6 update 16.
    Can anyone help me? wt is problem here?
    C:\Middleware11g2\jdk160_14_R27.6.5-32\bin\javaw.exe -client -classpath C:\JDeveloper\mywork\MyTestApp\classes;C:\JDeveloper\mywork\WorkFlow\javamail-1.4.3\javamail-1.4.3\mail.jar "-Djavax.net.ssl.trustStore=D:\Middleware11g 2\wlserver_10.3\server\lib\DemoTrust.jks" com.company.jsf.TestSmtp
    javax.mail.MessagingException: Could not connect to SMTP host: smtp.gmail.com, port: 465;
    nested exception is:
    javax.net.ssl.SSLException: java.lang.RuntimeException: Unexpected error: java.security.InvalidAlgorithmParameterException: the trustAnchors parameter must be non-empty.
    package com.company.jsf;
    import com.sun.mail.smtp.SMTPTransport;
    import java.util.Properties;
    import javax.mail.Message;
    import javax.mail.MessagingException;
    import javax.mail.SendFailedException;
    import javax.mail.Session;
    import javax.mail.internet.InternetAddress;
    import javax.mail.internet.MimeMessage;
    public class TestSmtp {
    public TestSmtp() {
    super();
    public static void main(String[] args){
    String to="[email protected]", subject = "Hi", from = "[email protected]", cc = null, bcc = null, url = null;
    String mailhost = "smtp.gmail.com";
    String user = "myid", password = "mypassword";
    boolean debug = false;
    boolean verbose = false;
    boolean auth = true;
    String prot = "smtps";
    try{
    Properties props = System.getProperties();
    if (mailhost != null)
    props.put("mail." prot ".host", mailhost);
    if (auth)
    props.put("mail." prot ".auth", "true");
    // Get a Session object
    Session session = Session.getInstance(props, null);
    if (debug)
    session.setDebug(true);
    Message msg = new MimeMessage(session);
    if (from != null)
    msg.setFrom(new InternetAddress(from));
    msg.setRecipients(Message.RecipientType.TO, InternetAddress.parse(to, false));
    if (cc != null)
    msg.setRecipients(Message.RecipientType.CC, InternetAddress.parse(cc, false));
    if (bcc != null)
    msg.setRecipients(Message.RecipientType.BCC, InternetAddress.parse(bcc, false));
    msg.setSubject(subject);
    String text = "hr u?";
    msg.setText(text);
    SMTPTransport t = (SMTPTransport)session.getTransport(prot);
    try {
    if (auth)
    t.connect(mailhost, user, password);
    else
    t.connect();
    t.sendMessage(msg, msg.getAllRecipients());
    finally {
    if (verbose)
    System.out.println("Response: " + t.getLastServerResponse());
    t.close();
    System.out.println("\nMail was sent successfully.");
    catch(Exception e){
    e.printStackTrace();
    Edited by: user7895 on May 20, 2010 12:47 AM

    Hi,
    is this topic *"Customizing JSSE-The Installation Directory <java-home>"* enough or there is something more? bz it's not a small doc, worthless to read whole, i believe.
    i ll appreciate if u specify related topics, other than above.
    thanks with regards
    Edited by: user7895 on May 20, 2010 11:44 AM
    Edited by: user7895 on May 20, 2010 11:45 AM

  • Can I use data guard to create a RAC standby database for a non RAC primary

    Hi,
    we need to RAC our production database but the normal methods will mean a long outage. It is possible to create a standby as a single node RAC database and when ready do a graceful failover to the standby database and open it for business. The next step would be to create another RAC node from this on the original server.
    servers are already cluster aware, using ASM etc
    Oracle 10.2

    Yes, you will be able to setup RAC stnadby for a non-RAC Primary. For primary it just needs a available destination for redo shipping it doesn't matter whether it's RAC enabled or not. And ofcourse you are using 10.2 anyway only one node will be running MRP and that is too in standby mount mode.
    However since you have are using You may follow below sequence.
    1. Setup a new standby as RAC enabled.
    2. Perform a switchover.
    3. Shutdown the Old primary (which is standby now).
    4. Install CRS and RDBMS on the old primary and it's new node.
    5. Modify the cluster_database=TRUE and cluster_database_instances=<required number of instances>.
        With above modification mount the standby database in standby mode and start MRP.
    6. Introduce the database and instances to the OCR using SRVCTL add command.
    7. Once you your database is synchronized with Primary do a switchover.
    9. Now you can repeat step 3 to 6 on the other site too.   <- if you need your secondary site to be RAC enabled too
    10. Finally both the sites should be RAC enabled.
    Hope this is helpful!!!
    Thanks,
    Asif Haliyal

  • Best way to have resizeable LUNS for datafiles - non RAC system

    All,
    (thanks Avi for the help so far, I know its a holiday there so I wait for your return and see if any other users can chip in also)
    one of our systems (many on the go here) is being provided by an external vendor - I am reviewing their design and I have some concerns about the LUN's to house the datafiles:
    they dont want to pre-assign full size luns - sized for future growth - and want more flexibility to give each env less disk space in the beginning and allocate more as each env grows
    they are not going to use RAC (the system has nowhere near the uptime/capacity reqs - and we are removing it as it has caused enormous issues with the previous vendors and their lack of skills with it - we want simplicity)
    They have said they do not want to use ASM (I have asked for that previously I think they have never used it before - I may be able to change their minds on this but they are saying as not RAC not needed)
    but they are wondering how they give smaller luns to the env and increase size as they grow - but dont want to forever be adding /u0X /u0Y /u0Z extra filesystems (ebusiness suite rapidclone doesn't like working with many filesystem anyway and I find it unelegant to have so many mount points)
    they have suggested using large ovm repo's and serving the data filesystems out of that (i have told them to use the repo's just for the guest OS's and use direct phy's attached luns for datafiles (5TB of them)
    now they have suggested creating a large LUN (large enuogh for many envs at the same time [dev / test1 / test2 etc]) .... and putting OCFS2 on it so that they can mount it to all the domU/guest's and they can allocate space as needed uot of that:
    so that they have guests/VM's (DEV1 - DEV2 - TEST1 say) (all seperate vm's) and all mounting the same OCFS2 cluster filesystem (as /u01 maybe) and they can share that for the datafiles under a sep dir so that each DB VM would see:
    /u01/ and as subdirectories to that DEV1 DEV2 TEST1 so:
    /u01/DEV1
    /u01/DEV2
    /u01/TEST1
    and only use the right directory for each guests datafiles (thus sharing the space in u01(the big LUN) as needed per env)....
    i really dont like that as each guest is going to have the same oracle unix user details and able to write to each other dir's - id prefer dedicated LUNS for each VM - not mounted to many VM's
    so I am looking for a way to suggest something better....
    should I just insist on ASM (but this is a risk as I fear they are not experienced with it)
    or go with OEL/RHEL LVM and standard ext filesystems that can be extended - what are the risks with this? (On A Linux Guest For OVM, Which Partitions Can Be LVM? [ID 1080783.1]) - seems to say there is little performance impact
    or is there another option?
    Thanks all
    Martin
    Edited by: Martin Brambley on 11-Jun-2012 08:53

    Martin, what route did you end up going?
    We are about to deploy several hundred OEL VMs that are going to run non-RAC database instances. We don't plan to use ASM either. Our plan right now is to use one large 3TB LUN virtual disk to carve out the operating system space for the VMs and then have a separate physical attached LUN for each VM that will host a /u01 filesystem using LVM. I have concerns with this as we don't know how much space /u01 will ultimately need and if we end up having to extend /u01 on all of these VMs, that sounds like it will be messy. Right now I've got 400 separate 25gb LUNs presented to all of my OVM servers that we plan to use for /u01 filesystems.

  • Is X-windows and GUI desktops supported on the ODA "engineered system" running a RAC database?  If it is, what is the yum command needed to install the X-windows, Gnome, and KDE package groups?

    Is X-windows and GUI desktops supported on the ODA "engineered system" running a RAC database?  If it is, what is the yum command needed to install the X-windows, Gnome, and KDE package groups?

    While I agree with the direction of the suggestions with installing packages for X-windows, we do not have a blanket 'apply any package' recommendation.
    In particular we do not support altering the kernel (although we do have exceptions which we review on a case by case basis).
    Basically, if the you want to alter functionality that would not impact core functionality you are usually fine.
    A good guideling is : The more dependencies that there are between the package / rpm you are considering using the higher the potential impact on functionality - meaning higher chance for problems
    Note: We do use VNC including Real and Tiger regularly , but we have no hard recommendation on how you may want to use X-windows. I have never seen a limitation other than comments on bugs
    or incompatibility within the X-window product itself with certain kernel levels.
    Patching may overwrite some packages that you may install, however,  _depending on packages/rpms added_ there is also the possibility that you will break existing functionality to the point
    that patching itself will fail ( we have already seen a few cases of this in which case the proper mitigation is to remove / roll-back any alterations to the ODA before patching, and then adding the packages/rpms
    back after the patching is completed.
    From what you are discussing the impact should be low without conflicts, but please consider the above, and if you have specific packages which you consider potential problems
    please create an SR so that we can review packages / rpms on an individual basis.
    Once again: the main criteria for not supporting rpms is regarding the kernel itself
    Chuck

  • Changing the Physical IP address of Oracle RAC database Server

    Hi All,
    We are planning to change the Physical IP address of Oracle RAC database Server. I would like know to, what all are the changes need to be done the from Oracle part.
    Thanks in Advance

    Check document 283684.1 on metalink and/or
    http://orcl-experts.info/index.php?name=FAQ&id_cat=9

  • What value should be passed to the ColumnSize parameter for SQLBindParameter() for a nvarchar(n) column?

    I have a table with a field "f1" which type is nvarchar(1000). Now I want to update a record through binding variable method. Before I bind the parameter, I call SQLDescribeCol to get the length of f1 - 1000.
    The data for updating f1 is prepared and is coverted to UTF16, so the buffer length is 2*(character length). When I call SQLBindParameter() to bind the variable for f1, what length value should be passed to ColumnSize parameter? Some said the Columnsize
    should be in byte. In MSDN, I can't find any posts or examples for my cases. MSDN also mentioned that the limit for this argument is 4000.
    I try 1000 and 2000, both works. Which one is the exactly correct one? I use SQL Server 2008 and native client 10.0.
    Thanks in advance.
    Liu Peng

    Hello,
    Thank you for your question. I am trying to involve someone more familiar with this topic for a further look at this issue. Sometime delay might be expected from the job transferring. Your patience is greatly appreciated. 
    Thank you for your understanding and support.
    Elvis Long
    TechNet Community Support

Maybe you are looking for

  • Searching for Workflow for Purchase Order Change

    Hi experts! We are using SRM5.0 / extended classic scenario and backend R/3 4.7 Enterprise. We should implement an approver process for the purchase order changes. Does any one of you know whether there exists a standard SAP workflow? Many thanks for

  • The on the go list, doesn't work as it must be.

    This is my problem: if i choose a song like tiesto it will show in my play list, choosen bij on the go. armin van buuren. It's stupid :s.. and if i try an other song it allways shows an other song in my play list. Somebody else with the same problem?

  • Upload printer driver in SAP

    Hi , I am trying to upload the new printer driver in SAP. I downloaded the new SAP printer driver and trying to load into SAP thru SPAD. when I am trying to import the driver its asking for Development Class. PLEASE let me know which Development clas

  • Trouble with loadMasters function.

    Hey everybody, I'm trying to automate the replacement of a number of master spreads across many documents. Unfortunately, I keep getting the following error: Error Number: 48 Error String: Cannot find the folder " ". Both of these code snippets give

  • BLOB causing FRM-40505

    Any clue why a blob field in a datablock would cause an FRM-40505 (query failed). This only happens only after I upload to the server (Forms Servlet and 9iDB). Works fine in the developer (client/server mode). HELP!