Performance implications of using multiple rules

Hi:
1.  I have a question about the performance implications of multiple rules through a rule strategy sequence.  Can you share your experience on this?
2. Suppose that I have three rules that are accessed through a rule strategy sequence.  Would the increase in time to perform these three rules linear or some other function?
Thanks.

Satish,
Depends on the rule control you are using.
help.sap.com has given a really good example. Check this out
http://help.sap.com/saphelp_scm41/helpdata/en/26/c2d63b18bc7e7fe10000000a114084/frameset.htm
Let me know If you have trouble understanding or need some more help.
Regards
Kumar Ayyagari

Similar Messages

  • Performance Implications When Assigning Multiple VCPUs

    A wise man once told me not to assign more than two virtual CPUs to a VM due to CPU reservations, which in turn means slower performance. Does Oracle VM manage CPU requests better than VMware so multiple CPUs can be assigned to the VMs? Anyone have any benchmarking?

    "1. The only option within the Assign Task operation that lets you assign the user dynamically is by using an XPath expression."
    Yes
    "If I were to use this, should I be embedding the logic within the form itself and then use the XPath expression to find the user by dynamically updating the form after the first user completes the task?"
    I depends. If you can get to it server side, it woulb probably better. If you need to make a change to something, you don't want to change the form.
    "Is there another approach to achieve this other than having the logic for selecting the user defined in the form itself?"
    I'm not sure where the logic for the next user is located, but you could make a database or web service call from the process to get the next user instead.
    "I believe embedding the logic in the form might not be a good idea as it can have performance implications and also having user information in the form itself doesn't seem right."
    I agree.
    "Is there an operation that I could use, which would contain the logic that I would need to determine the next user and then assign the task to the user based on the result of the operation?"
    Again, I need a bit more info as to where the info (logic) is located.
    Jasmin

  • Performance Issue in using Oracle Rules SDK!!!!

    Hi,
    I am using Oracle rules SDK. I have created a dictionary and declared 9 global variables in it. Now, before testing my ruleset in my code, I populate those variables and update datamodel.
    Date startDate = new Date();
    try
    this.dataModel.update();
    *}catch (Exception e) {*
    e.printStackTrace();
    Date endDate = new Date();
    long duration = endDate.getTime() - startDate.getTime();
    System.out.println("Time Taken : "+duration);
    Now the issue is that time taken by dataModel.update() is freaking 4210 milliseconds, i.e 4.2 seconds. Any idea why this is slow??
    oracle.rules.sdk.editor.datamodel.DataModel = new DataModel(ruleDictionary);

    When your query takes too long ...
    Thanks,
    Karthick.

  • Implications of Using a Rule to Populate Periodic Data in a YTD app

    Is it bad practice to specify and populate an intersection with periodic data using rules when the application is set up as YTD?
    For example:
    Entity A has a $50 intercompany expense with Entity B. Entity B is reporting a $48 intercompany income with Entity A. We have a rule set up to populate a statistical "counterparty" account to show what the ICP is reporting. This way we can create custom SV and Financial Reports and also allow users to view info from entities (ICPs) in which they might not have access.
    So the YTD line items are...
    Entity A, ICP B shows interco_exp $50 and counterparty_interco_inc $48
    Entity B, ICP A shows interco_inc $48 and counterparty_interco_exp $50
    The question on Periodic vs YTD comes in when one entity is a different functional currency than the icp. Writing the rule is much easier using periodic data for P&L accounts (using the period fx rate to back into the functional currency of the entity).
    Thoughts/suggestions?
    Thanks.

    This would be done in InfoPath.
    Click on your Matter Name field and then select properties. For default value, select the data source as your other data source and then highlight the Matter name field. Click filter data and use the form field for Matter number.
    Check out this article:
    http://blogs.msdn.com/b/bharatgupta/archive/2013/03/07/create-cascading-dropdown-in-browser-enabled-infopath-form-using-infopath-2010.aspx
    Andy Wessendorf SharePoint Developer II | Rackspace [email protected]

  • LAPTOP PERFORMANCE AFFECTED WHEN USING MULTIPLE DISPLAYS

    Using a 15" MacBook Pro with 2.4GHz and 8 megs of ram running Lion, I created and ran a video-intensive Keynote file successfully from beginning to end.  But when connected to an external monitor, the entire laptop became sluggish, slow and non-responsive. Keynote would start and stop during playback, and even Desktop minimize a nimations ran very slowly... Any ideas WHY???

    How much ram in the MB?
    How much free space on the HD?

  • Export to shared buffer performance Implications

    All,
    I am using the following statement in one of the programs to export an internal table
    EXPORT it_dtab TO SHARED BUFFER indx(st) ID 'ZMME'.
    I like to know what the performance implications by using this ?

    Helo,
    You could check the current memory limit set for SAP shared space @ transaction RZ11 .
    Pass the profile parameter as rsdb/obj/buffersize and check the field 'Current Value' .
    You could discuss with your Basis counterpart if there's a need to increase the shared memory space which
    is generally not favourable as performance of other sap applns can come down on doing so.
    Regards
    Dedeepya C

  • What are the performance implications moving apps using cloud drive to Azure File Services?

    I run a number of cloud services with 5 or more nodes in using cloud drives. Cloud drive is scheduled to be deprecated in 2015. So I am thinking of replacing the cloud drive with Azure Files service.
    For each cloud service I am using one storage account to create all the the VHD/cloud drives. Some people at the time when cloud drive first appeared, told me that to get better performance, I should create only one VHD/Cloud Drive
    under only one storage account. For example, if I have five instances under a worker role then I should create 5 storage accounts and create one VHD/Cloud Drive under each storage account to be used by each node. I didn't follow that route because I was satisfied
    with the performance of the apps under cloud services having all VHD/Cloud Drives under one storage account.
    My question is, if I replace cloud drive with Azure file services, will my apps perform well having all shares under one storage account or create one storage account for each share?
    Thanks,
    @nazik_huq

    Thanks Obama for replying.
    Here is the comment from @jaiharidas of MSFT if anyone's interested:
    @Naziq, It is better to have multiple shares under single storage account and there is no perf implications. However, please ensure that your ingress/egress and request/sec is within
    the limits of a single storage account (seemsdn.microsoft.com/.../dn249410.aspx)
    and use multiple storage accounts if you need to scale beyond the limits.
    See the original comment  on Azure Storage Team here: http://ow.ly/ChPNf 
    @nazik_huq

  • MAXL - using multiple Dimension build rules in single script...

    I seem to be having problems running multiple Dimension build rules in a single MAXL script, is this a known problem or just an issue that I'm having? The build rules independently work as designed but when joined into a single script the script hangs...
    So, this is a 2 part question...
    1) Does MAXL not play nice when using multiple Dimension load rules in a single script?
    2) Do I need to run the build rules independent of each other?
    Also, is there some type of refresh/update on the cube that you can use in the script and would it be necessary...?
    Thanks for any light you can shed on the issue.
    Adam,

    I've written multiple dimension build statements as a part of 1 long MaxL statement. I'm executing this MaxL via a shell script (that replaces the variables such as username, password, servername, file path etc.) I've not seen any issues doing so.
    However, I've noted that, number of imports statements in 1 single MaxL command does not bode well for more than *60* import statements at once. So, if you're using more than 60 source files to build your dimensions, please consider breaking up your MaxL into multiple statements.
    Note: All but last of the import statements have 'deferred' re-structure specified. It speeds up the build process.

  • How to "Performing an Action on Multiple Selected Records" using Struts-JSP

    Hi all:
    How to do "Performing an Action on Multiple Selected Records" through using Struts/JSP solution that is original MVC/UIX customizing example by Jheadstart.
    I know how to reason by analogy through tutorial guide on
    Model and Control layer of MVC , but for View layer that is torment to me !
    have any idea ?
    thanks a lot .

    Ting Rung,
    In the table in the JSP page you add a checkbox like this:
    <td class="fixed">
    <input type="checkbox" name="multiselect" value="<%=index%>"/>
    </td>
    This will render a checkbox in each row, with index number of that row as value.
    When you submit the page the request contains the multiselect parameter when at least one row has the checkbox checked.
    You can get the value of the multiselect parameter by calling sessionData.getRequestParameterValues("multiselect") in your action. The value is a string array containing the index numbers of the rows that have the checkboc checked.
    Steven Davelaar,
    JHeadstart Team.

  • SSP5: Using multiple JVM for load balance performance?

    Sun 12 MAY 2002
    Apps 11.0.3
    SSP5 patchset I
    HP UX 11.0
    db 8.1.7.2 (64-bit)
    Load 60 concurrent sessions, each spawning 2.5 - 3 http connections.
    CPU 3:750Mhz
    RAM 8G
    Is anybody using multiple JVM for load balance?
    What is your ratio of JVM to concurrent iP sessions?
    Are you running Apache/Jserv on a server with any other applications, or is Apache by itself?
    If you are not using JVM, how many httpd processes do you get before Apache implodes? We stopped in the water at 90 httpd processes, but performance degraded starting at 70 sessions.
    Thx - Don

    Using Web Cache to load balance servlet-based Forms (6i and 9i) is unofficially supported. I say "unofficially" because we have actual customers doing it and getting support, but the 2 development teams (Forms and Web Cache) haven't actually done any integration testing of this sort of configuration yet. For your case, please contact your Support rep and ask what was done to use Web Cache as a load balancer for Forms6i at METRO in Germany. The Forms product managemment team is writing up a white paper to describe how to do it, but until then, you'll need to go through Support. Please contact me if you want more information.

  • Performance implications when working in Session.Record mode

    Performance implications when working in Session.Record mode
    Hi,
    Iu2019m considering working in GuiSession.Record mode (in SAPGui API) all the time for all SAP client applications, as part of a solution aimed to detect UI changes.
    I wanted to know if working in GuiSession.Record carries significant performance implications? Other drawback? Is there any reason why working in record mode would not be advisable?
    Thanks.

    Hello again,
    After a bit of search, I found it.
    I tried Method #1, it wasn't working with a corrupt cache at all for me.
    I had to do Method #2, manual mode, I entered the registry key manually.
    Let me fix it myself
    If you cannot access the Offline
    Files tab, use this method to re-initialize the Offline Files (CSC) cache on the system by modifying the registry. Use this
    method also to re-initialize the offline files database/client-side cache on multiple systems. Add the following registry subkey: 
    HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows\CurrentVersion\NetCache
    Key Name: FormatDatabase 
    Key Type: DWORD 
    Key Value: 1
    Note The actual value of the registry key is ignored. This registry change requires a restart. When the computer is
    restarting, the shell will re-initialize the CSC cache, and then delete the registry key if the registry entry exists. 
    Warning All cache files are deleted and unsynchronized data is lost.
    http://support.microsoft.com/kb/230738/en-us
    I want to hear back from you, tell me if it worked.
    Thanks !
    Konnan
    Understanding XP Mode, the key to success

  • Performance issues when using AQ notification with one consumer

    We have developed a system to load data from a reservation database to a reporting database
    A a certain point in the proces, a message with the identifier of the reservation is enqueued to a queue (multi-consumer) on the same DB and then propagated to a similar queue on the REP database.
    This queue (multi-consumer) has AQ notification enabled (with one consumer) which calls the queue_callback procedure which
    - dequeues the message
    - calls a procedure to load the Resv data into the Reporting schema (through DB link)
    We need each message to be processed ONLY ONCE thus the usage of one single subscriber (consumer)
    But when load testing our application with multiple threads, the number of records created in the Reservation Database becomes quite large, meaning a large number of messages going through the first queue and propagating to the second queue (very quickly).
    But messages are not processed fast enough by the 2nd queue (notification) which falls behind.
    I would like to keep using notification as processing is automatic (no need to set up dbms_jobs to dequeue etc..) or something similar
    So having read articles, I feel I need to use:
    - multiple subscribers to the 2nd queue where each message is processed only by one subscriber (using a rule : say 10 subscribers S0 to S10 with Si processing messages where last number of the identifier is i )
    problem with this is that there is an attempt to process the message for each subscriber, isn't there
    - a different dequeuing method where many processes are used in parallel , with each message is processed only by one subscriber
    Does anyone have experience and recommendations to make on how to improve throughput of messages?
    Rgds
    Philippe

    Hi, thanks for your interest
    I am working with 10.2.0.4
    My objective is to load a subset of the reservation data from the tables in the first DB (Reservation-OLTP-150 tables)
    to the tables in the second DB (Reporting - about 15 tables at the moment), without affecting performance on the Reservation DB.
    Thus the choice of advanced queueing (asyncronous )
    - I have 2 similar queues in 2 separate databases ( AND Reporting)
    The message payload is the same on both (the identifier of the reservation)
    When a certain event happens on the RESERVATION database, I enqueue a message on the first database
    Propagation moves the same message data to the second queue.
    And there I have notification sending the message to a single consumer, which:
    - calls dequeue
    - and the data load procedure, which load this reservation
    My performance difficulties start at the notification but I will post all the relevant code before notification, in case it has an impact.
    - The 2nd queue was created with a script containing the following (similar script for fisrt queue)
    dbms_aqadm.create_queue_table( queue_table => '&&CQT_QUEUE_TABLE_NAME',
    queue_payload_type => 'RESV_DETAIL',
    comment => 'Report queue table',
    multiple_consumers => TRUE,
    message_grouping => DBMS_AQADM.NONE,
    compatible => '10.0.0',
    sort_list => 'ENQ_TIME',
    primary_instance => '0',
    secondary_instance => '0');
    dbms_aqadm.create_queue (
    queue_name => '&&CRQ_QUEUE_NAME',
    queue_table => '&&CRQ_QUEUE_TABLE_NAME',
    max_retries => 5);
    - ENQUEUING on the first queue (snippet of code)
    o_resv_detail DLEX_AQ_ADMIN.RESV_DETAIL;
    o_resv_detail:= DLEX_AQ_ADMIN.RESV_DETAIL(resvcode, resvhistorysequence);
    DLEX_RESVEVENT_AQ.enqueue_one_message (o_resv_detail);
    where DLEX_RESVEVENT_AQ.enqueue_one_message is :
    PROCEDURE enqueue_one_message (msg IN RESV_DETAIL)
    IS
    enqopt           DBMS_AQ.enqueue_options_t;
    mprop           DBMS_AQ.message_properties_t;
    enq_msgid           dlex_resvevent_aq_admin.msgid_t;
    BEGIN
    DBMS_AQ.enqueue (queue_name => dlex_resvevent_aq_admin.c_resvevent_queue,
    enqueue_options => enqopt,
    message_properties => mprop,
    payload => msg,
    msgid => enq_msgid
    END;
    - PROPAGATION: The message is dequeued from 1st queue and enqueued automatically by AQ propagation into this 2nd queue.
    (using a call to the following 'wrapper' procedure)
    PROCEDURE schedule_propagate (
    src_queue_name IN VARCHAR2,
    destination IN VARCHAR2 DEFAULT NULL
    IS
    sprocname dlex_types.procname_t:= 'dlex_resvevent_aq_admin.schedule_propagate';
    BEGIN
    DBMS_AQADM.SCHEDULE_PROPAGATION(queue_name => src_queue_name,
                                            destination => destination,
    latency => 10);
    EXCEPTION
    WHEN OTHERS
    THEN
    DBMS_OUTPUT.put_line (SQLERRM || ' occurred in ' || sprocname);
    END schedule_propagate;
    - For 'NOTIFICATION': ONE subscriber was created using:
    EXECUTE DLEX_REPORT_AQ_ADMIN.add_subscriber('&&STQ_QUEUE_NAME','&&STQ_SUBSCRIBER',NULL,NULL, NULL);
    this is a wrapper procedure that uses:
    DBMS_AQADM.add_subscriber (queue_name => p_queue_name, subscriber => subscriber_agent );
    Then notification is registered with:
    EXECUTE dlex_report_aq_admin.register_notification_action ('&&AQ_SCHEMA','&&REPORT_QUEUE_NAME','&&REPORT_QUEUE_SUBSCRIBER');
    - job_queue_processes is set to 10
    - The callback procedure is as follows
    CREATE OR REPLACE PROCEDURE DLEX_AQ_ADMIN.queue_callback
    context RAW,
    reginfo SYS.AQ$_REG_INFO,
    descr SYS.AQ$_DESCRIPTOR,
    payload RAW,
    payloadl NUMBER
    IS
    s_procname CONSTANT VARCHAR2 (40) := UPPER ('queue_callback');
    r_dequeue_options DBMS_AQ.DEQUEUE_OPTIONS_T;
    r_message_properties DBMS_AQ.MESSAGE_PROPERTIES_T;
    v_message_handle RAW(16);
    o_payload RESV_DETAIL;
    BEGIN
    r_dequeue_options.msgid := descr.msg_id;
    r_dequeue_options.consumer_name := descr.consumer_name;
    DBMS_AQ.DEQUEUE(
    queue_name => descr.queue_name,
    dequeue_options => r_dequeue_options,
    message_properties => r_message_properties,
    payload => o_payload,
    msgid => v_message_handle
    -- Call procedure to load data from reservation database to Reporting DB through the DB link
    dlex_report.dlex_data_load.load_reservation
    ( in_resvcode => o_payload.resv_code,
    in_resvHistorySequence => o_payload.resv_history_sequence );
    COMMIT;
    END queue_callback;
    - I noticed that messages are not taken out of the 2nd queue,
    I guess I would need to use the REMOVE option to delete messages from the queue?
    Would this be a large source of performance degradation after just a few thousand messages?
    - The data load through the DB may be a little bit intensive but I feel that doing things in parallel would help.
    I would like to understand if Oracle has a way of dequeuing in parallel (with or without the use of notification)
    In the case of multiple subscribers with notification , does 'job_queue_processes' value has an impact on the degree of parallelism? If not what setting has?
    And is there a way supplied by Oracle to set the queue to notify only one subscriber per message?
    Your advice would be very much appreciated
    Philippe
    Edited by: user528100 on Feb 23, 2009 8:14 AM

  • OracleAS J2EE Stretch Cluster ... performance implications ...

    ... reposting here ...
    Hi,
    Customer is looking at putting in a J2EE Type installation of OracleAS 10g.
    To provide high availabilty and disaster recovery, looking at having multiple OracleAS middle tiers split across sites. Initially 2 middle tiers located in different data centres, looking to expand middle tiers as capacity demands increase.
    Also looking at clustering these in an OracleAS cluster configuration.
    This is for a BPEL implementation.
    Are there any performance implications that I should be aware of in such a configuration?
    Any implications of creating an OracleAS Cluster using servers located in physically different locations?
    Any other implications I should be aware of?
    Thanks in advance

    Hi gs86,
    Good questions :).
    To have a BPEL HA configuration you do not require a clustered J2EE & WC. All you need is BPEL on a 10.1.2.0.2 instance pointing to the same dehydration store.You do not get any of the HA benefits for BPEL when using a 10G AS J2EE cluster. It is not managed by DCM and none of the suitcases are centrally deployed so you do not get single point of administration.
    Do note that if you are also planning to use the J2EE and WC instances as a normal J2EE cluster then you can have BPEL in this cluster as well, just note that the BPEL instance won't "listen" to cluster wide DCM commands.
    If you want to just have a BPEL cluster then the configuration should be a Load Balancer, multiple J2EE AS instances (not clustered) and BPEL installed on these AS instances, pointing to the same deyhydration store.
    Any implications of creating an OracleAS Cluster using servers located in physically different locations?
    No, besides the usual network latency there "should" be no performance implications. Hopefully they will be on the same subnet so that they can access each other.
    Hope this helps.
    Deepak

  • Slow performance refreshing report with multiple subreports in main report

    I am using multiple subreports in my main report that appear to be slowing down the performance of the report considerably. Is there something I can do to increase performance with subreports.

    Hi Brian,
    Subreports are a known performance hog.
    I would suggest going back to the drawing board and rethinking on the layout of the report. The best thing, of course, would be to completely get rid of the subreports by creating a single SQL statement that takes care of all the report's requirements.
    Sometimes it is imperative to have subreports, in which case, you should try to make sure any selection filters from both the Main Report and the Subreport are passed back to the database.
    Also, avoid using Page N of M for better performance - with or without Subreports.
    -Abhilash

  • How to tune data loading time in BSO using 14 rules files ?

    Hello there,
    I'm using Hyperion-Essbase-Admin-Services v11.1.1.2 and the BSO Option.
    In a nightly process using MAXL i load new data into one Essbase-cube.
    In this nightly update process 14 account-members are updated by running 14 rules files one after another.
    These rules files connect 14 times by sql-connection to the same oracle database and the same table.
    I use this procedure because i cannot load 2 or more data fields using one rules file.
    It takes a long time to load up 14 accounts one after other.
    Now my Question: How can I minimise this data loading time ?
    This is what I found on Oracle Homepage:
    What's New
    Oracle Essbase V.11.1.1 Release Highlights
    Parallel SQL Data Loads- Supports up to 8 rules files via temporary load buffers.
    In an Older Thread John said:
    As it is version 11 why not use parallel sql loading, you can specify up to 8 load rules to load data in parallel.
    Example:
    import database AsoSamp.Sample data
    connect as TBC identified by 'password'
    using multiple rules_file 'rule1','rule2'
    to load_buffer_block starting with buffer_id 100
    on error write to "error.txt";
    But this is for ASO Option only.
    Can I use it in my MAXL also for BSO ?? Is there a sample ?
    What else is possible to tune up nightly update time ??
    Thanks in advance for every tip,
    Zeljko

    Thanks a lot for your support. I’m just a little confused.
    I will use an example to illustrate my problem a bit more clearly.
    This is the basic table, in my case a view, which is queried by all 14 rules files:
    column1 --- column2 --- column3 --- column4 --- ... ---column n
    dim 1 --- dim 2 --- dim 3 --- data1 --- data2 --- data3 --- ... --- data 14
    Region -- ID --- Product --- sales --- cogs ---- discounts --- ... --- amount
    West --- D1 --- Coffee --- 11001 --- 1,322 --- 10789 --- ... --- 548
    West --- D2 --- Tea10 --- 12011 --- 1,325 --- 10548 --- ... --- 589
    West --- S1 --- Tea10 --- 14115 --- 1,699 --- 10145 --- ... --- 852
    West --- C3 --- Tea10 --- 21053 --- 1,588 --- 10998 --- ... --- 981
    East ---- S2 --- Coffee --- 15563 --- 1,458 --- 10991 --- ... --- 876
    East ---- D1 --- Tea10 --- 15894 --- 1,664 --- 11615 --- ... --- 156
    East ---- D3 --- Coffee --- 19689 --- 1,989 --- 15615 --- ... --- 986
    East ---- C1 --- Coffee --- 18897 --- 1,988 --- 11898 --- ... --- 256
    East ---- C3 --- Tea10 --- 11699 --- 1,328 --- 12156 --- ... --- 9896
    Following 3 out of 14 (load-) rules files to load the data columns into the cube:
    Rules File1:
    dim 1 --- dim 2 --- dim 3 --- sales --- ignore --- ignore --- ... --- ignore
    Rules File2:
    dim 1 --- dim 2 --- dim 3 --- ignore --- cogs --- ignore --- ... --- ignore
    Rules File14:
    dim 1 --- dim 2 --- dim 3 --- ignore --- ignore --- ignore --- ... --- amount
    Is the upper table design what GlennS mentioned as a "Data" column concept which only allows a single numeric data value ?
    In this case I cant tag two or more columns as “Data fields”. I just can tag one column as “Data field”. Other data fields I have to tag as “ignore fields during data load”. Otherwise, when I validate the rules file, an Error occurs “only one field can contain the Data Field attribute”.
    Or may I skip this error massage and just try to tag all 14 fields as “Data fields” and “load data” ?
    Please advise.
    Am I right that the other way is to reconstruct the table/view (and the rules files) like follows to load all of the data in one pass:
    dim 0 --- dim 1 --- dim 2 --- dim 3 --- data
    Account --- Region -- ID --- Product --- data
    sales --- West --- D1 --- Coffee --- 11001
    sales --- West --- D2 --- Tea10 --- 12011
    sales --- West --- S1 --- Tea10 --- 14115
    sales --- West --- C3 --- Tea10 --- 21053
    sales --- East ---- S2 --- Coffee --- 15563
    sales --- East ---- D1 --- Tea10 --- 15894
    sales --- East ---- D3 --- Coffee --- 19689
    sales --- East ---- C1 --- Coffee --- 18897
    sales --- East ---- C3 --- Tea10 --- 11699
    cogs --- West --- D1 --- Coffee --- 1,322
    cogs --- West --- D2 --- Tea10 --- 1,325
    cogs --- West --- S1 --- Tea10 --- 1,699
    cogs --- West --- C3 --- Tea10 --- 1,588
    cogs --- East ---- S2 --- Coffee --- 1,458
    cogs --- East ---- D1 --- Tea10 --- 1,664
    cogs --- East ---- D3 --- Coffee --- 1,989
    cogs --- East ---- C1 --- Coffee --- 1,988
    cogs --- East ---- C3 --- Tea10 --- 1,328
    discounts --- West --- D1 --- Coffee --- 10789
    discounts --- West --- D2 --- Tea10 --- 10548
    discounts --- West --- S1 --- Tea10 --- 10145
    discounts --- West --- C3 --- Tea10 --- 10998
    discounts --- East ---- S2 --- Coffee --- 10991
    discounts --- East ---- D1 --- Tea10 --- 11615
    discounts --- East ---- D3 --- Coffee --- 15615
    discounts --- East ---- C1 --- Coffee --- 11898
    discounts --- East ---- C3 --- Tea10 --- 12156
    amount --- West --- D1 --- Coffee --- 548
    amount --- West --- D2 --- Tea10 --- 589
    amount --- West --- S1 --- Tea10 --- 852
    amount --- West --- C3 --- Tea10 --- 981
    amount --- East ---- S2 --- Coffee --- 876
    amount --- East ---- D1 --- Tea10 --- 156
    amount --- East ---- D3 --- Coffee --- 986
    amount --- East ---- C1 --- Coffee --- 256
    amount --- East ---- C3 --- Tea10 --- 9896
    And the third way is to adjust the essbase.cfg parameters DLTHREADSPREPARE and DLTHREADSWRITE (and DLSINGLETHREADPERSTAGE)
    I just want to be sure that I understand your suggestions.
    Many thanks for awesome help,
    Zeljko

Maybe you are looking for

  • I cannot install itunes. The error message telss me I need a CD

    I am unable to install itunes 10.5.  I am asked for the CD to provide a needed file.  I cannot uninstall the prior version of itunes. 

  • PL/SQL scrpt to automate DATABASE RECOVERY using logical backups com, cum, inc export

    Hellow friends, I am writing a script TO RECOVER MY DATABASE FROM EXPORT BACKUPS IN PL/SQL, which are stored in my folder F:\ORACLE\ORADATA\BACKUPS. There are many export backups in this folder comprising of some COMPLETE, CUMULATIVE and INCREMENTAL

  • Apple TV will recognize my YouTube login credentials

    I've had Apple TV for several months and it was working great until recently. Suddenly it no longer recognizes my YouTube credentials. When I try to login I get a message stating: "This account cannot be used with Apple TV. Try using a different acco

  • Mp4 video isn't burning in my idvd project

    I've been trying to burn a mp4 vid into my idvd project, which is fine.  But after I begin to burn the finished product, "along with the slideshow which loads up just fine" the whole process comes to a stop when iDVD  get to loading video part it.  t

  • Releasing a Recall

    Good evening: We are implementing the transaction WTY and according the SAP documentation after the creation of the Recall using the transaction WTYRCL the Recall should be released using the Action Code A400, however in the transaction WTRCL there i