Calculating % Growth dynamically.

Dear Experts,
I have a requirement in a report, mentioned below
                    Week               Week
Products  Sales   %Change  Sales %Change
Prod1        10       NA            15     50%
Prod2         30      NA            25     -16.67%
Selction::   0CALMONTH
In my query definition, I have mentioned the following attributes
Rows:: Products
Columns:: 0calWeek
                    Structure
                        Sales
                        %Change
Question:: I need to calculate the % of growth dynamcially based on selected month.
Please provide me inputs for this requirement.
Thanks in advance.
Regards,
-Vb

Hi Vj,
You can use %G or %A based on the requirement.
Regards,
aanand

Similar Messages

  • Calculating Average dynamically in BW Report

    Hello experts,
    I need to create a report where in I need to display month wise billing qty along with the average qty,
    here 'billing qty' is a key figure and "avgqty" should be calculated dynamically.
    The formula for avgqty is :
    "average qty" = ("beginning qty" + "n" stock at month's end)/(n+1) where n is the monthly duration that the user gives in the selection.
    ex: if the user put from "01/2006" to "03/2006" in the selection screen for the  Calendar Year/Month  then  here n becomes 2, so:
    "avgqty" = (total qty for jan + total qty for feb + total qty for mar )/(2+1)
    Can some please let me know the procedure on how to pass this "n" value to the formula that I need to use for my "avgqty" calculation.  
    Thanks & Regds

    Hi Murali,
    i assuming You have your Month variable (MONTH)( user/entry).
    First create 1 Customer Exit variable (CUMMONTH)with interval value. Go to CMOD and write this code.
    DATA: L_S_RANGE TYPE RSR_S_RANGESID.
    DATA: LOC_VAR_RANGE LIKE RRRANGEEXIT.
    CASE I_VNAM.
    WHEN 'CUMMONTH'.
    IF I_STEP = 2.
    "after the popup
    LOOP AT I_T_VAR_RANGE INTO LOC_VAR_RANGE WHERE VNAM = 'MONTH'.
    CLEAR L_S_RANGE.
    L_S_RANGE-LOW = LOC_VAR_RANGE-LOW."low value = input L_S_RANGE-HIGH = LOC_VAR_RANGE-HIGH. "high value = input L_S_RANGE-SIGN = 'I'.
    L_S_RANGE-OPT = 'BT'.
    APPEND L_S_RANGE TO E_T_RANGE.
    EXIT.
    ENDLOOP.
    ENDIF.
    ENDCASE.
    Now create a selection which have this variable CUMMONTH.( so that you get the summation of total qty from jan to qty of mar)
    Now, come to your (n+1) part.
    Create a formula variable (FUMMONTH)again customer exit having single value.
    write this code. (not full code ..try in your system)
    char l(1),
    char h(1).
    CASE I_VNAM.
    WHEN 'FUMMONTH'.
    IF I_STEP = 2.
    "after the popup
    LOOP AT I_T_VAR_RANGE INTO LOC_VAR_RANGE WHERE VNAM = 'MONTH'.
    CLEAR L_S_RANGE.
    lm = LOC_VAR_RANGE-LOW+4(2)."low value = input month
    hm = LOC_VAR_RANGE-HIGH+4(2). "high value = input month
    ly = LOC_VAR_RANGE-LOW(4)."low value = input year
    hy = LOC_VAR_RANGE-HIGH(4). "high value = input year
    'write logic
    'for if user enter from 2005(ly) to 2006(hy)..
    ' subtract h-l and give it L_S_RANGE-LOW
    L_S_RANGE-LOW = subtract
    L_S_RANGE-SIGN = 'I'.
    L_S_RANGE-OPT = 'EQ'.
    APPEND L_S_RANGE TO E_T_RANGE.
    EXIT.
    ENDLOOP.
    ENDIF.
    ENDCASE.
    after this..
    Make a Formula named Avg Qty
    =NODIV((That selection)(name of formula variable +1))
    Hope i m clear.
    Regards,
    San!
    Message was edited by: San!

  • Calculating growth percentage in a pivot table

    I have a pivot table with the following data generated by a pivot table of customer and sales data. The pivot table is showing the 2010 and 2009 data.
    Sales
    Customer 2010 2009
    ABC 1000 800
    DEF 2000 1000
    How would I get the calculation to show the % growth from 2009 to 2010 shown below ??
    Sales
    Customer 2010 2009
    ABC 1000 800 25%
    DEF 2000 1000 100%
    Thanks for any help

    Re: how to Make a report from two cubes
    This link will help you.

  • Dynamic calc problem while calculation another dynamic calc member

    I have 6 dimensions. Two of them are dense and are tagged as Accounts and Time. The dimension I am having problem with is called "Metrics". It was originally dense. But because the calc scripts were executing slowly, I decided to make "Metrics" a sparse dimension. Now the calc scripts are all executing pretty fast. However, I now have data issues. For example: Within Metrics, I have a hierarchy that looks as follows: Metrics Input.Amount ~ BW.YAG.EQU Dynamic Calc, Two Pass Pct.YAG.EQU Dynamic calc, Two Pass. Both of these Members have a member formula associated with them. "Pct.YAG.EQU" member refers to "BW.YAG.EQU" within it's member formula. Also, notice "Pct.YAG.EQU" comes after "BW.YAG.EQU" Any other thoughts. Really appreciate all your help in this matter. Thanks

    I have 6 dimensions. Two of them are dense and are tagged as Accounts and Time.The dimension I am having problem with is called "Metrics". It was originally dense. But because the calc scripts were executing slowly, I decided to make "Metrics" a sparse dimension. Now the calc scripts are all executing pretty fast. However, I now have data issues.For example:Within Metrics, I have a hierarchy that looks as follows:Metrics Input.Amount ~ BW.YAG.EQU Dynamic Calc, Two Pass Pct.YAG.EQU Dynamic calc, Two Pass.Both of these Members have a member formula associated with them. "Pct.YAG.EQU" member refers to "BW.YAG.EQU" within it's member formula. Also, notice "Pct.YAG.EQU" comes after "BW.YAG.EQU"Any other thoughts.Really appreciate all your help in this matter.Thanks

  • Can I use a Sum Calculation accross dynamically added rows?

    Well I have a working script that can add/delete extra rows to a table.
    And I have a working script that can calculate the sum of the columns.
    But when I put them together, the add row function stops working.
    Is there a way to reconcile the two?
    See attached example. It's actually not a real table, but subforms mimicking one.

    You keep making the same mistake. You cannot just copy the code from one form to another. Each form has a different structure and the code is showing the path to the object. Have a look at the HIerarchy view and see how to get to the object that you want from the root node. I have updated your sample to show this.
    Paul

  • Forecasting Based Planning with Dynamic Safety stock Calculation?

    Hi Guys:
    Based on the client requirement we are going with forecast based planning i.e based on Sales historical data , rather then buyer determining how much amount of safety stock need to be maintained ,
    Can system calculate/propose safety stock based on lead time and service levels set for that particular material?.
    how to load historical sales data into the system before go Live , so when they run MRP Run ( using forecast based Planning ) -> Planning orders>PO are triggered? or is there any alternative ?
    Thanks
    Sweth.

    Hi.
    Range of coverage profile in material master MRP2 view
    The range of coverage profile contains the parameters for calculating the dynamic safety stock. This is a statistical calculation on the basis of average daily requirements.
    Use
    The dynamic safety stock is calculated depending on the requirements within a specified period and the range of coverage.
    Dependencies
    The range of coverage profiles (period length for calculating the average consumption, range of coverage, and so on) are maintained in Customizing in the MRP screen.
    SPRO>MM>CBP>MRP CALCULATION>Define rage of coverage profile
    or
    Production>MRP>MRP Calculation>Define rage of coverage profile
    Thanks
    Diwakar

  • How to pass dynamic parameter to a database function in OBIEE

    Hi,
    I have a requirement like this. I have to create one report in OBIEE which was in Discoverer. Now in discoverer report there are some calculated item in the worksheet based on database pkg.functions. The parameter which user gives at run time that parameters are then passed to the discoverer calculated items dynamically. But I am not able to do this in OBIEE answers.
    Can anyone tell me step by step how I can able to pass the user selected parameter values in OBIEE answer level.
    The example:
    GET_COMM_VALUE_PTD("AFE Cost & Commitment".Afe Id,:"Period Name(AFE)","AFE Cost & Commitment".Data Sel,"AFE Cost & Commitment".Org Id)
    GET_COMM_VALUE_PTD --- Function database
    ("AFE Cost & Commitment".Afe Id,:"Period Name(AFE)","AFE Cost & Commitment".Data Sel,"AFE Cost & Commitment".Org Id --- Parameters... :"Period Name(AFE)" is the dynamic parameter selected run time by user.
    Please help.
    Thanks
    Titas

    Hi,
    I already did that. But the existing discoverer re value report is showing correct value but the OBIEE report with EVALUATE function shows incorrect value.
    The requirement is to create a OBIEE Dashboard from a Discoverer existing report. Now in discoverer report theere are several calculated items in worksheet level where database custom function is used and user selected parameter value is passed run time in that function. But when that is created in OBIEE with EVALUATE function in RPD, the report shows incorrect data. Could not understand how to pass the parameter value runtime.
    Below is the discoverer 'Show SQL' query and EVALUATE function syntax which has been used.
    SELECT APPS.XXBG_PL_PROJ_ANALYSIS_AFE_PKG.GET_COMM_VALUE_PTD(XXBG_PL_PROJ_AFE_V.AFE_ID,
    :"Period Name(AFE)",
    XXBG_PL_PROJ_AFE_V.DATA_SEL,
    XXBG_PL_PROJ_AFE_V.ORG_ID),
    XXBG_PL_PROJ_AFE_V.AFE_DESC,
    XXBG_PL_PROJ_AFE_V.AFE_NUMBER,
    XXBG_PL_PROJ_AFE_V.APPROVED_AFE_AMOUNT
    FROM APPS.PA_PERIODS_ALL PA_PERIODS_ALL,
    APPS.XXBG_PL_PROJ_AFE_V XXBG_PL_PROJ_AFE_V
    WHERE ((XXBG_PL_PROJ_AFE_V.ORG_ID = PA_PERIODS_ALL.ORG_ID))
    AND (XXBG_PL_PROJ_AFE_V.DATA_SEL = :"Data Selection(AFE)")
    AND (PA_PERIODS_ALL.PERIOD_NAME = :"Period Name(AFE)")
    AND (XXBG_PL_PROJ_AFE_V.AFE_NUMBER = :"AFE Number(AFE)")
    AND (XXBG_PL_PROJ_AFE_V.OPERATING_UNIT = :"Operating Unit(AFE)")
    The EVALUATE function syntax is as below:
    EVALUATE('XXBG_PL_PROJ_ANALYSIS_AFE_PKG.GET_COMM_VALUE_PTD(%1,%2,%3,%4)' AS FLOAT , "BG PL Project Analysis Report_1"."AFE Cost & Commitment"."AFE Cost & Commitment.Afe Id", "BG PL Project Analysis Report_1"."Periods 1"."Periods 1.Period Name", "BG PL Project Analysis Report_1"."AFE Cost & Commitment"."AFE Cost & Commitment.Data Sel", "BG PL Project Analysis Report_1"."AFE Cost & Commitment"."AFE Cost & Commitment.Org Id")
    The PERIOD needs to be passed at run time which user will select.
    Please help to solve the issue.

  • Java variables calculation on change in JSP (Excel replica)

    I have a excel sheet that I need to replicate in JSP (using struts 1 framework). Now everything is done but I am stuck at the calculation part. As you might have seen in excel, some columns are calculations based on other columns.
    So if I change field 'a', 'd' field should change automatically based on its formula (let's say, d=a*1.1 +b*2.1).
    Now I tried to use the funstion in javascript where I read 'a' and 'b' values from JSP, then I calculate and everything works fine. But when I try to assign the calculation result for field 'd' in JSP throush javascript, 'd' does not change.
    function adjustCalcs ()
    var a = eval("document.form.al.value");
    var b = eval("document.form.b.value");
    var dCalc = ((a * 1.05) + (b * 1.1));
    document.form.d.value = dCalc ; (this is where it should change the value of field 'd' in JSP, but ti won't...donno why?)
    Or is there any better way to do these dynamic changes in JSP ? Would appreciate the insight.
    Thanks,

    Fine. What's your problem in JSP side then? Do you want to let Java take over the calculation and "dynamic stuff" ? Then you need to let Javascript submit the form to the server side during the change event so that JSP can then display the result using taglibs/EL. But this is less good for user experience as this costs effectively one HTTP request and the user would see a "flash of content". Alternatively you can use Ajax for this to do it all asynchronously.

  • Dynamic Selection of 'To' date

    Hi,
    I have Planned Order Start date 'From' and 'To' option in our selection screen of a customized program. I want 'From' fiield to be blank and 'To' field mandatory. The calcuation of 'To' date should be dyanamic in nature (depending upon Selection variable and Name of Variable maintained in the Variant)
    If I use the variant option for the same then the system proposes the date in 'From' field and not 'To' field after calculating it dynamically.
    This is required so that the system should fetch all the past Planned Orders plus the Orders coming within 'To' date option in the selection screen.
    Thanks in Advance.
    Regards,

    In the 'variable name' select the indicator (first column) as 'I' (Select specified values) and select 'LE' (less or equal) from 'option' (second column). After selecting the said options upon hitting enter you'll get a new pop up where you need to enter the number of days further into the future that you want the dynamic date should be calculated (a positive means future, a negative  means in the past). Once you save the variant, you'll see the setting appearing in the 'from'. Do not worry about the setting that appears in the 'from' instead of 'to'. When you select this variant or execute this program via this variant the date will be calculated based on the current date + future number of days and all past dates from that future date. Again do not worry about the setting in the 'from'  - this is how it is set up.

  • Error in emagent.trc

    Hi
    got error in emagent.trc
    2010-03-05 15:27:38,062 Thread-1900551 ERROR pingManager: nmepm_pingReposURL: Cannot connect to https://192.168.21.243:1159/em/upload: retStatus=-1
    2010-03-05 15:27:38,104 Thread-1900551 ERROR ssl: 9: Common Name = "localhost" Does not Match Hostname = "192.168.21.243"
    2010-03-05 15:27:38,104 Thread-1900551 ERROR http: 9: Unable to initialize ssl connection with server, aborting connection attempt
    2010-03-05 15:27:38,104 Thread-1900551 ERROR pingManager: nmepm_pingReposURL: Cannot connect to https://192.168.21.243:1159/em/upload: retStatus=-1
    2010-03-05 15:28:08,324 Thread-1966087 ERROR ssl: 9: Common Name = "localhost" Does not Match Hostname = "192.168.21.243"
    2010-03-05 15:28:08,341 Thread-1966087 ERROR http: 9: Unable to initialize ssl connection with server, aborting connection attempt
    2010-03-05 15:28:08,342 Thread-1966087 ERROR pingManager: nmepm_pingReposURL: Cannot connect to https://192.168.21.243:1159/em/upload: retStatus=-1
    2010-03-05 15:28:08,391 Thread-1966087 ERROR ssl: 9: Common Name = "localhost" Does not Match Hostname = "192.168.21.243"
    2010-03-05 15:28:08,391 Thread-1966087 ERROR http: 9: Unable to initialize ssl connection with server, aborting connection attempt
    2010-03-05 15:28:08,391 Thread-1966087 ERROR pingManager: nmepm_pingReposURL: Cannot connect to https://192.168.21.243:1159/em/upload: retStatus=-1
    I have installed OMS Grid on one server (192.168.21.243) and agent is installed on other server(192.168.21.244)
    entry in /etc/hosts of 21.244 server
    192.168.21.244 isgmmlds244.insolutionsglobal.local localhost
    192.168.21.243 isgmmlds243.insolutionsglobal.local isgmmlds243
    entry in emd.properties
    # perl executable directory
    perlBin=/u01/app/oracle/product/agent10g/perl/bin
    # script directory
    scriptsDir=/u01/app/oracle/product/agent10g/sysman/admin/scripts
    # emd Root directory(read-only location). Metrics should not create files
    # under this directory
    emdRoot=/u01/app/oracle/product/agent10g
    # agent Root directory(writeable).
    # Use this property to base any temporary file creation.
    agentStateDir=/u01/app/oracle/product/agent10g
    # Chronos root directory
    chronosRoot=/u01/app/oracle/product/agent10g/sysman/emd/chronos
    # OMS Upload URL
    # if there is no receiving OMS or if you wish to disable the UploadManager
    # please set this value to empty or comment out below line
    REPOSITORY_URL=https://192.168.21.243:1159/em/upload
    # Proxy settings used for fetchlets. not applicable for uploader.
    #proxyHost=www-proxy.us.oracle.com
    #proxyPort=80
    #dontProxyFor=.us.oracle.com
    # If it is necessary to go through an http proxy server to get to the
    # repository, uncomment the following two lines
    #REPOSITORY_PROXYHOST=
    #REPOSITORY_PROXYPORT=
    #REPOSITORY_PROXYREALM=%EM_REPOS_PROXYREALM%
    #REPOSITORY_PROXYUSER=%EM_REPOS_PROXYUSER%
    #REPOSITORY_PROXYPWD=%EM_REPOS_PROXYPWD%
    # This string is used by the agent to connect to remote targets
    # ***IMPORTANT*** Do not change the contents of this setting. Only the
    # install should modify this value.
    agentSeed=267923712
    # This string is used by the agent to determine which algorithm to use for encrypted data
    # The string value will be same as the release version
    agentVersion=10.2.0.0.0
    # How long, in minutes, collector will wait until next load.
    UploadInterval=15
    # If an upload fails, we want to backoff by UploadFailBackoff
    # percentage more each time for each successive upload failure
    # e.g. if we start the retry at 1 minute. If we fail in our next
    # upload attempt, we try again 1minute 12 seconds later etc.
    UploadFailBackoffPct= 20
    # Maximum amount of time, in seconds, upload will wait for response from
    # repository before timing out.
    # 0 means infinite.
    UploadTimeout=1800
    # Timeout for upload a file, HealthMonitor will abort the EMD process if
    # UploadManager does not finish uploading one file within this time
    # Unit is minite, the default value is 1 hour ( 15 Min)
    #UploadMaxTime=30
    # How big the file the collector will upload (in KB)
    # the actual file size could be slightly bigger.
    # This is only a guidance.
    UploadFileSize=2048
    # The maximum number of megabytes(MB) the upload manager will support in the
    # upload directory before temporarily disabling collections, logging and tracing
    UploadMaxBytesXML=50
    # The maximum number of files the upload manager will support in the
    # upload directory before temporarily disabling collections, logging and tracing
    UploadMaxNumberXML=5000
    # The maximum amount (%) of disk space that can be used on the EMD filesystem
    # before the following is disabled:
    # - Collection of data (upload manager)
    # - Logging and tracing
    UploadMaxDiskUsedPct=98
    # The amount (%) of disk space that can be used on the EMD filesystem
    # before the following is re-enabled after being disabled previously:
    # - Collection of data (upload manager)
    # - Logging and tracing
    UploadMaxDiskUsedPctFloor=95
    # How long, in minutes, upload manager waits until it schedules recount
    # of the XML files in upload directory (defaults to 60 mins)
    UploadFileRecountInterval=60
    # Time out for the Agent to think that the access to the database could be
    # hung. The Agent would not allow more threads to go and hang with a database
    # if any one access to the database does not return within DbHangTimeout
    # seconds.
    # The more the timeout more the possibility of allowing more threads
    # to be hung in the EMD with a bad database.it is specified in seconds.
    #DbHangTimeout=200
    # The script to run if disk space usage exceeds parameters or if certain other
    # error conditions occur.
    #emdFailureScript=emdfail.command
    # The location of the file that contains the root certificate.
    emdRootCertLoc=/u01/app/oracle/product/agent10g/sysman/config/b64LocalCertificate.txt
    internetCertLoc=/u01/app/oracle/product/agent10g/sysman/config/b64InternetCertificate.txt
    # The download URL for the EMD Oracle Wallet and its local file location.
    # Note: Ensure that this URL references a valid port number at which the
    # console is available on http
    emdWalletSrcUrl=http://isgmmlds243.insolutionsglobal.local:4889/em/wallets/emd
    emdWalletDest=/u01/app/oracle/product/agent10g/sysman/config/server
    # The email address for out-of-band notifications
    emd_email_address=_NOT_AVAILABLE_
    emd_email_gateway=_NOT_AVAILABLE_
    # The return email address for out-of-band notifications
    emd_from_email_address=_NOT_AVAILABLE_
    # EMD main servlet URL
    EMD_URL=https://isgmmlds244.insolutionsglobal.local:1831/emd/main/
    # Whether the EMD should listen on all NICs on the current host (the default)
    # or just the NIC associated with the hostname in EMD_URL
    AgentListenOnAllNICs=FALSE
    # This parameter indicates the thread model to use. uncomment one of the
    # following ThreadPoolModel line to indicate the model, agent need to use.
    # SMALL if number of targets is less than 11.
    # MEDIUM if number of targets is from 11 to 50.
    # LARGE if number of targets is greater than 50.
    # This parameter can be overriden by 'MaxThreads' which indicates the maximum
    # number of threads that CEMD can have running at any point of time. User can
    # mostly use the other parameter ThreadPoolModel to indicate what model for the
    # agent to use.
    # ThreadPoolModel = SMALL
    # ThreadPoolModel = MEDIUM
    # ThreadPoolModel = LARGE
    # This parameter indicates the stack size, threads have to be created with.
    # we can set it to 0 to use the OS default value.
    # In some version of solaris(such as 2.9 ) specify a stack size may cause Java Hotspot warning
    # when JVM is used.
    NormalThreadStackSize=0
    #Signals that users want to ignore.
    #It can be one or combination of SIGTERM, SIGINT
    # The signals can listed with ',' as delimiter.
    #IgnoreSignals=
    # This property is NO LONGER USED for host config collection (instead, hostConfigClasspath is used).
    # If noone else uses it, it can be removed.
    ouiLoc=/u01/app/oracle/product/agent10g/oui
    # Classpath for host config collection
    # VERY IMPORTANT NOTE: If you change the host config classpath entries in this section, you HAVE
    # to change the HOST_CONFIG_TEST_CLASSPATH entry in the utl/emdwqaenv file too.
    # NOTE: if the location of log4j-core.jar is changed
    # make sure that appropriate changes are also made
    # to the CLASSPATH variable in this file if necessary
    # PORTING NOTE: Check OUI-required jar files for OUI APIs to work. They could be different on your platform.
    # Also, our version of xmlparserv2.jar must be compatible with the one OUI expects (both OUI and ECM use XML parser).
    hostConfigClasspath=/u01/app/oracle/product/agent10g/oui/jlib/xmlparserv2.jar:/u01/app/oracle/product/agent10g/oui/jlib/OraInstaller.jar:/u01/app/oracle/product/agent10g/oui/jlib/srvm.jar:/u01/app/oracle/product/agent10g/oui/jlib/share.jar:/u01/app/oracle/product/agent10g/sysman/jlib/emcoreAgent.jar:/u01/app/oracle/product/agent10g/sysman/jlib/emagentRT.jar:/u01/app/oracle/product/agent10g/sysman/jlib/emagentSDK.jar:/u01/app/oracle/product/agent10g/sysman/jlib/log4j-core.jar
    # JAVA HOME required for JavaExt metric calls
    JAVA_HOME=/u01/app/oracle/product/agent10g/jdk/jre
    # CLASSPATH required for JavaExt metric calls
    CLASSPATH=/u01/app/oracle/product/agent10g/jdbc/lib/ojdbc14.jar:/u01/app/oracle/product/agent10g/lib/xmlparserv2.jar:/u01/app/oracle/product/agent10g/sysman/jlib/log4j-core.jar:/u01/app/oracle/product/agent10g/jlib/http_client.jar:/u01/app/oracle/product/agent10g/oc4j/j2ee/home/lib/http_client.jar:/u01/app/oracle/product/agent10g/lib/http_client.jar:/u01/app/oracle/product/agent10g/jlib/share.jar:/u01/app/oracle/product/agent10g/jlib/jssl-1_1.jar:/u01/app/oracle/product/agent10g/jlib/javax-ssl-1_1.jar:/u01/app/oracle/product/agent10g/jlib/ojmisc.jar:/u01/app/oracle/product/agent10g/jlib/repository.jar:/u01/app/oracle/product/agent10g/opmn/lib/optic.jar:/u01/app/oracle/product/agent10g/sysman/jlib/emagentRT.jar:/u01/app/oracle/product/agent10g/sysman/jlib/emagentSDK.jar:/u01/app/oracle/product/agent10g/lib/dmsEmd.jar:/u01/app/oracle/product/agent10g/sysman/jlib/emcoreAgent.jar
    # this is the java option
    #JAVA_OPTIONS=-Xmx128m
    #JAVA_OPTIONS=-Djava.awt.headless=true
    # These are the optional Java flags for the in-process Java
    # Native Interface (JNI)
    agentJavaDefines=-Doracle.dms.refresh.wait.time=1000 -DUrlTiming.UseJSSE=true -Dnetworkaddress.cache.ttl=1800 -Djava.awt.headless=true
    # When set to true, the emd.log and emd.trc file will be appended by a pid
    # For example:
    # emd_15789.log
    # emd_15789.trc
    # By default, it is not set for regression test purpose.
    # After installed, please uncomment this line because when EMD is restarted,
    # previous log information will be lost if not set to true.
    #LogFileWithPID=true
    # Default log file maximum size before rolling in Kilobytes
    LogFileMaxSize=4096
    # Maximum number of log file roll files to maintain before deletion
    LogFileMaxRolls=4
    # Default trace file maximum size before rolling in Kilobytes
    TrcFileMaxSize=4096
    # Maximum number of trace file roll files to maintain before deletion
    TrcFileMaxRolls=4
    # To enable the metric browser, uncomment the following line
    # This is a reloadable parameter
    #enableMetricBrowser=true
    # To disable Remote operations (non-reachable agent) uncomment the following
    # line:
    #disableRemoteOperations=true
    # How long (in secs) until a remote operation sumbitted via the dispatcher can
    # take before the agent in bounced by the watchdog. This timeout applied to only
    # those operation which ought to finish within a certain interval of time
    RemoteAPITimeout=300
    # add an additional admin path from which the agent will read additional
    # target metadata. The specified path should have the mandatory 4 directories
    # under it. 1. metadata 2. scripts 3. default_collection 4. discover
    #altAdminPath=<additional admin path>
    # The health monitor executes callbacks in a separate thread. if for some
    # reason the threads spawned from health monitor to execute callbacks are
    # hung, the agent could run out of resource and freeze.
    # so when the current thread count spawned from the Health monitor reaches
    # this limit, the agent is bounced. the default is set to 5.
    #MaxHealthMonitorThreads=5
    # SSL session cache flag
    SSLSessionCache=FALSE
    # emagent perl tracing levels
    # supported levels: DEBUG, INFO, WARN, ERROR
    # default level is WARN
    EMAGENT_PERL_TRACE_LEVEL=WARN
    # other optional parameters for perl tracing
    # EMAGENT_PERL_TRACE_DIR default to $ORACLE_HOME/sysman/log
    #EMAGENT_PERL_TRACE_DIR=
    # EMAGENT_PERL_TRACE_FILESIZE default to 5M
    #EMAGENT_PERL_TRACE_FILESIZE=5
    # this section define the logging level for each component,
    # it is "reloadable", you can change the level at runtime to reset
    # the logging level, but if you remove the line, it will not be
    # changed.
    tracelevel.main=WARN
    tracelevel.emSDK.xml=WARN
    tracelevel.emSDK.utl=WARN
    tracelevel.Dispatcher=WARN
    tracelevel.ThreadPool=WARN
    tracelevel.pingManager=WARN
    tracelevel.collector=WARN
    tracelevel.http=WARN
    tracelevel.http.client=WARN
    tracelevel.browser=WARN
    tracelevel.ssl=WARN
    tracelevel.ssl.io=WARN
    tracelevel.blackouts=WARN
    tracelevel.upload=WARN
    tracelevel.command=WARN
    tracelevel.reload=WARN
    tracelevel.scheduler=WARN
    tracelevel.Authentication=WARN
    tracelevel.metadata=WARN
    tracelevel.targets=WARN
    tracelevel.TargetManager=DEBUG
    tracelevel.engine=WARN
    tracelevel.javaproc=WARN
    tracelevel.vpxoci=WARN
    tracelevel.javavm=WARN
    tracelevel.fetchlets=WARN
    tracelevel.fetchlets.os=WARN
    tracelevel.fetchlets.osline=WARN
    tracelevel.fetchlets.oslinetok=WARN
    tracelevel.fetchlets.snmp=WARN
    tracelevel.fetchlets.UDM=WARN
    tracelevel.fetchlets.PropsFromAssocTgt=WARN
    tracelevel.fetchlets.sql=WARN
    tracelevel.fetchlets.url=WARN
    tracelevel.fetchlets.urllines=WARN
    tracelevel.fetchlets.urllinetoken=WARN
    tracelevel.fetchlets.URLTiming=WARN
    tracelevel.fetchlets.OJMX=WARN
    tracelevel.emd.fetchlets.JMX.generic=WARN
    tracelevel.fetchlets.propEcho=WARN
    tracelevel.fetchlets.readFromFile=WARN
    tracelevel.fetchlets.readMultFromFile=WARN
    tracelevel.fetchlets.throwable=WARN
    tracelevel.fetchlets.resourceGrab=WARN
    tracelevel.fetchlets.statusHang=WARN
    tracelevel.fetchlets.emSDK=WARN
    tracelevel.recvlets=WARN
    tracelevel.recvlets.snmp=WARN
    tracelevel.credproviders=WARN
    tracelevel.utl=WARN
    tracelevel.NLS=WARN
    tracelevel.util.dates=WARN
    tracelevel.util.emdprops=WARN
    tracelevel.util.files=WARN
    tracelevel.util.fileops=WARN
    tracelevel.resman.socket=WARN
    tracelevel.resman.process=WARN
    tracelevel.ResMonitor=WARN
    tracelevel.emtgtctl=WARN
    # The following three properties are for HTTP timeout
    # Timeout value for reading Http primary header, default is 30 seconds.
    httpTimeoutPrimaryHeader=30
    # Timeout value for reading http secondary headers, default is 30 seconds
    httpTimeoutSecondaryHeaders=30
    # Timeout value for reading http body, default is 60 seconds.
    httpTimeoutBody=60
    # Timeout used by ping manager
    #PingTimeout=30
    #Timeout for Dynamic property evaluation. This is provided in seconds.
    #The dynamicPropsComputeTimeout value applies to all target types unless if a
    #target type specific value is provided using the
    #dynamicPropsComputeTimeout_<targetType>= syntax
    #dynamicPropsComputeTimeout=30
    #Requested by RAC team in bug 4423072
    dynamicPropsComputeTimeout_rac_database=120
    #bug 4595094
    dynamicPropsComputeTimeout_oracle_database=120
    #bug 5016243
    dynamicPropsComputeTimeout_oc4jjvm=120
    #bug 5647863
    #@description=DynamicProperty evaluation timeout value of Microsoft SQL Server
    #@default=120
    dynamicPropsComputeTimeout_microsoft_sqlserver_database=120
    # changes in context of bug id 4591002
    # Need to recompute the values of dynamic properties in case of failure to do so
    # at startup and reload. Currently this applies only to "Critical" marked dynamic
    # properties. The following two values are applicable per target and not per
    # dynamic property
    # dynamicPropReComputeInterval --> time difference between a failed dynamic property
    # computation and the next try to compute the property in seconds. The default value is 120 seconds.
    # dynamicPropReComputeMaxTries --> maximum number of reties for calculating failed
    # dynamic properties. The default value is 4 retires.
    agentTZRegion=Asia/Calcutta
    output of status of agent
    [oracle@isgmmlds244 patchset10.2.0.4]$ /u01/app/oracle/product/agent10g/bin/emctl status agent
    Oracle Enterprise Manager 10g Release 4 Grid Control 10.2.0.4.0.
    Copyright (c) 1996, 2007 Oracle Corporation. All rights reserved.
    Agent Version : 10.2.0.4.0
    OMS Version : 10.2.0.4.0
    Protocol Version : 10.2.0.4.0
    Agent Home : /u01/app/oracle/product/agent10g
    Agent binaries : /u01/app/oracle/product/agent10g
    Agent Process ID : 12468
    Parent Process ID : 12377
    Agent URL : https://isgmmlds244.insolutionsglobal.local:1831/emd/main/
    Repository URL : https://192.168.21.243:1159/em/upload
    Started at : 2010-03-05 15:34:24
    Started by user : oracle
    Last Reload : 2010-03-05 15:34:24
    Last successful upload : (none)
    Last attempted upload : (none)
    Total Megabytes of XML files uploaded so far : 0.00
    Number of XML files pending upload : 129
    Size of XML files pending upload(MB) : 38.17
    Available disk space on upload filesystem : 29.76%
    Last attempted heartbeat to OMS : 2010-03-05 15:35:17
    Last successful heartbeat to OMS : unknown
    Agent is Running and Ready
    error for upload agent[oracle@isgmmlds244 patchset10.2.0.4]$ /u01/app/oracle/product/agent10g/bin/emctl upload agent
    Oracle Enterprise Manager 10g Release 4 Grid Control 10.2.0.4.0.
    Copyright (c) 1996, 2007 Oracle Corporation. All rights reserved.
    EMD upload error: uploadXMLFiles skipped :: OMS version not checked yet..
    Please provide solution
    Regards
    shd

    shd wrote:
    entry in /etc/hosts of 21.244 server
    192.168.21.244 isgmmlds244.insolutionsglobal.local localhost
    192.168.21.243 isgmmlds243.insolutionsglobal.local isgmmlds243
    Remove the localhost alias from this above enty.
    localhost should be - and probably also is - pointing to 127.0.0.1
    HTH

  • In Infopath, how do I create automatic reminders if a form is coming up for expiry?

    Hi, we have an InfoPath 2010 form that has a "regularenddate", "temporaryenddate", "newextensiondate", and a "terminationdate".
    I would like to know if it's possible to program into the workflow for this form using SharePoint Designer 2010, that 3 days before the date in each of these fields, that a reminder is sent to the person who submitted the form, reminding them that the form
    is expiring.
    Selecting this does not work:  "If Current Item:DepartureDate equals Today"/Send an email
    If this could be set:  "If CurrentItem:DepartureDate is 3 days from the date" (in the field)/Send an email, that would really help!!!!
    Thanks for any help you can provide.  I'm very familiar with InfoPath forms and using SharePoint Designer, so if you want to get technical, go ahead.
    Thanks,
    Dan

    Hello,
    You can create variables for date columns and add -3 for each date in WF to send email before 3 days. Refer below links: 
    http://wonderlaura.com/Lists/Posts/Post.aspx?ID=64
    http://www.documentmanagementworkflowinfo.com/sample-sharepoint-workflows/date-calculations-construct-dynamic-strings-sharepoint-designer-workflows.htm
    Unfortunately "Lakshmanan Sethu" link is not working for me, getting "sever error".
    Let me know in case any doubt
    Hemendra:Yesterday is just a memory,Tomorrow we may never see
    Please remember to mark the replies as answers if they help and unmark them if they provide no help

  • Fill in spool options without dialog in Smartforms

    Hi,
    I´m trying to set the spool parameters before printing in Smartforms. Therefore I don´t want a dialog to choose the print options.
    There are too structures, which are going to be passed to the smartform function: OUTPUT_OPTIONS and CONTROL_PARAMETERS.
    I have the shortname of the device (CHAR4 field), like 'LOCL' or a print device.
    How can I now set the parameters before calling the function of the smartform? When I fill the 4char device field into output_options-tddest, it´s not working! It´s always printing on the default device!
    What else should I fill to get it printed right?
    Cheers Arne

    Hi Vijay,
    thanks for Your answer. The background is, that I need to change all production forms form DIN A4 to DIN A5 and make bigger changes in the layout.
    Therefore I decided to create it new in Smartforms, it should be better than calculating the dynamic boxes again in Sapscript.
    If You want to print the forms, You can do that about CO01/02/03. Therefore You have to set "lists" where You can customize the printer device, the sapscript form and the sapscript program.
    To do it in Smartforms, I´m checking the list parameters in my print program and, in case which list was choosen, I call the specific smartform.
    No I want to use the device options set in the list. I´m using the function CO_PRINT_GET_INFO_LIST and I get back two structures with the list information. PRINT_CO and PRINT_OPTS.
    In PRINT_CO there is a field called DESTI. It contains the shortname of the device.
    Now I want to pass this one to the smartform. The device is valid, but my smartform always uses the default device.
    My code:
    CALL FUNCTION 'CO_PRINT_GET_INFO_LIST'
           IMPORTING
                print_co_exp   = print_co
                print_opts_exp = print_opts
           EXCEPTIONS
                OTHERS         = 0.
        wa_control_parameters-device = 'PRINTER'.
        wa_control_parameters-preview = ''.
        wa_control_parameters-no_dialog = 'X'.
        wa_output_options-tddest = print_co-desti.
    *    wa_output_options-tdprinter = print_co-desti.
        wa_output_options-tdarmod = '1'.
        wa_output_options-tdcopies = '001'.
        wa_output_options-tddelete = 'X'.
        wa_output_options-tdlifetime = '2'.
        wa_output_options-tdtitle = print_opts-tdtitle.
        CALL FUNCTION '/1BCDWB/SF00000029'
          EXPORTING
    *       ARCHIVE_INDEX              =
    *       ARCHIVE_INDEX_TAB          =
    *       ARCHIVE_PARAMETERS         =
            CONTROL_PARAMETERS         =  wa_control_parameters
    *       MAIL_APPL_OBJ              =
    *       MAIL_RECIPIENT             =
    *       MAIL_SENDER                =
           OUTPUT_OPTIONS             =  wa_output_options
           USER_SETTINGS              = 'X'
          IMPORTING
            JOB_OUTPUT_OPTIONS        =  wa_job_output.
    I hope You understand my problem.
    Cheers Arne

  • Oracle 11g OLAP & SQL

    Hi All
    Our company is in the process of doing a POC warehouse where we are using Oracle OLAP extensively for summary management. I have been tasked with porting all our existing reports (Cognos) from using an Informix backend to start using Oracle. The OLAP team has created some cube views for me but Im struggling to get my around how Im going to use them for reporting purposes.
    Example
    1) Im using the following sql (abbreviated) to get my data:
    select
    v_product.product_description.
    v_product.level_name,
    v_sales.sales,
    v_sales.calc_measure
    from v_product, v_sales, v_location, v_time
    where ... all the joins....
    v_product.level_name in ('DEPARTMENT', 'CLASS')
    and v_location.level_name = 'TOTAL'
    and v_time.level_name = 'TOTAL'
    2) This brings back data that looks like:
    product_description level_name sales calc_measure
    MEAT DEPARTMENT 232323 23.56
    POULTRY DEPARTMENT 43444 35.23
    BEEF CLASS 232323 23.56
    CHICKEN CLASS 67455 35.23
    LAMB CLASS 73444 23.56
    PORK CLASS 55555 35.23
    3) I need to create a list report thats grouped by department and for each department shows all the classes but off the data above is very difficult. I cannot just select the all the class values and then do the aggregation in the report as there is a calculated measure so I need to select the value for that level from the cube view. Is it possible in one sql statement or will I need more?
    Thanks for any ideas

    Dave thanks for your reply. Please excuse my poor example this was my first day using cube views and I cannot login to my work setup from home so going by memory alone.
    To answer your question
    1. We are using 11g AW
    2. I don't remember the exact cube view names but not relevant to my question (I think)
    3. Alas the oracle forums don't support much formatting else I would have provided an ascii example. I have uploaded the sample report output here http://i279.photobucket.com/albums/kk145/angusgoosemiller/sample.gif
    Better Example
    1. From what I can gather if your query more that one level in the same dimension from a cube view you get the results denormalized as rows. So effectively for my report I what the department and class levels from the product hierarchy where class is a child of department and some relevant measures one of which is a calculated measure. If I select this from the cube view I am getting results in the form:
    DEPARTMENT LEVEL ... row values
    DEPARTMENT LEVEL ... row values
    CLASS LEVEL ... row values
    CLASS LEVEL ... row values
    2) My report is a list report that is grouped by department and for each department all the class records are displayed with the measures. There must also be a department total for every department level and a grand total for the report. If the calculated measure was not included I could just return all the class records as the there is a department attribute defined that is also in the cube view and calculate all the department values dynamically in the report. However due to the calculated measure and probably as a best practice from a performance/redundancy perspective I only want to select from the cube view in its aggregated form like is currently happening.
    3) From a report design perspective this provides some challenges as relationally hierarchy levels are normally modeled as columns and we use to process calculated measures dynamically in the report. Going forward we would like all the calculations etc happening in the OLAP engine.
    4) So basically the way I see it I need the following from the cube:
    4.1) The department records
    4.2) The class records
    4.3) the department total records
    4.4) the grand total record for all departments
    Can I get that in one SQL statement in such a manner that I can produce a report? How would an oracle based reporting solution get the data, via sql or directly from the cube via the olap api.
    Thanks for your help I really appreciate any advice!
    Cheers
    Angus

  • Dual SSD/Platter Set up

    My initial goal with this project is to increase the performance of AE by lower render and editing time. My first thought with this project was to replace the standard HDD in my MacBook Pro with a faster drive. Faster drive=faster project completion.
    Upon further investigation running through the forums a lot of people are running multipule drive machines a spectating there read and write drives to spectate ones to maximize the performance. I know some of the people who have thunderbolt ports are doing raid0 to accomplish this sadly I do not have thunder bolt.
    What I would like to do is pull out the super drive and replace with a second HDD to accomplish this same type of idea. The HDD drive in the optical bay would be a traditional 7200rpm drive in the range from 750gb to 1TB to store the majority of the large files I have. The main drive I would change to a smaller SSD drive for my OS and apps.
    All that being said I want the best return on my investment.
    Would I see significant improvement in my work flow? Or is it just a waste of time/money?
    If  I did do that how would I set the settings in AE to take advantage of the dual drive correctly?
    Here are the specs of my MacBook
    2.8ghz core 2 duo
    8GB ram (max)
    500gb 5400rpm drive
    OS Snow Leopard
    Thanks for the help in advance

    Unless you make excessive use of footage sources that require such massive file I/O, the advantage will be zero. Likewise, any advantages with using an SSD or RAID for disk cache will only show up, if the project is so complex, that using cached data will be faster than calculating stuff dynamically. AE has always worked this way and it still does, even if CS6 takes a much more aggressive approach to some of those things. It all comes down to what you do in your projects and it's the same like when to use MP rendering and when not or al lthose endless threads on the best computer for AE. There are no simple answers here. In the end, you could indeed spend a lot of money for nothing.
    Mylenium

  • Multiple hierarchies single dimension single report

    I am trying to add multiple hierarchies from a single dimension in a single report. To create the hierarchy in BI admin I followed the instructions found here
    Oracle BI EE 10.1.3.3/2 &amp;#8211; One Dimension &amp;#8211; Multiple Hierarchies &amp;laquo; Business Intelligence &am…
    and here
    https://forums.oracle.com/thread/2440305
    My Patient hierarchy looks like the this
    All Patients->Gender->Patient ID
    All Patients->Age Group->Patient ID
    When I drag and drop the second hierarchy in the Criteria Oracle BI Answers says the following "...you can only add one hierarchy per dimension to a report". I deduct that what I am trying to do is not possible!:-)
    Q1: Is this feature indeed not supported or am I doing something wrong?
    Q2: If this is not supported can someone explain why?
    Version Oracle Business Intelligence 11.1.1.6.0

    a) Reporting across levels of 2 hierarchies is not supported. As explained below, you cannot use OLAP to get aggregate values at cross combination of levels from hierarchies: H1 and H2.
    However if you wish to, you can report at the lowest level, Patient, common to both hierarchies and display the Age Group as well as Gender attributes corresponding to each patient. This becomes reporting at Patient/lowest level. You can of course use the capabilities of the reporting tool like obiee to use a Pivot object (say) and exclude Patient columns like Id, name etc. and get obiee to perform localized summaries within the reporting layer and get summarizations of interest. OBIEE Pivot measures allow for aggregation using Min, Max, Sum, Count, Running Sum etc. OLAP facilitates this solution but does not particularly help in the aggregation (since aggregation is now being done in obiee and not as per cube definition rules in olap aw). It acts as a source for information at the lowest level stored in the cube. Some performance degradation would result but it may be acceptable/manageable depending on the data volumes/report criteria.
    NOTE: For workaround/solution, use the regular fields, not hierarchical columns, from OBIEE presentation layer.
    b) Not supported: OLAP Cubes contain pre-calculated (or dynamically calculated) summaries at various level of each hierarchy. This aids in faster reporting performance. Allowing 2 hierarchies at the same time defeats this feature/goal. A cross combination of levels across different hierarchies requires a dynamic recalculation of the detail records to get the right values for, say, Age Group: 30-39 and Gender: Male. H1 summaries at AgeGroup level for Age Group: 30-39 cover all genders and cannot be split up into male/female w/o recalculation. Similarly H2 summaries at Gender level for Gender: Male cover all age groups. Splitting it up is not possible except via workaround/ solution explained in (a).
    Hope that helps.

Maybe you are looking for