Pivoting with non measurement data

I am BIP 11g newbie and have a problem...
I have the following data:
PersonID Attrib_desc Attrib_value
Ck001 Sex Female
Ck001 DOB 11/12/2005
Ck002 Sex Male
Ck002 DOB 06/06/1970
and I wish to create a BI Publisher report that looks like this:
PersonID Sex DOB
Ck001 Female 11/12/2005
Ck002 Male 06/06/1970
I have tried everything and can't work out how to 'pivot' the data when it is not none measurement! And am not allowed to alter data structure in underlying source - can anyone help?!

Tim Dexter kindly answered this issue for me - see http://blogs.oracle.com/xmlpublisher/entry/tricky_grouping for how to solve this issue. Thanks Tim!

Similar Messages

  • Working with NON-Structured Data

    Dear Colleague,
    I am preparing to build a Forms application (using Oracle 9i Developer Suite, Release 2, version 9.0.2.0.1).
    The application (and Data model) will need to associate SOPs (standard operating procedures) with certain measurements and treatments. It is planned to have the SOPs be, for example, MS Word documents, i.e. the SOPs will be non-structured data.
    What is best practice when handling this situation? Please advise.
    Assume I would like to view, print and (maybe) update the SOPs (Word files).
    1. Should I just store a link to the Word file and a corresponding description or is it better to store the Word file in the DB?
    2. If in the DB, which data type?
    3. If SOP not in the DB, but a file, what is the mechanism for retrieval and display?
    4. Printing the Word file via Forms - how? problems to avoid?
    5. Update of the Word document via Forms - possible? How?
    Best regards,
    Randy

    Hi Vijay,
    sorry
    please rewrite the formula for the button disable as::
    NOTE:: in the formula i have used '<''>'  which nothing but a lessthan symbol and greater than symbol without single quotes (i.e. not equal to)
    BOOL(IF(LEN(@Emp_code) '<''>'0 AND LEN(@Earea)'<''>'0 AND LEN(@Edept)'<''>'0 AND DSUB(DVAL(@Joining_date), DVAL(@Ending_date),'D')<0 ,false,true))
    and make sure you have placed the date picker UI controls for your date functionality.
    this will not work for calender UI control on the form.
    and othe important point is the joining date should be less than ending date then only the button will be enable..
    if you want your joining date should be greater than ending date then write the formaula like this::
    BOOL(IF(LEN(@Emp_code)<>0 AND LEN(@Earea)'<''>'0 AND LEN(@Edept)'<''>'0 AND DSUB(DVAL(@Joining_date), DVAL(@Ending_date),'D')>0 ,false,true))
    i tried this
    its working for me
    Regards
    Srinivas

  • Process not writing to JMS queue with non-xa data source

    I have a process reading from a JMS-AQ with non XA, but it does not seem to be performing a commit and writing to the queue. If I use XA all the way ut works, but I don't get the desired error handling which I have built. Anything obvious I am missing? 11.1.1.5 SOA

    Hi,
    Under Weblogic JDBC Datasources select the datasource and then select Transaction tab there you can find Use XA datasource interface.
    and In DB adapter ... go to Configuration Tab ---Outbound Connection Pool--- Unhide the pool group -Select JNDI name ( e.g eis/DB/local ) ---enter the JDBC datasource created earlier in weblogic JDBC datasource against xADataSourceName..
    Now Click on Transaction tab & select Transaction Support: as XA Transaction .
    Regards,
    Abhinav Gupta

  • Transfer an internal table with non char data type to an application server

    Here is a part of my code. lv_string variable was type string initially but i got an error Not mutually convertible in a Unicode Program. I changed it to type ty_afko. Now the program works but the data it puts out has just a bunch of  #####. I cannot make all value in the internal table type char because then my select statement is giving errors..  Please help...
    Type: BEGIN OF ty_afko,
           rsnum LIKE afko-rsnum,          " Number of reservation
           aprio LIKE afko-aprio,          " Order Priority
           maufnr LIKE afko-maufnr,          " Number of superior order
           lead_aufnr LIKE afko-lead_aufnr,     " Leading order in current processing
         END OF ty_afko.
    data: gt_afko         TYPE STANDARD TABLE OF ty_afko INITIAL SIZE 0,
          wa_afko         TYPE ty_afko,
    select  rsnum aprio maufnr lead_aufnr
    from afko into table gt_afko.
    IF gv_error = 'X'.
      EXIT.
    ELSE.
    <b>DATA: Lv_string type ty_afko.</b>
      loop at gt_afko into wa_afko.
        clear lv_string.
    <b>  move wa_afko to lv_string.</b>
        perform f_transfer_dataset using gc_f_afko lv_string changing gv_error.
      endloop.
    endif.
    FORM f_transfer_dataset USING         p_file
                                          p_data
                            CHANGING      p_error TYPE c.
      Data: lv_file(25) type c.
      CONCATENATE gc_path p_file INTO lv_file.
      TRANSFER p_data TO lv_file.
    ENDFORM.                    "F_TRANSFER_DATASET

    The key is to create another empty structure with fields of type c.
    TYPES:       BEGIN OF ty_resb,
                bdter TYPE RESB-BDTER,     " This field is type d
                bdmng type resb-bdmng,  " This field is type p with 3 decimals - total 14
              END OF ty_resb,
              BEGIN OF ty_resbstr,     " This structure is the same as above but the field types will be char.
               bdter(10),            
               bdmng(14),                
              END OF ty_resbstr.
    DATA:     gt_resb         TYPE STANDARD TABLE OF ty_resb,
              wa_resb         TYPE ty_resb,
           structure type  ty_resbstr. " Empty fields with the same structure as the work area.
    SELECT   bdter bdmng  FROM resb INTO corresponding fields of TABLE gt_resb.
          open dataset pfile for output in text mode encoding default.
          loop at gt_resb into wa_resb.
              clear gv_string.
            move-corresponding wa_resb to gv_string.
             perform f_transfer_dataset using gc_f_resb gv_string.     
            endloop.
         FORM f_transfer_dataset USING p_file p_data.         
           TRANSFER p_data TO lv_file.     
         ENDFORM.

  • Dynamic style with non spatial data?

    Hi all,
    Here is the situation : I have two spatial tables with attributes.
    Each row in the first one will be affected to a row or several rows in the second using a third table that links the primary keys of both tables. (the third table is actually a temporary table loaded from an xml file)
    I would like to represent that on the map using a dynamic style. My column on which is based the affectation is varchar2 column. So I can't use numeric buckets.
    Is it possible to do that? And how?
    Thank you for your help.
    Max

    Hi Max,
    for varchar2 columns you can use Collection bucket styles where the values can numbers of strings. Post the general sturcture of each table (the attributes), and how you would like to link them. You may be able to define a view linking the tables, and then use one attribute of the view for the collection style.
    Joao

  • Measurement data and problem with path selection

    Hello,
    I have a problem with the measurement data function; My instrument is the "Hyoki 3532-50 LCR meter".
    I´m tring to save the data in a text file using the "write to measurement file VI" . From the instrument i receive an array containing the data of the measure and then i use a for loop to obtain the single data obtainig the dynamic data necessary like input in the VI function. 
    The problem is that i receive ( for example) the Rp data and the Cp data in two separate moment and the result using the VI is something like
    1      56
    2      68
    3      95
    1      0,12
    2      0,30
    3      0,56
    (this is a stupid example to explain my situation)
    What i would like to obtain is 
    1      56   0,12
    2      68   0,30
    3      95   0,56
    The fact that i can+t receive all the three dynamic data in the same moment present also the problem that i don´t have the possibility to save in a different file every time i start the measure but to do it i have to reset the labview program ( this is due to the fact that in the property  i select "ask only once" because otherwise if i select " ask at each iteration" the second data flow overwrite the file) .
    If you need some image from code i can provide them 

    You really need to attach some code. We are graphical programmers and don't undestand longwinded text explanation
    Also the code images you attached earlier are useless. The abundance of local variables, stacked sequences, express VIs, and dead code make them impossible to debug by looking at a picture.
    LabVIEW Champion . Do more with less code and in less time .

  • Is it possible to upload non-numeric data into Planning from ODI?

    Dear All,
    I have problem to upload non-numeric data into planning?
    Regards,
    Thomas

    I am not sure what that has to do with non-numeric data and planning.
    If you want to move your ODI environment then have a look at this post :- Re: move full ODI environment to another Machine...
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Are Cube organized materialized view with Year to Date calculated measure eligible for Query Rewrite

    Hi,
    Will appreciate if someone can help me with a question regarding Cube organized MV (OLAP).
    Does cube organized materialized view with calculated measures based on time series  Year to date, inception to date  eg.
    SUM(FCT_POSITION.BASE_REALIZED_PNL) OVER (HIERARCHY DIM_CALENDAR.CALENDAR BETWEEN UNBOUNDED PRECEDING AND CURRENT MEMBER WITHIN ANCESTOR AT DIMENSION LEVEL DIM_CALENDAR."YEAR")
    are eligible for query rewrites or these are considered advanced for query rewrite purposes.
    I was hoping to find an example with YTD window function on physical fact dim tables  with optimizer rewriting it to Cube Org. MV but not much success.
    Thanks in advance

    I dont think this is possible.
    (My own reasoning)
    Part of the reason query rewrite works for base measures only (not calc measures in olap like ytd would be) is due to the fact that the data is staged in olap but its lineage is understandable via the olap cube mappings. That dependency/source identification is lost when we build calculated measures in olap and i think its almost impossible for optimizer to understand the finer points relating to an olap calculation defined via olap calculation (olap dml or olap expression) and also match it with the equivalent calculation using relational sql expression. The difficulty may be because both the olap ytd as well as relational ytd defined via sum() over (partition by ... order by ...) have many non-standard variations of the same calculation/definition. E.g: You can choose to use or choose not to use the option relating to IGNORE NULLs within the sql analytic function. OLAP defn may use NASKIP or NASKIP2.
    I tried to search for query rewrite solutions for Inventory stock based calculations (aggregation along time=last value along time) and see if olap cube with cube aggregation option set to "Last non-na hierarchical value" works as an alternative to relational calculation. My experience has been that its not possible. You can do it relationally or you can do it via olap but your application needs to be aware of each and make the appropriate backend sql/call. In such cases, you cannot make olap (aw/cubes/dimensions) appear magically behind the scenes to fulfill the query execution while appearing to work relationally.
    HTH
    Shankar

  • OBIEE Date Comparison report with multiple measures

    Hello everyone...
    I have a requirement where I need to create a date comparison report (Current Month vs. Prior Month vs Prior Month Year Ago vs YTD vs Prior YTD vs Rolling 3 months, etc) with multiple Measures (Quote count, Quote offered count, New Business count, etc). I am to generate this report in this fashion:
    Quotes Current Month vs PM vs YoY Prior
    Month Current Month Prior Year Current
    YTD vs PM Prior
    YTD Rolling 3 Months vs PM Prior Rolling 3 Months
    Issued 120 120 123 1400 1020 1700 1580
    Unique Risks
    Offered 556 449 571 5881 6220 6101 5892
    Unique Risk Close Ratio
    Offered Close Ratio 21.6% 26.7% 21.5% 23.8% 16.4% 27.9% 26.8%
    Bitmap
    Avg Issued Premium 240 238 222 2801 2630 3104 2892
    Avg Offered Premium
    New Business
    Count
    Avg Full Term Premium
    I understand that time series is an option, and have created the time series (which show up as measures).
    My problem is:
    when creating the report (even as a pivot table) , OBIEE is not allowing me to drag the measure to rows, nor is it allowing me to drag the time series to columns. Also, I need to do conditional formating on some of the columns showing an up arrow for increased vs. prior month and down arrow for decreased vs. prior month.
    Does anyone have a solution for this issue? Any assistance would be greatly appreciated...
    Please advise and thanks!!!

    Jason,
        You've been very helpful, but I guess I'm just not explaining myself that well. The Credit Memo's throw things off because they do not contain a "Paid Date" value. That field in the record is blank. Since it is blank the report query ignores them. Therefore when the report query returns all records that have a "Paid date" range of "x to y" they are never returned. This is a problem whenever a credit memo exists for an invoice because the "invoice total" never changes. So when a customer pays on their account for a particular invoice, they only pay what is due which is the "invoice total" minus any "credit" from the "Credit Memo". So when my A/R person receives payment they are showing the invoice paid in full because MAS automatically includes all invoices and credit memo's. My report does not because there is no data in the "Paid Date" field of the record for Credit Memos.
    I've been doing alot research in trying to resolve this "exception" to my report. I think I may be on to something, but my knowledge of arrays is extremely limited. My resolution involves populating an array with the "invoice numbers" that are returned from the initial query from the supplied date range. Then using the "invoice numbers" from the array have it then populate my details section of the report with all of "invoice records" both the actual invoice record and any "credit memo" records. I can then group the records returned  based upon the "invoice number". then within that grouping I can performing my calculations to show whether or not the invoice has been paid in full. What do you think?  I've been able to populate an array with the invoice numbers based upon my date search range, but I have yet to figure out how to take that information and perform another lookup to pull in the rest of the data I need. Do you have any ideas?
    Thanks.  Bill

  • Help with seting up a Data Sorce can't be created with non-existent Pool

    I am wanting to use an Oracle 9i with WebLogic 7
    I have the following in my config.xml:
    <JDBCConnectionPool DriverName="oracle.jdbc.driver.OracleDriver"
    Name="Thin.Pool" Password="{3DES}C3xDZIWIABA="
    Properties="user=SYSTEM" TestTableName="OID"
    URL="jdbc:oracle:thin:@localhost:1521:DB_SID"/>
    <JDBCDataSource JNDIName="DB_DS" Name="DB_DS" PoolName="Thin.Pool"/>
    The console seems happy, no error mesages but in the log I get:
    ####<Mar 31, 2003 6:33:45 PM MST> <Info> <HTTP> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <101047>
    <[ServletContext(id=4110316,name=console,context-path=/console)]
    FileServlet: Using standard I/O>
    ####<Mar 31, 2003 6:35:37 PM MST> <Info> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001082> <Creating Data Source named DB_DS for pool
    Thin.Pool>
    ####<Mar 31, 2003 6:35:37 PM MST> <Error> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001059> <Error during Data Source creation:
    weblogic.common.ResourceException: DataSource(DB_DS) can't be created
    with non-existent Pool (connection or multi) (Thin.Pool)
         at weblogic.jdbc.common.internal.JdbcInfo.validateConnectionPool(JdbcInfo.java:127)
         at weblogic.jdbc.common.internal.JdbcInfo.startDataSource(JdbcInfo.java:260)
         at weblogic.jdbc.common.internal.JDBCService.addDeploymentx(JDBCService.java:293)
         at weblogic.jdbc.common.internal.JDBCService.addDeployment(JDBCService.java:270)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentTarget.java:375)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentTarget.java:154)
         at java.lang.reflect.Method.invoke(Native Method)
         at weblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl.java:732)
         at weblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:714)
         at weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBeanImpl.java:417)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
         at weblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerImpl.java:952)
         at weblogic.management.internal.ConfigurationMBeanImpl.updateConfigMBeans(ConfigurationMBeanImpl.java:578)
         at weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBeanImpl.java:419)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
         at weblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerImpl.java:952)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:470)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:198)
         at $Proxy16.addDeployment(Unknown Source)
         at weblogic.management.internal.DynamicMBeanImpl.unprotectedUpdateDeployments(DynamicMBeanImpl.java:1784)
         at weblogic.management.internal.DynamicMBeanImpl.access$0(DynamicMBeanImpl.java:1737)
         at weblogic.management.internal.DynamicMBeanImpl$1.run(DynamicMBeanImpl.java:1715)
         at weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:780)
         at weblogic.management.internal.DynamicMBeanImpl.updateDeployments(DynamicMBeanImpl.java:1711)
         at weblogic.management.internal.DynamicMBeanImpl.setAttribute(DynamicMBeanImpl.java:1035)
         at weblogic.management.internal.ConfigurationMBeanImpl.setAttribute(ConfigurationMBeanImpl.java:353)
         at com.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:1358)
         at com.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:1333)
         at weblogic.management.internal.RemoteMBeanServerImpl.setAttribute(RemoteMBeanServerImpl.java:898)
         at weblogic.management.internal.MBeanProxy.setAttribute(MBeanProxy.java:324)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:193)
         at $Proxy13.setTargets(Unknown Source)
         at java.lang.reflect.Method.invoke(Native Method)
         at weblogic.management.console.info.FilteredMBeanAttribute.doSet(FilteredMBeanAttribute.java:92)
         at weblogic.management.console.actions.mbean.DoEditMBeanAction.perform(DoEditMBeanAction.java:145)
         at weblogic.management.console.actions.internal.ActionServlet.doAction(ActionServlet.java:171)
         at weblogic.management.console.actions.internal.ActionServlet.doPost(ActionServlet.java:85)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at weblogic.servlet.internal.ServletStubImpl$ServletInvocationAction.run(ServletStubImpl.java:1058)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:401)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:306)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:5445)
         at weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:780)
         at weblogic.servlet.internal.WebAppServletContext.invokeServlet(WebAppServletContext.java:3105)
         at weblogic.servlet.internal.ServletRequestImpl.execute(ServletRequestImpl.java:2588)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:213)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:189)
    Why does it say:
    can't be created with non-existent Pool
    Thanks,

    Add "Targets" attribute to the connection pool. You may
    get an idea how it looks like by searching config.xml
    for "targets". If target servers are not set, the pool won't be
    deployed and can not be used to a create datasource.
    Regards,
    Slava Imeshev
    "BBaker" <[email protected]> wrote in message
    news:[email protected]...
    I am wanting to use an Oracle 9i with WebLogic 7
    I have the following in my config.xml:
    <JDBCConnectionPool DriverName="oracle.jdbc.driver.OracleDriver"
    Name="Thin.Pool" Password="{3DES}C3xDZIWIABA="
    Properties="user=SYSTEM" TestTableName="OID"
    URL="jdbc:oracle:thin:@localhost:1521:DB_SID"/>
    <JDBCDataSource JNDIName="DB_DS" Name="DB_DS" PoolName="Thin.Pool"/>
    The console seems happy, no error mesages but in the log I get:
    ####<Mar 31, 2003 6:33:45 PM MST> <Info> <HTTP> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <101047>
    <[ServletContext(id=4110316,name=console,context-path=/console)]
    FileServlet: Using standard I/O>
    ####<Mar 31, 2003 6:35:37 PM MST> <Info> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001082> <Creating Data Source named DB_DS for pool
    Thin.Pool>
    ####<Mar 31, 2003 6:35:37 PM MST> <Error> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001059> <Error during Data Source creation:
    weblogic.common.ResourceException: DataSource(DB_DS) can't be created
    with non-existent Pool (connection or multi) (Thin.Pool)
    atweblogic.jdbc.common.internal.JdbcInfo.validateConnectionPool(JdbcInfo.java:
    127)
    atweblogic.jdbc.common.internal.JdbcInfo.startDataSource(JdbcInfo.java:260)
    atweblogic.jdbc.common.internal.JDBCService.addDeploymentx(JDBCService.java:29
    3)
    atweblogic.jdbc.common.internal.JDBCService.addDeployment(JDBCService.java:270
    atweblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentT
    arget.java:375)
    atweblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentT
    arget.java:154)
    at java.lang.reflect.Method.invoke(Native Method)
    atweblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl
    .java:732)
    atweblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:7
    14)
    atweblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBea
    nImpl.java:417)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
    atweblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerI
    mpl.java:952)
    atweblogic.management.internal.ConfigurationMBeanImpl.updateConfigMBeans(Confi
    gurationMBeanImpl.java:578)
    atweblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBea
    nImpl.java:419)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
    atweblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerI
    mpl.java:952)
    at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:470)
    at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:198)
    at $Proxy16.addDeployment(Unknown Source)
    atweblogic.management.internal.DynamicMBeanImpl.unprotectedUpdateDeployments(D
    ynamicMBeanImpl.java:1784)
    atweblogic.management.internal.DynamicMBeanImpl.access$0(DynamicMBeanImpl.java
    :1737)
    atweblogic.management.internal.DynamicMBeanImpl$1.run(DynamicMBeanImpl.java:17
    15)
    atweblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManage
    r.java:780)
    atweblogic.management.internal.DynamicMBeanImpl.updateDeployments(DynamicMBean
    Impl.java:1711)
    atweblogic.management.internal.DynamicMBeanImpl.setAttribute(DynamicMBeanImpl.
    java:1035)
    atweblogic.management.internal.ConfigurationMBeanImpl.setAttribute(Configurati
    onMBeanImpl.java:353)
    atcom.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:135
    8)
    atcom.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:133
    3)
    atweblogic.management.internal.RemoteMBeanServerImpl.setAttribute(RemoteMBeanS
    erverImpl.java:898)
    atweblogic.management.internal.MBeanProxy.setAttribute(MBeanProxy.java:324)
    at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:193)
    at $Proxy13.setTargets(Unknown Source)
    at java.lang.reflect.Method.invoke(Native Method)
    atweblogic.management.console.info.FilteredMBeanAttribute.doSet(FilteredMBeanA
    ttribute.java:92)
    atweblogic.management.console.actions.mbean.DoEditMBeanAction.perform(DoEditMB
    eanAction.java:145)
    atweblogic.management.console.actions.internal.ActionServlet.doAction(ActionSe
    rvlet.java:171)
    atweblogic.management.console.actions.internal.ActionServlet.doPost(ActionServ
    let.java:85)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
    atweblogic.servlet.internal.ServletStubImpl$ServletInvocationAction.run(Servle
    tStubImpl.java:1058)
    atweblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java
    :401)
    atweblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java
    :306)
    atweblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(W
    ebAppServletContext.java:5445)
    atweblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManage
    r.java:780)
    atweblogic.servlet.internal.WebAppServletContext.invokeServlet(WebAppServletCo
    ntext.java:3105)
    atweblogic.servlet.internal.ServletRequestImpl.execute(ServletRequestImpl.java
    :2588)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:213)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:189)
    Why does it say:
    can't be created with non-existent Pool
    Thanks,

  • Pie chart with two measures and date dimension navigation not working

    Hi Experts,
    Pie chart with two measures and date dimension navigation not working. Any help is appreciated.
    Thanks
    V

    Hi Deepak,
    I had time dimension in the RPD.
    I have stacked bar chart with same time dim like year & month in the report. when I go to legand and set navigation it is working fine. But not with pie chart.
    I am not not sure what is the problem. When I click on Pie chart it is not navigating to the target report. Can it be any other issues..???

  • Ale master data scenario integration with non sap syetem i.e seebeyound

    hi
    we have a business process scenario integrating with non sap system (i.e seebeyound)
    action code      itemnumber    plant code
    a(add)          material1     plant 2
    d(delete)       material 3    plant 6
    etc
    we have upto one million materials assigned to 300 plants
    we have to send all the above data to non-sap system (see beyound)
    should we go for a custom idoc and setup the rfc configuration for the two sytems.
    or can you please help how to go for this scaneriao

    this is not one time.it is on daily basis.
    please look at this scenario.what i need is how should i go for ale/idoc scenario for this one
    Business case for Heiler Part/Plant Interface:
    A key function of the Heiler PBC tool is to provide users a choice
    between a "Global Catalog View" (e.g. all parts in the catalog) or a
    "Local Catalog View" (e.g. the subset of parts in the catalog that have
    a Material Master established for the user's plant in SAP).  The "Local
    Catalog View" is the default condition and encourages users to procure
    parts that are on contract for their plant and potentially already
    stocked as an inventory item at their plant.  Parts that are not in a
    user's "Local Catalog View" can only be procured as a spot buy purchase
    order, requiring processing by a buyer.
    The new Part/Plant interface is required in order for the Heiler PBC
    tool to provide this "Local Catalog View" function.  The Heiler
    Part/Plant table will document what plants in SAP have a Material Master
    record established for each part.  SAP will maintain the Heiler
    Part/Plant table using this interface.  When a user performs a part
    search in the Heiler PBC tool, the Part/Plant table determines if a part
    is included in the user's "Local Catalog View".

  • Using Non-destructive filter with Nested XML data

    Hi,
    How do you use Non-destructive filter with Nested XML data?
    I am using the non-destructive filter sample with my own xml which is setup to search for the <smc></smcs> in my xml below. But when i test it it only searches the last row of the "smc". How can i make it work so it can search for repeating nodes? or does it have something to with how my xml is setup?
        <ja>
            <url>www.sample.com</url>
            <jrole>Jobrole goes here</jrole>
            <prole>Process role goes here...</prole>
            <role>description...</role>
            <prole>Process role goes here...</prole>
            <role>description....</role>
            <prole>Process role goes here...</prole>
            <role>description...</role>
            <sjc>6K8C</sjc>
            <sjc>6B1B</sjc>
            <sjc>6B1F</sjc>
            <sjc>6B1D</sjc>
            <smc>6C9</smc>
            <smc>675</smc>
            <smc>62R</smc>
            <smc>62P</smc>
            <smc>602</smc>
            <smc>622</smc>
            <smc>642</smc>
            <smc>65F</smc>
            <smc>65J</smc>
            <smc>65L</smc>
            <smc>623</smc>
            <smc>625</smc>
            <smc>624</smc>
            <smc>622</smc>
            <audience>Target audience goes here....</audience>
        </ja>
    here is the javascript that runs it.
    function FilterData()
        var tf = document.getElementById("filterTF");
        if (!tf.value)
            // If the text field is empty, remove any filter
            // that is set on the data set.
            ds1.filter(null);
            return;
        // Set a filter on the data set that matches any row
        // that begins with the string in the text field.
        var regExpStr = tf.value;
    if (!document.getElementById("containsCB").checked)
            regExpStr = "^" + regExpStr;
        var regExp = new RegExp(regExpStr, "i");
        var filterFunc = function(ds, row, rowNumber)
            var str = row["smc"];
            if (str && str.search(regExp) != -1)
            return row;
            return null;
        ds1.filter(filterFunc);
    function StartFilterTimer()
        if (StartFilterTimer.timerID)
            clearTimeout(StartFilterTimer.timerID);
        StartFilterTimer.timerID = setTimeout(function() { StartFilterTimer.timerID = null; FilterData(); }, 100);
    I really need help on this, or are there any other suggestions or samples that might work?
    thank you!

    I apologize, im using Spry XML Data Set. i guess the best way to describe what im trying to do is, i want to use the Non-desctructive filter sample with the Spry Nested XML Data sample. So with my sample xml on my first post and with the same code non-destructive filter is using, im having trouble trying to search repeating nodes, for some reason it only searches the last node of the repeating nodes. Does that make sense? let me know.
    thank you Arnout!

  • Streaming data from another PC, combining it with DAQ measurement

    Hi
    I'm just getting back into LabVIEW after 2 years without using any NI software, so apologise if I'm missing something really obvious and this is a simple question. I have had a search of the forums, but can't find what I'm looking for. Perhaps I just don't know what to look for.
    I'm starting to plan a combined data acquisition system for an engineering test, which will involve two PCs. I would like some advice on how to set up a LabVIEW system on one of them.
    I need to produce a system which will:
    (a) log measurement data using normal DAQmx method - that bit should be straightforward enough.
    (b) import independent measurement data from the other PC, probably via an Ethernet LAN. What’s a good way to get data into LabVIEW?
    (c) correlate the 2 sets of data. What I really want to do is feed the input data into my measurement system in real time.
    (d) produce data files incorporating all data
    I'm not involved in programming the system for the other PC, but can have some input into how it’s arranged.
    Is TCP the best way to go? I would be very grateful for any guidance anyone can offer on how to do steps (b) and (c).
    Thanks.

    MRH wrote:
    Hi
    I'm just getting back into LabVIEW after 2 years without using any NI software, so apologise if I'm missing something really obvious and this is a simple question. I have had a search of the forums, but can't find what I'm looking for. Perhaps I just don't know what to look for.
    I'm starting to plan a combined data acquisition system for an engineering test, which will involve two PCs. I would like some advice on how to set up a LabVIEW system on one of them.
    I need to produce a system which will:
    (a) log measurement data using normal DAQmx method - that bit should be straightforward enough.
    (b) import independent measurement data from the other PC, probably via an Ethernet LAN. What’s a good way to get data into LabVIEW?
    (c) correlate the 2 sets of data. What I really want to do is feed the input data into my measurement system in real time.
    (d) produce data files incorporating all data
    I'm not involved in programming the system for the other PC, but can have some input into how it’s arranged.
    Is TCP the best way to go? I would be very grateful for any guidance anyone can offer on how to do steps (b) and (c).
    Thanks.
    You will have to give us more info about the dat on the other machine for us to be able to help.
    Is the data on a web-page (DataSocket reads)
    In a DAQ sub-system on the other machine (Action Engine Served to the clinet via VI server)
    In a file on the other machine (map the drive and file read)
    In memory on the other machine (SCRAMNet)
    So tell us more about the data.
    Ben
    Ben Rayner
    I am currently active on.. MainStream Preppers
    Rayner's Ridge is under construction

  • Obiee spurious pivot table non-total figures

    Hi,
    My problem is that a pivot table in conjunction with non-hierarchical columns presents figures where no figures exist at data level rather than pivot table totals.
    OBIEE version 11.1.1.7.0
    Essbase version 11.1.2.3
    I have an Essbase data source for my subject area.
    I have a column (dimension) 'Account' which has the structure (in Essbase) Account is parent of 'Total Income and Expenditure' is parent of 'Total Income' and 'Total Expenditure'. I filter to only include 'Total Income' and 'Total Expenditure' in my query.
    I have a column (dimension) 'Entity' which has a number of tiers of which I am including in the query only the first level of children of Entity.
    My problem is that in a number of rows I get figures against 'Total Income' where when I check against the source this should be null, all of the way down the hierarchy - i.e. there are no figures at all so this cannot be an aggregation issue.
    I am not doing anything exotic on the business layer in terms of alternative aggregation sources.
    If I use a hierarchical column I get the correct numbers, but I do not want to in this case.
    Any input appreciated.
    thanks,
    Robert.

    Use expression sum(Marks by Name) in the FX of mark measure in the criteria tab and set the property to suppress the values for measure column.
    appreciate if you mark as correct.
    Regards,
    Kashi

Maybe you are looking for