Query is not hitting in detail Level

Hi,
1.Why the query is hitting the detailed level source when aggregate source is available ? BI Server is confused and hitting the detail level. Even after setting the levels at content tab it is not hitting the aggregate source.
Thanks In Advance,
Raji......

Make sure your detail level table and agg tables are not set to same level, these suppose to be different levels from same dim.
Pls mark if helps

Similar Messages

  • Query Designer not working with Aggregation Levels on BW 7.30

    Hi,
    Every time I try to create a query on top of an Aggregation Level that I created with the new ABAP based RSPLAN transaction.
    If somebody can help me out I'd appreciate it.
    By the way, we don't have the Java stack, but we are currently using BO4's JVM as we're on ramp up.
    This is the error that I get when trying to create the query:
    QD Revision 667
    ERROR ID: E-ONGUIUNHANDLEDEXCEPTION
    And this is the error log:
    1:42:13 p.m..319: Info: Query Designer Start. Revision: 667
    QDbCommandBase::Execute  - Standard View
    QDbCommandBase::Execute  - Table View
    QDbCommandBase::Execute  - Rows/Columns
    QDbCommandBase::Execute  - Cells
    QDbCommandBase::Execute  - Conditions
    QDbCommandBase::Execute  - Exceptions
    QDbCommandBase::Execute  - InfoProvider
    QDbCommandBase::Execute  - Filter
    QDbCommandBase::Execute  - Documents
    QDbCommandBase::Execute  - Where-Used List
    QDbCommandBase::Execute  - Properties
    QDbCommandBase::Execute  - Properties
    QDbCommandBase::Execute  - Messages
    QDbCommandManager::ItemClickedHandler - Bar clicked: NewQuery
    QDbCommandBase::Execute  - New...
    QDbCommandBase::Execute  - Table View
    QDbCommandBase::Execute  - Rows/Columns
    QDbCommandBase::Execute  - Cells
    QDbCommandBase::Execute  - Conditions
    QDbCommandBase::Execute  - Exceptions
    QDbCommandBase::Execute  - InfoProvider
    QDbCommandBase::Execute  - Filter
    QDbCommandBase::Execute  - Documents
    QDbCommandBase::Execute  - Where-Used List
    QDbCommandBase::Execute  - Messages
    QDbCommandBase::Execute  - Table View
    QDbCommandBase::Execute  - Cells
    QDbCommandBase::Execute  - Conditions
    QDbCommandBase::Execute  - Exceptions
    QDbCommandBase::Execute  - Rows/Columns
    QDbCommandBase::Execute  - Filter
    QDbCommandBase::Execute  - InfoProvider
    QDbCommandBase::Execute  - Properties
    -EXCEPTION-START- 1:42:52 p.m..416: TRACE EXCEPTION  ---
    Exception Name: TargetInvocationException
    Exception Message: Exception has been thrown by the target of an invocation.
    Exception    at System.RuntimeMethodHandle._InvokeMethodFast(Object target, Object[] arguments, SignatureStruct& sig, MethodAttributes methodAttributes, RuntimeTypeHandle typeOwner)
       at System.RuntimeMethodHandle.InvokeMethodFast(Object target, Object[] arguments, Signature sig, MethodAttributes methodAttributes, RuntimeTypeHandle typeOwner)
       at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags invokeAttr, Binder binder, Object[] parameters, CultureInfo culture, Boolean skipVisibilityChecks)
       at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)
       at System.RuntimeType.InvokeMember(String name, BindingFlags bindingFlags, Binder binder, Object target, Object[] providedArgs, ParameterModifier[] modifiers, CultureInfo culture, String[] namedParams)
       at System.Type.InvokeMember(String name, BindingFlags invokeAttr, Binder binder, Object target, Object[] args)
       at com.sap.bi.et.QueryDesigner.QDpPropertyBase.PropertyGet(Object iPropertySource, String iPropertyName)
       at com.sap.bi.et.QueryDesigner.QDpPropertyBase.ValueGetDefault(Object iPropertySource)
       at com.sap.bi.et.QueryDesigner.QDpPropertyBase.pValueDefault()
       at com.sap.bi.et.QueryDesigner.QDpPropertyBase.ValueDefault()
       at com.sap.bi.et.QueryDesigner.QDpPropertyBase.ValueResolved()
       at com.sap.bi.et.QueryDesigner.QDpPropertyMulti.SetFirstProperty(QDpPropertyBase iProperty)
       at com.sap.bi.et.QueryDesigner.QDpPropertyMulti.Add(QDpPropertyBase iProperty)
       at com.sap.bi.et.QueryDesigner.QDuPropPageQuery.ToDialog()
       at com.sap.bi.et.QueryDesigner.QDuPropPageBase.ToDialogCore()
       at com.sap.bi.et.QueryDesigner.QDuPropPageBase.set_CommandContext(QDbCommandContext Value)
       at com.sap.bi.et.QueryDesigner.QDuPropPages.EnablePage()
       at com.sap.bi.et.QueryDesigner.QDuPropPages.ContextChangedHandler(Object iSender, QDbCommandContext iCommandContext)
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.ContextChangedEventHandler.Invoke(Object iSender, QDbCommandContext iCommandContext)
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.OnContextChanged(QDbCommandContext iCommandContext)
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.CalculateContext(QDcView iViews, QDeAreaType iAreaType)
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.CalculateContext(QDbElement iElement)
       at com.sap.bi.et.QueryDesigner.QDbCommandPropertiesQuery.ExecuteCommand()
       at com.sap.bi.et.QueryDesigner.QDbCommandBase.Execute()
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.CommandExecute(QDbCommandBase iCommand)
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.InitialCommandExecute(QDbCommandBase iCommand)
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.DoExecuteCommandInternal()
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.DoExecuteCommand(QDbCommandBase iCommand)
       at com.sap.bi.et.QueryDesigner.QDbCommandManager.ItemClickedHandler(Object sender, BarItemClickedEventArgs args)
       at com.sap.bi.et.QueryDesigner.QDdEventDispatcher.MenuItemClickedHandler(Object iSender, BarItemClickedEventArgs iE)
       at Syncfusion.Windows.Forms.Tools.XPMenus.BarManager.OnItemClicked(BarItemClickedEventArgs args)
       at Syncfusion.Windows.Forms.Tools.XPMenus.BarItem.OnItemClicked(EventArgs args)
       at com.sap.bi.et.QueryDesigner.QDiBarItem.OnItemClicked(EventArgs args)
       at Syncfusion.Windows.Forms.Tools.XPMenus.BarItem.PerformClick()
       at Syncfusion.Windows.Forms.Tools.XPMenus.BarRenderer.OnMouseUp(MouseEventArgs e)
       at Syncfusion.Windows.Forms.Tools.XPMenus.BarControlInternal.OnMouseUp(MouseEventArgs e)
       at System.Windows.Forms.Control.WmMouseUp(Message& m, MouseButtons button, Int32 clicks)
       at System.Windows.Forms.Control.WndProc(Message& m)
       at Syncfusion.Windows.Forms.Tools.XPMenus.BarControlInternal.WndProc(Message& m)
       at System.Windows.Forms.Control.ControlNativeWindow.OnMessage(Message& m)
       at System.Windows.Forms.Control.ControlNativeWindow.WndProc(Message& m)
       at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
    Full Stack:   at com.sap.bi.et.common.appl.Log.Debug.WriteTraceToFile(Level lLevel, String lString, Exception ex)
       at com.sap.bi.et.common.appl.Log.Trace.Exception(Exception ex, String iAdditionalInformation)
       at com.sap.bi.et.QueryDesigner.QDbApplicationData.OnGuiUnhandledException(Object iSender, ThreadExceptionEventArgs iEventArgs)
       at System.Windows.Forms.Application.ThreadContext.OnThreadException(Exception t)
       at System.Windows.Forms.Control.WndProcException(Exception e)
       at System.Windows.Forms.Control.ControlNativeWindow.OnThreadException(Exception e)
       at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
       at System.Windows.Forms.UnsafeNativeMethods.DispatchMessageW(MSG& msg)
       at System.Windows.Forms.Application.ComponentManager.System.Windows.Forms.UnsafeNativeMethods.IMsoComponentManager.FPushMessageLoop(Int32 dwComponentID, Int32 reason, Int32 pvLoopData)
       at System.Windows.Forms.Application.ThreadContext.RunMessageLoopInner(Int32 reason, ApplicationContext context)
       at System.Windows.Forms.Application.ThreadContext.RunMessageLoop(Int32 reason, ApplicationContext context)
       at System.Windows.Forms.Application.Run(Form mainForm)
       at com.sap.bi.et.QueryDesigner.QDbApplicationData.Run(Boolean iAsApplication)
       at com.sap.bi.et.QueryDesigner.QDbQueryDesigner.Run(Boolean iAsApplication)
       at com.sap.bi.et.QueryDesigner.QDStarter.QDStartup.Main()
    -EXCEPTION-END----
    -CALLING-FROM- 1:42:52 p.m..416: TRACE EXCEPTION  ---
       at com.sap.bi.et.common.appl.Log.Debug.WriteTraceToFile(Level lLevel, String lString, Exception ex)
       at com.sap.bi.et.common.appl.Log.Trace.Exception(Exception ex, String iAdditionalInformation)
       at com.sap.bi.et.QueryDesigner.QDbApplicationData.OnGuiUnhandledException(Object iSender, ThreadExceptionEventArgs iEventArgs)
       at System.Windows.Forms.Application.ThreadContext.OnThreadException(Exception t)
       at System.Windows.Forms.Control.WndProcException(Exception e)
       at System.Windows.Forms.Control.ControlNativeWindow.OnThreadException(Exception e)
       at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
       at System.Windows.Forms.UnsafeNativeMethods.DispatchMessageW(MSG& msg)
       at System.Windows.Forms.Application.ComponentManager.System.Windows.Forms.UnsafeNativeMethods.IMsoComponentManager.FPushMessageLoop(Int32 dwComponentID, Int32 reason, Int32 pvLoopData)
       at System.Windows.Forms.Application.ThreadContext.RunMessageLoopInner(Int32 reason, ApplicationContext context)
       at System.Windows.Forms.Application.ThreadContext.RunMessageLoop(Int32 reason, ApplicationContext context)
       at System.Windows.Forms.Application.Run(Form mainForm)
       at com.sap.bi.et.QueryDesigner.QDbApplicationData.Run(Boolean iAsApplication)
       at com.sap.bi.et.QueryDesigner.QDbQueryDesigner.Run(Boolean iAsApplication)
       at com.sap.bi.et.QueryDesigner.QDStarter.QDStartup.Main()

    Hi,
    it seems that you are already using the most recent query designer version. Is your SAPGUI also the most recent version? If yes, open a OSS ticket. If not reinstall the SAPGUI and the BEx tools. I am also working with this query desinger version and I don't have this kind of problem.
    Regards,
    Gregor

  • Named query cache not hit

    Hi,
    I'm using Toplink ORM 10.1.3.
    I have a table called STORE_CONFIG which has a primary key called KEYWORD (a VARCHAR2). The POJO mapped to this table is called StoreConfig.
    In the JDeveloper (10.1.3.1.0) mapping workbench I've defined a named query to query by the PK called "getConfigPropertyByKeyword". The type of the named query is ReadObjectQuery.
    SELECT keyword, key_value
    FROM STORE_CONFIG
    WHERE (keyword = #key)
    Under the options tab I have the following settings:
    Cache Statement: true
    Bind Parameters: true
    Cache Usage: Check Cache by Primary Key
    The application logs show that the same database queries are executed multiple times for the same PK (keyword)! Why is that? Shouldn't it be checking the Object Cache rather than going to the DB?
    I've tried it with "Cache Statement: false" and "Bind Parameters: false" with the same problem.
    If I click the Advanced tab and check "Cache Query Results" then the database is not hit twice for the same record. However it was my understanding that since I am querying by PK that I wouldn't need to set "Cache Query Results".
    Doesn't "Cache Query Results" apply to the Query Cache and not the Object Cache?

    Your issue seems to be that you are using custom SQL for the query, not a TopLink expression. When you use an Expression query TopLink's know if the query is by primary key and can get a cache hit.
    When you use custom SQL, TopLink does not know that the SQL is by primary key, so does not get a cache hit.
    You could either use an Expression for the query,
    or when using custom SQL you should be able to name your query argument the same as your database field defined as the primary key in your descriptor (case sensitive).
    i.e.
    SELECT keyword, key_value
    FROM STORE_CONFIG
    WHERE (keyword = #KEYWORD)

  • When KF value only provided at higher level and not at more detailed level

    Dear Experts
    I have a multiprovider consisting 2 DSOs:
    1.  DSO for price.
    Eg. 
    Char-A__Char-sub1___Char-sub2___KF_PRICE
    AAAA__ AA11________AA22________$10
    AAAA__AA11_________AA23_______ $17
    As shown above,  AAAA is made up of 2 sub-items AA22 and AA23 , each priced differently.
    2.  DSO for Qty
    Char-A____KF_QTY
    AAAA_____200
    As shown above, this DSO for quantity does not have sub-item qty. What should be the quantities for each sub-item? Its always exactly the same as the Qty at Char-A level . Eg.  200 pieces of AAAA is made up of 200 pieces of AA22 and AA23.  
    The user wants to drilldown to sub-item levels to see Price X Qty.
    I tested, when not drilled down to these levels, the display and figures are correct.
    But when drilled down , the report will have additional line for each record due to the fact that the Qty is not at sub-item level.
    So, qty will be shown in another line and the price will be on the line with the sub-item.
    How can this situation be resolved using  BEx? Else, I would use staging to resolve this. But I am hoping BEx level can resolve it. If staging is used, I would need to copy the same qty to each sub-item for each record by adding qty field into Price DSO or union the 2 DSOs into 1 DSO. But I thought Multiprovider is already a Union?
    Kindly could you provide me some advice, clues or opinion?
    Thanks in advance.
    My Best regards
    Pascal

    Hi,
    This issue can be resolved by using constant selection
    For step by step approach please refer the below pdf.
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/70dcd7b5-6a3d-2d10-c984-e835e37071a2?QuickLink=index&overridelayout=true
    Hope it helps.
    Regards,
    AL

  • Query not hitting the Aggregate

    HI Gurus,
    I have created a Aggregate on one info cube. now when i am executing a query which was created before the creation of aggregate then that query is hitting the Aggregate.
    But i created one more new query (after the creation of aggregate) on the same infoprovider.
    Now when i am checking that query is not hitting aggregate..
    I am checking through RSRT --> Execute and Debug --> Display Aggregate.
    Its quite strange for me...
    Please help..

    Hi,
    Thanks for your quick reply.
    Let me explain my problem clearly. I am aware of that when the Aggregate contain all the characteristics that are in the query only that Aggregate is used by the report while executing it. I have few cubes on which I have created Aggregates which includes all the characteristics which are used in all the reports. Now I have created a multi provider on these cubes and created a query on that multi provider. Now when I execute the query does it hit the Aggregate or not.
    For Example.
    I have cubes A, B and C. A contain 1,2,3 characteristics and B contains 1,4,5 and  contain 1,4,6,7. I have created Aggregates on these cubes. I have created the Multi Provider on these and executing the query. This query contain all these characteristics 1,2,3,4,5,6,7. when I run the query does it hit the any of the Aggregates.
    Regards
    Sankar.

  • The Event read query was not supported

    I have an error during a Search request in my collab suite calendar.
    Here the soap message :
    <SOAP-ENV:Envelope xmlns:SOAP-ENV="http://schemas.xmlsoap.org/soap/envelope/"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xmlns:xsd="http://www.w3.org/2001/XMLSchema">
    <SOAP-ENV:Header>
    <auth:BasicAuth xmlns:auth="http://soap-authentication.org/2002/01/">
    <Name>lde</Name>
    <Password>xxxx</Password>
    </auth:BasicAuth>
    </SOAP-ENV:Header>
    <SOAP-ENV:Body>
    <cwsl:Search xmlns:cwsl="http://www.oracle.com/WebServices/Calendaring/1.0/">
    <CmdId>test</CmdId>
    <vQuery>
    <From>VEVENT</From>
    <Where>DTSTART >= '20061002T000000Z' AND DEND <= '20061002T235959Z'</Where>
    <x-oracle-searchhandle></x-oracle-searchhandle>
    <x-oracle-timestamp>20061002T113300Z</x-oracle-timestamp>
    </vQuery>
    </cwsl:Search>
    </SOAP-ENV:Body>
    </SOAP-ENV:Envelope>
    And the reply is :
    <soap:Envelope xmlns:soap="http://schemas.xmlsoap.org/soap/envelope/" soap:encodingStyle="http://schemas.xmlsoap.org/soap/encoding/">
    <soap:Body>
    <soap:Fault>
    <faultcode>soap:Server.Error::Data::CalEvent</faultcode>
    <faultstring>The Event read query was not supported</faultstring>
    <detail>
    <cwsl:Error xmlns:cwsl="http://www.oracle.com/WebServices/Calendaring/1.0/">
    <Class>Error::Data::CalEvent</Class>
    <Code>000C-07-00-00000023</Code>
    <Line>306</Line>
    <FileName>UniapiQuery.cpp,v</FileName>
    <Version>1.2</Version>
    <LastMod>2005/04/26 16:47:36</LastMod>
    <Author>kelsaid</Author>
    <Date>Mon Oct 02 11:48:16 2006</Date>
    <PID>30834</PID>
    <TID>1124678576</TID>
    </cwsl:Error>
    </detail>
    </soap:Fault>
    </soap:Body>
    </soap:Envelope>
    I haven't any errors with this sample.
    <SOAP-ENV:Envelope xmlns:SOAP-ENV="http://schemas.xmlsoap.org/soap/envelope/"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xmlns:xsd="http://www.w3.org/2001/XMLSchema">
    <SOAP-ENV:Header>
    <auth:BasicAuth xmlns:auth="http://soap-authentication.org/2002/01/">
    <Name>lde</Name>
    <Password>xxxxx</Password>
    </auth:BasicAuth>
    </SOAP-ENV:Header>
    <SOAP-ENV:Body>
    <cwsl:Search xmlns:cwsl="http://www.oracle.com/WebServices/Calendaring/1.0/">
    <CmdId>test</CmdId>
    <vQuery>
    <From>VEVENT</From>
    </vQuery>
    </cwsl:Search>
    </SOAP-ENV:Body>
    </SOAP-ENV:Envelope>
    Message was edited by:
    Ludovic DESSEMON

    I resolved my problem. It was a syntax error :
    <Where>DTSTART >= '20061002T000000Z' AND DTEND <= '20061002T235959Z'</Where>
    and not :
    <Where>DTSTART >= '20061002T000000Z' AND DEND <= '20061002T235959Z'</Where>

  • SSRS,MDX- Passed parameters are not hitting the main query in Dataset

    I have created a SSRS report which has to accept two parameters i.e., store name, Date(to show sales on particular date) in report. which are working fine for me. but, those passed parameters are not hitting the main query in main Dataset. what ever i
    pass in parameters, it is displaying all the rows(every city values). i think i am not correct in creating the statement in  Main dataset query, to receive those 2 parameters into the main query. can anybody please help me how to create the statement
    in main date set to display the results for the exact parameter i am passing(which has to be in MDX).
    ram

    Hi ramprasad74,
    According to your description, you want to use parameter in the report, then after parameter values are selected, corresponding data will be displayed in the report. But the parameters do not work and all the rows are displayed.
    To pass parameter to main report dataset, we need to define parameter for MDX query and assign parameter name same as SSRS report parameter. For detail information, please refer to the following steps:
    Create a Dataset to retrieve data for Store name parameter.
    Create Store name parameter, type name and prompt, set Date type to Text, check Allow multiple values check box, then select get values from the dataset.
    Create a Dataset for Date parameter.
    Create Date parameter, type name and prompt, set Date type to Date/Time, then select get values from the dataset.
    Create a new dataset used to retrieve data for the report, add statement like below to the mdx query: {STRTOMEMBER(@Date)} * {STRTOSET (@Store_name)}.
    In Query Designer pane, click Query Parameters and assign parameter name same as SSRS report parameter.
    For more information about SSRS Report with Single and Multi Selection Parameter using MDX Query, please refer to the following blog:
    http://www.codeproject.com/Articles/799265/SSRS-Report-with-Single-and-Multi-Selection-Parame
    If you have any more questions, please feel free to ask.
    Thanks,
    Wendy Fu
    If you have any feedback on our support, please click
    here.

  • The SharePoint list query is not valid: Data at the root level is invalid. Line 1, position 1. (rsSPDataProviderError)

    Hi there,
    All my reports are throwing below error in a sharepoint site configured with Access services. Does anyone have a clue what's causing it?
    An error has occurred during report processing. (rsProcessingAborted)
     Query execution failed for dataset 'Default'. (rsErrorExecutingCommand)
     The SharePoint list query is not valid: Data at the root level is invalid. Line 1, position 1. (rsSPDataProviderError)
    Thanks
    Manvir

    Hi Manvir,
    For troubleshooting your issue, please go to C:\Inetpub\wwwroot\wss\VirtualDirectories\ find your Web Application (represented by a port number), then delete an unknown folder created by sharepoint, namely vti_cnf from all the folders and sub-folders in
    your web application.
    Reference:
    http://forums.asp.net/t/985790.aspx?Data+at+the+root+level+is+invalid+Line+1+position+1
    Best Regards,
    Eric
    TechNet Community Support
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
    [email protected]

  • Grand total values are not matching with Detail report

    Report has grand totals and when I drill to the detail report, grand total values are NOT matching with parent report totals, I did some analysis but I'm clueless on this issue.
    Please provide your thoughts and insight on this issue..
    Thanks

    is your summary and detail reports hitting different facts, like summary hitting aggregate and detail report hitting it's corresponding detail level fact..?
    if then,
    From Front-end:
    Fix the filter values in detail report that are passing from master report then try delete each columns then check the grand total. If you found your values is matching by deleting particular column then you need to investigate what is the issue around with that dimension table..
    From Database side:
    1. check first aggregate table has proper aggregate data of it's detail..
    2. Take the detail report obiee generated query and try to comment each dimension table and it's corresponding joins to the facts, (before, this delete all the dimensional columns and other measures from select statement and put only that measure where you are getting wrong value, so that you need not to comment all the select and group by columns which saves your time.. ). Need to check by commenting each dimensional wid and it's table from clause, if you found that values is matching then there is some problem with wid columns data population in your ETL.
    Is that BI-Apps project?
    btw, whtz ur name?

  • Report not hitting pre seed cahe created using iBot by Admin

    Hi,
    We created an iBot to pre seed the cache using Administrator and it is creating the cache but when normal user run this particular report for the 1st time it's not hitting the cache(created using iBot). all the subsequest users request hitting the cache created by normal user. I still dont understand why it is not hitting the cache.
    Did anyone comeacross this issue before?
    what are the reasons not hitting cache with above scenario??
    does cache created by administrator using ibot work for other users? we actually running whole query with no filters on it using admin and normal users run the same report with some filters on it. as per oracle cache strategy it should work but it's not.
    any help appreciated.
    Thanks
    Jay.
    Edited by: Jay on May 16, 2012 7:23 PM
    Edited by: Jay on May 17, 2012 5:42 AM

    Hi Jay,
    I have given some inputs.
    The way the OBIEE server creates the SQL (both logical and physical) for a request can be a bit funky sometimes, the cache seeding only really works in fairly simple cases. Does your request have a pivot table in it by any chance? These are notorious for not caching properly, if you look at the log for the request you can see why as the server adds this strange "aggregate by" to the request (why this can't be done at the presentation level since the only change you are asling for is in the presentation of the content is beyond me). Those "aggregate by"s tend to stop a request being a cache hit unless it is identical to the one that seeded that cache, any change in parameters, columns etc (even if a logical subset) will not get a cache hit.
    2. "Oracle BI Server Cache" Cache Seeding option present in Destination tab of Delivers -- Please check this option.
    3. Caching is one of the many approaches to improve performance but it's not a magic solution. You need to understand that you can't cache everything. In particular you won't be able to cache reports that are driven by parameters and have facts that are too granular and exceed the number of rows each cache entry can have. If your fact is of a small size then you can get around the parameters issue by caching a report without any filters. The BI Server should able to derive subsequent queries as long as they meet the cache hit criteria. Have a look at the administration manual for all the rules a cache hit must meet.
    Hope it helps.
    Thanks,
    Satya

  • Aggregate not hitting

    Hello Friends,
    I am trying to create an aggregate for Income Statement Query. 
    On the filter I have Fiscal Year Variant, Currency type and Chart of Accounts
    On the Free Characteristics I have Profit center, company code and Fiscal year period.
    I tried different ways on the Characteristics with fixed values and hierarchy(account no) but the aggregate is never used when executed the query. Can anybody give some input
    thanks

    Hi
    If your aggregate are not hitting
    Please ask the User to send me Book mark how he is drill down the query
    Check with rsrt tcode how you aggregate is defined
    copy the aggregate which is hitting ur aggregate and
    apply Fixed for Company code value type ,version,Hierarchy level set and
    chech
    rsddstat,rsddstataggrdef and RSDDAGGRDIR tables
    rgrds
    Reddy

  • Querying objects not in the NamedCache

    The wiki topic on querying (http://wiki.tangosol.com/display/COH32UG/Querying+the+Cache ) points out that a query will "apply only to currently cached data".
         This seems fairly logical because it seems unreasonable to expect the cache to hold onto information that it has already evicted.
         If you were to design a DAO layer (following http://wiki.tangosol.com/display/COH32UG/Managing+an+Object+Model ) using the first of the following architectures:
         1. Direct Cache access:
         App <---> NamedCache <---> CacheStore <---> DB
         2. Direct Cache and DB-DAO access:
         App
         |
         CacheAwareDAO <---> CacheStore <---> DB
         |
         NamedCache
         |
         CacheStore
         |
         DB
         you would then have a situation where you would not be able to query evicted data.
         So by using the 2nd strategy I assume you would probably always want to bypass the cache for all queries other than by primary key, to ensure that you are always querying the entire persistent population.
         This seems a little coarse grained and also reduces the utility of the Coherence cache (unless the bulk of your queries are by primary key).
         Can anybody tell me if my assumption is wrong and if there are any usage strategies the mitigate this aspect?
         Thx,
         Ben

    Hi Rob,     >
         > Why would you need 2 separate caches?
         the first cache would have eviction policy, and caches values, but does not have indexes
         the second would not have eviction, does not store data, but has index updates on changes.
         This way you have a fully indexed but not stored data-set, similarly to the difference between stored and indexed attributes Lucene.
         > Why not just
         > maintain a index within each cache so that every
         > entry causes the index to get updated inline (i.e.
         > synchronously within the call putting the data into
         > the cache)?
         >
         You cannot manually maintain an index, because that is not a configurable extension point (it is not documented how an index should be updated manually). You have to rely on Coherence to do it for you upon changes to entries in the owned partitions.
         And since Coherence code does remove index references to evicted or removed data, therefore the index would not know about the non-cached data.
         Or did I misunderstood on how you imagine the indexes to be maintained? Did you envision an index separate from what Coherence has?
         > (You may have to change Coherence to do this.....)
         Changing Coherence was exactly what I was trying to avoid. I tried to come up with things within the specified extension points, and the allowed things, although it might be possible that I still did not manage to remain within the allowed set of operations.
         Of course, if changing Coherence is allowed, allowing an option of filtering index changes to non-eviction events is probably the optimal solution.
         > And I don't think that the write-behind issue would
         > be a problem, as the current state cache of the cache
         > (and it's corresponding index) reflects the future
         > state of the backing store (which accordingly to
         > Coherence's resilience guarantee will definitely
         > occur).
         >
         The index on the second cache in the write-behind scenario would be out-of-synch only if the second cache is updated by invocations to the cache-store of the first cache. If it is updated upon changes to the backing map, then it won't. Obviously if you don't have 2 caches, but only one, it cannot be out-of-synch.
         > So you would have a situation where cache evictions
         > occur regularly but the index just overflows to disk
         > in such a fashion that relevant portions of it can be
         > recalled in an intelligent fashion, leveraging some
         > locality of reference for example.
         >
         I don't really see, how this could be done. AFAIK, all the indexes Coherence has are maintained in memory and does not overflow to disk, but I may be wrong on this, but again, I may have misunderstood what you refer on index handling.
         > a) you leverage locality of reference by using as
         > much keyed data access as possible
         > b) have Coherence do the through-reading
         > c) use database DAO for range querying
         > d) if you were to use Hibernate for (c), you might be
         > able to double dip by using Coherence as an L2 cache.
         > (I don't know if this unecessarily duplicates cached
         > data....)
         >
         > Any thoughts on this?
         a: if you know ids on your own, then this is the optimal solution, provided cache hit rates can be driven high enough. if you have to query for ids, the latency might be too high.
         b: read-through can become suboptimal, since AFAIK, currently the cache store reads all rows one by one, only read-ahead uses loadAll, but I may be wrong on this. Loading from database can be optimized for multiple id loading as well, to be faster than the same via cache store. So it is very important that the cache hit rate be very high for performance-relevant data in case of read-through.
         c: use database dao for complex querying, possibly almost anything more complex than straight top-down queries. make performance tests for both solutions, try to rely on partition affinity, and try to come up with data structures that help with making indexes which can be queried with as few queries as possible, and with not too high index access count.
         d: you cannot query by Coherence on Hibernate second-level cache, as Hibernate second-level caches do not contain structured data, but contain byte[][]s or byte[], holding the column values serialized to it (separately or the same byte[], I don't remember which).
         Best regards,
         Robert

  • No detail level  objects in Voyager report

    Hi All,
    I created a report in Voyager workspace....Voyager does not show Detail level objects from the selection panel in the left and shows only Dimensions, and i want to use deatil level objects in reports.
    I also see no formula tool bar. I am using an SAP BW query.
    Please help. we are on BO xi 3.1 sp2, FP2.5
    Edited by: pali1010 on Apr 26, 2010 10:47 PM
    Edited by: pali1010 on Apr 26, 2010 10:48 PM

    Display attributes have several string properties which are visible in Voyager:
    u2022 Short description
    u2022 Medium description
    u2022 Long description
    u2022 Key and short description
    u2022 Key and medium description
    u2022 Key and long description
    These string properties can be viewed and changed from the Properties tab of the Tab panel.
    Key and text properties are concatenated into a single string, and are therefore displayed in a single column or row caption field. They do not appear as separate columns or rows.
    Now come to your question where youwant to se detail level.
    Can you expand the dimension by clicking on the parent member?
    It should expand.
    Bashir Awan

  • Input ready query is not showing loaded data in the cube

    Dear Experts,
    In Input ready query we have problem that it is not showing the values which was not entered by throught hat query. Is any settings in input ready query that we can do to populate the loaded data on the cube as well as data entered through Input ready query itself.
    Thanks,
    Gopi R

    Hi,
    input ready queries always should display most recent data (i.e. all green and the yellow request). So you can check the status of the requests in the real-time InfoCube. There should exist only green requests and maybe at most one yellow request.
    In addition you can try to delete the OLAP cache for the plan buffer query: Use RSRCACHE to do this. The technical names of the plan buffer query can be found as follows:
    1. InfoCube\!!1InfoCube, e.g. ZTSC0T003/!!1ZTSC0T003 if ZTSC0T003 is the technical name of the InfoCube
    2. MPRO\!!1MPRO, e.g. ZTSC0M002/!!1ZTSC0M002 if ZTSC0M002 is the technical name of the multiprovider
    If the input ready query is defined on an aggregation level using a real-time InfoCube, the first case is relevant; if the aggregation level is defined on a multiprovider the second case is relevant. If the input-ready query is defined on a multiprovider containing aggregation levels again the first case is relevant (find the real-time InfoCubes used in the aggregation level).
    Regards,
    Gregor

  • Text data is not displaying at Report level for an Infoobject Item name

    Hi All,
    Texts data is maintained at Info object level but text information is not displaying at Report level for an object.  I have set the property at query level as "Text", but at report level it is displaying key values of the object, not displaying texts information.  This problem occured in QA(Quality) system but not replicated in Production system.  I have checked all properties at backend and query level and compared the same between QA and Production system didn't find any difference.
    Please suggest what might be the cause: at report level it is displaying key values instead of the textual information even text data is mainintained at backend.  In production it is displaying textal data for the IO.
    Thanks in advance.

    Hi All,
    Thanks for your valuable information.  I have changed the setting under Advanced tab as "Master data" for Access type for results values, but didn't solved the problem still it showing key information for the IO.
    In Production system the report displaying texts information for the IO, but in QA its displaying key information. The problem exits only in QA, its working fine in DEv and Production.
    As per my understanding might be the issue with Patches will be checking with Basis team on the same.
    Please suggest if you have any alternatives to fix the issue.

Maybe you are looking for