BIA Server problem - Data package not consistent

Dear All,
I have a strange issue with my BIA server. When I run the report by selecting top most nodes of the hierarchy.
Report ends up with message: No applicable data, and below error messages
Aggregates of ZFICSC22 were compressed. Data package is not consistent
Error while reading data, navigation is possible
The report works fine when I select the subsequent nodes (immediate lower nodes) of a hierarchy.
The report also works when I execute bypassing the BIA
I have cleaned up all the hierarchy buffer but the problem persists
As of now I am left out with last option which is rebuilding the indexes unless any one of you comes up with a better solution.
Thanks & Regards
Ziya
Edited by: Ziyauddin Mohammed on Dec 5, 2008 4:02 PM

Hi,
Have you implemented OSS Note 1163991?
There is nothing like data package compression in BIA.
Regards,
-Vitaliy
Edited by: Vitaliy on Dec 5, 2008 9:08 AM

Similar Messages

  • Data Packages not hitting server

    Hi,
    I have an issue where the users log in remotely and submit data packages. When they run the package they get a successfully submitted message, but when they have a look on the View Status box the package never arrived.
    I have to admit this only happens when they run the packages for a lot of entities, but this should not be happening. Have anyone encountered this problem before?
    Regards,
    Andries

    I have faced this same problem, now running on BPC 5.1 SP8 (on SQL 2005) but first starting back on 5.1 SP3 or thereabouts. Support notes / release notes claims it was fixed sometime around SP4, but I am still facing it on SP8.
    It's sporadic, and I find it happens on both a wireless and ethernet LAN connection from the client. I even find it happens when I'm running Excel on the server, connecting via an RDP connection.
    I've filed a message with SAP support, but haven't yet had the time to reproduce it in ApShell (which is difficult to do, since it's sporadic.)
    I've set up data manager debugging logs on both client and server, and find that they are of very little use, since when this problem occurs, exactly nothing is logged to either log.
    Restarting the COM apps on the server always seems to resolve the problem; afterward the packages always seem to run on the first attempt.
    I've also noticed that while one user cannot run a package (even after 5 or 10 or 20 attempts), other users can run packages on their first attempt. Then an hour later when the admin bounces the COM apps, the first user can run their package fine.
    The problem seems to occur regardless of the type of package. Most packages are simple script logic, calling a custom logic file (normally LGF, not LGX) but some are calling the standard copy package, or custom DTS packages that run a stored procedure directly against the database.

  • BIA Server problem

    Hi All,
    In the monitor screen,under the details tab  i  see 'BIA server is overloaded '  but the requests are rolled up and available for reporting.
    will it have an impact on the data in the reporting??
    Thanks,
    V.Singh.

    Hi,
    You have an elevated risk according to BWA best practices. You are receiving this alert because you are using more than 50% of the allocated memory in a given check point ( even if you pay 100% license you are supposed to leave 50% of the memory idle for temp indexes, query joins etc ) . This warning usually comes with many "number of memory unloads in the past 24 hours etc" warnings. But there is practically no impact to your reporting - this claim can be challenged by theoreticians.
    You either take a risk and just keep monitoring it so that it won't get worse, or remove a cube or two from BWA. But definitely consider planning "logically partition" your cubes . This will allow you to control the growth of BWA. Say you have a revenue cube; every physical/fiscal year copy the cube to a new cube and start loading to new cube and add the new cube under multiproviders. The next year you remove the oldest cube from BWA and introduce the new one..
    Besides this, look at the standalone Python tool to see the peek times. You may be just getting this warning in certain time window e.g while loading data to BWA and running information broadcasting queries at the same time..Or you may be running APD queries that used to cause a lot of issues - i am not sure if it's still the case.
    Cheers
    Tansu

  • File to RFC problem- data is not inserting into ztable in R3 system

    Hi
    i have done File to RFC scenario which picks data from flat file and inserts into ztable via RFC in R3. but while testing my scenario everything is successful in XI monitoring (Successful flag in MONI and RWB) and in auditlog message status is DLVD.
    it seems to be everything successful in XI and RFC call also successful in R3 system.
    but for some reason data is not inserting into table (RFC is used to insert data into ztable)
    Is there any way to debug RFC call excecution in XI..?
    RFC code is like this:
    insert ZMM_AUTO_GR from INPUT_TABLE.
    commit work.
    END FUNCTION.
    please advice what could be the reason not inserting into table.
    Help would be appreciated.
    Regards,
    Rajesh

    Hi Praveen,
    please see audit log- from communication channel monitoring..
    Receiver channel 'CC_INCA_RFC_SAPECC_Receiver' for party '', service 'R3DCLNT210' (internal name 'RfcClient[CC_INCA_RFC_SAPECC_Receiver]')
    Client data: {jco.client.lang=EN, jco.client.snc_mode=0, jco.client.client=210, jco.client.passwd=******, jco.webas.ignore_jdsr_error=1, jco.client.user=jsaha, jco.client.sysnr=00, jco.client.ashost=ausr3devdc02}
    Repository data: {jco.client.lang=EN, jco.client.snc_mode=0, jco.client.client=210, jco.client.passwd=******, jco.webas.ignore_jdsr_error=1, jco.client.user=jsaha, jco.client.sysnr=00, jco.client.ashost=ausr3devdc02}
    Current pool size: 1, maximum pool size : 1
    Channel History
    - OK: 2008-07-28 04:32:04 PDT: Message processed for interface ZAUTO_GR_STAGE_INCA
    - OK: 2008-07-28 04:31:04 PDT: Message processed for interface ZAUTO_GR_STAGE_INCA
    - OK: 2008-07-28 03:56:56 PDT: Message processed for interface ZAUTO_GR_STAGE_INCA
    - OK: 2008-07-28 03:49:04 PDT: Message processed for interface ZAUTO_GR_STAGE_INCA
    - OK: 2008-07-28 03:48:04 PDT: Message processed for interface ZAUTO_GR_STAGE_INCA

  • Bridge CS4 Server problem- slide show not loading

    Hi
    I have created a slide show in CS4 and emdedded the index code in my gallery page, created in Dreamweaver CS4. I have used an iframe within a DIV tag.
    Preview works fine in both Chrome and IE. However, once uploaded to Server, all that can be seen is a grey screen.
    I have contacted the server Heart Internet to see if they can help but as yet no idea. 
    I was just wondering if anyone has come across this before and if so, is there a quick fix I can implement.
    I am basic but ok with HTML, and can insert code, if given it, but no good at the problem solving side of code writing.
    Any ideas would be really appreciated
    Thanks

    It sounds as if for some bizarre reason, the transition is causing the player buffer to empty out.
    This happens on certain - usually older - players when sequential cells are set as non sequential in compiling.
    (I had this one bite me - hard - recently, although from a different application).
    Best way to check this is to go get a copy of the excellent PGCEdit (shareware) and load the compiled disc.
    http://download.videohelp.com/r0lZ/pgcedit/index.html
    Find the main timeline for the slideshow, and double click it. (see attached image)
    The "chapters" will be listed, and we need to see these ticked as sequential playback, or seamless.
    If there is no tick, add one between each track, save the disc & reburn with IMGBurn (freeware/Donationware)
    http://www.imgburn.com/
    For some reason, certain chipsets seem to create a gap when transitioning a chapter marker on a timeline with a single sequential audio stream when the chapter is incorrectly set to non seamless.
    This may not be what is happening here, but I think it is worth a look.

  • Out-of-date package not been updated in almost a year

    The question I have is: what should I do when I see a package that has not been updated in almost a year (after its flag-out-of-date)? I can't seem to find any info on how such a situation is handled. I can't reflag and sending an email to the maintainer seems intrusive and/or rude. There may be an underlying reason to not upgrade, or the flag was lost in time and forgotten.
    The package I have in mind is opencl-headers. It got stuck on v1.1, and never updated to v1.2, and now still not to v2.0. There is an AUR package for the 1.2 headers, but that just seems silly to have around.
    Thanks for any input!

    I see, there is a backstory to all this. It seems though momentum was lost when ocl-icd devs did not respond to the email requesting info. The last commit to their repo was for 2.0 support, back in 2013. It seems abandoned since then, maybe a new spark of activity will come when a new OpenCL spec comes out.
    The current situation is quite terrible really: libcl in the repos is from the NVIDIA package, so this means a hard dependency on a vendor not willing to follow the latest version of the spec. Way back when it was probably a sane choice, but now, as the email linnked by Allan shows, the situation is different.
    Either choice, be it ocl-icd or Khronos, will improve the situation (OpenCL version and license-wise). Currently ocl-icd has one advantage (detailed in the POCL developer docs): choosing the icd at runtime.
    As any CUDA or AMDAPP or Intel OpenCL package will include their libOpenCL, and it can be selected by LD_LIBRARY_PATH if required, I hope a sane libOpenCL package will soon make it into the repo.
    As for backward compatibility: I believe the headers are backwards compatible, and it is up to the applications to verify the device they want to use supports the functionality they use. As for forward compatibility: if one solution fails and an alternative is better: just replace the package providing the ICD library by a better one.
    Note that both the OpenCL headers and the ICD (libOpenCL) are designed to be platform-agnostic and should in principle work with all versions of the spec and devices, limited only by the device capabilities as reported through OpenCL functions. Currently, libcl and the old opencl-headers version is the limiting factor, which is Terrible®, given the (open source) alternatives.

  • Data Could not be retrieved from database...

    Hi All,
    I have a problem on Excel 2007 while refreshing (for pivot table) data from SQL Server view:
    "Data could not be retrieved from database. Check the database server or contact your database administrator. Make sure that external database is availiable, and then try the operation again"
    Excel file is distributed on about 20 workstations in network with the same Excel version, and problem occurs only on one station. I have checked credentials and connection between workstation and server and it seems to be ok.
    Any ideas?

    it is problem, because sql server is not member of domain... so I'm using sql aunthentication.
    strange thing is that from any other workstations (aldo in domain) it works fine when providing SQL user. I'm supposing that it is problem with some excel settings

  • Data Package is not consistent error

    Hi All,
    We have a Bex issue the screenshot of which is given in the link below. We have an info cube which has huge data volume. We load it very frequently (every hour from 2 sources). This info cube has been loading from say last 5 years. We stopped compressing the data about an year ago for some reason. We are thinking of compressing the data sometime soon. All of a sudden, users are getting the error shown in the link below more frequently. No much information is available at Service Market place about this issue. Can somebody please throw some light on this?
    [Error Screenshot|https://photos-1.dropbox.com/i/l/orsd6TKDlIBVoH4JO79GWQAPNFPlTa88QJ0zkV7Lcc0/53945346/1324418400/3799537/Bex%20Error.JPG#1]
    Thanks and Regards
    Subray Hegde

    Hi Deepak,
    I am not able to attach a screenshot. Somebody directed me to servimg.com but, I do not know how to create an account in this site. The error reads like this:
    "Error Aggregates of <Infocube> were compressed. Data package is not consistent. Error. Error while reading data; navigation possible.".
    Thanks and Regards
    Subray Hegde

  • Problems with Sun One Web Server 6.1 javax.xml.xpath package not found

    I used myeclipse to build an xml app and tested on jboss. it worked perfectly. However when i deployed it to our solaris sun one web server the app fell apart completely with the following error. I m unable to figure out what went wrong. Any help will be much appreciated.
    Thanks.
    [11/Dec/2007:22:12:37] failure (13539):      for host 121.247.233.169 trying to GET /feeds/rss.jsp, service-j2ee reports: StandardWrapperValve[jsp]: WEB2792: Servlet.service() for servlet jsp threw exception
         org.apache.jasper.JasperException: WEB4000: Unable to compile class for JSP
         /opt/SUNWwbsvr/test/ClassCache/test/_jsps/_feeds/_rss_jsp.java:8: package javax.xml.xpath does not exist
         import javax.xml.xpath.*;
         ^

    Thanks for the response. I tried to use xalan package which resolved the javax.xml.xpath package not found error (xalan.jar in WEB-INF/lib folder). However I m now getting the following error. Probably incompatible version is the reason. Please advise!
    [11/Dec/2007:23:46:28] failure (17028):      for host 121.247.233.169 trying to GET /feeds/rss.jsp, service-j2ee reports: StandardWrapperValve[jsp]: WEB2792: Servlet.service() for servlet jsp threw exception
         javax.servlet.ServletException: org.apache.xpath.XPathContext.<init>(Z)V
         at org.apache.jasper.runtime.PageContextImpl.handlePageException(PageContextImpl.java:536)
         at _jsps._feeds._rss_jsp._jspService(_rss_new_jsp.java:627)
         at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:107)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:908)
         at com.iplanet.ias.web.jsp.JspServlet$JspServletWrapper.service(JspServlet.java:687)
         at com.iplanet.ias.web.jsp.JspServlet.serviceJspFile(JspServlet.java:459)
         at com.iplanet.ias.web.jsp.JspServlet.service(JspServlet.java:375)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:908)
         at org.apache.catalina.core.StandardWrapperValve.invokeServletService(StandardWrapperValve.java:771)
         at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:322)
         at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
         at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:212)
         at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
         at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:209)
         at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
         at com.iplanet.ias.web.connector.nsapi.NSAPIProcessor.process(NSAPIProcessor.java:161)
         at com.iplanet.ias.web.WebContainer.service(WebContainer.java:580)
         ----- Root Cause -----
         java.lang.NoSuchMethodError: org.apache.xpath.XPathContext.<init>(Z)V
         at org.apache.xpath.jaxp.XPathImpl.eval(XPathImpl.java:207)
         at org.apache.xpath.jaxp.XPathImpl.evaluate(XPathImpl.java:281)
         at _jsps._feeds._rss_new_jsp._jspService(_rss_jsp.java:165)
         at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:107)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:908)
         at com.iplanet.ias.web.jsp.JspServlet$JspServletWrapper.service(JspServlet.java:687)
         at com.iplanet.ias.web.jsp.JspServlet.serviceJspFile(JspServlet.java:459)
         at com.iplanet.ias.web.jsp.JspServlet.service(JspServlet.java:375)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:908)
         at org.apache.catalina.core.StandardWrapperValve.invokeServletService(StandardWrapperValve.java:771)
         at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:322)
         at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
         at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:212)
         at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
         at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:209)
         at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:509)
         at com.iplanet.ias.web.connector.nsapi.NSAPIProcessor.process(NSAPIProcessor.java:161)
         at com.iplanet.ias.web.WebContainer.service(WebContainer.java:580)

  • Error trying to run SSIS Package via SQL Server Agent: DTExec: Could not set \Package.Variables[User::VarObjectDataSet].Properties[Value] value to System.Object

    Situation:
    SSIS Package designed in SQL Server 2012 - SQL Server Data Tools
    Windows 7 - 64 bit.
    The package (32 bit) extracts data from a SQL Server db to an Excel Output file, via an OLE DB connection.
    It uses 3 package variables:
    *) SQLCommand (String) to specify the SQL Statement to be executed by the package
    Property path: \Package.Variables[User::ExcelOutputFile].Properties[Value]
    Value: f:\Output Data.xls
    *) EXCELOutputFIle (String) to specify path and filename of the Excel output file
    Property path: \Package.Variables[User::SQLCommand].Properties[Value]
    Value: select * from CartOrder
    *) VarObjectDataSet (Object) to hold the data returned by SQL Server)
    Property path: \Package.Variables[User::VarObjectDataSet].Properties[Value]
    Value: System.Object
    It consists out of 2 components:
    *) Execute SQL Task: executes the SQL Statement passed on via a package variable. The resultng rows are stored in the package variable VarObjectDataSet
    *) Script Task: creates the physical output file and iterates VarObjectDataSet to populate the Excel file.
    Outcome and issue:The package runs perfectly fine both in SQL Server Data Tools itself and in DTEXECUI.
    However, whenever I run it via SQL Server Agent (with 32 bit runtime option set), it returns the errror message below.
    This package contains 3 package variables but the error stating that a package variable can not be set, pops up for the VarObjectDataSet only.  This makes me wonder if it is uberhaupt possible to set the value of a package variable
    of type Object.
    Can anybody help me on this please ?
    Message
    Executed as user: NT Service\SQLSERVERAGENT. Microsoft (R) SQL Server Execute Package Utility  Version 11.0.2100.60 for 32-bit  Copyright (C) Microsoft Corporation. All rights reserved.    Started:  6:40:20 PM  DTExec: Could
    not set \Package.Variables[User::VarObjectDataSet].Properties[Value] value to System.Object.  Started:  6:40:20 PM  Finished: 6:40:21 PM  Elapsed:  0.281 seconds.  The package execution failed.  The step failed.
    Thank you very much in advance
    Jurgen

    Hi Visakh,
    thank you for your reply.
    So, judging by your reply, not all package variables used inside a package need to be set a value for when run in DTEXEC ?
    I already tried that but my package ended up in error (something to do with "... invocation ...." and that error is anything but clearly documented. Judging by the error message itself, it looks like it could be just about anything. that is why I asked my
    first question about the object type package variable.
    Now, I will remove it from the 'set values' list and try another go cracking the unclear error-message " ... invocation ...". Does an error message about " ... invocation ..." ring any bells, now that we are talking about it here ?
    Thx in advance
    Jurgen
    Yes exactly
    You need to set values only forthem which needs to be controlled from outside the package
    Any variable which gets its value through expression set inside package or through a query inside execute sql task/script task can be ignored from DTExec
    Ok I've seen the invocation error mostly inside script task. This may be because some error inside script written in script task. If it appeared after you removed the variable then it may because some reference of variable existing within script task.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Creation of data packages due to large amount of datasets leads to problems

    Hi Experts,
    We have build our own generic extractor.
    When data packages (due to large amount of datasets) are created, different problems occur.
    For example:
    Datasets are now doubled and appear twice, one time in package one and a second time in package two. Since those datsets are not identical, information are lost while uploading those datasets to an ODS or Cube.
    What can I do? SAP will not help due to generic datasource.
    Any suggestion?
    BR,
    Thorsten

    Hi All,
    Thanks a million for your help.
    My conclusion from your answers are the following.
    a) Since the ODS is Standard - within transformation no datasets are deleted but aggregated.
    b) Uploading a huge amount of datasets is possible in two ways:
       b1) with selction criteria in InfoPackage and several uploads
       b2) without selction criteria in InfoPackage and therefore an automatic split of datasets in data packages
    c) both ways should have the same result within the ODS
    Ok. Thanks for that.
    So far I have only checked the data within PSA. In PSA number of datasets are not equal for variant b1 and b2.
    Guess this is normal technical behaviour of BI.
    I am fine when results in ODS are the same for b1 and b2.
    Have a nice day.
    BR,
    Thorsten

  • Problems with the O2 blackberry data package on my Curve 3G.

    I have already informed O2 about this but they claim that I should be used the blackberry support services, but nothing there helps me!
    I got my Blackberry Curve 3G on September 9th this year and I put on the Blackberry Data Package bolt-on onto my phone on September 16th. I then received a text to say they've taken £5 from my credit and it will be up and running in the next 24 hours. Its now September 19th and my BBM is not working at all and I am extremely upset with the services and behaviour I have received from both O2 and Blackberry.
    Is there any way you can help? If this fails, I shall be forced to go back to the shop from where I got my Blackberry from and ask for their help.
    Many thanks, Jade.

    Can a bubble whistle Problems with the O2 blackberry data package on my Curve 3G.? The seat matures in your oar. The lad ices the pursuing method inside a resident. A judge spins against the vendor! The rose wows the hello. 
    filipina heart

  • HT4623 I want to purchase a ipad 32gb in the larger size but have read many bad reviews that the wifi does not stay connected and I don't want to have to buy another data package as I already own an iphone that i pay data for, what are my concerns

    I want to purchase an ipad 32gb in the larger size buy have read many bad reviews that they have problems staying connected to wifi.  I don't want to have to buy a data package so I only want the wifi one.  I already pay apple for my iphone data and I can't afford more money.  Why are there so many bad reviews and are the newer ones that much better, according to many reviews they are not.Please help

    I already pay apple for my iphone data
    You pay your carrier for data, not Apple.
    Why are there so many bad reviews
    Where? Regardless, there are always going to be a handful of people having some sort of issue with any given device, but I would not therefore assume that you would be one of them.

  • How to Troubleshoot why data is not moving over into the Data Warehouse after Sql Server Agent Job Run

    Hello,
    Here is my problem:
    Data was imported into the staging area. After resolving some errors and running the job, I got the data to move over to the next area. From there, data should be moving over into the DW.  I have been troubleshooting for hours and cannot reslove this
    issue. I have restarted the sql service services, I have ran a couple packages manually, and the job is running successfully. 
    What are some reasons why data is not getting into the data warehouse? Where should I be looking? 
    Your help is greatly appreciated!!

    Anything is possible.
    So, just to reiterate, running the job manually works, running the scheduled job does not result in errors neither data arriving to the DW, right? And it used to, correct?
    If so, the 1st step would be to examine the configuration(s). But not before you inspect the package. Do you have an ability to export it to a file system and open in BIDS?
    Arthur My Blog

  • Query referencing session state not consistently returning data

    Hi,
    I have a form that summarizes expense data by category for employees, per fiscal year, fiscal period, and project.
    Clicking on a button next to a particular category takes the user to a new page that should display the detail of those expenses, based on expense_code.
    The expense_code is passed via URL to the second page, and the query region for the detail has the following source:
    select exp_code, JRNL_ID_NO, DESCRIPTION, REFERENCE, CUR_PER_TRAN, CUM_YTD
    from fgspec_ac where exp_code between :P118_EXPENSE_CODE_FROM and :P118_EXPENSE_CODE_TO and project=:P114_PROJECT and fis_year=:P114_FIS_YEAR and fis_period=:P114_FIS_PERIOD
    :P118_EXPENSE_CODE_FROM and :P118_EXPENSE_CODE_TO are passed via URL, and the :P114_% fields reference data on the initial form.
    The problem is that when I click on the button to bring up the detail for the first time, the query returns no rows. If I return to the original form and click on the button a second time, the query the displays the required data. This behavior is consistent, regardless of expense category, fiscal period, etc. The first time I try to access detail, no records returned, but the second time, it works.
    I have used the 'SESSION' link to debug and view the session state in each case, and the results are IDENTICAL whether the query returns data or not. That is, all 5 variables contain the exact same data.
    Do you know what I might be doing wrong?
    Thanks,
    Michelle

    Hi guys,
    Thanks very much for your replies. In fact, I have tried both methods. Yes, I am navigating from page 114 to 118, and passing values to page 118 from page 114 (ie: P114_PROJECT passes to P118_PROJECT, etc...). In my query, I have tried referencing both, with the same results.
    When checking values in session state (by using the 'SESSION' link), values from page 114 are available in both cases (when the query returns rows and when it does not). It is very strange behavior.
    I have been developing with HTMLDB for over a year now, and never seen anything like it!!! I will try loading this on HTMLDB.oracle.com, and see if it replicates, and if so, maybe you can have a look...............
    Thanks!
    Michelle

Maybe you are looking for