Download Interactive Report data does not match on-screen data

Hi,
I have a page which has a process which runs a query and returns a random % of rows from a table, and saves this into a collection.
Then, based on the random rows, I want to select data from another table, where the rows match certain columns in the collection.
I have created a query to select the data I want, and have it displayed in an Interactive Report region.
The data on the screen is fine, but when I go to 'Dowload' the data to csv or excel from the Interactive report, the data is different.
I have also created a normal report region with the same query, and when I download it from here, it works fine, so I know it has nothing to do with the query.
Any ideas?
Amanda.
The process details are:
Process Point: On Load - Before Regions
Run Process: Once per page visit
Then in the Interactive report query is something like:
Select ..
from
DATA_TABLE c
select
c001 User_Id,
c002 Account_No,
c003 Transaction_Date
from htmldb_collections where collection_name='PX_COLLECTION'
*) d*
where
c.user_id = d.user_id AND
c.account_no = d.account_no AND
c.transaction_date = d.transaction_date

Hi,
This is because of web caching. You should invalidate the page caching in the access tab of page properties or you just change the page url with a dummy paramater like this
http://domain/portal/page?_pageid=6,1,6_11&_dad=<dad>&_schema=<schema>&1=1
Here &1=1 is a dummy parameter
This will invalidate the cache and the page will be refreshed.
Thanks,
Sharmila

Similar Messages

  • Job Cancelled with an error "Data does not match the job def: Job terminat"

    Dear Friends,
    The following job is with respect to an inbound interface that transfers data into SAP.
    The file mist.txt is picked from the /FI/in directory of the application server and is moved to the /FI/work directory of application server for processing. Once the program ends up without any error, the file is moved to /FI/archive directory.
    The below are the steps listed in job log, no spool is generated for this job and it ended up with an error "Data does not match the job definition; job terminated".Please see below for more info.
    1.Job   Started                                                                               
    2.Step 001 started (program Y_SAP_FI_POST, variant MIST, user ID K364646)                           
    3.File mist.txt copied from /data/sap/ARD/interface/FI/in/ to /data/sap/ARD/interface/FI/work/.
    4.File mist.txt deleted from /data/sap/ARD/interface/FI/in/.                                   
    5.File mist.txt read from /data/sap/ARD/interface/FI/work/.                                    
    6.PD-DKLY-Y_SAP_FI_POST: This job was started periodically or directly from SM36/SM37 (Message Class: BD, Message Number : 076)  
    7.Job PD-DKLY-Y_SAP_FI_POST: Data does not match the job definition; job terminated (Message Class : BD, Message No. 078)    
    8.Job cancelled after system exception
    ERROR_MESSAGE                                                
    Could you please analyse and come up about under what circumstance the above error is reported.
    As well I heard that because of the customization issues in T.Code BMV0, the above error has raised. 
    Also please note that we can define as well schedule jobs from the above transaction and the corresponding data is stored in the table TBICU
    My Trials
    1. Tested uplaoding an empty file
    2. Tested uploading with wrong data
    3. Tested uploading with improper data that has false file structue
    But failed to simulate the above scenario.
    Clarification Required
    Assume that I have defined a job using BMV0. Is that mandatory to use the same job in SM37/SM36 for scheduling?
    Is the above question valid?
    Edited by: dharmendra gali on Jan 28, 2008 6:06 AM

    Dear Friends,
    _Urgent : Please work on this ASAP _
    The following job is with respect to an inbound interface that transfers data into SAP.
    The file mist.txt is picked from the /FI/in directory of the application server and is moved to the /FI/work directory of application server for processing. Once the program ends up without any error, the file is moved to /FI/archive directory.
    The below are the steps listed in job log, no spool is generated for this job and it ended up with an error "Data does not match the job definition; job terminated".Please see below for more info.
    1.Job Started
    2.Step 001 started (program Y_SAP_FI_POST, variant MIST, user ID K364646)
    3.File mist.txt copied from /data/sap/ARD/interface/FI/in/ to /data/sap/ARD/interface/FI/work/.
    4.File mist.txt deleted from /data/sap/ARD/interface/FI/in/.
    5.File mist.txt read from /data/sap/ARD/interface/FI/work/.
    6.PD-DKLY-Y_SAP_FI_POST: This job was started periodically or directly from SM36/SM37 (Message Class: BD, Message Number : 076)
    7.Job PD-DKLY-Y_SAP_FI_POST: Data does not match the job definition; job terminated (Message Class : BD, Message No. 078)
    8.Job cancelled after system exception
    ERROR_MESSAGE
    Could you please analyse and come up about under what circumstance the above error is reported.
    As well I heard that because of the customization issues in T.Code BMV0, the above error has raised.
    Also please note that we can define as well schedule jobs from the above transaction and the corresponding data is stored in the table TBICU
    My Trials
    1. Tested uplaoding an empty file
    2. Tested uploading with wrong data
    3. Tested uploading with improper data that has false file structue
    But failed to simulate the above scenario.
    Clarification Required
    Assume that I have defined a job using BMV0. Is that mandatory to use the same job in SM37/SM36 for scheduling?
    Is the above question valid?

  • Regarding "Data does not match the job definition; job terminated"

    Dear Friends,
    The following job is with respect to an inbound interface that transfers data into SAP.
    The file mist.txt is picked from the /FI/in directory of the application server and is moved to the /FI/work directory of application server for processing. Once the program ends up without any error, the file is moved to /FI/archive directory.
    The below are the steps listed in job log, no spool is generated for this job and it ended up with an error "Data does not match the job definition; job terminated".Please see below for more info.
    1.Job Started
    2.Step 001 started (program Y_SAP_FI_POST, variant MIST, user ID K364646)
    3.File mist.txt copied from /data/sap/ARD/interface/FI/in/ to /data/sap/ARD/interface/FI/work/.
    4.File mist.txt deleted from /data/sap/ARD/interface/FI/in/.
    5.File mist.txt read from /data/sap/ARD/interface/FI/work/.
    6.PD-DKLY-Y_SAP_FI_POST: This job was started periodically or directly from SM36/SM37 (Message Class: BD, Message Number : 076)
    7.Job PD-DKLY-Y_SAP_FI_POST: Data does not match the job definition; job terminated (Message Class : BD, Message No. 078)
    8.Job cancelled after system exception
    ERROR_MESSAGE
    Could you please analyse and come up about under what circumstance the above error is reported.
    As well I heard that because of the customization issues in T.Code BMV0, the above error has raised.
    Also please note that we can define as well schedule jobs from the above transaction and the corresponding data is stored in the table TBICU
    My Trials
    1. Tested uplaoding an empty file
    2. Tested uploading with wrong data
    3. Tested uploading with improper data that has false file structue
    But failed to simulate the above scenario.
    Clarification Required
    Assume that I have defined a job using BMV0. Is that mandatory to use the same job in SM37/SM36 for scheduling?
    Is the above question valid?

    Hi dharmendra
    Good Day
    How are you
       I am facing the same problem which you have posted
    By any chance have you got the soultion for this
    If so please let me know the solution for this.
    Thanks in advance.
    Cheers
    Vallabhaneni

  • Background Job cancelling with error Data does not match the job definition

    Dear Team,
    Background Job is getting cancelled when I run a Job on periodically but the same Job is executing perfectly when I run it manually(repeat scheduling) .
    Let me describe the problem clearly.
    We have a program which picks up files from an FTP server and posts the documents into SAP. We are scheduling this program as a background Job daily. This Job is running perfectly if the files contain no data. But if the file contains data the JOb is getting cancelled with the following messages.
    And also the same Job is getting executed perfectly when repeat scheduling is done ( even for files with data).
    Time     Message text                                                                       Message class Message no. Message type
    03:46:08 Job PREPAID_OCT_APPS2_11: Data does not match the job definition; job terminated        BD           078          E
    03:46:08 Job cancelled after system exception ERROR_MESSAGE                                      00           564          A
    Please help me in resolving this issue.
    Thanks in advance,
    Sai.

    hi,
    If you have any GUI function modules used in the program of job
    you cannot run it in background mode.

  • Stock transport order requiremnet date does not match in tranRRP3) &MD04

    STO delivery date does not match in transaction MD04 (ECC 6.0) and RRP3 (SCM5.1). This situation occurs when delivery date in STO is past due.  During Shipment Scheduling & avaibilty check, materail staging/availability date(EKET-MBDAT) get updated with current date. This EKET -MBDAT date is reflecting in the stock requirement list as a requirement date which is correct. Still APO product view screen dispalys the past due date as a requiremnet date instead of date in MBDAT.
    Please let me know in case of any more clarification on this.

    Dear Sanjay,
    STO Delivery Date is receipt date at destination location. This minus Transportation Lead time is STO requirement date at source.
    When you conduct ATP check for this PO/STO, your committed date is updated at source and delivery date will be updated as committed date plus TL lead time.
    Do check your GR/GI times in Product Master as primary date mismatch issue.
    If you are using any Customized program for updating EKET-MBDAT, it wont automatically trigger the change date in APO.
    In such cases you need to run CCR in APO and reconcile.
    The better way is to periodically schedule IM initialization (RIMODINI). This maps all R3/ECC data in your IM for PO/STOs and you would see the consistency in APO and ECC.
    Regards,
    Bipin

  • ORA-23327: imported deferred rpc data does not match GLOBAL NAME of ..

    Hi,
    We took an export of a production database in Oracle 9i (9.2.0.7) and created a new database on a test server with Oracle 10g (10.1.0.5)
    The original production database was called DC20 and our new test database is called SD20.
    Looking at the logs after the import we're seeing this error:
    "ORA-23327: imported deferred rpc data does not match GLOBAL NAME of importing db"
    We've looked into it a little and it seems to make sense because the names of the databases are different, and from what we've read, we really didn't need to do a full import including SYSTEM anyway. We also came across a solution telling us:
    "If you are not using Replication on the target database, these errors can be ignored. However, you should subsequently remove all the replication objects by running:
    $ORACLE_HOME/rdbms/admin/catrepr.sql, then catproc.sql"
    What I was wondering was if someone could validate that running these scripts and ignoring the errors will be ok, or if they'll cause us issues further down the line.
    In case its needed, heres my Export parameter file
    "FULL=Y
    DIRECT=Y
    LOG=E:\Exports\DC20\Log\expDC20.LOG
    FILE=E:\Exports\DC20\Data\DC20.DMP"
    and heres my Import parameter file
    "FULL=Y
    COMMIT=Y
    GRANTS=N
    IGNORE=Y
    BUFFER=500000
    FILE=E:\Imports\SD20\data\SD20.dmp
    LOG=E:\imports\SD20\log\impSD20.LOG"
    Thanks for your help!

    Thanks Neil, we did already see this note and added it to our list of possible solutions. After we found this one, we also found the one I quoted, which would seem to be the path of least resistance (it would save a good 10 hours in import time if nothing else).
    However, if we can't do what I'm asking about in my original post however, the note you mentioned will be our next route..

  • Inventory in warehouse report value does not match balance sheet account

    Dear experts,
    When I run the inventory in warehouse report, the value (using last purchase price) does not match the inventory finished goods account. Can you please advise why?
    Thank you for any advice.
    Jane

    Dear Jane,
    The only report for inventory to link to financial statement would be Inventory Audit Report.
    All other reports are just for your convenience. They are not financial based.
    Thanks,
    Gordon

  • Query data does not match with Multiprovider

    hello,
    I am as confused as my subject.
    I executed the query from Multiprovider, which getting the data from sales and inventory cube.
    Everything works fine until I try to validate some of the data.
    I can validate some early records but I cannot see later ones in any of the cubes or multiprovider.
    I deleted the data and loaded all over again but didnt work.
    I am trying to display the record from Plant, material and calday ... i do see the date and plant # but not material #.
    I hope I am making sense to you guys.
    Any input would be appreciated.
    Khalid

    Hello Kalyan,
    Material is part of both cubes (inventory and sales) and I did flag it in multiprovider when selecting "display data" > dimensions.
    I am not sure if it makes difference ... there is another cube (store planning)  attached to same multiprovider ... which not feed data but as you know does not report on material level.
    Do you think it could create the problem. Again (store planning) cube is not feeding data to multiprovider at the moment.
    Thank You ... any other suggestion.
    Khalid

  • Interactive Report - search does not work when using custom authentication

    Apex 3.2.x
    I can authenticate fine with my custom authentication and all of my pages work okay except for one page that uses the Interactive Report feature. When I click 'Filter' then enter the column name, operation (contains, =, like, etc.) and the expression, then click the 'Apply' button, the page just re-displays and my filter information is missing?
    If I first login to Apex, select and run my application, the Interactive Report features work just fine. What's missing?

    More information:
    After login into my Apex workspace (development environment), when I display the Interactive Report and click debug I see this debug message:
    "using existing session report settings"
    When I login using my application's custom authentication and click debug I see this debug message:
    "creating session report settings as copy of public saved report"
    Based on this, it appears that my session info in not set correctly when using custom authentication... but I'm not sure what needs to be set.
    Edited by: user9108091 on Oct 22, 2010 6:44 AM

  • Event date does not agree with photo dates in the event

    Mavericks 
    Iphoto 9.5.1
    I have made a new event and put lots of genealogy pictures for one family member in it.  I changed all the photos to Jan 4, 2002.  These are docs, pics, obits, etc. Thus I have given this one person this date.  Other family names  have another date.  I modified all the original files within each event.
    Beside this couples name on the "Event" label it says Jan 3, 2002 - August 23, 2014.  Therefore it shows itself in the Jan 3, 2002 group of people.  That is about 20 family groups up the page.   
    I have done the following:
         --Made sure that there are no pictures with the date Jan 3, 2002 in the event
         --Have taken all the pictures in the event and separated them out into new events--census, certificates, pictures, obits, etc.  Thought I would trick the event to restarting! 
         --Did a "select-all"  of that event and created a new event--same thing.
         --Closed down iPhoto and started over again.
    I have no idea why the Jan 3, 2002 is showing up.  These are one side of the family.  The January 4 is the other side!
    On August 23, 2014, I did take a couple of the cemetery pictures in this event.  However, they were given the new date of January 4 and Modified the original files using the checkbox.  I put the same picture in another family member's event and it is OK.  I
    Hope I have given enough details.

    I think I have solved it!  After much frustration and then writing for help, I was not to be stopped.  I kept looking up other helps and one person suggested that one of the pictures might be a problem.  So I tried that approach.  I made an event just out of the "possible" problem pix (the August  2014 date.)  It was not causing a problem in another event, however.   From that  event with the same pix, I exported the pix to the desktop, and then imported the pix back into iPhoto.  Then I put them into the event that was having the problem.  So far so good.  The date of the event matches all the dates of the photos.  I will now delete the "problem children."
    Maybe this will help someone else. 

  • Database Export exporting data does not export time of date columns

    Hi,
    I'm trying to export some data using Database Export and for the date columns the resulting scripts truncates dates showing only the date part not the time one. If export from a query result the resulting script include all the information (date and time).
    Does anyone has an idea on how to export all the date information using database export.
    thanks

    Yep, as Jim said, under Tools / Preferences / Database / NLS Parameters I changed Date Format to DD-MON-YYYY HH24:MI:SS
    Seems to have done the trick. What's the significance of DD-MON-RR versus DD-MON-YY?

  • Outline Agreement (Contract) data does not match with EKPO table.

    Hi,
    I have created a outline agreement(Contract) with net price as 100 / 1 Kg. But when I go to EKPO table in front of Net Price it is reflecting net price as 1 / 1 Kg. So it shows that it is dividing the net price by 100 in Table EKPO. This is happening with the currencies which are maintained in T Code OY04 (Currencies like JPY, CLP, ESP, ITL) with decimals as 0.
    Rest all curriencies which have not maintained in OY04, there is no problem in EKPO as the data matches with Outline Agreement Net Price.
    Request you to please provide the input to resolve the same.
    Thanks & Regards,
    Neeraj Singh

    Hi Neeraj Singh,
    The value of data stored in EKPO depends on the decimal places defined
    for the currency. In standard the decimal places is taken as 2 . Therefore the value will be
    the same as order price if the currency is defined for two decimal
    places ( default is 2 ).
    However, if the decimal places for JPY is defined as zero the table
    will show a value of price /100. If you choose a currency with
    3 decimal places the value in the table will be price * 10 and so on.
    Example:
    Currency                          Decimal places      Order price      Value in EKPO
    SGD                                               2               2000               2000
    NLG                                               2               2000               2000
    JPY                                               0               2000               20
    KWD (Kuwati Dinar)                     3               2000               20000
    ITL                                                 0               2000               20
    All currencies that do not have any decimal places at all
    (TCURX-CURRDEC = 0) must have EKPO-NETPR multiplied by 100.
    Please check SAP notes which may be of help to understand this issue 53206 and 137626
    The issue you reported in not an error but standard.  I hope this information is of help.
    Regards,
    Lorraine
    Edited by: Lorraine Donnelly on Jul 12, 2010 6:57 PM

  • Monitoring data does not match

    Hallo! Has the following problem. Processor of a virtual machine loaded 100% but in dashboard azure  peak loading of this VM 6-10% What does it mean? Why monitoring displaying incorrect data?

    Hi,
    Thanks for your reply.
    Since it is an issue related to Azure management portal, I recommend you to ask in the forum below for assistance if the issue persists:
    https://social.msdn.microsoft.com/Forums/en-US/home?forum=windowsazuremanagement
    Best regards,
    Susie
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact [email protected]

  • The version of this vi downloaded on the target does not match the version you have in the project

    I am using Labview 8.6 and field point cFP-2220
    I have a project that have some VIs running on the development computer and some VIs running on the field point (one top level VI and a few sub Vis).
    When I run the program every thing is working, I disconnect from the field point, still every things is running; however when I reconnect I get the error above.
    For FP Write(float-IO).vi and FP Write(Boolean-IO)
    error message attached. Any help will be greatly appreciated.
    Thanks.
    Attachments:
    error.JPG ‏42 KB

    I am having this same issue, but am running LV2010. The conflict, for me, is always with NI_LVConfig.lvlib
    If I hit the Apply button, the code deploys and runs on my cFP-2120 without issue. I disconnect from the target in Project Explorer, reconnect without any code changes and I get the same Conflict Resolution popup.
    The code I ran on my RT target was a Get Date/Time in Seconds writing to an indicator - that's it.

  • Sort data does not match song info

    I've just discovered that some of my songs have sort data from different files. Short of going though all my music song by song is there a way to clear all sort data?

    Generally setting a common Album Artist will fix split albums.
    For deeper problems see Grouping tracks into albums.
    tt2

Maybe you are looking for