Problems with queries

hello, i created 2 queries, one is summarized level and the second is detailed. when i pull up both queries i get correct data but when i try to go through "goto" i get nothing for the second queyr. i checked the variable values and those are the exact values that i enter when i call up the query separately. can someone help me please? thanks

i think you dont have the correct assignment in the query... there should not be any reason that this doesnt work It should work.
Please check the rri assignment once again and then check.
Regards,
Den .
Close this thread if its already answered.

Similar Messages

  • Problem with queries while refreshing materialized views

    We are in oracle 8.1.7. We have some materialized views that are refreshed with DMBS_REFRESH.REFRESH in a Job. The problem is that when the refresh occurs while a query is executing we get a "Invalid Rowid" or "0RA-08103 Object not longer exists" in the session executing the query.
    We already have reproduced the problem. Is this a bug ? Any idea?
    Thanks.

    Check if the objects referred in the MV defination has undergone any reorg recently.

  • Problem with Queries Structures in Bex Query Designer and Universe Designer

    Hi everyone,
    I have a problem when i try to do a Universe from a Bex Query Designer.
    When I use a Characteristic Restrictions in a Structure, the universe only show me the structure without the objects inside it.
    EJ:
    1) Bex Query Designer
    Colunms
    (Structure) ZSRT_123
                          - (Characteristic Restrictions) Reg Guayana
    >          (InfoObject)     0Product = 'XXX'
    >          (InfoObject)     0Country = 'YYY'
         - (Characteristic Restrictions) Reg Caracas
    >          (InfoObject)     0Product = 'AAA'
    >          (InfoObject)     0Country = 'BBB'
    Rows
    (Structure) ZSRT_456
                   (Key Figure)     Net Sales
    2) Universe designer
    (Class) ZSRT_456
    >(Measure)      Net Sale
    (Class)     ZSRT_123
    >(Dimension)      ZSRT_123
    The universe dont translate the two Characteristic Restrictions in the structure ZSRT_123.
    Any information would be much appreciated. Thank you.

    Hi,
    I assume that you are using BOE XI 3.1 SP2.
    We have added an option in SP2 to generate or not the L00 level.
    Unfortunately this option applies not only on characteristics but also on hierarchies, hierarchies variables and structures: that's the reason why you have no elements belonging to the structure generated.
    I recommend to always generate L00 in order to have your structure generated correctly.
    This issue will be fixed in SP3.
    Regards,
    Didier

  • Problem with performance of a query having order by, distinct clause

    Hi,
    I have a problem with queries having order by, distinct clause.
    While its executing its taking lot of time. With DBMS_PROFILER identified the queries taking long time.
    The table is having approximately 70 million rows.
    Problem -1
    select * from table_name order by col1;
    select distinct col1,col2 from table_name;
    Here i am having 2 solutions request to let me know whether i am right if not suggest me right solution.
    Solution1:
    Max parallel servers is 8.
    select /* + parallel(table_name,8) */ * from table_name order by col1;
    select /* + parallel(table_name,8) */ distinct col1, col2 from table_name ;
    Solution-2:
    select /* + first_rows */ * from table_name order by col1;
    select /* + first_rows */ distinct col1, col2 from table_name ;
    Problem-2
    I am having a query with where condition on columns.
    Select * from table_name where col1='value1' and col2!='value2';
    Index created on col1 and col2.
    As we no that not equal won't use index as it is a composite index it should use the lead column. but its not using the index.
    Should i forcibly use index with hint or suggest me better solution.
    Any help really appreciated.
    Thanks in advance

    unique wrote:
    The table is having approximately 70 million rows.
    select * from table_name order by col1;Do you really want 70,000,000 rows from your table without any restrictions ? And furthermore ordered output ? I honestly understand the slowness of that query.
    Here i am having 2 solutions request to let me know whether i am right if not suggest me right solution.Who knows if you choosed the right solution. I would suggest to reconsider your query and the need of 70,000,000 returned rows.
    Problem-2
    I am having a query with where condition on columns.
    Select * from table_name where col1='value1' and col2!='value2';Please, provide the explain plan, eventually the tkprof output could also help. And tables descirption AND indexes description.
    And OS and Oracle version.
    Nicolas.

  • ALERT PROBLEM WITH QUERY OUTPUT

    I have a few queries that were originally created using the B1 Query Wizard, that I then attached to Alerts.  Unfortunately when the Alert runs, the query output is different either in layout or records returned (usually less) when compared to running the query directly.  Is this a glitch in B1 as some folks have told me or is there a problem with queries created by the Query Wizard?

    Hi Frank,
    This is a limitation.
    The note 874145 is written about it. It also offers some workarounds.
    I would think that you can also create a 'Select count(*)' query that would also run as an alert that would tell the user how many records are affected. I did not test this, just an idea.
    Hope it helps,
    Jesper

  • Problem with the result of multiple queries in Web Intelligence (BO XI R2)

    We have a problem with Web Intelligence reports. Behavior of the results is like if we selected the function Count in Designer the properties indicators. Same queries separated one for one report give correct results and together 7 queries in one report give correct results only for the first query and for others (6 others) same results for each date (if report is monthly) or the same result for each hour (if the result is daily) for all Queues tested (other which were not tested have result 0). Even the function Sum inserted into the report structure behaves weird u2013 we have the same result as in the filed (it is a copy of that result, not a sum). We tested queries together in SQL Developer and they give correct results, while in BO big query which contains script from Query 3, Query 2  and Query 1 cannot be completely executed (it executes without end).
    We use Business Objects XI R2 SP3.
    Thanks in advance.

    Hi Peter,
    I guess it is the problem due to the merged dimension and different granurality level in the BO queries. From the document properties uncheck the option of auto-merge common dimension objects or manually unmerge all the common dimensions
    than give it a try.
    I guess this would solve your problem. If not can you provide more query details like what objects you have used in the query.

  • Problem with a ColdFusion (Queries spend a long time to return)

    Hy guys
    My team have a problem with a Coldfusion (version 8,0,0,176276). Apparently this issue has a relation with queries that spend a long time to return some result, making the poll to open more connexions with the database until stop the application.
    Actually the high number of connexions is a reflex of this issue, a consequence and the same way that is the fail on the application.
    I would like to know if exist a way to debug the calls to database? this information will help my team a lot.
    Thank you.
    Herbert Dias

    Reader has a different forum http://forums.adobe.com/community/adobe_reader_forums

  • Problem with slash character in CONTAINS queries

    Hi there,
    i've got a problem with a CONTAINS query using Oracle Text.
    The following query works fine:
    select col1 from my_table_text where contains(text,'%02%')>0;
    When i'm trying the query
    select col1 from my_table_text where contains(text,'%02/%')>0;
    the following error occurs:
    ORA-29902: error in executing ODCIIndexStart() routine
    ORA-20000: Oracle Text error:
    DRG-50937: query too complex
    I didn't find anywhere that the slash has a special meaning (escape character or something else). This occurs only when the slash character is positioned immediately after or before the % operator. The search string '%0/2%' works.
    Oracle version: 9.2.0.1.0
    OS: Solaris 8 SPARC
    Maybe someone can help me.
    Thanks in advance,
    Chris

    Hi Chris,
    I think you want to search for all words having '02/'.
    Am I right? In that case try using '02//'
    Yes. '/' means escape character. So when it finds '/%' it tries to escape '%'.
    Hope this helps.
    Regards,
    Anupama

  • Problems with HDV footage from sony HDR - HC9

    I have a new sony HDR-HC9 cam corder. I want to shoot in hi def (HDV) format for best quality, but I produce standard def DVDs as a finished product. I was hoping to use the built in iLink conversion from HDV to DV when downloading to the computer into iMovie. I have had many problems with this.
    Usually, when I set it up to do the downconversion, I see a blue screen in iMovie, although iMovie will ff and rew the camera, no images come across. I experimented with setting the camera in DV record mode at the same time as I had the iLink down conversion set, and this appeared to work. I started downloading an hour tape, and when I came back 65minutes later (after it should have been done), saw the message that iMovie had quit. Reopened the project, and saw an hour worth of footage in the trash, but it was too crazy to reorder the clips from the trash back to the clips pane.
    I resigned to import in hdv (1080i) format and not mess with hardware downconversion. This worked fine. I then added some chapter markers and shared the project with iDVD. I saved the project as an image file and then mounted it to check it for errors. The video was OK, but towards the end of the project, the sound was out of sync by a few seconds with the video.
    I deleted encoded assets from the iDVD project, then reburned (save as image file). Same defect.
    I then started a new iMovie project in standard DV widescreen, and dragged all of the clips from the HDV (1080i) project into the time line. It took over an hour for my mac mini to run the conversion, but the movie in iMovie appeared OK. I redid the iDVD project, and saved as image file. It appears to have worked correctly (although I have not looked at it on my widescreen TV yet to assess image quality.
    Questions:
    1) Has anyone had any luck getting HDR-HC9 iLink HDV-DV hardware conversion to work with iMovie, consistently?
    2) Can anyone comment on image quality issues with converting HDV to DV footage using iMovie software.
    3) Has anyone seen audio sync problems with iDVD from HDV original footage.
    Thanks

    Hi - you may get a more exact answer in the BlackMagic Macintosh forum, since there aren't a lot of BlackMagic queries on the Apple Discussions forums:
    http://forum.blackmagicdesign.com/viewforum.php?f=11

  • Problem with Latches and Parallelising PL/SQL Cursors

    In 9i i have process that do the following:
    PROCEDURE LoadClients IS
      CURSOR cChanges IS
        SELECT ...
        WHERE MOD(PK,pnTotalJobs>) = pnThisJob To call this I do the following
      nTOTALJOBS CONSTANT := 2;
    BEGIN
      FOR i IN 1..nTOTALJOBS LOOP
        DBMS_JOB.SUBMIT(LoadClient(pnTotalJobs=>nTOTALJOBS, pnThisJob=>i-1));
      END LOOP;
      COMMIT;So by changing the value of the constant - nTOTALJOBS - I can control the number of instances of LoadClient that run. This means I only need one procedure which can be called multiple times, making maintenance simpler than having multiple procedures.
    This process has been running successfully on several procedures of over a year. However recently I have been having trouble with latches in memory. I think these are related to using the variables in the cursor and part of the oracle treating the different queries as being identical. This results in locks in memory and none of the jobs running. In most cases I'm setting nTOTALJOBS to 2. I'm looking into the further with my DBAs.
    My questions are:
    1. Has anyone come across this sort of problem before?
    2. If so is there something I can do in my code or DB parameters to alleviate or test this?
    Notes.
    1. The process uses row by row processing in PL/SQL This done to get row by row error handling. I'm not going to change this in the short-term
    2. The SQL does a call over a database link, I don't know if this has any effect.
    3. This method does have a proven performance benefit
    4. The code snippets above are just illustrative, I'm not asking for syntax corrections.
    Thank you for taking the time to read through this.

    I would first investigate the "problem with latches" to be sure that it has anything to do with these jobs that you are submitting.
    Your concern seems to be "because I am submitting multiple copies of the same PLSQL procedure to run concurrently, am I running into latching issues ?".
    Take a step back and look at
    a. MultiUser OLTP environments : At any time you can have more than a few users running the same SQL queries submitted through Forms / Client Frontends.
    b. ERP Report Extractions (eg in Oracle EBusiness Suite or Peoplesoft) : At any time (say during the daily batch runs ?) you can have multiple copies of the same report (with different bind variables for different business_units / countries etc) running concurrently and extracting reports.
    The question would be : How many concurrent executions of the same code is too much ?
    I would say that latching issues come up on :
    a. Shared Pool : Multiple sessions attempting to load SQL statements and Parse SQL statements
    You could reduce some of this with reusable SQLs and bind variables instead of literals in quick running jobs / user front end screens etc
    b. Buffer Cache : Hot Blocks being frequently pinned -- eg index root blocks or "busy" table blocks.
    If you are running, say 4-8 concurrent copies of the the same code on a 4 CPU or 8 CPU machine, you may not be running into latching issues on the code -- "may not be" -- we can't be too sure, until we identify which latches are under "issue".
    However, if the queries are executing nested loops and frequently accessing the same index blocks, you may more likely be having issues with latches on cache buffer chains.
    So, this is what I would suggest :
    a. Identify the size and scale of the "issues" -- how much CPU time / wait time is accounted for by the latches in relation to the total response time of each of the jobs
    b. Identify which latches are the ones under contention
    c. On the basis of which latches are the holding up performance, decide your next course of action.
    Hemant K Chitale
    http://hemantoracledba.blogspot.com

  • Problem with QofQ and Null Values

    Just converted from CF5 to CF7. Finally figured out what was
    causing the error message :
    "The value "" cannot be converted to a number" when running a
    QofQ. It happens whenever there is a NULL (and I think a ZERO)
    value in the QofQ. I'm using the QofQ in a seperate chart program.
    The main query merges data from a store and a region table.
    I'm doing a LEFT OUTER JOIN to combine the REGION with the STORE,
    AND so that stores with NO SALES for a given day will show up as
    ZERO. On the main report, when I don't use a QofQWhen grouping by
    DAY or MONTH, I can test the value in the CFOUTPUT for NULL or ZERO
    and display "0" or "N/A" on the report.
    But when I try to chart it, using a QofQ to read in the
    values from the main query, and format the numbers, I get this
    error message. I've tried using the CAST function, using FLOAT,
    DOUBLE, BIGINT, etc., but still get same error message.
    The only way I could fix this is to add a WHERE statement to
    the QofQ, that only includes sales values greater than ZERO. But
    then, I have gaps in my chart because stores with NO SALES for a
    given day or MONTH don't appear. Because they are removed from the
    query. The whole point of LEFT JOINS is to include items with no
    values (or NULL) values.
    Does anyone know if it's possible to keep these values in a
    QofQ? So that items (e.g., store locations in my case) with NULL or
    ZERO values for some or all days or months can be charted? Even
    with a ZERO value? I've read everything I can get my hands on
    regarding QofQ, the CAST function, etc., but nothing seems to work.
    Thanks for any help. (FYI, didn't have this problem with CF5. It's
    QofQ could handle NULL or ZERO values without providing an error).
    Gary

    Well, I tried the IsNull function on the 3 SUM() values I was
    calculating in my SELECT statement (on a LEFT OUTER JOIN QUERY),
    and I STILL got another "cannot convert "" to a number" but this
    time the error message was more direct, and pointed me to a DATE
    field in my query, where I was SORTING AND GROUPING either by DAY
    or MONTH (depending on what user selected on form).
    In my QofQ, I'm not just reformating the integer SUM()
    values, but also the date values. So, I applied the IsNull()
    function to the DATE values in my main query. I still kept getting
    errors. But after experimenting ((SQL Server BOL docs doesn't give
    ISNULL() examples for DATE FIELDS, only NUMBER fields), I tried
    putting the dates in SINGLE QUOTES in my IsNull() function, the
    QUERY RAN. Then I wasn't sure WHAT date to enter, e.g., 1/1/1889 or
    1/1/2001, etc.
    Then, I discovered, after experimenting, that you cannot just
    blindly enter ANY date when using IsNull in a date field,
    especially if you are using CFGRAPH to chart the results by day.
    You have to select a date within the date range the user selected,
    so this means using the FORM.DATE (or URL.DATE) value from the
    form. Here's the working example from my SELECT main query:
    SELECT SUM(ISNULL(d.ttldb,0)) AS SumOfDB,
    SUM(ISNULL(d.ttldbv,0)) AS SumOfDBV,
    SUM(ISNULL(d.ttldbi,0)) AS SumOfDBI,
    <CFIF '#url.reptype#' IS "DAILY">
    ISNULL(d.depday, '#url.date1#') AS depday
    <CFELSE>
    month(ISNULL(d.depday, '#url.date2#')) AS TranMonth,
    year(ISNULL(d.depday, '#url.date2#')) AS TranYear
    </CFIF>
    Then I discovered that whatever date was used in the URL.DATE
    field (in my case it would be one of the two date field from the
    form, and depended on whether you coded DATE1 or DATE2, the chart
    would always show ZERO values for that date. So I tried putting the
    IsNull() date values in the GROUP BY and ORDER BY statements, e.g.,
    <CFIF '#url.reptype#' IS "DAILY">
    GROUP BY ISNULL(d.depday, '#url.date1#')
    ORDER BY ISNULL(d.depday, '#url.date1#') DESC
    <CFELSE>
    GROUP BY year(ISNULL(d.depday, '#url.date2#')),
    month(ISNULL(d.depday, '#url.date2#'))
    ORDER BY year(ISNULL(d.depday, '#url.date2#')),
    month(ISNULL(d.depday, '#url.date2#'))
    </CFIF>
    And this worked. The report ran just fine, and the chart
    charted every value correctly.
    I guess you have to learn all the little "quirks" about CF7
    and how it's QofQ and chart programs work. But so far, so good.
    Thank you all for your help. Now I've got to update other old
    CF5 programs that are returning errors when there are null values
    from my LEFT OUTER JOIN queries. Will take some time, but at least
    I know how to do it. Thanks to your help.
    I really appreciate your time and efforts to help with this
    issue.
    Gary

  • Problem with IPSec on  solaris 9

    Hi all
    I'm facing a problem with IPSec on solaris 9 that I didn't have with Solaris 8 (With the Security package installed).
    I've an application that creates SA's by using the pf-key interface.
    What it does is first doing a GETSPI to a specific SPI and a specific Destination IP Address.
    This will create an SA and put it in a LARVAL state. After about a minute my application will do an UPDATE to this SPI and that command should change the state of the SA from LARVAL to MATURE but instead I get an error saying that this SPI & IP address already exist (errno = 17).
    Well of course it's already exist that's the all point it should just change the state of an existing SA.
    This exact scenarion was is working fine on Solaris 8.
    Am I doing somthing wrong (maybe there is a package on the solaris 9 that I need to install ?)
    or is this a bug in solaris 9.
    If anyone has any idea on how to do that (without using a one step ADD for a new SA) I will be very thankfull.

    Sorry for using reply for querying.
    I got a problem in creating a Security Association using the PF_KEY Socket (first used SADB_GETSPI and got SPI,with SPI tried to update SADB_UPDATE).
    Getting this problem on Sun Solaris 8.
    It returns errno 122 . operation not supported.
    Here is my mailId [email protected]
    I got few more queries regarding PF_KEY socket.
    Not much directions are available also for pf_key socket in internet.
    Monitor produces the following error.
    # ipseckey monitor
    "Base message (version 2) type UPDATE, SA type AH.
    Error Operation not supported on transport endpoint from PF_KEY.
    Message length 16 bytes, seq=4294967294, pid=450."
    Here is my mailId [email protected]
    Thanks in Advance.
    ssundar.

  • Problem with logminer in Data Guard configuration

    Hi all,
    I experience strange problem with applying of the logs in DataGuard configuration on the logical standby database side.
    I've set up the configuration step by step as it is described in documentation (Oracle Data Guard Concepts and Administration, chapter 4).
    Everything went fine until I issued
    ALTER DATABASE START LOGICAL STANDBY APPLY;
    I saw that log applying process was started by checking the output of
    SELECT NAME, VALUE FROM V$LOGSTDBY_STATS WHERE NAME = 'coordinator state';
    and
    SELECT TYPE, HIGH_SCN, STATUS FROM V$LOGSTDBY;
    but in few minutes it stoped and quering DBA_LOGSTDBY_EVENTS I saw the following records:
    ORA-16111: log mining and apply setting up
    ORA-01332: internal Logminer Dictionary error
    Alert log says the following:
    LOGSTDBY event: ORA-01332: internal Logminer Dictionary error
    Wed Jan 21 16:57:57 2004
    Errors in file /opt/oracle/admin/whouse/bdump/whouse_lsp0_5817.trc:
    ORA-01332: internal Logminer Dictionary error
    Here is the end of the whouse_lsp0_5817.trc
    error 1332 detected in background process
    OPIRIP: Uncaught error 447. Error stack:
    ORA-00447: fatal error in background process
    ORA-01332: internal Logminer Dictionary error
    But the most useful info I found in one more trace file (whouse_p001_5821.trc):
    krvxmrs: Leaving by exception: 604
    ORA-00604: error occurred at recursive SQL level 1
    ORA-01031: insufficient privileges
    ORA-06512: at "SYS.LOGMNR_KRVRDREPDICT3", line 68
    ORA-06512: at line 1
    Seems that somewhere the correct privileges were not given or smth like this. By the way I was doing all the operations under SYS account (as SYSDBA).
    Could smb give me a clue where could be my mistake or what was done in the wrong way?
    Thank you in advance.

    Which is your SSIS version?
    Please Mark This As Answer if it solved your issue
    Please Vote This As Helpful if it helps to solve your issue
    Visakh
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • Problem with unit in Query

    Hallo Experts,
    We encounter a problem with unit in our Query.
    Szenario:
    We use the Unit like PC, SET in Query.
    In Table T006 those Units have value 0 in the field ANDEC, so they will be rounded up and rounded down in Query.
    Our Problem is some Queries need the rounding up and down but others don't need this function.
    Where and how should we change to solve the problem.
    thanks & Best regards

    hi,
    if you want to control the rounding off while unit conversion you can do it from SPRO
    In the BW system, goto SPRO -> SAP NetWeaver -> General Settings -> Check Units of Measurement -> Units of Measurement -> Choose the unit corresponding to 'KM' -> Click on 'Details' button in the toolbar.
    Look the field 'Decimal Pl rounding'. Change this to 3 or whatever value you choose and see if it works.
    in above link i have taken example of KM, you need need to select the required unit.
    let us know if this is solves your problem. I am not sure if this can be done for individual queries but this is controlled by UOM wise as mentioned above.
    Regards,
    Rk.

  • Problem with Date Parameter

    Hi
    I am new to reporting services/report designer but have created a number of simple reports with Parameters using data from queries that have worked but I have a problem with this using Date and Time. 
    I am using MS Visual Studio 2008 (BIDS) and our SQL Server is Windows Server 2008 R2
    I need to allow the users of the report to choose a start date and an end date
    The Start Dates are held in a field called PhaseStartDate
    The End Dates are held in a field called PhaseEndDates
    I have tried multiple ways to try to get this to work with no success and with the latest attempt I get the message  
    "An error occurred during local report processing. The definition of the report 'JKVoidLoss' is invalid. The report parameter 'StartDate' has a default value or ValidValue that depends on the report parameter"Startdate". Forward dependancies
    are not valid"
    I will explain how I set up the Startdate. The EndDate is set up the same with the relevant values and fields.
    I created a new data source called StartDate. Under QUERY I have Select PhaseStartDate from QLHRA_VoidPhases. This runs okay when tested in Query designer and RUN.
    Under FIELDS I have Field name as PhaseStartDate and Field Source as PhaseStartdate
    Under PARAMETERS I have Parameter Name as PhaseStartDate and Parameter Value as [@StartDate]
    I set up a Parameter called StartDate with Data type as Date/Time
    Under Available Values I selected Get values from a query with Dataset being StartDate, Value field being PhaseStartdate and Label field being PhaseStartDate
    There are no Default Values
    I have tried moving the parameters up and down using the blue arrows under report data. I have tried multiple different methods to resolve this by changing values in the Properties screens.
    I also tried removing the DataSets and just using the Parameters with default settings. in preview, this ran the report but no matter what dates where entered the output was always the same.
    Any assistance will be greatly appreciated as I have spent hours on this so far
    Regards
    John

    Hi
    I am new to reporting services/report designer but have created a number of simple reports with Parameters using data from queries that have worked but I have a problem with this using Date and Time. 
    I am using MS Visual Studio 2008 (BIDS) and our SQL Server is Windows Server 2008 R2
    I need to allow the users of the report to choose a start date and an end date
    The Start Dates are held in a field called PhaseStartDate
    The End Dates are held in a field called PhaseEndDates
    I have tried multiple ways to try to get this to work with no success and with the latest attempt I get the message  
    "An error occurred during local report processing. The definition of the report 'JKVoidLoss' is invalid. The report parameter 'StartDate' has a default value or ValidValue that depends on the report parameter"Startdate". Forward dependancies are not valid"
    I will explain how I set up the Startdate. The EndDate is set up the same with the relevant values and fields.
    I created a new data source called StartDate. Under QUERY I have Select PhaseStartDate from QLHRA_VoidPhases. This runs okay when tested in Query designer and RUN.
    Under FIELDS I have Field name as PhaseStartDate and Field Source as PhaseStartdate
    Under PARAMETERS I have Parameter Name as PhaseStartDate and Parameter Value as [@StartDate]
    I set up a Parameter called StartDate with Data type as Date/Time
    Under Available Values I selected Get values from a query with Dataset being StartDate, Value field being PhaseStartdate and Label field being PhaseStartDate
    There are no Default Values
    I have tried moving the parameters up and down using the blue arrows under report data. I have tried multiple different methods to resolve this by changing values in the Properties screens.
    I also tried removing the DataSets and just using the Parameters with default settings. in preview, this ran the report but no matter what dates where entered the output was always the same.
    Any assistance will be greatly appreciated as I have spent hours on this so far
    Regards
    John
    As I understand what you need to do is to just remove parameter from the dataset query for populating the date values and then it should work fine
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

Maybe you are looking for

  • Can't Transfer Video to iPod

    I have an 160 gb iPod Classic (format Windows 8) and cannot transfer video podcasts from iTunes to my iPod.  I have verified I have the current version of iTunes and iTunes says my iPod has version 2.0.4 which it says is the current version.  I can p

  • What version was this PDF created in?

    I have a PDF that was uploaded to our website.  If I save it to my computer and look at the properties I only see detail reflecting my version of Adobe Reader .  Is there a way for me to see what version of full Adobe was used to create this file?

  • Grid in Oracle Forms

    wil it be possiblt to create a form where we can drag the change the width of the columns in the form. How can this be accomplished?

  • How to change language version of Photoshop?

    I'm CC subscriber since few days ago, and after installing different language version of PS, I now want English Photoshop instead. The first installed version is deleated, but I can't find out how to change to English version. Can anyone help?

  • Error 9 while updating to v3.1

    Using iTunes 9, when I try to update my 1G iPod Touch to the latest version I get an error 9. I can downgrade/restore to v3.0 no problem. The size of the firmware file for v3.1 is 243903kB. Any ideas?