Help needed in storing data in DSO or InfoCube

Hi Experts ,
   I have a set of data which comes from R/3 in the following Format:
FG     Base Model     Option 1     Option 2     Option 3
FG1     BM 1                       Opt1       Opt2         opt3
FG2     BM1                       Opt 1       Opt2        Opt 4
FG3     BM3                       Opt5       Opt6        Opt7
The data gets stored in the PSA in the above format.
But i want to load the data from the PSA to the DSO or InfoCube in the following format.I want that if i do a display data in the DSO or the Infocube the output should look like the below :
FG     Base Model     Option
FG1     BM1                      Opt1
FG1     BM1                      Opt2
FG1     BM1                      Opt3
FG2     BM1                      Opt1
FG2     BM1                      Opt2
FG2     BM1                      Opt4
FG3     BM3                      Opt5
FG3     BM3                      Opt6
FG3     BM3                      Opt7
Is there any way to do this please help. Thanks in advance...

Hi Samir,
Use transformation rule group in the transformations.
Here is one of the examples...the way you wanted.
http://help.sap.com/saphelp_nw70/helpdata/EN/44/32cfcc613a4965e10000000a11466f/content.htm
Thanks
Ajeet

Similar Messages

  • Problem in transfering data from DSO to InfoCube

    Hi,
    I am trying to transfer data from DSO to Infocube. But it is giving the following message.
    Last delta update is not yet completed
    Therefore, no new delta update is possible.
    You can start the request again
    if the last delta request is red or green in the monitor (QM activity)
    It is not allowing to load data to InfoCube from DSO. I checked the request number. But there is no request number in InfoCube. If request no is there I can delete the request from InfoCube and delete the data mart status in DSO and reload again.But there is no request id in IC. So plz guide me how I can tranfer data to InfoCube overcoming this problem.
    Actually further updata was running from DSO to IC. But it failed. Executing Function Module : RSPC_PROCESS_FINISH, I set the chain  green before loading
    data to InfoCube. So it starts executing the next chain. So there is no request id. So plz guide me how I can retain this request id.
    Edited by: Purushotham Reddy on Jan 28, 2008 11:19 AM

    Hi Reddy,
    Make the QM Status Red of the failed request even if it is red.
    and then run the same infopackage used in the chain. This will ask for repeat of last delta.
    If this process is understood to u then its fine otherwise read below:
    the other option is check if this failed request is present in PSA. If it is present in PSA then update the failed request from PSA, here u wont need to do a repeat of last delta. This Option is easier. this u can do from the ,manage tab of the data source I think this u might be surely knowing....
    Assign Points if it helps u.....
    Thanks,
    Nagesh.

  • Error in uploading data from DSO to Infocube in 3.5(rsa1old)

    While uploading data from DSO to Infocube, I selected Initial update option
    In schedule tab when I clicked start, an error occured it says:
    Delete init request
    REQU_D4YTAGX8PEOUQJLSKOEPH9EV before running
    init. again with same selection.
    Can anyone let me know how should I delete that request.
    Thanks,
    Soujanya.

    Hi,
    Go to RSRQ and enter the request then execute. This should bring you to Monitor u2013 Administration Workbench. From there you will see the InfoSource (usually has u201C8u201D as a prefix). Get the InfoSource name. Go to RSA1 to find the InfoSource. From here it gets a bit tricky. Go to the datasource and drill down to the related InfoPackage. Go to Infopackage scheduling. Then go  to Scheduler Menu, the Initialization Options for Source System should be open. You can see from there the request with error. Let me know if this answers your question.
    Regards,
    JD Datuin

  • Help needed with missing data problem in CRVS2010

    We recently upgraded the reporting engine in our product to use Crystal Reports for Visual Studio 2010 (previously engine was CR9). Our quote report, which has numerous subreports and lots of conditional formatting, started losing data when a quote took more than a single page to be printed. We knew the SQL results included the data, but the report was not printing those lines at all or sometimes printing a partial line. In addition, the running total on the report would exclude the lines that were being missed on the next page. In one example submitted by a customer, 3 lines were skipped between pages.
    I think I have identified two potential issues that document the possibility of data not being included in the report.
    The first potential issue is an issue with the "suppress blank section" option being checked. This issue is supposedly fixed with ADAPT01483793, being released someday with service pack 2 for CRVS2010.
    The second potential issue is using shared variables. This issue is supposedly fixed with ADAPT01484308, also targeted for SP2.
    Our quote report does not explicitly use shared variables with any of the subreports, but it does have several subreports, each in its own section that has the "supress blank section" option checked. We have other reports that use this feature, as well, and they are not exhibiting the problem.
    One different thing about the quote report is that it has a section with multiple suppression options selected. The section has a conditional suppression formula, which controls whether the section is included at all within the report. The section also has the suppress blank section option selected. There are multiple fields within the report that are each conditionally suppressed. In theory, the section's suppress formula could evaluate to true, yet all of the fields within the section are suppressed (due to null values), and then the "suppress blank section" option would kick in.
    The missing data only seems to happen when the section is not being suppressed, and at least one of the fields is being included in the report. If I clear the "suppress blank section" check box, and change the section formula to also include the rules applied to the fields in the section, the missing data problem seems to be resolved.
    Is this related to ADAPT01483793? Will it be fixed in service pack 2?
    If more details are needed, I would be happy to provide a sample report with stored data.

    Hi Don,
    Have a look at the Record Selection formula in CR Designer ( stand alone ) and when exported to RPT format opening that report in the Designer also. 
    There's been a few issues with => logic in the record selection formula. It could be you are running into this problem. Look for NOT inserted into your selection formula.
    Oh and SP2 is coming out shortly so it may resolve the issue. But if you want you could purchase a support, or if you have a support contract then create a case in SMP and get a rep to work with you to debug the issue.
    If you have not try the Trial Version of CR 2011, put it on a VM-ware image or Test PC so you don't corrupt anything for production and have a look at and test it in that designer also. If you purchase a case and it is a bug then you'll get a credit back for the case.
    Don
    Edited by: Don Williams on Oct 26, 2011 7:40 AM

  • Help needed on meta data export

    I need a meta data export dump of a source database of two schemas and import to a fresh new database.
    COuld you please help with the steps to be followed.

    Assuming you are using 10g and exp utility
    You can say rows=N
    exp help=yes
    Export: Release 10.2.0.1.0 - Production on Thu Jul 23 13:36:59 2009
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    You can let Export prompt you for parameters by entering the EXP
    command followed by your username/password:
    Example: EXP SCOTT/TIGER
    Or, you can control how Export runs by entering the EXP command followed
    by various arguments. To specify parameters, you use keywords:
    Format: EXP KEYWORD=value or KEYWORD=(value1,value2,...,valueN)
    Example: EXP SCOTT/TIGER GRANTS=Y TABLES=(EMP,DEPT,MGR)
    or TABLES=(T1:P1,T1:P2), if T1 is partitioned table
    USERID must be the first parameter on the command line.
    Keyword Description (Default) Keyword Description (Default)
    USERID username/password FULL export entire file (N)
    BUFFER size of data buffer OWNER list of owner usernames
    FILE output files (EXPDAT.DMP) TABLES list of table names
    COMPRESS import into one extent (Y) RECORDLENGTH length of IO record
    GRANTS export grants (Y) INCTYPE incremental export type
    INDEXES export indexes (Y) RECORD track incr. export (Y)
    DIRECT direct path (N) TRIGGERS export triggers (Y)
    LOG log file of screen output STATISTICS analyze objects (ESTIMATE)
    {color:red}*ROWS export data rows (Y)* {color} PARFILE parameter filename
    CONSISTENT cross-table consistency(N) CONSTRAINTS export constraints (Y)
    OBJECT_CONSISTENT transaction set to read only during object export (N)
    FEEDBACK display progress every x rows (0)
    FILESIZE maximum size of each dump file
    FLASHBACK_SCN SCN used to set session snapshot back to
    FLASHBACK_TIME time used to get the SCN closest to the specified time
    QUERY select clause used to export a subset of a table
    RESUMABLE suspend when a space related error is encountered(N)
    RESUMABLE_NAME text string used to identify resumable statement
    RESUMABLE_TIMEOUT wait time for RESUMABLE
    TTS_FULL_CHECK perform full or partial dependency check for TTS
    TABLESPACES list of tablespaces to export
    TRANSPORT_TABLESPACE export transportable tablespace metadata (N)
    TEMPLATE template name which invokes iAS mode export
    Export terminated successfully without warnings.

  • Help needed to insert data from different database

    Hi ,
    I have a requirement where i need to fetch data from different database through database link .Depending on user request , the dblink needs to change and data from respective table from mentioned datbase has to be fetched and populated .Could i use execute immediate for this, would dblink work within execute immediate .If not , could pls let me know any other approach .

    What does "the dblink needs to change" mean?
    Are you trying to dynamically create database links at run-time? Or to point a query at one of a set of pre-established database links at run-time?
    Are you sure that you really need to get the data from the remote database in real time? Could you use materialized views/ Streams/ etc to move the data from the remote databases to the local database? That tends to be far more robust.
    Justin

  • Mockup data into DSO or InfoCube.

    Hi Experts,
    I am working on BI 7.0 within Solution Manager. There are only about 3 Chars and 4 KeyFigures. Still there seems to be some problem while activating the DSO which Basis is looking into at the moment.
    But I have to do something very urgently. To complete the BEx Query Unit Testing, I only need to somehow load 5 records of data into an InfoCube or DSO and then assign it to the MultiProvider which is existing already for the BEx person to do his Unit Testing.
    I tried using Real-time Cube but still there is no option to add data into it.
    Has anyone faced a similar situation? Is there a way to mock up sample data into a Standalone InfoCube or DSO?
    Any help is very much appreciated. Points will be assigned.
    Thanks,
    Chandu

    Hi Daya,
    The error am getting while trying to activating the DSO is as follows:
    Runtime Errors         DATA_OFFSET_TOO_LARGE
    Exception              CX_SY_RANGE_OUT_OF_BOUNDS
    Date and Time          16.02.2009 11:24:02
    Short text
    Invalid subfield access: Offset too large
    What happened?
    Error in the ABAP Application Program
    The current ABAP program "CL_RSODSO_SID_HANDLER=========CP" had to be
    terminated because it has
    come across a statement that unfortunately cannot be executed.
    What can you do?
    Note down which actions and inputs caused the error.
    To process the problem further, contact you SAP system
    administrator.
    Using Transaction ST22 for ABAP Dump Analysis, you can look
    at and manage termination messages, and you can also
    keep them for a long time.
    Error analysis
    An exception occurred that is explained in detail below.
    The exception, which is assigned to class 'CX_SY_RANGE_OUT_OF_BOUNDS', was not
    caught in
    procedure "RECEIVE_DATA_PACKAGE" "(METHOD)", nor was it propagated by a RAISING
    clause.
    Since the caller of the procedure could not have anticipated that the
    exception would occur, the current program is terminated.
    The reason for the exception is:
    In the running program "CL_RSODSO_SID_HANDLER=========CP", the field "L_CHAVL"
    of the type
    "C" and length 60 was to be accessed with the offset 60.
    However, subfield accesses with an offset specification that is not
    smaller than the field length are not permitted.
    How to correct the error
    Reduce the offset with which the field is to be accessed.
    If the error occurred in your own ABAP program or in an SAP
    program you modified, try to remove the error.
    If the error occures in a non-modified SAP program, you may be able to
    find an interim solution in an SAP Note.
    If you have access to SAP Notes, carry out a search with the following
    keywords:
    "DATA_OFFSET_TOO_LARGE" "CX_SY_RANGE_OUT_OF_BOUNDS"
    "CL_RSODSO_SID_HANDLER=========CP" or "CL_RSODSO_SID_HANDLER=========CM002"
    "RECEIVE_DATA_PACKAGE"
    If you cannot solve the problem yourself and want to send an error
    notification to SAP, include the following information:
    1. The description of the current problem (short dump)
    To save the description, choose "System->List->Save->Local File
    (Unconverted)".
    2. Corresponding system log
    Display the system log by calling transaction SM21.
    Restrict the time interval to 10 minutes before and five minutes
    after the short dump. Then choose "System->List->Save->Local File
    (Unconverted)".
    3. If the problem occurs in a problem of your own or a modified SAP
    program: The source code of the program
    In the editor, choose "Utilities->More
    Utilities->Upload/Download->Download".
    4. Details about the conditions under which the error occurred or which
    actions and input led to the error.
    The exception must either be prevented, caught within proedure
    "RECEIVE_DATA_PACKAGE" "(METHOD)", or its possible occurrence must be declared
    in the
    RAISING clause of the procedure.
    To prevent the exception, note the following:
    I can send you the full short dump if you could give me your email id.
    I tried t-code RSINPUT for a Real-time Cube, but seems like it is not working. How would it be possible to enter KeyFigures in this case though.
    Any help is much appreciated.
    Thanks,
    Chandu

  • Error while loading data from DSO to Infocube for 0Fiscvarnt

    Hi All,
    Could you please help me  any one to resolve the below error.
    *"Fixed value discrepancy with InfoObject 0FISCVARNT value 'K4 ', expected 'K2 "  (MessageClass RSAU).
    We have two  Staging DSO's upto Second level DSO FiscalyearVariant  both K2 and K4 updating, but second level DSO to infocube data loading we had a above problem. all are direct mapping, no routiunes or constants at all transformation level.
    But we checked R/3 side both K2 and K4 variants are available.
    Thanks,
    Ramesh

    Hi,
    It's a constant issue I met before. I guess you have assigned a constant value K2 in your target ODS.
    Please change the constant back. Go to RSA1, change target ODS, find 0FISCVARNT, right click it and choose 'Provider-specific Properties', remove K2 from 'Constant'.
    Now try the reload and I'm sure no more issues. If you really want the constant then you have to change data or transformation.
    p.s. Don't forget re-activating transformations and DTP
    Regards,
    Frank
    Edited by: Frank Lee on Jul 14, 2009 3:13 PM

  • Issue while loading of data from DSO to InfoCube

    Hi Experts,
    Can you tell me what might root casue if data is coming into DSO from R3 its correct and fine as required but while loading it to InfoCube from DSO its showing wrong data like some of Line Items that were closed were shown open in Cube AND also KF values were not right
    Also there is no Routine code involved b/w DSO and InfoCube.
    Thanks in adv .
    NP

    Hope you didnt delete some req from DSO without deleting change log . This might cause inconsistency.
    If so , delete data from dso by right click delete data  and reload .

  • Pros help needed with post data code

    i needed to change the code below to post the userAnswer from radio button group,  to an ASPX page so i can read the data in and post it to a database. can anyone help. thanks
    btnCheck.addEventListener(MouseEvent.CLICK, checkAnswer);
    function checkAnswer(evt:MouseEvent):void {
    userAnswer = String(rbg.selectedData);
        messageBox.text =  userAnswer + " has been clicked";

    //Create the loader object
    var prodloader:URLLoader = new URLLoader ();
    //Create a URLVariables object to store the details
    var variables: URLVariables = new URLVariables();
    //Createthe URLRequest object to specify the file to be loaded and the method ie post
    var url:String = "url here";
    var prodreq:URLRequest = new URLRequest (url);
    prodreq.method = URLRequestMethod.POST;
    prodreq.data = variables;
    function submitHandler(event:Event):void {
        variables.productId = whatever;
        prodloader.load(prodreq);
        btnSubmit.addEventListener(MouseEvent.CLICK, submitHandler);
        function contentLoaded(event:Event):void {
           //Stuff here
            prodloader.addEventListener(Event.COMPLETE, contentLoaded);

  • Urgent.... Help Needed for Storing file in Application Server

    Hi All,
    I have a requirement where in I have to save a file in Application Server.
    File is of 141 Character Length, so there can be spaces left at the end.
    When I am downloading the file into Application Sever and then downloading the same file from Application Server to Presentation into a notepad then the spaces at the end are truncating.
    I want that the file that is generated should be of 141 character though if there may be spaces at the end.
    It will be great if you can help me out in this.
    Thanks in Advance.
    Jayant Sahu.

    when downloading to App server, you need to specify the LENGTH option on your TRANSFER ststement to keep fixed length.  You must also use a STRING type variable in your transfer.
    From Documentation on Transfer command:
    <b>If the file was opened as a text file or a legacy text file, the trailing blank characters are deleted for all data objects, except for those of data type string. The line end marker defined when the file was opened is then added to the remaining content of the data object or to the result of the conversion, and the final result is written byte-by-byte to the file.</b>
    There also used to be a problem with WS_DOWNLOAD - you had to do the following:
    Maintaining Trailing spaces when downloading to PC
    Before calling DOWNLOAD or WS_DOWNLOAD, do a perform SET_TRAIL_BLANKS(saplgrap) using 'X'
    To set the length of each record including your blanks add this code: perform SET_FIXLEN(saplgrap) using '0' '100'
    Don't think this is needed any more.
    Andrew

  • Help Needed in persisting data in cache from CEP on to a database

    Hi,
    We are trying to create a Oracle Complex Event Processing (CEP) application in which persist the data stored in the cache (Oracle Coherence) to a back end database
    Let me provide you the steps that I have followed:
    1)     Created a CEP project with cache implementation to store the events.
    2)     Have the following configuration in the context file:
    <wlevs:cache id="Task_IN_Cache" value-type="Task" key-properties="TASKID" caching-system="CachingSystem1">
    <wlevs:cache-store ref="cacheFacade"></wlevs:cache-store>
    </wlevs:cache>
    <wlevs:caching-system id="CachingSystem1" provider="coherence">
    <bean id="cacheFacade" class="com.oracle.coherence.handson.DBCacheStore">
    </bean>
    3)     Have the following in the coherence cache config xml:
    <cache-mapping>
    <cache-name>Task_IN_Cache</cache-name>
    <scheme-name>local-db-backed</scheme-name>
    </cache-mapping>
    <local-scheme>
    <scheme-name>local-db-backed</scheme-name>
    <service-name>LocalCache</service-name>
    <backing-map-scheme>
    <read-write-backing-map-scheme>
    <internal-cache-scheme>
    <local-scheme/>
    </internal-cache-scheme>
    <cachestore-scheme>
    <class-scheme>
    <class-name>com.oracle.coherence.handson.DBCacheStore</class-name>
    </class-scheme>
    </cachestore-scheme>
    </read-write-backing-map-scheme>
    </backing-map-scheme>
    </local-scheme>
    4)     Have configured tangosol-coherence-override.xml to make use of coherence in my local machine.
    5)     Have written a class that implements com.tangosol.net.cache.CacheStore
    public class DBCacheStore implements CacheStore{
    But when I try to deploy the project on to the CEP server getting the below error:
    org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'Task_IN_AD': Cannot resolve reference to bean 'wlevs_stage_proxy_forTask_IN_Cache' while setting bean property 'listeners' with key [0]; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'wlevs_stage_proxy_forTask_IN_Cache': Cannot resolve reference to bean '&Task_IN_Cache' while setting bean property 'cacheFactoryBean'; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'Task_IN_Cache': Invocation of init method failed; nested exception is java.lang.RuntimeException: Error while deploying application 'AIT_Caching'. Cache store specified for cache 'Task_IN_Cache' does not implement the required 'com.tangosol.net.cache.CacheStore' interface.
    Can you please let me know if I am missing any configuration. Appreciate your help.

    Hi JK,
    Yes my class com.oracle.coherence.handson.DBCacheStore implements the com.tangosol.net.cache.CacheStore interface. I am providing you with the DBCacheStore class.
    package com.oracle.coherence.handson;
    import com.tangosol.net.CacheFactory;
    import com.tangosol.net.NamedCache;
    import com.tangosol.net.cache.CacheStore;
    import java.sql.DriverManager;
    import java.sql.Connection;
    import java.sql.PreparedStatement;
    import java.sql.ResultSet;
    import java.sql.SQLException;
    import java.util.Collection;
    import java.util.Iterator;
    import java.util.LinkedList;
    import java.util.List;
    import java.util.Map;
    import oracle.jdbc.pool.OracleDataSource;
    public class DBCacheStore implements CacheStore {
    protected Connection m_con;
    protected String m_sTableName;
    private static final String DB_DRIVER = "oracle.jdbc.OracleDriver";
    private static final String DB_URL = "jdbc:oracle:thin:@XXXX:1521:XXXX";
    private static final String DB_USERNAME = "XXXX";
    private static final String DB_PASSWORD = "XXXX";
    public DBCacheStore()
    m_sTableName = "TASK";
    System.out.println("Inside constructor");
    init();
    //store("10002", "10002");
    public void init()
    try
         OracleDataSource ods = new OracleDataSource();
    /* Class.forName(DB_DRIVER);
    m_con = DriverManager.getConnection(DB_URL, DB_USERNAME, DB_PASSWORD);
    m_con.setAutoCommit(true);*/
         ods.setURL(DB_URL);
         ods.setUser(DB_USERNAME);
         ods.setPassword(DB_PASSWORD);
         m_con = ods.getConnection();
    System.out.println("Connection Successful");
    catch (Exception e)
         e.printStackTrace();
    //throw ensureRuntimeException(e, "Connection failed");
    public String getTableName() {
    return m_sTableName;
    public Connection getConnection()
    return m_con;
    public Object load(Object oKey)
    Object oValue = null;
    Connection con = getConnection();
    String sSQL = "SELECT TASKID, REQNUMBER FROM " + getTableName()
    + " WHERE TASKID = ?";
    try
    PreparedStatement stmt = con.prepareStatement(sSQL);
    stmt.setString(1, String.valueOf(oKey));
    ResultSet rslt = stmt.executeQuery();
    if (rslt.next())
    oValue = rslt.getString(2);
    if (rslt.next())
    throw new SQLException("Not a unique key: " + oKey);
    stmt.close();
    catch (SQLException e)
    //throw ensureRuntimeException(e, "Load failed: key=" + oKey);
    return oValue;
    public void store(Object oKey, Object oValue)
         System.out.println("Inside Store method");
         NamedCache cache = CacheFactory.getCache("Task_IN_Cache");
         System.out.println("Cache Service:" + " "+ cache.getCacheService());
         System.out.println("Cache Size:" + " "+ cache.size());
         System.out.println("Is Active:" + " "+ cache.isActive());
         System.out.println("Is Empty:" + " "+ cache.isEmpty());
         //cache.put("10003", "10003");
         //System.out.println("Values:" + " "+ cache.put("10003", "10003"));
    Connection con = getConnection();
    String sTable = getTableName();
    String sSQL;
    if (load(oKey) != null)
    sSQL = "UPDATE " + sTable + " SET REQNUMBER = ? where TASKID = ?";
    else
    sSQL = "INSERT INTO " + sTable + " (TASKID, REQNUMBER) VALUES (?,?)";
    try
    PreparedStatement stmt = con.prepareStatement(sSQL);
    int i = 0;
    stmt.setString(++i, String.valueOf(oValue));
    stmt.setString(++i, String.valueOf(oKey));
    stmt.executeUpdate();
    stmt.close();
    catch (SQLException e)
    //throw ensureRuntimeException(e, "Store failed: key=" + oKey);
    public void erase(Object oKey)
    Connection con = getConnection();
    String sSQL = "DELETE FROM " + getTableName() + " WHERE id=?";
    try
    PreparedStatement stmt = con.prepareStatement(sSQL);
    stmt.setString(1, String.valueOf(oKey));
    stmt.executeUpdate();
    stmt.close();
    catch (SQLException e)
    // throw ensureRuntimeException(e, "Erase failed: key=" + oKey);
    public void eraseAll(Collection colKeys)
    throw new UnsupportedOperationException();
    public Map loadAll(Collection colKeys) {
    throw new UnsupportedOperationException();
    public void storeAll(Map mapEntries)
         System.out.println("Inside Store method");
    throw new UnsupportedOperationException();
    public Iterator keys()
    Connection con = getConnection();
    String sSQL = "SELECT id FROM " + getTableName();
    List list = new LinkedList();
    try
    PreparedStatement stmt = con.prepareStatement(sSQL);
    ResultSet rslt = stmt.executeQuery();
    while (rslt.next())
    Object oKey = rslt.getString(1);
    list.add(oKey);
    stmt.close();
    catch (SQLException e)
    //throw ensureRuntimeException(e, "Iterator failed");
    return list.iterator();
    }

  • Help needed in inserting data using a query

    I need to have some data as a result of the following query:
    select sched_num,load_id,ord_id,split_id from ord_load_seq
    where (ord_id,split_id,sched_num) in (select ord_id,split_id,sched_num from ord_load_seq where seq_num = '2'
    group by ord_id,split_id,sched_num having count(1) >1)
    order by ord_id,split_id,sched_num;
    But currently it retunrns no rows. The problem is in the having count(1)> 1 clause.
    When i make =1, it returns rows. But no rows on > 1. I even tried inserting some rows to get the result >1
    But still the query on a whole returns no rows.Please help.

    ohhh... lets start our lesson children:
    here is code to consider:
    Connected to Oracle Database 10g Enterprise Edition Release 10.2.0.2.0
    SQL>
    SQL> --Q1
    SQL> with tbl as
      2  (select 1 fld1, 1 fld2, 1 fld3 from dual
      3  union all
      4  select 1 fld1, 1 fld2, 2 fld3 from dual
      5  union all
      6  select 1 fld1, 1 fld2, 3 fld3 from dual
      7  union all
      8  select 1 fld1, 2 fld2, 3 fld3 from dual
      9  union all
    10  select 1 fld1, 3 fld2, 3 fld3 from dual)
    11  select  fld1, fld2, fld3 from tbl
    12  order by fld1, fld2, fld3
    13  /
          FLD1       FLD2       FLD3
             1          1          1
             1          1          2
             1          1          3
             1          2          3
             1          3          3
    SQL> --Q2
    SQL> with tbl as
      2  (select 1 fld1, 1 fld2, 1 fld3 from dual
      3  union all
      4  select 1 fld1, 1 fld2, 2 fld3 from dual
      5  union all
      6  select 1 fld1, 1 fld2, 3 fld3 from dual
      7  union all
      8  select 1 fld1, 2 fld2, 3 fld3 from dual
      9  union all
    10  select 1 fld1, 3 fld2, 3 fld3 from dual)
    11  select  fld1, fld2, fld3, count(1) from tbl
    12  group by fld1,fld2,fld3
    13  order by fld1, fld2, fld3
    14  /
          FLD1       FLD2       FLD3   COUNT(1)
             1          1          1          1
             1          1          2          1
             1          1          3          1
             1          2          3          1
             1          3          3          1
    SQL> --Q3
    SQL> with tbl as
      2  (select 1 fld1, 1 fld2, 1 fld3 from dual
      3  union all
      4  select 1 fld1, 1 fld2, 2 fld3 from dual
      5  union all
      6  select 1 fld1, 1 fld2, 3 fld3 from dual
      7  union all
      8  select 1 fld1, 2 fld2, 3 fld3 from dual
      9  union all
    10  select 1 fld1, 3 fld2, 3 fld3 from dual)
    11  select  fld1, fld2, fld3 from tbl
    12  group by fld1,fld2,fld3
    13  having count(1) > 1
    14  order by fld1, fld2, fld3
    15  /
          FLD1       FLD2       FLD3
    SQL> --Q4
    SQL> with tbl as
      2  (select 1 fld1, 1 fld2, 1 fld3 from dual
      3  union all
      4  select 1 fld1, 1 fld2, 2 fld3 from dual
      5  union all
      6  select 1 fld1, 1 fld2, 2 fld3 from dual
      7  union all
      8  select 1 fld1, 1 fld2, 3 fld3 from dual
      9  union all
    10  select 1 fld1, 2 fld2, 3 fld3 from dual
    11  union all
    12  select 1 fld1, 3 fld2, 3 fld3 from dual)
    13  select  fld1, fld2, fld3 from tbl
    14  order by fld1, fld2, fld3
    15  /
          FLD1       FLD2       FLD3
             1          1          1
             1          1          2
             1          1          2
             1          1          3
             1          2          3
             1          3          3
    6 rows selected
    SQL> --Q5
    SQL> with tbl as
      2  (select 1 fld1, 1 fld2, 1 fld3 from dual
      3  union all
      4  select 1 fld1, 1 fld2, 2 fld3 from dual
      5  union all
      6  select 1 fld1, 1 fld2, 2 fld3 from dual -- inserted duplicate row
      7  union all
      8  select 1 fld1, 1 fld2, 3 fld3 from dual
      9  union all
    10  select 1 fld1, 2 fld2, 3 fld3 from dual
    11  union all
    12  select 1 fld1, 3 fld2, 3 fld3 from dual)
    13  select  fld1, fld2, fld3, count(1)   from tbl
    14  group by fld1,fld2,fld3
    15  order by fld1, fld2, fld3
    16  /
          FLD1       FLD2       FLD3   COUNT(1)
             1          1          1          1
             1          1          2          2
             1          1          3          1
             1          2          3          1
             1          3          3          1
    SQL> --Q6
    SQL> with tbl as
      2  (select 1 fld1, 1 fld2, 1 fld3 from dual
      3  union all
      4  select 1 fld1, 1 fld2, 2 fld3 from dual
      5  union all
      6  select 1 fld1, 1 fld2, 2 fld3 from dual -- inserted duplicate row
      7  union all
      8  select 1 fld1, 1 fld2, 3 fld3 from dual
      9  union all
    10  select 1 fld1, 2 fld2, 3 fld3 from dual
    11  union all
    12  select 1 fld1, 3 fld2, 3 fld3 from dual)
    13  select  fld1, fld2, fld3 from tbl
    14  group by fld1,fld2,fld3
    15  having count(1) > 1
    16  order by fld1, fld2, fld3
    17  /
          FLD1       FLD2       FLD3
             1          1          2
    SQL> Q1. As you may see we have an bunch of data where each row is unique combination of the columns value.
    Q2. lets try to group it by fld1 and fld2 and fld3 columns. We don't expect any miracle and got the same data as Q1. Why? Because each row is unique combination(group) in scope of fld1, fld2, fld3 - count(1) exactly shows us that.
    Q3. Q2 is explanation why we got no rows filter (having count(1) > 1)because the result of the clause is always false.
    Q4. Lets put some duplication.
    Q5. Now we got some new result. Count shows us group with more then one rows.
    Q6. Shows us exact group which we've found in Q5.

  • Help needed in converting date to number

    Hi,
    I am trying to execute the below query, its failing with "invalid number" error.
    select * from process_status
    where time_process > to_char('&3','yyyymmdd')
    time_process is the "number" datatype.
    Details:
    SQL> select * from process_status
    2 where time_process > to_char('&3','yyyymmdd');
    Enter value for 3: 01-MAY-09
    old 2: where time_process > to_char('&3','yyyymmdd')
    new 2: where time_process > to_char('01-MAY-09','yyyymmdd')
    where time_process > to_char('01-MAY-09','yyyymmdd')
    ERROR at line 2:
    ORA-01722: invalid number
    If I execute the below query its working fine, I am facing problem only when I pass the date explicitly to the query. Please help me on converting the date to number format.
    select * from etl_process_status
    where time_process > '20090501'
    and rownum < 3;
    Thanks

    as others have said, it would be best to have your time_process column stored as a date in oracle, if it contains dates.
    The reason for this is because if you store your dates as a number, you have removed information from Oracle that says "This is a date", so when Oracle comes to try and estimate the number of rows greater than a specific value, it can no longer guess correctly, since it's basing its guess on that column being a number.
    There are lots more numbers between '20091201' and '20091101' than there are days between 1st November and 1st December.
    By hiding the fact that your data is a date, you could throw your execution plan off and end up with sub-optimal performance.
    In short, store your data in the correct column format!

  • Help needed in stored procedures

    Hi,
    I wrote a stored procedure when i execute it. i get the error
    ORA-00942: table or view does not exist
    ORA-06512: at "SANDEEP.PROC_INS_TEMP_NAC_GL", line 18
    ORA-06512: at line 9
    I was unable to resolve this error please help.
    CREATE OR REPLACE PROCEDURE PROC_INS_TEMP_NAC_GL(
    p_bu IN STRING,
    p_period IN STRING
    IS
    sql_stmt varchar2(1000);
    cnt number;
    BEGIN
         SELECT count(view_name) INTO cnt FROM sys.all_views where view_name = UPPER(p_bu||'_v_ldga') and owner = 'SUN';
         sql_stmt := 'INSERT INTO temp_nac_gl'||
         ' SELECT ''' || p_bu ||'''BU, gl.ACCNT_CODE , gl.ANAL_T0 , gl.ANAL_T1 , gl.PERIOD , gl.MEMO_AMT , gl.AMOUNT , gl.REGION, gl.ANAL_T3'||
                   ' FROM sun.'|| p_bu ||'_v_ldga gl, temp_acnt_grp acnt' ||
                   ' WHERE gl.accnt_code = acnt.acnt_code ' ||
                   ' AND acnt.bu = '''|| p_bu ||'''' ||
                   ' AND gl.period <= '''|| p_period || '''' ||
                   ' ORDER BY gl.REGION';
    EXECUTE IMMEDIATE sql_stmt;
    END;
    /

    If you are accessing a object from a Oracle procedure/function then you need direct rights and not through roles.
    You can check this by executing a SELECT from SQL plus , if you are able to see the table then type the following commond
    SQL> set role none;
    SQL> SELECT statement
    If it gives an error then you dont have direct rights.
    Thanks
    Edited by: Himanshu Kandpal on May 14, 2009 5:04 PM

Maybe you are looking for

  • SL & USB error message

    I use this particular Mac as a workstation for Logic Pro. I have two USB cables connected to the computer. One is for a powered hub (hosting a midi interface and an iLok key) and the second is for an audio interface. These devices are plugged into a

  • JBI Installation Error on Windows

    The problem: JBI default install fails on windows. We have solved the problem with a workaround, listed below. Our configuration is: * JBI 1.0 * Application server Q2 * Windows 2000 We get the following symptoms: * After JBI is installed, the applica

  • To Load ALV Layout in report selection screen

    Hi... I want the selection screen option for user so that he can select any of the saved layouts. I have done some coding after refering to some threads in here.But I can't proceed further...plz help <b>selection screen</b> SELECTION-SCREEN BEGIN OF

  • I am having trouble receiving large files using a T1 line.

    I can receive large files (3-10meg) using any connection other than T1.  Is there a setting I need to change on my Macbook pro?

  • Crystal Reports Server 2008 - Cant run Oracle Reports

    Hey All I have setup Crystal Reports Server 2008 on Microsoft Windows Server 2003 SP2. Oracle 9 is currenlty installed on the server and the TNSNAMES.ORA file has been configured so that it points to the correct server. The ODBC connection has been s