FI Query on 0FIGL_C01

Dear BI Gurus,
This is my first SAP BI project and I'm responsible for creating reports based on InfoCube  0FIGL_C01 : General Ledger :Transaction Figures using DataSource 0FI_GL_1.
Now my question is , I've replicated the DataSource on the BI side , created Transformation and created InfoPackage and DTP successfully, the Data is in the InfoCube but the GL account is not displaying the data.
1.How do I display the GL account details , they are only showing in number EX: GL Account 110100
I want the actual detail like  Expenses  , Salaries etc.
Or if this scenario is not feasible I'm fine with any other scenario as well , this is for Proof of Concept ( POC )
Regards,
-Neha

Hey Prasad ,
I did some research yesterday and find out the following.
1.I need to load the following Masterdata :
0GL_ACCOUNT_ATT
0GL_ACCOUNT_TEXT
0GL_ACCOUNT_T011_HIER
in order to run my query (0FIGL_VC1_Q0001: Balance Sheet and P&L: Actual/Actual Comparison  ) based on 0FIGL_VC1 InfoCube.
I've already loaded the Data into InfoCube 0FIGL_C01 , can you suggest the following.
1.Loading the above mentioned masterdata to SAP BI
2.How to load the data into Virtual InfoCube 0FIGL_VC1 in order to satisfy my requirement.
I really appreciate your time and help
Best Regards,
-Neha

Similar Messages

  • Query regarding Opened/Closed GL periods

    Hello Gurus,
    ive got an issue, when using 0FIGL_C01 Cube.
    The query ive created balances on cumulative balance (as key figure) and GL account names (as hierarchy).
    Now i want to show the data of cumulative balance regarding closed periods.
    I.e. if we have current April month the balance should be 0, but all other in the past January, February, March should be displayed.
    How it is possible to do this?
    Ive tryed to create a calculated key figure, whith code something like:
    IF debit and credit == 0 then Cummulative balance=0 else Cumulative balance.
    It works fine but it shows also the postings of current period, which isnt realy closed.
    Tried to search the indicators of FI open/closed periods, but the only thing i found is S_ALR_87003642 - Open and Close Posting Periods transaction showing Company code, GL account ranges and from to period data.
    What could be the sollution for this? Should i create new DS for the mensioned TC and then to create a multicube or perhaps is a better and faster way in doing it?
    Thanks a lot in advice
    Laurynas Prikockis

    Hi Shrikant,
    >
    Shrikant Varma wrote:
    > Yes, I could consider using belnr selection to reduce the size of the selection. This will again be a selection more
    > based on knowledge of application(of course with fine tuning of packet size).
    >  But then this approach will not be as efficient and as easy as letting database take control of the iteration of this
    > huge dataset - as is intended by the use of cursors with the help of packet size.
    >
    I'm not sure what you mean here: "let the database contorl the iteration of this huge dataset - with cursors".
    Generally speaking you have 2 options when it comes to parallel processing:
    Parallelize processes in the ABAP stack. These processes access different, non-overlaping, parts of your restult set
    in parallel. Using the BELNR could be an approach to build different, non-overlapping paackages.
    Database side parallelization. This is possible for every action that uses multiblock I/O (full table scans and index fast full scans) and for partitioned segments. For non-partitioned segments and single block I/O database side parallelization is not possible.
    So, generally speaking, you have 3 options:
    1.) parallelize in ABAP and build idependent packages
    2.) force the DB to do multiblock I/O (full table scan, index fast full scan) in parallel
    3.) partition your table and indexes.
    I think (it's just my opinion) for an ERP (not BW / BI) system
    - option 3 is not a good option for OLTP systems (several reasons). I see it very, very, very rarely or close to never in OLTP systems
    - option 2 may read (much) more data then you would need to read, although a very efficient I/O method would be used (multibock I/O)
    - option 1 requires programming in ABAP but will read exactly the data you need to read and can be controlled very well in the ABAP stack (parallel single block I/O will be used)
    In alamost all cases (that are similar to your case) i have see option 1 was used.
    Kind regards,
    Hermann

  • How to Include 0GLACCEXT characteristic (FSV) into Query?

    I am trying to produce a Balance Sheet, I have the data coming in from GL and I have used Z cubes but I have copied all the data from Standard BI Content Cubes,  I have also created a Z Multi Provider based on the 0FIGL_VC1 virtual cube.
    In the query designer, I can see all the chars but the 0GLACCEXT (which is the FSV), I have uploaded the master data to this CHAR and I was going to include it in the rows section, but when I look through all the dimentions of the InfoProvider on the left hand pane ( of the BI7 Query designer) I can not find this CHAR, any clues or ideas as to what has happened?
    Please let me know if you have come accross it and how you resolved it... thanks

    0GLACCEXT does not exists in 0FIGL_C01, but exists in 0FIGL_VC1. There is a function module (someting like this /xxxxx/BCT_FIGL_DATA_GET_VC01) between these two, that has the logic to feed data for 0GLACCEXT from 0FIGL_C01.
    I have used Z cubes but I have copied all the data from Standard BI Content Cubes, I have also created a Z Multi Provider based on the 0FIGL_VC1 virtual cube.
    Looks like u have ...  Z cube --> 0FIGL_VC1 --> Z Multi Provider 
    In that case, u have to copy that function module between Z cube & 0FIGL_VC1, so that 0GLACCEXT (0FIGL_VC1) gets filled.
    To see that function module..
    Double click on 0FIGL_VC1 --> goto change mode --> click on Information (blue ' i ' ) icon --> click on "Type/Namespace" tab --> click "details" tab
    Here u can see function module ( someting like this /xxxxx/BCT_FIGL_DATA_GET_VC01).
    If u check standard queries... all the queries related to balance sheet, defined on virtual provider (0FIGL_VC1, 0FIGL_VC2 & 0FIGL_VC10), but not on the cubes. The reason behind this is, only the virtual provider consists of 0GLACCEXT

  • Doubt in Cash flow Statement Query

    Hi all,
        In FI module i am trying to create query 'Cash flow statement' using infocube 0figl_c01.
    For 'cash inflow' i restricted to debit postings for G/L account char. with two account number (200602 & 200611 as given by finance dept.) But the values are not matching as it extracts all the entries which come to that account for the posting period. But my requirement is to take only the entries related to document type 'DZ - Customer Payments' (for which there is no char. in the infocube or in the R3 Extractor).
    How to solve the issue. do i have to create a generic data source and change the infocube?
    Thanks,
    Prakash

    Hi Vachan,
    Thx for the reply.
    The entries in the InfoCube is correct. Is there any other method by which i can solve this problem. Because changing the datasource and InfoCube is tedious at this time.
    Regards,
    Prakash

  • Cant see the infocube in the Query Designer

    Hi,
    I am working with 0FI_GL_1 datasource. Activated the respective business content in BW and loaded the data into 0FIGL_C01 infocube.
    when I go to query designer to create a query, I dont see this infocube. Is this permission (security) issue? Or Do i need to anythingelse?
    Thanks a lot
    Sri

    Hi Sri
    after clicking on new(blank page icon) to create new query, if ur not able to see ur cube, just click on binaculor icon i.e to to find and type ur cube tech. name and search, then u can find ur cube.
    Hope it helps..........let me know
    Regards
    NR
    *Assign points for helpfull answers
    Message was edited by: NR

  • 0FIGL_C01 Financial Statement Hierarchy 0GLACCEXT

    Hi Guys,
    I am using 0FIGL_C01 to develop a projectwise P&L report.
    I have loaded the 0GLACCEXT hierarchy and have added the object in cube. I have designed the query using FSV restriction.
    But when I execute the report I am not getting the data against any node.
    When I see the data for any GL, I am getting the data in the report. Could you please help me identify & rectify the problem.
    Regards
    Dolly

    P&L report should always be created on 0FIGL_V01, not on 0FIGL_C01.
    0FIGL_V01 has 0GLACCEXT in it and u should not include 0GLACCEXT in 0FIGL_C01.
    There is a function module RS_BCT_FIGL_DATA_GET_VC1. when u execute P&L reports, data would get pulled from 0FIGL_C01 to 0FIGL_V01 through this FM where the 0GLACCEXT logic is defined.
    So when u install 0FIGL_V01 (indataflow before) make sure this function module gets installed.

  • Problem when activating Query from Business Content

    Hi Gurus,
    I'm currently encountering problems when activating a query (Financial Statement and P&L: Actual/Actual Comparison) in the business content under these 2 cubes (0FIGL_VC1 and 0FIGL_VC2). Since I was told that these cubes are Virtual Cubes I also have already activated Basis Cubes (0FIGL_C01 and 0FIGL_C02)  as a prerequisite of the 2 Virtual Cubes. I have also loaded the data for the 2 cubes and have encountered no problem at all.
    But while activating the said Query (Financial Statement and P&L: Actual/Actual Comparison). I have encountered a problem during activation and only this error message appears 1. Error when activating element DO9NE34DKNREWIBWNVAW8BS7M (Technical Name of the Query) and 2. BEx transport request '' is not available or not suitable.
    What other possible steps am I missing here?Is there anything else I should do or activate before activating the Query. Thank you again for all the help.
    - Kit

    Hi Kristian,
    Do the following ->
    Goto transaction RSA1 -> Transport Connection -> Edit -> Transport -> Switch On Standard.
    After this try activating your Query again.
    Hope this helps!
    Regards,
    Saurabh

  • Output coming wrong when Query  executed

    Hi,
    I Extracted herarchial data TO  infocube 0FIGL_C01.
    I also installed Standard Query 0FIGL_VC1_Q0001.
    Now, all data is coming correctly upto infocube 0FIGL_C01 AND TO 0FIGL_VC01.
    (as you know ,the data flow to virtual cube ,once we  load it to standard info cube ).
    But when query is executed ,It is displaying wrong output .
    Can any one tell why ?
    Can you tell what internal execution take place.ie..whether any  fumnction module is executed when the standard query is
    executed .How the data  flow from infocube to virtual infocube and to query?
    (all objects i used are standard )
    Thanks,
    Naresh.

    Hi,
    It should give correct values, see if you are missing something while reconciling.
    Have a look at this FM:RS_BCT_FIGL_DATA_GET_VC1
    Go through this document:
    [Handling Balance Sheets and Profit and Loss Statements in SAP BW 3.0|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/2f6d9790-0201-0010-dcbc-e099c7c049a2]
    Hope it helps...
    Regards,
    Ashish

  • Error while running a query-Input for variable 'Posting Period is invalid

    Hi All,
    NOTE: This error is only cropping up when I input 12 in the posting period variable selection. If I put in any other value from 1-11 I am not getting any errors. Any ideas why this might be happening?
    I am getting the following error when I try and run a query - "Input for variable 'Posting Period (Single entry, mandatory)' is invalid" - On further clicking on this error the message displayed is as follows -
    Diagnosis
    Variable Posting Period (Single Value Entry, Mandatory) is used as a lower limit (X) and an upper limit () in an interval selection. This limit has the value #.
    System Response
    Procedure
    Enter a different value for variable Posting Period (Single Value Entry, Mandatory). If the value of the other limit is determined by another variable, you can change its value also.
    Procedure for System Administration

    OK.
    Well, if the variable is not used in any interval selection, then I would say "something happened to it".
    I would make a copy of the query and run it to check if I get the same problem with period 12.
       -> If not, something is wrong in the original query (you can proceed as below, if changes to original are permitted).
    If so, then try removing the variable completely from the query and hardcode restriction to 12.
       -> If problem still persists, I would have to do some thinking.
    If problem is gone, then add the variable again. Check.
       -> If problem is back, then the variable "is sick". Only quick thing to do, is to build an identical variable and use that one.
    If problem also happens with the new variable, then it's time to share this experience with someone else and consider raising an OSS.
    Good luck!
    Jacob
    P.S: what fisc year variant are you using?
    Edited by: Jacob Jansen on Jan 25, 2010 8:36 PM

  • Logical database in adhoc query

    Hello All,
    Can anyone tell me what is the logical database in adhoc query?

    Hi
    When you create a query , you have to select an infoset. Infoset can be considered as a source from which data is populated in the Query Fields.
    Infosets are created from Transaction SQ02.
    There can be four methods through which an Infoset can become a source of data:
    1.  Table join ( By joining two or more tables from Data dictionary)
         example: Joining tables PA0001 and PA0006 on Pernr to get a one resultant dataset
    2. Direct read of Basis Table ( Like PA0001 as a source for data in Infoset )
    3. Logical Database ( A Pre-written Program by SAP that extract data from clusters, tables taking care of authorizations and validity periods)
    Example : Logical database PNP, PNPCE (Concurrent Employement),PCH ( LDB for Personnel Development Data)
    Custom Logical DBs can be created in T_Code SE-36.
    4. Data Retrieval by a Program ( Custom code written by ABAP developers which will collect and process data) . This program has a corresponding Structure in data dictionary and the fields of this structure will be used in query)
    Reward Points, if helpful.
    Regards
    Waseem Imran

  • Query help on Goods Receipt Query with AP Invoice

    Looking for a little help on a query.  I would like to list all the goods receipts for a given date range and then display the AP Invoice information (if its been copied to an AP Invoice).  I think my problem is in my where clause, I plagerized an SAP query to show GR and AP from a PO as a start.  SBO 2005 SP01.  Any help would be great appreciated.  Thanks
    SELECT distinct 'GR',
    D0.DocStatus,
    D0.DocNum ,
    D0.DocDate,
    D0.DocDueDate,
    D0.DocTotal,
    'AP',
    I0.DocStatus,
    I0.DocNum ,
    I0.DocDate,
    I0.DocDueDate,
    I0.DocTotal,
    I0.PaidToDate
    FROM
    ((OPDN  D0 inner Join PDN1 D1 on D0.DocEntry = D1.DocEntry)
    full outer join
    (OPCH I0 inner join PCH1 I1 on I0.DocEntry = I1.DocEntry)
    on (I1.BaseType=20 AND D1.DocEntry = I1.BaseEntry AND D1.LineNum=I1.BaseLine))
    WHERE
    (D1.BaseType=22 AND D1.DocDate>='[%0]' AND D1.DocDate<='[%1]')
    OR (I1.BaseType=20 AND I1.BaseEntry IN
    (SELECT Distinct DocEntry
    FROM PDN1 WHERE BaseType=22 AND DocDate>='[%0]' AND DocDate<='[%1]'))

    Hi Dalen ,
    I  believe it is because of the condition
    (D1.BaseType=22 AND D1.DocDate>='%0' AND D1.DocDate<='%1')
    OR (I1.BaseType=20 AND I1.BaseEntry IN
    (SELECT Distinct DocEntry FROM PDN1 WHERE PDN1.BaseType=22 AND DocDate>='%0' AND DocDate<='%1'))
    Try changing
    D1.BaseType=22 OR D1.DocDate>='%0' AND D1.DocDate<='%1
    PDN1.BaseType=22 OR DocDate>='%0' AND DocDate<='%1'))
    Lets see what would be the result . Lets have some fun with troubleshooting
    See what would be the difference in the result .
    Thank you
    Bishal

  • Can you check for data in one table or another but not both in one query?

    I have a situation where I need to link two tables together but the data may be in another (archive) table or different records are in both but I want the latest record from either table:
    ACCOUNT
    AccountID     Name   
    123               John Doe
    124               Jane Donaldson           
    125               Harold Douglas    
    MARKETER_ACCOUNT
    Key     AccountID     Marketer    StartDate     EndDate
    1001     123               10526          8/3/2008     9/27/2009
    1017     123               10987          9/28/2009     12/31/4712    (high date ~ which means currently with this marketer)
    1023     124               10541          12/03/2010     12/31/4712
    ARCHIVE
    Key     AccountID     Marketer    StartDate     EndDate
    1015     124               10526          8/3/2008     12/02/2010
    1033     125               10987         01/01/2011     01/31/2012  
    So my query needs to return the following:
    123     John Doe                        10526     8/3/2008     9/27/2009
    124     Jane Donaldson             10541     12/03/2010     12/31/4712     (this is the later of the two records for this account between archive and marketer_account tables)
    125     Harold Douglas               10987          01/01/2011     01/31/2012     (he is only in archive, so get this record)
    I'm unsure how to proceed in one query.  Note that I am reading in possibly multiple accounts at a time and returning a collection back to .net
    open CURSOR_ACCT
              select AccountID
              from
                     ACCOUNT A,
                     MARKETER_ACCOUNT M,
                     ARCHIVE R
               where A.AccountID = nvl((select max(M.EndDate) from Marketer_account M2
                                                    where M2.AccountID = A.AccountID),
                                                      (select max(R.EndDate) from Archive R2
                                                    where R2.AccountID = A.AccountID)
                   and upper(A.Name) like parameter || '%'
    <can you do a NVL like this?   probably not...   I want to be able to get the MAX record for that account off the MarketerACcount table OR the max record for that account off the Archive table, but not both>
    (parameter could be "DO", so I return all names starting with DO...)

    if I understand your description I would assume that for John Dow we would expect the second row from marketer_account  ("high date ~ which means currently with this marketer"). Here is a solution with analytic functions:
    drop table account;
    drop table marketer_account;
    drop table marketer_account_archive;
    create table account (
        id number
      , name varchar2(20)
    insert into account values (123, 'John Doe');
    insert into account values (124, 'Jane Donaldson');
    insert into account values (125, 'Harold Douglas');
    create table marketer_account (
        key number
      , AccountId number
      , MktKey number
      , FromDt date
      , ToDate date
    insert into marketer_account values (1001, 123, 10526, to_date('03.08.2008', 'dd.mm.yyyy'), to_date('27.09.2009', 'dd.mm.yyyy'));
    insert into marketer_account values (1017, 123, 10987, to_date('28.09.2009', 'dd.mm.yyyy'), to_date('31.12.4712', 'dd.mm.yyyy'));
    insert into marketer_account values (1023, 124, 10541, to_date('03.12.2010', 'dd.mm.yyyy'), to_date('31.12.4712', 'dd.mm.yyyy'));
    create table marketer_account_archive (
        key number
      , AccountId number
      , MktKey number
      , FromDt date
      , ToDate date
    insert into marketer_account_archive values (1015, 124, 10526, to_date('03.08.2008', 'dd.mm.yyyy'), to_date('02.12.2010', 'dd.mm.yyyy'));
    insert into marketer_account_archive values (1033, 125, 10987, to_date('01.01.2011', 'dd.mm.yyyy'), to_date('31.01.2012', 'dd.mm.yyyy'));
    select key, AccountId, MktKey, FromDt, ToDate
         , max(FromDt) over(partition by AccountId) max_FromDt
      from marketer_account
    union all
    select key, AccountId, MktKey, FromDt, ToDate
         , max(FromDt) over(partition by AccountId) max_FromDt
      from marketer_account_archive;
    with
    basedata as (
    select key, AccountId, MktKey, FromDt, ToDate
      from marketer_account
    union all
    select key, AccountId, MktKey, FromDt, ToDate
      from marketer_account_archive
    basedata_with_max_intervals as (
    select key, AccountId, MktKey, FromDt, ToDate
         , row_number() over(partition by AccountId order by FromDt desc) FromDt_Rank
      from basedata
    filtered_basedata as (
    select key, AccountId, MktKey, FromDt, ToDate from basedata_with_max_intervals where FromDt_Rank = 1
    select a.id
         , a.name
         , b.MktKey
         , b.FromDt
         , b.ToDate
      from account a
      join filtered_basedata b
        on (a.id = b.AccountId)
    ID NAME                     MKTKEY FROMDT     TODATE
    123 John Doe                  10987 28.09.2009 31.12.4712
    124 Jane Donaldson            10541 03.12.2010 31.12.4712
    125 Harold Douglas            10987 01.01.2011 31.01.2012
    If your tables are big it could be necessary to do the filtering (according to your condition) in an early step (the first CTE).
    Regards
    Martin

  • Query help : Query to get values SYSDATE-1 18:00 hrs to SYSDATE 08:00 hrs

    Hi Team
    I want the SQl query to get the data for the following comparison : -
    Order Created is a Date Column , and i want to find out all the values from (SYSDATE-1) 18:00 hours to SYSDATE 08:00 hours
    i.e.
    (SYSDATE-1) 18:00:00 < Order.Created < SYSDATE 08:00:00.
    Regards

    Hi, Rohit,
    942281 wrote:
    If i want the data in the below way i.e.
    from (SYSDATE-1) 18:00 hours to SYSDATE 17:59 hours ---> (SYSDATE-1) 18:00:00 < Order.Created < SYSDATE 07:59:00.If you want to include rows from exactly 18:00:00 yesterday (but no earlier), and exclude rows from exatly 08:00:00 today (or later), then use:
    WHERE   ord_dtl.submit_dt  >= TRUNC (SYSDATE) - (6 / 24)
    AND     ord_dtl.submit_dt  <  TRUNC (SYSDATE) + (8 / 24)
    So can i use the below format : -
    ord_dtl.submit_dt BETWEEN trunc(sysdate)-(6/24) and trunc(sysdate)+(7.59/24) . Please suggest . .59 hours is .59 * 60 * 60 = 2124 seconds (or .59 * 60 = 35.4 minutes), so the last time included in the range above is 07:35:24, not 07:59:59.
    If you really, really want to use BETWEEN (which includes both end points), then you could do it with date arithmentic:
    WHERE   ord_dtl.submit_dt  BETWEEN  TRUNC (SYSDATE) - (6 / 24)
                      AND         TRUNC (SYSDATE) + (8 / 24)
                                               - (1 / (24 * 60 * 60))but it would be simpler and less error prone to use INTERVALs, as Karthick suggested earlier:
    WHERE   ord_dtl.submit_dt  BETWEEN  TRUNC (SYSDATE) - INTERVAL '6' HOUR
                      AND         TRUNC (SYSDATE) + INTERVAL '8' HOUR
                                               - INTERVAL '1' SECONDEdited by: Frank Kulash on Apr 17, 2013 9:36 AM
    Edited by: Frank Kulash on Apr 17, 2013 11:56 AM
    Changed "- (8 /24)" to "+ (8 /24)" in first code fragment (after Blushadown, below)

  • Query help, subtract two query parts

    Hi,
    I am beginner of PL/SQL and have a problem I couldn’t solve:
    Table (op_list):
    Item     -     Amount -     Status
    Item1     -     10     -     in
    Item2     -     12     -     in
    Item3     -     7     -     in
    Item1     -     2     -     out
    Item2     -     3     -     out
    Item1     -     1     -     dmg
    Item3     -     3     -     out
    Item1     -     2     -     out
    Item2     -     5     -     out
    Item2     -     2     -     in
    Item3     -     1     -     exp
    Would like to get result of query (subtract amount of 'out/dmg/exp' from 'in' ):
    Item - Amount left
    Item1     -     5
    Item2     -     6
    Item3 -     3
    I wrote code that returns sum of all incoming items and sum all out/dmg/exp items, but couldn’t solve how to subtract one part of querry from another. Or maybe there is a better way. Also worried what happens if there is no 'out/dmg/exp' only 'in'
    select item.name, sum(op_list.item_amount)
    from op_list
    inner join item
    on op_list.item = item.item_id
    where op_list.status = 'in'
    group by item.name
    union
    select item.name, sum(op_list.item_amount)
    from op_list
    inner join item
    on op_list.item = item.item_id
    where op_list.status = 'out'
    or op_list.status = 'dmg'
    or op_list.status = 'exp'
    group by item.name
    Return:
    Item1     -     10      [10 in]
    Item1     -     5     [2+1+2]
    Item2     -     14     [12+2]
    Item3     -     7
    Item3     -     4     [3+1]
    Thanks in advance

    Hi,
    We can also use simple inline views to get what we need.
    select a.item,a.amount-b.amount Balance from
    (select item,sum(amount) Amount from op_list
    where status = 'in'
    group by item) a,
    (select item,sum(amount) Amount from op_list
    where status in ('out','dmg','exp')
    group by item) b
    where
    a.item=b.item
    order by item;
    ITEM       BALANCE
    Item1                      5
    Item2                      6
    Item3                      3Regards,
    Prazy

  • Query help: query to return column that represents multiple rows

    I have a table with a name and location column. The same name can occur multiple times with any arbitrary location, i.e. duplicates are allowed.
    I need a query to find all names that occur in both of two separate locations.
    For example,
    bob usa
    bob mexico
    dot mexico
    dot europe
    hal usa
    hal europe
    sal usa
    sal mexico
    The query in question, if given the locations usa and mexico, would return bob and sal.
    Thanks for any help or advice,
    -=beeky

    How about this?
    SELECT  NAME
    FROM    <LOCATIONS_TABLE>
    WHERE   LOCATION IN ('usa','mexico')
    GROUP BY NAME
    HAVING COUNT(DISTINCT LOCATION) >= 2Results:
    SQL> WITH person_locations AS
      2  (
      3          SELECT 'bob' AS NAME, 'USA' AS LOCATION FROM DUAL UNION ALL
      4          SELECT 'bob' AS NAME, 'Mexico' AS LOCATION FROM DUAL UNION ALL
      5          SELECT 'dot' AS NAME, 'Mexico' AS LOCATION FROM DUAL UNION ALL
      6          SELECT 'dot' AS NAME, 'Europe' AS LOCATION FROM DUAL UNION ALL
      7          SELECT 'hal' AS NAME, 'USA' AS LOCATION FROM DUAL UNION ALL
      8          SELECT 'hal' AS NAME, 'Europe' AS LOCATION FROM DUAL UNION ALL
      9          SELECT 'sal' AS NAME, 'USA' AS LOCATION FROM DUAL UNION ALL
    10          SELECT 'sal' AS NAME, 'Mexico' AS LOCATION FROM DUAL
    11  )
    12  SELECT  NAME
    13  FROM    person_locations
    14  WHERE   LOCATION IN ('USA','Mexico')
    15  GROUP BY NAME
    16  HAVING COUNT(DISTINCT LOCATION) >= 2
    17  /
    NAM
    bob
    salHTH!
    Edited by: Centinul on Oct 15, 2009 2:25 PM
    Added sample results.

Maybe you are looking for

  • Adobe Photoshop Elements 11 People Recongnition

    So maybe I'm using this wrong or maybe this is just how it works.  I have about 5,000 photos, I imported them all into Photoshop Elements 11.  On the main screen, I click Add People and it starts going through all my photos,  but I seem to be getting

  • US Sales Tax Report is running for more than 10 hours

    Hi All, Could any one please suggest us on below issue. We have observed that "U.S. Sales Tax Report" Concurrent Program is running for more than 10 hours in R12 Production. The system has been upgraded from 11.5.8 to R12. Earlier in 11i, it was runn

  • REMOVING  ADMINISTRATOR FROM iMac G4 Flat Screen

    I have a new MacBook Pro and have given my iMac G4 Flatscreen to my wife for her personal use. I was the Original Administrator on the machine and she was a user. She is now also an Administrator and I wan't to delete all reference to my account and

  • Payment Cards - Orders with Billing Plan - Authorization Logic

    Hello Experts, Generally, payment card authorization requests are triggered when an order is saved.  How does authorization work for orders with Billing Plans?  Thanks. Regards, Jess A.

  • OBIEE Conformed dimension with Bridge Table

    Hi, I have an issue and tried all the links from Mark and Gerad regarding bridge tables. But probably this is a bit different. I have to extend the OOB data model for OBIA - where the relationship between group account and gl account dimensions are 1