Writing PS Queries against the PS_JOB table results in runaways

When we use PS query to write a query against the PS_JOB table it adds in joins to security tables that causes the query to become a runaway. Has anyone else noticed this?
Has anyone found a workaround ?
Using HCM 9.0 with Oracle DB
Edited by: allenc on Feb 10, 2011 10:42 AM

Are all of your users having the problem or just the power users?
How many active/total job rows do you have?
We have about 20,000 active job rows and 750,000 total job rows. Our departmental clerks don't have any problems going against job data but the power users have issues. We cloned the job record as a view. Then we removed the query security from our view and added this record to the query tree that only the power users had access to. That fixed our problems.

Similar Messages

  • Is it possible to filter the nested table result set of table column

    Hi
    Create or replace type address_record
    as object
    ( address_no number,
    address_name varchar2,
    address_startDate date,
    address_expiryDate date
    create or replace type address_rec_tab
    as table of address_record;
    Create table employee
    (emp_no number,
    emp_name varchar2,
    adresses address_rec_tab
    1st approach
    ==========
    <pre>
    select
    emp.emp_no,
    emp.emp_name,
    emp.addresses
    from employee emp,
    table(*emp.addresses*) add
    where add.address_expiryDate >=sysdate
    </pre>
    In the above example my SQL query address collection object is not returning filtered or current address list.
    I suppose this is due to fact taht my where clause is not attached to the nested table.
    Through my reading I gather that I can only use the following query to filter the address collection.
    2nd approach
    ==========
    <pre>
    select
    emp.emp_no,
    emp.emp_name
    cursor(select address_no,
    address_name,
    address_startDate,
    address_expiryDate
    from employee emp,
    table (*emp.addresses*) add
    where add.address_expiry_date >=sysdate)
    from employee emp,
    table (*emp.addresses*) add
    where add.address_expiry_date >=sysdate) -- probably this redundent
    </pre>
    But this approch forces me to rebuild addresses collection object.
    I was wondering anybody can suggest me a way so that 1st approach works? I do not have to rebuild collection object in this way.
    Thanks for your help in advance
    Regards
    Charan

    Create statements have been slightly modified;
    Create or replace type address_record as object
    ( address_no          number,
      address_name        varchar2(20),
      address_startDate   date,
      address_expiryDate  date
    create or replace type address_rec_tab as table of address_record;
    Create table employee
    (emp_no number,
    emp_name varchar2(20),
    add_list address_rec_tab
    nested table add_list store as a_list
    insert into employee values (1, 'KMCHARAN', address_rec_tab ( address_record(1, 'NORTH POLE', trunc(sysdate-1), trunc(sysdate+10) ) ,
                                                                  address_record(1, 'SOUTH_POLE', trunc(sysdate-1), trunc(sysdate+10) )
    insert into employee values (2, 'ME', address_rec_tab ( address_record(2, 'EAST', trunc(sysdate-2), trunc(sysdate+12) ) ,
                                                            address_record(2, 'WEST', trunc(sysdate-2), trunc(sysdate+12) )
    SQL> l
      1  select *
      2  from employee
      3      ,table(add_list) a
      4* where a.Address_StartDate = trunc(sysdate-1)
    SQL> /
        EMP_NO EMP_NAME
    ADD_LIST(ADDRESS_NO, ADDRESS_NAME, ADDRESS_STARTDATE, ADDRESS_EXPIRYDATE)
    ADDRESS_NO ADDRESS_NAME         ADDRESS_S ADDRESS_E
             1 KMCHARAN
    ADDRESS_REC_TAB(ADDRESS_RECORD(1, 'NORTH POLE', '08-APR-10', '19-APR-10'), ADDRESS_RECORD(1, 'SOUTH_
             1 NORTH POLE           08-APR-10 19-APR-10
             1 KMCHARAN
    ADDRESS_REC_TAB(ADDRESS_RECORD(1, 'NORTH POLE', '08-APR-10', '19-APR-10'), ADDRESS_RECORD(1, 'SOUTH_
             1 SOUTH_POLE           08-APR-10 19-APR-10

  • Two queries against aud$ with different results

    hi guys
    I'm not so good with queries and that is the reason for my question:
    We have audit activated and we want to get the following information (monthly):
    - how many times the user logon to database.
    - every logon of each user in a month.
    For the first (1) requirement we have the following query:
    select USERID "Cuenta", USERHOST, TERMINAL, nombres||' '||PRIMER_APELLIDO||' '||SEGUNDO_APELLIDO "WhiteList"
    , count(*) "TOTAL"
    from aud$, ab.usuarios
    where (ACTION# = 100) and
    (NTIMESTAMP# between (to_date(to_char('01092013 00:00:00'),'ddmmyyyy HH24:MI:SS'))
        and (to_date(to_char('30092013 23:59:00'),'ddmmyyyy HH24:MI:SS'))) and
    USERID = CODIGO_USUARIO(+) and
    USERID not in ('DBSNMP','SYSMAN')
    group by USERID, USERHOST, TERMINAL,  nombres||' '||PRIMER_APELLIDO||' '||SEGUNDO_APELLIDO
    order by USERID, USERHOST, TERMINAL;
    the result of this query shows like this:
    Header 1
    Cuenta               USERHOST                  TERMINAL        WhiteList                                     TOTAL                                                                                     
    ASEGUR               MAQUINADIAB\SSSSS        IUOOO          PEPITO GARCIA Y GARCIA SOCIED                    10                                                                                     
    ASEGUR               POSADASDE\MNNN4550094  MNNN4550094    PEPITO GARCIA Y GARCIA SOCIED                     1                                                                                     
    ASEGUR               POSADASDE\YUMI          YUMI            PEPITO GARCIA Y GARCIA SOCIED                    10                                                                                     
    ASEGUR               YUH                                       PEPITO GARCIA Y GARCIA SOCIED                    20                                                                                     
    ASEGUR               SDFRG                                    PEPITO GARCIA Y GARCIA SOCIED                    13                                                                                     
    ASEGUR               signy                                     PEPITO GARCIA Y GARCIA SOCIED                    29                                                                                     
    ASEGUR               sigurd                                    PEPITO GARCIA Y GARCIA SOCIED                    32                                                                                     
    ASEGUR               valhalla-Legacy                           PEPITO GARCIA Y GARCIA SOCIED                    12                                                                                     
    ADMIN                MAQUINADIAB\SSSSS        IUOOO          USUARIO ADMINISTRADOR NETWORKING                3                                                                                     
    SPRINGUSR            bragi                                                                                      98                                                                                     
    SPRINGUSR            hermod                                                                                     59                                                                                     
    SPRINGUSR            YUH                                                                                        49
    So, is the total logons in a month by user.
    for the second requirement we are using the follow query:
    select USERID "Cuenta", USERHOST, TERMINAL, to_char(NTIMESTAMP#,'YYYYMMDD HH24:MI:SS') "Fec Ing", nombres||' '||PRIMER_APELLIDO||' '||SEGUNDO_APELLIDO "WhiteList"
    --, count(*) "TOTAL"
    from aud$, ab.usuarios
    where (ACTION# = 100) and
    --to_char(NTIMESTAMP#,'dd-mm-yy')=to_char(sysdate-50,'dd-mm-yy') and
    (NTIMESTAMP# between (to_date(to_char('01062013 18:00:00'),'ddmmyyyy HH24:MI:SS'))
        and (to_date(to_char('30062013 23:59:00'),'ddmmyyyy HH24:MI:SS'))) and
    USERID = CODIGO_USUARIO(+) and
    USERID not in ('DBSNMP','SYSMAN')
    group by USERID, USERHOST, TERMINAL,to_char(NTIMESTAMP#,'YYYYMMDD HH24:MI:SS'), nombres||' '||PRIMER_APELLIDO||' '||SEGUNDO_APELLIDO
    order by USERID, USERHOST, TERMINAL;
    Header 1
    Cuenta               USERHOST                  TERMINAL        Fec Ing              WhiteList                                                                                                                                                            
    USER12               DOMINIODDD\IOIPOP        IOIPOP          20130930 12:08:33    ANGEL ROBERTO GARCIA Y GARCIA S                                                                                                                                        
    USER12               DOMINIODDD\IOIPOP        IOIPOP          20130930 14:28:47    ANGEL ROBERTO GARCIA Y GARCIA S                                                                                                                                        
    USER12               DOMINIODDD\IOIPOP        IOIPOP          20130930 16:24:43    ANGEL ROBERTO GARCIA Y GARCIA S 
    so, shows each logon done by user in a month.
    But in the both queries, the results are different. It is not suppose that they have to be the same number of logons?
    I mean, If I sum the numbers in TOTAL column I have less that I get in the second query. Always !!!
    could you help us?
    thank you

    Your timeframes are different:
    (NTIMESTAMP# between (to_date(to_char('01092013 00:00:00'),'ddmmyyyy HH24:MI:SS')) 
    and (to_date(to_char('30092013 23:59:00'),'ddmmyyyy HH24:MI:SS'))) and 
    versus
    (NTIMESTAMP# between (to_date(to_char('01062013 18:00:00'),'ddmmyyyy HH24:MI:SS')) 
        and (to_date(to_char('30062013 23:59:00'),'ddmmyyyy HH24:MI:SS'))) and 

  • SQL Server Profiler does not capture cube processing queries against the relational database

    Hello, experts,
    This my first attempt to use SQL Server Profiler to find out why my cube processing is so slow. Here is what I did:
    My cube sources its data from a SQL Server database. While the cube processing is going, I launched the profiler connecting to the database engine (note: not the Analysis Services). I  choose the "Tuning" template and started the trace. However,
    I do not see any query activities (such as, SELECT DISTINCT ...) in the profiler trace window.
    Did I do something wrong, or do I have the wrong expectation?
    Thank you much in advance.

    Hi QQFA,
    Agree with Olaf.
    Please make sure that you choose the appropriate events to monitor the cube processing queries when using SQL Server Profiler.
    And I recommend you review the following links to get more information about tuning Analysis Services with SQL Profiler.
    Monitoring and Tuning Analysis Services with SQL Profiler:
    http://www.informit.com/articles/article.aspx?p=1745747
    Analysis Services Trace Events:
    http://msdn.microsoft.com/en-us/library/ms174867.aspx
    Thanks,
    Lydia Zhang

  • How to Query the Audit Tables

    Our Production system is suffering performance issues which we attribute to the large volume of records in the audit tables. We want to implement an archival solution to alleviate this, which would involve transferring older audit records to a separate database. To achieve this, we need to write SQL queries against the <app>DATAAUDIT and <app>TASKAUDIT tables in the HFM database. We are aware that we need to join other tables to see descriptions, e.g. <app>SCENARIODESC. We have also deduced the translation of Activity Codes from "HFMAuditExtractSchemaDefs.xml", which accompanies the HFM Audit Extract Utility.
    Unfortunately, there are some things which we do not know how to translate:
    1) Period (lPeriod) on the <app>DATAAUDIT table, e.g. 67108869 but need to show "Jun" etc.
    2) User (ActivityUserId) on the <app>TASKAUDIT table, e.g. -329543984
    3) Module (strModuleName) on the <app>TASKAUDIT table, e.g. #@RID@#19
    Can anyone advise how we should translate these pieces of data in order to give similar results to those seen on the Data Audit and Task Audit screens in Workspace?

    1.) Period - I'm not sure where this is coming from, but I'm pretty confident it is not sitting in the database which I find odd. I want to find this as well, but haven't had any free time to search around, sorry.
    2.) User - This one I can help you with.
    --Get Task Item Data by Hour for a given day by user
    select datepart(HOUR, cast(endtime-2 as smalldatetime)), count(activitycode) from <APP_NAME>_task_audit where activitycode in (1,4,6,8,9,12,14,15,16,21,23,44) and cast(endtime-2 as smalldatetime) between '2011-12-7 00:00:00' and '2011-12-7 23:59:59' and ActivityUserID in (select lUserID from HSV_ACTIVITY_USERS where sUserName = '<USER_NAME>') group by datepart(HOUR, cast(endtime-2 as smalldatetime))In the sample SQL Query above, I am looking for all of the task audit activity (for the specified tasks) for a specific user by hour. The part of the query you want to focus on is
    and ActivityUserID in (select lUserID from HSV_ACTIVITY_USERS where sUserName = '<USER_NAME>')The ActivityUserID in the task audit table will match up to the lUserID field in the HSV_ACTIVITY_USERS table. sUserName will be the actual text username. Depending on how you are configured, the name may look like johndoe or johndoe@Native Directory or johndoe@AD, etc. If you query that table, you'll figure it out pretty quickly.
    3.) Module ID. I can only partially assist on this one.
    This is not also in the HFM database. I suspected it might be floating in workspace; however, can't really find it there either.
    I can tell you a couple things though....
    #@RID@# Is just a placeholder for Resource ID. The important part is the number behind it. If you do a search on your folders for @RID@, you'll get one hit where it is defined as a constant, but that is it. (I was hoping a file search would reveal a table of the ID's; however, I wasn't that lucky). There is probably a table of values that correspone somewhere, but I haven't found it.
    If you do a distinct select, you'll see there's not too many ...
    select distinct strmodulename from <APP_NAME>_task_audit#@RID@#10
    #@RID@#13
    #@RID@#14
    #@RID@#16
    #@RID@#18
    #@RID@#23
    #@RID@#30
    #@RID@#32
    #@RID@#4294967295
    #@RID@#45
    #@RID@#46
    #@RID@#49
    #@RID@#52
    #@RID@#58
    #@RID@#59
    #@RID@#65
    #@RID@#66
    #@RID@#67
    #@RID@#68
    #@RID@#69
    #@RID@#7
    #@RID@#9
    HyperionReports
    How can you determine these values?
    The quick and dirty way would be to run a Task Audit Extract for a given date range as this will reveal the Module Name.
    Then execute a SQL Query on the Task Audit Table for the same data range. You could then match up the records and deduce the values.
    --Get Task Items by Date Range
    select (select sUserName from HSV_ACTIVITY_USERS where lUserID = ActivityUserID) as UserName, ActivityCode, ServerName, AppName, cast(StartTime-2 as smalldatetime), cast(EndTime-2 as smalldatetime), strDescription, strModuleName from <APP_NAME>_task_audit where cast(endtime-2 as smalldatetime) between '2011-12-7 00:00:00' and '2011-12-7 23:59:59' order by endtime ascOnce you determine the values, I would create a table in your database to hold them so that you could use them in future queries. I would also recommend you do the same with the activitycodes.

  • Writing Custom Queries (Logical SQL to Physical SQL)

    What is the recommended method for writing cross-subject area custom queries against the HCM database? Is there a published data model available for reference?

    Hi Experts,
    Please help on this.

  • Related Content links in the repository tables

    We have 150 FR reports and many of them are using Related Content links. I am trying to find the easiest way to view all the Reports that have Related Content links and which reports they are linking to. I was trying to see about writing a sql against the tables but need to know which tables to join and what fields to select.
    If anyone has a sql they have built to do similar selects or a feature in EPM that I am missing let me know.
    Thanks!
    Robert

    Indeed, we have the same issue with our Solaris 10 containers.
    None of the network interface are visible since we upgraded to Solaris 11.1.
    Highly likely that we will have to fallback to 11.0
    root%A-infmagt004[8] zoneadm -z udsr02 boot
    Dec 7 09:55:41 gvas-infmagt004 mac: [ID 469746 kern.info] NOTICE: net0 registered
    Dec 7 09:55:41 gvas-infmagt004 mac: [ID 469746 kern.info] NOTICE: dmz0 registered
    Dec 7 09:55:41 gvas-infmagt004 mac: [ID 469746 kern.info] NOTICE: vd155 registered
    Dec 7 09:55:41 gvas-infmagt004 mac: [ID 469746 kern.info] NOTICE: vd157 registered
    Dec 7 09:55:41 gvas-infmagt004 mac: [ID 469746 kern.info] NOTICE: vd153 registered
    Dec 7 09:55:41 gvas-infmagt004 mac: [ID 469746 kern.info] NOTICE: vd301 registered
    Dec 7 09:55:43 gvas-infmagt004 dlmgmtd[65]: [ID 183745 daemon.warning] Duplicate links in the repository: net0
    Dec 7 09:55:43 gvas-infmagt004 dlmgmtd[65]: [ID 183745 daemon.warning] Duplicate links in the repository: dmz0
    Dec 7 09:55:43 gvas-infmagt004 dlmgmtd[65]: [ID 183745 daemon.warning] Duplicate links in the repository: vd155
    Dec 7 09:55:43 gvas-infmagt004 dlmgmtd[65]: [ID 183745 daemon.warning] Duplicate links in the repository: vd157
    Dec 7 09:55:43 gvas-infmagt004 dlmgmtd[65]: [ID 183745 daemon.warning] Duplicate links in the repository: vd153
    Dec 7 09:55:43 gvas-infmagt004 dlmgmtd[65]: [ID 183745 daemon.warning] Duplicate links in the repository: vd301
    root%A-infmagt004[4] zlogin udst02 ifconfig -a
    lo0: flags=2001000849<UP,LOOPBACK,RUNNING,MULTICAST,IPv4,VIRTUAL> mtu 8232 index 1
    inet 127.0.0.1 netmask ff000000
    useless.

  • There are no queries in the log for selected measure group

    I configured SSAS 2008 R2 Server to dump every 5th query to olapquerlog table in a new logging database. Data is coming to the log table fine. I have been running it from last two weeks. I have about 175 queries against the cube which i
    want to aggregate. When i run the usage based aggregation wizard, ssas never finds any query for all measure group. i see a warning below that, "There are no queries in the log for selected measure group".
    Can anyone tell me why?
    Aneel Ismaily

    ID and Database Name look same when i open code from cube;
    <ID>Spend
    Analytics DW</ID>
    <Name>Spend
    Analytics DW</Name>
    this is how data looks like in sql table;
    MSOLAP_Database
    MSOLAP_ObjectPath
    MSOLAP_User
    StartTime
    Duration
    1
    SpendAnalyticsDW
    NT1101112\BICORP.SpendAnalytics.Spend Analytics DW.Account Ledger
    ROCKTENN\!sharepoint2010
    6/22/2011
    718
    2
    SpendAnalyticsDW
    NT1101112\BICORP.SpendAnalytics.Spend Analytics DW.Account Payable 1
    ROCKTENN\SDouglas
    6/20/2011
    359
    Note: I have removed actual dataset from the table, because of its length.Aneel Ismaily

  • Updating the DBSTATC table

    When I run brconnect via db13 I noticed that a number of tables are defined as with "c" and stats are run against the entire table.  After looking at the documentation and recomendations I have noticed that several tables are well beyond the number of rows reccomended for "c" and should be set to various estimate levels (e.g. e3, e30...etc).
    We can go in on a table by table basis and set these, but I was wondering if there is some sort of automated way of doing this.  What I have in mind is "script" that would check the number of rows and automatically update the table to the correct parameter ("c", or "e")?
    Thanks,
    Jeff Henke

    Hi Jeff
    Please check sap note 403704 first. There are dbstat sql scripts attached which reset your dbstatc. BR*Connect then chooses the sample size according to the table sizes.
    Regards
    Michael

  • Run queries against system tables for oracle BPEL processes

    I want to run queries against system tables for oracle BPEL processes. It is becoming very difficult for me to us EM as it is very slow,
    and not all the time, it is sufficient for our needs.
    We are doing load testing and we want find out info like how many requests came in and how many faulted and what time is taken by each request...
    So do any of you have the query that I can use and tables that I need to go against?

    Use the BPEL hydration database table "cube_instance".
    There should be plenty of example in the forum regarding this table.

  • Hello Anybody, I have a question. Can any of you please suggest me how to make an xml file from the database table with all the rows? Note:- I am having the XSD Schema file and the resulted XML file should be in that XSD format only.

    Hello Anybody, I have a question. Can any of you please suggest me how to make an xml file from the database table with all the records?
    Note:- I am having the XSD Schema file and the resulted XML file should be in that XSD format only.

    The Oracle documentation has a good overview of the options available
    Generating XML Data from the Database
    Without knowing your version, I just picked 11.2, so you made need to look for that chapter in the documentation for your version to find applicable information.
    You can also find some information in XML DB FAQ

  • How to create Payment against the invoices in R12 using API/Interface table

    Hi Experts,
    I need your help.
    As per my requirement , we need to create the payments against the invoices using the payment information provided.
    Earlier the Payments against the invoice is done by (3rd party system) in 11i.
    But, as a part of R12 implementation,
    The requirement is like now they are going to provide us and flat file(having below payment information) and we need to create the payment with these information.
    COLUMNS PROVIDED
    SOURCE
    INVOICE INTERNAL NUMBER #
    ACCOUNTING SEGMENTS
    DEBIT/CREDIT
    AMOUNT
    CURRENCY
    BANK ACCOUNT
    PAYMENT DATE
    DUE DATE
    PAYMENT METHOD
    PAYMENT TERMS
    BATCH ID
    We need to create an inbound interface to achieve the below requirement.
    1. Create the payments in oracle R12 with the above payment information. (unable to find the way please suggest).
    2. Creating the Journal entries. -> We are having GL_INTERFACE & Journal Import program to import the journals.
    3. Finally we have to posting the same into GL.-- Interface & Import programs available.
    Is there any api or interface tables can be used to create the payment against the invoices in r12?
    Could you please help me to achive my above requirements.
    Any sample code or reference will be a great help for me.
    Waiting for your response.
    Thanks,
    Thiru.

    Hi Thiru,
    Perhaps these links can help:
    http://ccchuanblog.blogspot.in/2010/12/oracle-ap-payment-api.html
    http://oracle.ittoolbox.com/groups/technical-functional/oracle-apps-l/ap-invoice-payment-need-help-4037407
    See "Understanding Oracle Payments APIs" here:
    http://docs.oracle.com/cd/B40089_10/current/acrobat/120ibyig.pdf
    If this answers your question, please mark the post. Thanks.
    Kristofer Cruz
    Edited by: Kristofer Cruz on Sep 5, 2012 2:41 PM

  • Which table i will  get the  status of the User decision Result status

    Hi Experts
    Which table i will  get the  status of the User decision Result status
    (eg : Inspection completion forced )
    This i can see in QA33

    Hi jag
    Ok fine actually it is not storing in any table...these r all text symols...if u see this program MQEVAF22...U EILL GET SOME IDEA
    CALL FUNCTION 'QAST_STATUS_CHECK'
           EXPORTING
                I_OBJNR          = QALS-OBJNR
                I_STATUS         = C_STATUS_PAKO
           EXCEPTIONS
                STATUS_NOT_ACTIV = 01.
      IF SY-SUBRC NE C_RC01.
        MOVE C_KREUZ TO L_STAT_PAKO.
      ELSE.
        CLEAR L_STAT_PAKO.
      ENDIF.
    Prüfen, ob Los bestandsrelevant ist
      MOVE RQEVA-STAT_BERF TO L_STAT_BERF.
    Prüfen, ob Los auf SKIP sitzt
      CALL FUNCTION 'QAST_STATUS_CHECK'
           EXPORTING
                I_OBJNR          = QALS-OBJNR
                I_STATUS         = C_STATUS_SKIP
           EXCEPTIONS
                STATUS_NOT_ACTIV = 01.
      IF SY-SUBRC NE C_RC01.
        MOVE C_KREUZ TO L_STAT_SKIP.
      ELSE.
        CLEAR L_STAT_SKIP.
      ENDIF.
      IF QALS-STAT35 >< C_KREUZ.
    VE noch nicht getroffen !
        IF     ( QALS-OFFENNLZMK > CI_0 OR QALS-OFFEN_LZMK > CI_0 )
           AND QALS-STAT13     = C_PPKTTYP-PHPR
           AND G_ANZAHL_PPKT_LOEVM > CI_0.
    Es wurden physische Proben storniert !              PRIORITÄT 1A
            MOVE TEXT-I17 TO RQEVA-INFO_LINE.  L_TEXT_NR = 'I17'.
        ELSEIF ( QALS-STAT14 = C_KREUZ AND QALS-OFFENNLZMK > CI_0 )
          OR ( L_STAT_PAKO = C_KREUZ
               AND ( QALS-OFFENNLZMK > CI_0 OR QALS-OFFEN_LZMK > CI_0 ) ).
    Prüfung wurde abgebrochen !                         PRIORITÄT 1
          MOVE TEXT-I04 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I04'.
          MOVE C_KREUZ TO L_ICON.
        ELSEIF G_ANZ_RUECK_MK > 0.
    Es wurden Merkmale rückgewiesen  !                  PRIORITÄT 2
          MOVE TEXT-I01 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I01'.
          MOVE C_KREUZ TO L_ICON.
        ELSEIF QALS-OFFENNLZMK > CI_0 OR QALS-OFFEN_LZMK > CI_0 .
          IF QALS-OFFENNLZMK > CI_0.
    Es sind nicht alle Merkmale abgeschlossen !         PRIORITÄT 3
            MOVE TEXT-I02 TO RQEVA-INFO_LINE.  L_TEXT_NR = 'I02'.
          ELSEIF QALS-OFFEN_LZMK > CI_0.
            MOVE TEXT-I05 TO RQEVA-INFO_LINE.  L_TEXT_NR = 'I05'.
          ENDIF.
        ELSEIF QALS-WERK NE QALS-WERKVORG.
        Prüflosmenge wurde umgelagert auf anderes Werk   Priorität 4
          MOVE TEXT-I16 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I16'.
        ELSEIF QALS-LMENGEIST > QALS-LOSMENGE.
        Istmenge im Los größer als Losmenge!             Priorität 5
          MOVE TEXT-I15 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I15'.
        ELSEIF L_STAT_SKIP    = C_KREUZ.
    Prüfverzicht: Los ist geskippt !                    PRIORITÄT 6
          MOVE TEXT-I03 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I03'.
        ENDIF.
      ELSE.
    VE getroffen bzw. eingetragen
       IF     ( QALS-OFFENNLZMK > CI_0 OR QALS-OFFEN_LZMK > CI_0 )
           AND QALS-STAT13     = C_PPKTTYP-PHPR
           AND G_ANZAHL_PPKT_LOEVM > CI_0.
    Es wurden physische Proben storniert !              PRIORITÄT 1A
            MOVE TEXT-I17 TO RQEVA-INFO_LINE.  L_TEXT_NR = 'I17'.
        ELSEIF ( QALS-STAT14 = C_KREUZ AND QALS-OFFENNLZMK > CI_0 )
          OR ( L_STAT_PAKO = C_KREUZ
              AND ( QALS-OFFENNLZMK > CI_0  OR QALS-OFFEN_LZMK > CI_0 ) ).
    Prüfung wurde abgebrochen !                         PRIORITÄT 1
          MOVE TEXT-I04 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I04'.
          MOVE C_KREUZ TO L_ICON.
        ELSEIF L_STAT_PAKO EQ SPACE
           AND ( QALS-OFFENNLZMK > CI_0 OR QALS-OFFEN_LZMK > CI_0 ).
    Es sind noch Langzeitmerkmale zu bearbeiten !       PRIORITÄT 2
          MOVE TEXT-I05 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I05'.
        ELSEIF L_STAT_BERF = C_KREUZ AND QALS-STAT34 EQ SPACE.
    Los ist bestandsmäßig noch nicht entlastet !        PRIORITÄT 3
          MOVE TEXT-I06 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I06'.
        ELSEIF QALS-WERK NE QALS-WERKVORG.
        Prüflosmenge wurde umgelagert auf anderes Werk   Priorität 4
          MOVE TEXT-I16 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I16'.
        ELSEIF QALS-LMENGEIST > QALS-LOSMENGE.
        Istmenge im Los größer als Losmenge!             Priorität 5
          MOVE TEXT-I15 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I15'.
        ELSEIF L_STAT_SKIP = C_KREUZ.
    Prüfverzicht: Los ist geskippt !                    PRIORITÄT 6
          MOVE TEXT-I03 TO RQEVA-INFO_LINE.    L_TEXT_NR = 'I03'.
        ENDIF.
      ENDIF.

  • Loading the different result sets in the same sequence for the target table

    Dear all,
    I have 5 tables say A,B,C,D as my source and i made 3 joins P,Q,R .the result sets of these 3 joins are loading into a target table X but with 3 different targets with same table name.
    I created one sequence say Y as my target table has primary key and mapped to three different targets for the same target table which i need to load.
    But after deployed and executed successfully ,i am able to load the data from three join result sets with differeent sequence numbers.
    I am looking to load data like this.
    If First Result set P has 10 Records,SEcond Result Set Q Has 20 and the third result set has 30 records then while loading data into first target it creates the seq for the 10 records from 1..10 and while loading the data for second result set ,it creates the sequence from 11 ...20 and while loading the third target with the third result set it creates the sequence from 21 ----30.
    But i am looking to load the three result sets in the sequence 1to 10 but not like creating fresh sequence for each result set.
    how can we achieve this in owb?
    any solution for this will be appreciated.
    thank you
    kumar

    My design is like following
    SRC1
    ---->Join1--------------------------->Target1( Table X)<-----Seq1
    SRC2
    SRC3
    ----> Join2----------->Target2(Table X)<----Seq1
    SRC4
    -----> Join3 -------> Target3(Table X)<-----Seq1
    SRC5
    Here the three 3 targets are for the same Table X as well sequence is same i.e seq1
    If the First Join has 10 rows ,Seq1 generates sequence in 1 to 10 while loading target1
    But while loading second target,Same Seq1 is generating new sequence from 11 but i am looking to load target2 and target 3 starting from sequence 1 but not from 11 or so.
    As per your comments :
    you want to load 3 sources to one target with same sequence numbers?
    yes
    Are you doing match from the other two sources on first source by id provided by sequence (since this is the primary key of the table)?
    No
    can you please tell me how to approach for this?
    Thank You
    Kumar

  • Writing a trigger to update the QOH in a Parts table for for each item of the order whenever the order is filled. The trigger is for the OrderItem table when an item record is inserted into the OrderItem table.

    Any ideas?

    Hi,
    If orderitem and parts are the only tables involved, and each orderitem references only 1 part, then it's pretty simple: just update parts, like this
    UPDATE  parts
    SET     qoh       = qoh - :NEW.quantity
    WHERE   part_num  = :NEW.part_num;
    If you can modify filled orderitems (e.g., DELETE orderitems or change the part_num after the order is filled, or change the status from filled back to unfilled then it's more complicated; you might have UPDATE the :OLD.part_num as well as the :NEW.part_num
    I hope this answers your question.
    If not, post your best attempt, along with a little sample data (CREATE TABLE and INSERT statements, relevant columns only) for all tables involved, and also post the results you want from that data.
    Since you're asking about a DML statement, such as UPDATE, the sample data will be the contents of the table(s) before the DML, and the results will be state of the changed table(s) when everything is finished.
    Explain, using specific examples, how you get those results from that data.
    Always say which version of Oracle you're using (e.g., 11.2.0.2.0).
    See the forum FAQ: https://forums.oracle.com/message/9362002

Maybe you are looking for

  • IPhone 3GS not recognized by iTunes on Mac

    I have not been able to synch my phone since upgrading to most current version of iTunes. I have followed the instructions from apple help on how to uninstall and reinstall iTunes. I have done that twice now. Didn't work. I have also reset my phone p

  • Windows crashes when ipod syncs to itunes&blue screen appears

    Hello and thank you all in advance. This all began when I disconnected my ipod because it attempted to delete all of my songs. Long story short, I restored it because it kept saying the ipod was corrupt. when I tried to to connect to connect it to de

  • Sqlj runtime error with oracle 10G in Websphere6 or 5.1

    We are trying to migrate from oracle 9i to 10G and we are compiled all our sqlj code with codegen = -oracle option.(previously we used to use codegen=-jdbc which for some reason doesnot work).it compiles fine with out any erros.. But when we try to r

  • Switching SharePoint from Secured to Anonymous Access - Problems have Arisen

    Background: We had SharePoint setup by a professional SharePointer several years ago to be secured from anonymous Access.  Documentation was non-existant except showing us how to setup new sites and a few other things. Goal: We are trying to remove t

  • Multiple CVCs generated with same DP BOM

    Guys, In my example, I am using DP with BOM scenario. The "issue" I am facing is following: I have Characteristics Product and Location as part of POS; I created two CVCs with sames product different locations, saying: (COMPUTER, NEWYORK); (COMPUTER,