Simultaneous development on PY schemes and rules

Dear Colleagues,
I'm working for a big company in the IT department and I'm responsible for the process payroll. We have a lot of developers, who are working on payroll schemes and rules simultaneously for different demands and different go live dates.
The problem we are faced with is, that after go live of a demand or a project, in which a scheme or a calculation rule is included, sometimes an error occurs because of transported statements in schemes or rules, which are not already finished within another demand.
Example:
Project A and Project B is working on the same payroll scheme. Project A has been finished and went live (including the statements of project B), but project B has not finished the development and some new calculation rules or operations are missing on the productive system. Because of this, an error occurs when running the payroll.
Question:
Is it possible to deal with payroll schemes and rules similar like workbench objects? That means, is there any functionality availiable to block those objects for editing, when it is already included in a transport request?
Thanks in advance for any idea or help!
Kindly Regards
Adem Demir

No problem. At some point you will have taken a copy of iteration one's project and added that to CVS as iteration two. From that point onwards your iteration two index would be the same as iteration one. If you added keywords to it, it would be the index to use for iteration two. The index for iteration one would be unaffected as it would still be in CVS in a separate project.
Things only get a little more complicated if you are working on two different iterations of the project at the same time. If this is the case it may be better to finish one iteration first and copy the <projectname>.hhk file into the other project. Provided you haven't rrenamed or deleted any topics all references will be fine.

Similar Messages

  • Help on Time type 0903, Identifying Schema and Rule configuration?

    Can you please help me with the following:
    1. What is time type 0903?
    2. How to identify schema and rule configuration?
    3. How do I identify the number of employees affected in case I have to make a configuration change?
    Thanks

    Hello,
    1) Time type is identified according to what purpose it is used. Check table T555A for details on why it is used in your configuration.
    2) There is no configuration in schema and rule. Schema and rules has the programs in them.
    3)What configuration changes you are planning to do? Depending on this we can then try to identify who is affected.
    Hope this answers your questions.
    Let me know if you need more info on how to configure time management and write schema and rules.

  • What are schemas and rules in HR?How it will use in HR?

    Hi HR Proff,
    What are schemas and rules in HR?How it will use in HR?
    Give some documentation on this?
    Early reply is highely appriciable.
    Regards,
    Chow.

    Refer this link -
    http://www.sd-solutions.com/documents/SDS_Payroll_Schemas.html
    Regards,
    Amit

  • HR schema and rule function overview

    Hello, is there any HR schema and rule function overview there ? I mean for example function VARST, ADDDB, HRS, PCY etc., which are under PE01 and PE02 available to use.
    Thanks.

    Schemas can be viewed and modified through pe01.  Schemas use functions and Personnel Calculation Rules (PCR).
    PCRs can be viewed and modified through pe02.  Time Evaluation PCRs use operations such as HRS, VARST, ADDDB and OUTWP.  Payroll Calculation PCRs use operations such as RTE, NUM, AMT, VARGB, ADDWT and OUTWP.
    Functions and Operations can be viewed through pe04, along with their parameters.
    Documentation on Functions and Operations can be viewed either in the schema or rule by selecting the object + F1 (or the ? icon), or through pe04.

  • Undigesting XML with Schema and Rules

    We currently use a Digester with schema and rules files to digest an input XML and automatically build our Java objects. What I'd like to know is if there is a REVERSE process? Is there any class or utility that can take those Java objects and build an XML file?
    Thanks,
    Rich

    Ask Microsoft why IE6 does not validate. Normally server-side developers do
    not rely on a browser to validate XML against schema or DTD but validate XML
    on the server. The same is true for XSLT transformations...
    "Benoit Degreve" <[email protected]> wrote in message
    news:aqnihv$l75$[email protected]..
    Hi everybody,
    I'm trying to validate an XML file with a schema but it doesn't seem to
    work.
    I open the XML file (that doesn't match the rules that are in the schema)
    with a IE6, and there is no errors...
    I have the same problem with a DTD (its reference is in the XML file like
    shown in my book)...
    Does the validation process have to be done by a specific application ?For
    the schema ? For the DTD ? In other words, does IE6 make the validation
    process automatically ?
    Could someone help me ?

  • Need help in Time Evaluation  schema and rules

    Hi Gurus,
      1)Can anyone help me with Time Evaluation schemas and rules with some example?
      2)Where can I find good documentation on Time Schemas and rules?
      3)what are the practical Interview questions I can expect on Time management?
    I would really appreciate If you could guide me on this..
    Thanks
    Avy

    https://forums.sdn.sap.com/click.jspa?searchID=15025266&messageID=5372783
    https://forums.sdn.sap.com/click.jspa?searchID=15025266&messageID=5903057

  • Changes to Schema or rules

    Hi All,
    We are using customised schema which consists of some customised rules as well.
    1. How can I trace what changes are done to Schema or rule at what period and by whom?
    2. I wish to take whole back up of schema and rules into a PDF file for future reference. How can I do that.
    Thanks a lot.

    Tracing is not possible.Only the last changed user will appear in attributes.
    Go to Tcode SE38, Use the report RPDASC00
    Give the schema to be expanded saved,
    give all the sub schemas in that schema,
    Check the two boxes
    Execute and save in a notepad.

  • Delete blob column in source and rename schema using rule

    Hi
    I have blob column in source table, which i am deleting while capturing through rule based tranformation and also renaming the schema through the same rule.
    SET SERVEROUTPUT ON
    DECLARE
    v_dml_rule VARCHAR2(80);
    v_ddl_rule VARCHAR2(80);
    BEGIN
    DBMS_STREAMS_ADM.ADD_TABLE_RULES(
    table_name => 'ENGINE.ACTIVATIONS',
    streams_type => 'CAPTURE',
    streams_name => 'CAPTURE_TO_&destsid',
    queue_name => '&source_strmadmin_user..&sourcesid._TO_&destsid',
    include_dml => true,
    include_ddl => false,
    include_tagged_lcr => false,
    source_database => 'MMSTG.DEMOWORLD.COM',
    dml_rule_name => v_dml_rule,
    ddl_rule_name => v_ddl_rule,
    and_condition => ':lcr.get_command_type() != ''DELETE''',
    inclusion_rule => true);
    DBMS_STREAMS_ADM.DELETE_COLUMN(
    rule_name => v_dml_rule,
    table_name => 'ENGINE.ACTIVATIOINS',
    column_name => 'INSTANCE_DATA',
    value_type => '*',
    step_number => 0,
    operation => 'ADD');
    DBMS_STREAMS_ADM.RENAME_SCHEMA(
    rule_name => v_dml_rule,
    from_schema_name => 'ENGINE',
    to_schema_name => 'ENGINE_WEST',
    step_number => 0,
    operation => 'ADD');
    DBMS_OUTPUT.PUT_LINE('Capture DML Rule for ENGINE.ACTIVATIOINS => ' || v_dml_rule);
    END;
    I use step_number 0 for both rename schema and delete column. Some how, delete column transformation is not firing and it replicates the blob column with some bytes instead of zero bytes.
    How do we execute this both rule based transformation with correct step_number to fire both transformations. Please do let me know..
    Thanks
    Bala

    In what version?
    Step 0 is the default so that is not the issue.
    Can you also post the DDL for your table and the results of the following query:
    SELECT *
    FROM dba_streams_unsupported
    WHERE owner = <schema owner>
    AND table_name = <your table name>;

  • Understanding & developing the payroll Schema & Rules in India

    Hi
    Where do we get the complete details of the IN00 (with some discription)? How do we understand the coding ? If I navigate from the shcema to the rule , the rule appears as one line command and i enter the rule (pe03) i get a different format... 
    New to this area. Please help
    Thanks
    YGP

    Hi,
    >Adding to the post of Sikindar and Remi, you can also search for articles published on Schemas and PCRs.
    Regards
    Chandrashekar

  • Simultaneous development for desktop and web

    What I'm looking for is simultaneous development of a desktop client and a web client (as an applet). Basically, the desktop client and the web client should provide the same functionality. What I'd like to do is develop these two concurrently; with the same tool, using the same codebase with build-specific changes being kept to the absolute minimum. To estimate the size and difficulty of the project.. JDBC, sockets, graphics.
    From what I read, Java Studio Enterprise would be OK for working on both aspects. Functionality would be provided by a set of special Beans I'd write. What I'm wondering about is that how well could this succeed. Question time!
    * I'd like to have the desktop client use SWT and the web client Swing. It should be possible to wrap these to provide a common API for myself, but how much work am I looking at?
    * Regarding the question above - if I use a visual RAD tool like Sun Java Studio Enterprise, how easy/much work would it be replace the widget library with their "cross-platform" widget library equivalents (so that in the code, it'd reference my own versions and all)? This is about reducing the build-specific changes again - when I drag and drop a button to a form, I'd want it to be one of my own magical SWT/Swing buttons.
    * I'd imagine the application would get fairly large - how much unavoidable change should I expect between the desktop client and web client? Are there any serious pitfalls/caveats?
    * Would it be feasible to realize the desktop "client" through Web Start - just make the applet and have it launched through that?
    I'm evaluating different views and ideas for a possible project and would like to know how Java would fare in something like this. Let me know what you think. Thanks for your input!

    Maybe something like OpenID Connect built on top of oAuth 2.0 would be applicable?
    Users can manage their identities with an Identity Provider. Your Web and Desktop applications become Relying Parties and they will be given an ID Token and possibly an Access Token when the user is authenticated with the Identity Provider.
    The Identity Provider could be extended to federate onto other Identity Providers such as ADFS, Google, Facebook...

  • IT0071 Dynamic Action does not overwrite Pension Scheme Calculation Rule

    I am trying to write a dynamic action to create IT0071 with default values for the Employee contribution (0%) and Employer contribution (4%). However when IT0071 is created the default values from the associated Pension Scheme Calculation Rule (4% and 8%) are not being overwritten by the dynamic action. Is there a way to get the dynamic action to override the Pension Scheme Calculation Rule values?
    This is the dynamic action u2026
    0171              BENGR     06     10     P     T001P-MOLGA='08'
    0171              BENGR     06     15     P     PSPAR-MASSN='01'/X
    0171              BENGR     06     20     P     PSPAR-MASSN='12'/X
    0171              BENGR     06     25     P     P0171-BENGR<>NOBN
    0171              BENGR     06     30     P     P0171-BSTAT<>NOBN
    0171              BENGR     06     300     I     INS,0071,SS7,,(P0171-BEGDA),(P0171-ENDDA)
    0171              BENGR     06     315     W     P0071-PCTEE='0.0000'
    0171              BENGR     06     320     W     P0071-PCTER='4.0000'
    Any help much appreciated.
    Regards,
    Janet

    Hi Brian,
                  In standard SAP physical sample planning requires sample drawing procedure.
    Link: [http://help.sap.com/erp2005_ehp_05/helpdata/EN/2d/3521b6448c11d189420000e829fbbd/frameset.htm]
                  And drawing procedure is advised with partial sample numbers.
    Link: [http://help.sap.com/erp2005_ehp_05/helpdata/EN/2d/3521b6448c11d189420000e829fbbd/frameset.htm]
                 Generally when we use partial sample numbers, we also use a drawing procedure. And in that scenario SAP describes the calculation as shown in
    Link: [http://help.sap.com/erp2005_ehp_05/helpdata/EN/2d/3521b6448c11d189420000e829fbbd/frameset.htm]
                 What you have done (partial samples for MICs and not having for others, all without drawing procedure) is not generally done. I think our observations in this context should be confirmed only if we try more of such examples.
    Regards
    Sm.

  • Schemas and workspaces in Apex 3.0

    Hi Friends
    We are quite new to HTML DB. We have few queries pertaining to schemas and workspaces.
    We are developing an application in workspaces X which is associated with database user/schema X. We need to give this application for testing to the users. For that we need to have different set of environment so that it does not hamper our development.
    So what should be our plan of action.?
    Currently we have created a different database user/schema in the same database and associated a different workspace Y to it. But we are facing import/export issues across the workspaces. We have already raised this issue in the forum.
    Can we have same workspace associated with 2 different database schemas/users? If yes, then how does HTML DB handle the same?
    Thanks in advance.
    Regards
    Minakshi

    Hi Friends,
    As per Scott's suggestion, we attached our workspace X to 2 schemas A and B. Then we exported the app say 555(here the schema is A) and imported it using the parsing schema B in new application id say 500. But now the real problem starts. If we have added some new functionality in a page in our application 555 and now only that page we want to export and import it in application 500, then it gives error as "This page was exported from a different application or from an application in different workspace. Page cannot be installed in this application." So in the sql file that we exported, we changed few things as follows(marked in bold)
    --application/set_environment
    prompt APPLICATION 500 - Load_Factor_System
    -- Application Export:
    -- Application: 500
    -- Name: Load_Factor_System
    -- Date and Time: 12:01 Monday May 21, 2007
    -- Exported By: ADMIN
    -- Flashback: 0
    -- Export Type: Page Export
    -- Version: 3.0.0.00.20
    -- SET APPLICATION ID
    wwv_flow.g_flow_id := 500;
    wwv_flow_api.g_id_offset := 0;
    After we changed the export file as above and then tried to import it in application 500 it gives error as "ORA-20001: GET_BLOCK Error. ORA-20001: Execution of the statement was unsuccessful. ORA-00001: unique constraint (FLOWS_030000.WWV_FLOW_PAGE_PLUGS_PK) violated <pre>declare s varchar2(32767) := null; l_clob clob; l_length number := 1; begin s := null; wwv_flow_api.create_page_plug ( p_id=&gt; 6303905775095939 + wwv_flow_api.g_id_offset, p_flow_id=&gt; wwv_flow.g_flow_id, p_page_id=&gt; 1020, p_plug_name=&gt; 'User Master', p_region_name=&gt;'', p_plug_template="
    Can anybody please help us with this issue? The other solution for this is to delete the complete application 500 and then freshly install the application. But we do not want this since any change in few pages should not result in deletion of complete application.
    Also Scott, thanks a lot for your detailed solution as to how to put an example on apex.oracle.com. We shall definitely put an example and post it.
    Regards
    Minakshi

  • Duplicate data in front scheme and back scheme?

    Assume that a near cache has been defined with a front-scheme and a back-scheme. The front-scheme is a local-scheme and the back-scheme is a distributed-scheme.
    Now assume that there is a piece of data for which the current JVM is a master. That is, the data resides in the back cache.
    Assume that the JVM accesses that data.
    Will this populate the front cache even though the master is in the back cache on the same JVM?
    Does Coherence therefore store two copies of the data, one in the front cache, one in the back cache?
    Or does Coherence realize that the requested data is in the back cache on the same JVM, and therefore doesn't make another copy and populate the front cache, but instead simply redirects the get call to it's own back cache?
    This basically has memory implications and I'm trying to configure a cache that works on a single JVM as well as a cluster. I am trying to get away with having just the one configuration file for coherence irrespective of whether the application is on a single storage enabled node, or a cluster.
    I cannot tell from the diagram in the Developer Guide at http://docs.oracle.com/cd/E24290_01/coh.371/e22837/cache_intro.htm#BABCJFHE
    because I'm not sure whether the front/local cache for JVM1 has not been shown as populated with A because it is the master for A, or because JVM 1 never accessed A.

    926349 wrote:
    Assume that a near cache has been defined with a front-scheme and a back-scheme. The front-scheme is a local-scheme and the back-scheme is a distributed-scheme.
    Now assume that there is a piece of data for which the current JVM is a master. That is, the data resides in the back cache.
    Assume that the JVM accesses that data.
    Will this populate the front cache even though the master is in the back cache on the same JVM?
    Yes, it will, if access is via the near cache.
    Does Coherence therefore store two copies of the data, one in the front cache, one in the back cache?
    Yes, but with usual usage patterns the front-map is size limited, so it will hold the just pk-accessed entry in the front map, but it may evict other entries from the front map, so the front-map may not be a full duplicate of the entire data-set.
    Or does Coherence realize that the requested data is in the back cache on the same JVM, and therefore doesn't make another copy and populate the front cache, but instead simply redirects the get call to it's own back cache?
    Nope, no such logic is in place. On the other hand, as I mentioned front-map is usually size-limited, and also the storage nodes don't have to access the cache via the front-map, they can go directly to the distributed cache by getting the back-cache from the near cache instance.
    This basically has memory implications and I'm trying to configure a cache that works on a single JVM as well as a cluster. I am trying to get away with having just the one configuration file for coherence irrespective of whether the application is on a single storage enabled node, or a cluster.
    You should not try to. You should always choose the correct topology and configuration for the actual use case. There is no one-size-optimally-fits-all solution.
    Best regards,
    Robert

  • Compare 2 schemas and get the difference with .sql file.

    Hi,
    I am using ORACLE DATABASE 11g R2 and ORACLE Linux 5.
    I want to perform a very lengthy process and want to make it automated.
    I am having a software named as SVN. In which all the developers keep their updated scripts.
    We have 2 schema's one is used for developement and when all the development seems good we implement the scripts on the final schema which is used by testing people also.
    Now every day we need to check the modifications in the development schema, that we do by observing the updated scripts in the SVN software. Now we get the scripts which are modified.
    We will fire this scripts to a sample schema then we will compare the objects in this schema and the same objects in our final schema. If the objects in sample schema is different than the final schema it should give me the
    script to create the same type object in the final schema.
    Below is a proper example to explain :-
    Developement Schema :- 'DEV'
    SVN Script Schema :- 'SVNscript'
    Main Schema :- 'MAIN'
    1) On Monday developers modified/added 4 tables in the development schema names 'DEV' after working for a day they found that the changes are necessary and so checked-in in the SVN software.
    2) On tuesday morning we found 4 scripts which has been modified/added in the SVN.
    The Scripts were as follows :-
    1St table :- An extra column was added.
    2nd table :- An index was created on it.
    3rd table :- 2 Columns were dropped from it.
    4th table :- A new table is added in the schema.
    Now taking these 4 scripts i am going to execute it in a sample schema named :- SVNscript
    SVNscript schema will have now 4 tables with proper properties(columns,indexes,...). Now this is the final table structure as we want in our 'MAIN' schema .
    Coming to MAIN schema which is our most important schema and all the development work is finally posted here.Considering MAIN schema is having around 1000 tables,500 function/procedure/package and many more DB objects.
    I want to compare(one way compare) from SVNscript schema --> MAIN schema :- That the objects present in the SVNscripts schema is same or not in the MAIN schema. If its not same then this code should generate a .sql script for me which i should be able to fire on the MAIN schema.
    The output .sql scripts should be something like this :-
    1st Table :- Alter table add...
    2nd Table :- Create index on table...
    3rd Table :- Alter table drop...
    4th table :- Create table tablename...
    I found a link while trying this but it is not the perfect fit to my requirement
    http://www.dbspecialists.com/files/scripts/compare_schemas.sql
    Please let me know the best code to compare 2 schemas and get a .sql file as output of the difference.
    Thanks in advance.

    Yes I followed the tutorial properly this time. Still i have a few queries :-
    1) This is returning me the 'ALTER TABLE...' statement but how to make it more efficient to return all the DDL dependent on the table, Like index,trigger,view...
    2) I want to compare objects in a different schema on a different db. I have created an DBLINK but how to use it ... can i use it like
    DBMS_METADATA.OPEN('TABLE'," Network_link_name " );Can you please give me some clear about it.
    3) I created a table 'TAB1' with which i want to compare a 'TAB1' in other schema.I was not getting how to compare on remote dblink so,I created the same table with the following code in the same schema
    CREATE TABLE TAB1_OLD as select * from TAB1@DBLINK where 1=2;Now when i fire the below query to get the alter difference I get the following error :-
    SQL> SELECT get_table_alterddl('TAB1','TAB1_old') FROM dual;
    SELECT get_table_alterddl('TAB1','TAB1_old') FROM dual
    ORA-31603: object "TAB1_old" of type TABLE not found in schema "SVNCHECK"
    ORA-06512: at "SYS.DBMS_METADATA", line 5225
    ORA-06512: at "SYS.DBMS_METADATA", line 5189
    ORA-06512: at "SVNCHECK.GET_TABLE_SXML", line 17
    ORA-06512: at "SVNCHECK.COMPARE_TABLE_SXML", line 12
    ORA-06512: at "SVNCHECK.GET_TABLE_ALTERXML", line 11
    ORA-06512: at "SVNCHECK.GET_TABLE_ALTERDDL", line 11
    Can you please guide me why i am getting this error?
    I have made sure that TAB1_OLD table has been created and the entries are also present in the data dictionary tables.
    SQL> select * from tab1_old;
    EMPNO ENAME                  MGR DEPTNO
    ----- -------------------- ----- ------The only difference is that this table was created with CREATE TABLE AS SELECT statement ....
    Thanks.

  • Schemas and PCRs

    hi, experts.
    i need help in learning schemas and writing PCRs. please suggest me any book or send me any document that will help.
    regards,
    madhu

    hai..
    How to read rules and Schemas
    1.Structure of relationship in Rules and Schemas:
    a)Schemas consist of Functions and Sub-schemas
    b)Some functions have the rules attached to it as one of its parameter
    c)Rules consist of a set of operations to perform some actions
    d)Operations and Functions are the executable components
    As per the diagram Set of Operations make a Rule which in turn attached to some functions and those functions embedded in Schemas or Sub-schemas (Inside the main schemas) to decide the flow of the payroll program.
    Structure of a Schema:
    Func:
    This column is used to give a function name.
    Par1, Par2, Par3, Par4:
    1. Function has maximum four parameters.
    2. The function can have zero to four parameters as per the definition of the function.
    3. Pressing F4 we can get the list of all the values which can be used as parameters.
    4. For some functions in the first parameter is the name of the rule created.
    5. These parameter values are predefined the function is being created.
    D:
    1. This column is used to comment and uncomment a function.
    2. If (*) is being put then the line is commented and will not be executed in Payroll processing.
    3. If nothing is given then the line would be executable.
    Desc
    1. In this column description for the function is being used.
    Functions:
    Functions are used for;
    • Performing some payroll computations (E.g. INEPF function calculates the PF amount of an employee during payroll run)
    • Calling rules (E.g. P0045 function calls a rule INLN to compute the loan details of a personnel number).
    • Getting data from Infotypes (E.g. P0581 will get the data from Infotype-581 for payroll processing).
    • For some decisions. (E.g. IF & ENDIF function is used to execute as per the true and false decisions) etc.
    Rules:
    Rules are used for holding a set of operations for a particular requirement to be accomplished.
    Attributes:
    1. Program Class.
    a. There are two program classes to be assigned to the rule while creating.
    i. Payroll (C).
    ii. Time (T).
    2. Country Grouping.
    a. For Payroll program class the country grouping should be mentioned. (E.g. 40 for India).
    b. For Time management rule the country grouping should be (*).
    3. Employee Sub-grouping.
    a. All the wagetypes have an attribute of employee sub-grouping.
    b. It varies client to client.
    c. Value 3 means EE sub-grouping is 3.
    d. Value * means all EE Sun-groupings.
    4. Wagetypes.
    a. The wagetype in the internal table which is meant to be processed by the rule.
    b. If a value for Wagetype is being given then the particular wagetype is being queried for processing.
    c. If **** has been given then all the wagetypes present in the particular internal table will be processed.
    Structure of the Rule.
    Frequently Used Internal Tables in PY Processing:
    The Payroll driver uses lots of internal tables used for storing data temporarily in the program for processing. Some of the important internal tables are:
    The structure of internal tables:
    EE Sub-grouping:
    1. The Wagetype is assigned to a particular Employee sub-group.
    2. 3 is a particular employee subgroup for basic pay wagetype.
    3. * value means for all EE sub-grouping.
    Wagetype Code:
    1. The Wagetype code number is the number assigned to a particular pay component.
    Wagetype Description:
    1. The description for the wagetype code.
    NUM:
    1. If there is a split for the wagetypes then this NUM field will make them different.
    2. If A person’s basic salary is changed in the mid of the month then the wagetype will be split in two amounts with NUM = 01 and NUM = 02.
    RTE:
    1. The RTE column stores a value for;
    a. Rate of interest.
    b. Number of leave days.
    c. Projection factor. Etc.
    2. E.g. In the above diagram /401 wagetype has RTE value 10 which is the projection factor used for different calculations.
    AMT:
    1. This column possesses the amount of the particular wagetype.
    Use of the internal tables:
    1. The internal tables are being used for calculations on the NUM, RTE & AMT fields.
    2. These internal tables are being read by the rules row by row.
    3. The Calculation rule would be only processed for those wagetypes which are specified in the rule’s source code.
    4. The row in the internal table containing the wagetype which has been defined in rule attributes will be placed in the header of the internal table for processing.
    5. The processing would be as per the operations used in the rule.
    Some important Operations & Functions:
    • The editor for Operations and Functions is PE04.
    • While the payroll driver encounters a function and operation it would call a subroutine written in the payroll driver.
    • The subroutine name for the operation will be OP****, Where **** will be the name of the operation.
    • For E.g. for Operation AMT the subroutine name would be OPAMT.
    • The code written for the function will be found as a subroutine in payroll driver as FU****, where **** will be the name of the function.
    • For E.g. for function INEPF the subroutine name would be FUINEPF.
    Operations:
    AMT:
    1. This Operation would fetch the Amount of the wagetype queried in the rule, in the Variable AMT for processing.
    2. The Syntax of the Operation is ZZZOVVVV.
    3. ZZZ -> AMT.
    4. O -> Operand.
    5. VVV -> Value or Variable.
    6. For E.g. (AMT= 1000) will fetch the Amount of Wagetype 1000 into AMT.
    7. Pressing F1 on the AMT Operation will provide the documentation.
    RTE:
    1. This Operation would fetch the Rate of the wagetype queried in the rule, in the Variable RTE for processing.
    2. The Syntax of the Operation is ZZZOVVVV.
    3. ZZZ -> RTE.
    4. O -> Operand.
    5. VVV -> Value or Variable.
    6. For E.g. (RTE= /401) will fetch the Rate of Wagetype /401 into RTE.
    7. Pressing F1 on the RTE Operation will provide the documentation.
    NUM:
    1. This Operation would fetch the Split indicator of the wagetype queried in the rule, in the Variable NUM for processing.
    2. The Syntax of the Operation is ZZZOVVVV.
    3. ZZZ -> NUM.
    4. O -> Operand.
    5. VVV -> Value or Variable.
    6. For E.g. (NUM= 1000) will fetch the Split indicator of Wagetype 1000 into NUM.
    7. Pressing F1 on the NUM Operation will provide the documentation.
    ADDWT:
    1. This Operation will append one more row in the internal table (Processed by the rule).
    2. After calculating an amount from a particular wagetype the amount can be assigned to another wagetype and append to the internal table.
    3. For E.g. ADDWT 2050 will add the NUM, RTE, AMT calculated to the NUM, RTE & AMT of the wagetype 2050.
    4. If the Wagetype 2050 is not there before in the internal table then NUM, RTE & AMT would be 0.
    5. If any value of these three variables are present before then they will be added to the calculated ones.
    Functions:
    PIT:
    1. This function will read the internal table IT (Input table).
    2. The function will hold a rule as first parameter.
    PRT:
    1. This function will read the internal table RT (Results table).
    2. The function will hold a rule as first parameter.
    PDT:
    1. This function will read the internal table IT (Input table).
    2. The function will hold a rule as first parameter.
    PORT:
    1. This function will read the internal table ORT (Old results table).
    2. The function will hold a rule as first parameter.
    P0014:
    1. This function will read the Infotype 0014 for processing.
    2. The function will hold a rule as first parameter.

Maybe you are looking for