Reg:hr anlytics informatica mapping

I have one more question.the out of the box mappings calculates age of every person on first of every month.If a person has birth day in the middle of the month,it doesnot calculate and the person is still one year younger though he passed his birthday.I want add an event 'AGE' in w_wrkfc_evt_merge_f where for every year on a birth day the age of the person increases.Have you ever seen a requiremnt where a new chage event type has to be added to the tabeles.
for the above requirement,how can i change the existing code in w_wrkfc_evr_merge_f mapping in informatica to insert a new change event on person's birth day.
thanks

There are many ways to cater your requirements, I would say
customize the mapping to check the ddmm of birthdate with ddmm of current date or sysdate. if these matches add 1 to age else 0
If it is helpful, please mark as correct or helpful

Similar Messages

  • The Informatica maps are not taking the parameters from dac

    Hi All,
    The Informatica maps are not taking the parameters from dac. We have tried setting the parameters at DAC level, task level. The parameter file is getting generated by DAC and has correct values as we expect. But when we run a full load from DAC, the initial date in Informatica is being taken as 1/1/1753. Do we need to specify the parameter file name at workflow level?
    Is there any setting that we are missing. Please guide us and let us know the steps we should take to make sure that the parameters that we set at DAC level get passed to the Informatica maps. Also mention this is happening only with dates and other parameters like data source num , soruce and target database information is getting passed to Informatica maps.
    The version of Informatica as 7.4.1 and dac version AN 328, Build date Aug 8 2007.
    Please Help me in this issue.
    Thanks in advance for all members.

    Hi Ananth.v
    Try to make sure that your parameter file's path in Infa is in-sync with the path of DAC files being generated.
    The date you are getting seems strange - which are your Source/Target DB platforms?
    In general take a look at [this thread|   http://forums.oracle.com/forums/message.jspa?messageID=3405906#3405906] , it may make some thing clearer for you.
    Hope it helps,
    Please post if you found a solution,

  • Informatica Map details for Oracle Analytics

    We have installed Oracle BI Apps (Procurement and Spend) for the Oracle Apps container.
    It would be very helpful if we can have the following information :-
    1. Are the dependencies sorted out in the pre-built informatica repository?
    2. Are there any documentation available to provide us with a detail level information about the Informatica maps (source --> map_name --> target).
    Any help would be appreciated.

    Thanks for your reply.
    We have created the connections for the following
    1. Informatica Server
    2. Informatica Rep Server
    3. OLAP DB
    4. OLTP DB
    All the connections were checked and connectivity was successful.
    However, while starting the execution Plan for Supply Chain, the job is failing with the following error :-
    13 SEVERE Sat Jan 12 02:58:26 GMT 2008 Could not find session for SDE_ORA_Stage_GLAccount_SegmentConfig_Extract(ORA_11_5_10_Flatfile->DataWarehouse)
    14 SEVERE Sat Jan 12 02:58:26 GMT 2008
    ANOMALY INFO::: Error while loading nodes.
    EXCEPTION CLASS::: java.lang.NullPointerException
    com.siebel.etl.engine.core.SessionHandler.getNodeTables(SessionHandler.java:2999)
    com.siebel.etl.engine.core.SessionHandler.getNodes(SessionHandler.java:2754)
    com.siebel.etl.engine.core.SessionHandler.loadNodes(SessionHandler.java:450)
    com.siebel.etl.engine.core.ETL.thisETLProcess(ETL.java:322)
    com.siebel.etl.engine.core.ETL.run(ETL.java:604)
    com.siebel.etl.engine.core.ETL.execute(ETL.java:831)
    com.siebel.etl.etlmanager.EtlExecutionManager$1.executeEtlProcess(EtlExecutionManager.java:210)
    com.siebel.etl.etlmanager.EtlExecutionManager$1.run(EtlExecutionManager.java:164)
    Looks like it could not connect to one of the nodes. Any idea why this might have happened?
    Regards

  • I m trying to write to a Mainframe file from informatica mapping. I have created the data map for tarrget file and have imported it in informatica designer.After completing the mapping when i try to run it gives me below error. Please help me out in resol

    This is not a PC/PWX problem, but a mainframe security problem. TIDSS01.POS.IBD.DR0100.DAT is not a GDG file name.It might be the base for a GDG. A GDG file name would look something like eitherTIDSS01.POS.IBD.DR0100.DAT(+1)orTIDSS01.POS.IBD.DR0100.DAT.G001V00 So I suspect that you have the wrong file name.Please talk with your mainframe team.

    HI Dan, I am very new to using power exchnage please help me through this issue, Below is the details : My data map name :postest.test1_POS
    Copy book name which is used to create the power exchange Datamap : TIDSS01.ACTRLP.TEST(POSCPY)
    Main frame GDG name : TIDSS01.POS.IBD.DR0100.DAT Below are the Session properties i have set in informatica workflow:Schema Name Override : postest
    Map Name Override : test1_POS
    PWX Partition Strategy : Overrides driven
    Space : CYLINDER
    File Name Override :TIDSS01.POS.IBD.DR0100.DAT(+1)
    I am getting below error: PWXPC_12190
    Message: [ERROR] Transformation [test1_POS]: A permanent error has been encountered in PowerExchange: [
    [Informatica][SCLI PWX Driver] PWX-00267 DBAPI Error. DB_INSERT failed for file postest.test1_POS.
    [Informatica][SCLI PWX Driver] PWX-01279 DBNTC INSERT failed for file postest.test1_POS. Rcs 1274/2019/268.
    [Informatica][SCLI PWX Driver] PWX-01274 DBNTC INSERT Failed for file postest.test1_POS, rcs 260/2019/268.
    [Informatica][SCLI PWX Driver] PWX-02019 SQL insert failure. SQLCODE = 268.
    [Informatica][SCLI PWX Driver] PWX-00268 DBAPI Error. DB_OPEN failed for file TIDSS01.POS.IBD.DR0100.DAT.
    [Informatica][SCLI PWX Driver] PWX-00220 DYNALLOC failed for file TIDSS01.POS.IBD.DR0100.DAT RCs = 9700/0.
    [Informatica][SCLI PWX Driver] PWX-00221 DATA SET: TIDSS01.POS.IBD.DR0100.DAT WITH RETURN CODE 08 REASON CODE 00   
    [Informatica][SCLI PWX Driver] PWX-00221 RACF FUNCTION: RACDEF FOR   
    [Informatica][SCLI PWX Driver] PWX-00221 IGD308I DATA SET ALLOCATION REQUEST FAILED -   
    [Informatica][SCLI PWX Driver] PWX-00221 IKJ56893I DATA SET TIDSS01.POS.IBD.DR0100.DAT NOT ALLOCATED+   
    [Informatica][SCLI PWX Driver] PWX-07404 Permanent error set by Open call on file "TIDSS01.POS.IBD.DR0100.DAT" because dynamic allocation failed. rc=4
    [Informatica][SCLI PWX Driver] PWX-07515 Insert call for table postest.test1_POS met a permanent error. Return codes 267 2019 268.
    ]

  • DAC task with Informatica mapping and stored procedure (very slow)

    Hello,
    We have a DAC task that launch an Informatica Workflow with a simple query and stored procedure, like this:
    SQL QUERY
    ==========================
    SELECT
    W_ACTIVITY_F.ROW_WID,
    W_AGREE_D.AGREE_NUM,
    W_PRODUCT_D.ATTRIB_51,
    W_SRVREQ_D.ATTRIB_05,
    W_ORG_DH.TOP_LVL_NAME,
    W_ORG_D.ATTRIB_06,
    W_PRODUCT_GROUPS_D.PRODUCT_LINE,
    W_PRODUCT_D.PROD_NAME
    FROM
    W_AGREE_D,
    W_SRVREQ_F,
    W_ACTIVITY_F,
    W_PRODUCT_D LEFT OUTER JOIN W_PRODUCT_GROUPS_D ON W_PRODUCT_D.PR_PROD_LN = W_PRODUCT_GROUPS_D.PRODUCT_LINE,
    W_ORG_D,
    W_SRVREQ_D,
    W_ORG_DH
    WHERE
    W_SRVREQ_F.AGREEMENT_WID = W_AGREE_D.ROW_WID AND
    W_SRVREQ_F.SR_WID = W_ACTIVITY_F.SR_WID AND
    W_SRVREQ_F.PROD_WID = W_PRODUCT_D.ROW_WID AND
    W_SRVREQ_F.ACCNT_WID = W_ORG_D.ROW_WID AND
    W_SRVREQ_F.SR_WID = W_SRVREQ_D.ROW_WID AND
    W_ORG_D.ROW_WID = W_ORG_DH.ROW_WID
    STORED PROCEDURE
    ===========================
    ConvSubProy(W_AGREE_D.AGREE_NUM,
    W_PRODUCT_D.ATTRIB_51,
    W_SRVREQ_D.ATTRIB_05,
    W_ORG_DH.TOP_LVL_NAME,
    W_ORG_D.ATTRIB_06,
    W_PRODUCT_GROUPS_D.PRODUCT_LINE,
    W_PRODUCT_D.PROD_NAME)
    The mapping is very simple:
    Source Qualifier -> Stored procedure -> Update strategy (only two ports: ROW_WID and custom column) -> Target Table
    When I launch the DAC Execution Plan the corresponding task take much time (40 minuts). But when I launch the corresponding Workflow from Informatica PowerCenter Workflow Manager this only take 50 seconds... when I see the log session for the task I can see that much time is spent on the time of the updates. For example, when DAC is running the writer updates 10000 records every 6/7 minuts, but when Workflow Manager is running thw writer updates 10000 records every 8/9 seconds.
    So, what happens (in the DAC) to that so much time difference? Is there a way to reduce the execution time when the task is launched from DAC?
    Thanks
    Best Regards
    Benjamin Tey

    Have you tried using bulk load type?
    In Workflow Manager can you open the associated task, navigate to the mapping tab and seled the target table.
    What is the value for "Target load type" and which of the following boxes are checked: Insert, Update as Update, Update as Insert, Update else Insert, Delete?

  • Reg usage of XSLT Mapping

    Hi All,
    We need to build an interface from PI to some database using JDBC adapter to send out emails to business users of the exception records created in the table.please help in using the xslt mapping for this interface.
    Thanks in Advance,

    Hi,
    You can build your scenario as follows:
    1) If the emails have to be sent from PI then you need to get back a response from the database.....if this is failure message then initiate your send email process...the whole process can be implemented using a BPM: Source ---> PI ---> JDBC --->PI ---> Email
    2) You need to use a receiver mail adapter to send emails
    3) Refer this blog of michal in which it is mentioned how to use the XSLT mapping and configuration for sending the emails:
    /people/michal.krawczyk2/blog/2005/11/23/xi-html-e-mails-from-the-receiver-mail-adapter
    You can also refer: http://www.riyaz.net/blog/xipi-sending-emails-using-xi-mail-adapter/
    For more clarification you can also refer my answer in this thread:
    Re: xml in mail
    The XSLT mapping that you need will be only in the PI ---> Email scenario...
    Make sure that you download the desired email structure from service market place before you start with the scenario.......do let me know if my answer helps.
    Regards,
    Abhishek.

  • Reg error in interface mapping in RFC to JDBC scenario

    Hi Techies,
    Iam trying to map the scenario RFC to JDBC.
    In configuration window Iam testing the configuration.
    when Iam testing it is giving the error.
    " com.sap.aii.utilxi.misc.api.BaseRuntimeException thrown during application mapping com/sap/xi/tf/_mm_mapping_: Parsing an empty source. Root element expected! "
    My mapping is as follows
                       mt_receiver                    
                                                            statement               
                                                                ROW          
                                                     action---insert     
                 ZHRT002----
         Table     
         Item----
    access     
              MANDT               
              PERNR                                     PERNR
              PERID                            PERID
              SHOPN                           SHOPN
    Sender
    <?xml version="1.0" encoding="UTF-8"?>
    <ns0:ZTEST_PI_ZHRT002 xmlns:ns0="urn:sap-com:document:sap:rfc:functions">
       <ZHRT002>
          <item>
             <MANDT/>
             <PERNR/>
             <PERID/>
    </item>
       </ZHRT002>
    </ns0:ZTEST_PI_ZHRT002>
    Reciver
    <?xml version="1.0" encoding="UTF-8"?>
    <ns1:mt_receiver xmlns:ns1="urn:sap-com:RFC2JDBC"><Statement1><ROW action="INSERT"><TABLE>ZHRT002</TABLE><access><PERNR></PERNR><PERID></PERID></access></ROW></Statement1></ns1:mt_receiver>
    Please give suggestion how to go about this.
    Thanks in advance,
    Regards,
    Kiran

    Hi,
    Please check the payload you are using to test the configuration.
    The structure seems to be Incorrect.
    Regards,
    Deepak

  • Reg: Test an ABAP mapping

    Hi All,
                        When i go to test at interface mapping level there is no possiblity to test ABAP mapping. Is there any way to test ABAP mapping without executing scenario.
    Thanks,
    Satish.

    This is a link very useful for ABAP Mapping :
    http://help.sap.com/saphelp_nw04/helpdata/en/86/8280ba12d511d5991b00508b6b8b11/frameset.htm
    You should define a simple scenario. Sender Service, Receiver Service and Interface Mapping using your ABAP Mapping
    You can test your ABAP Mapping using the transaction,
    SXI_MAPPING_TEST
    Check my weblog as to how to use the transaction for testing,
    /people/sameer.shadab/blog/2005/09/29/testing-abap-mapping
    Check this thread:
    https://www.sdn.sap.com/sdn/collaboration.sdn?contenttype=url&content=https%3A//forums.sdn.sap.com/thread.jspa%3FforumID%3D44%26threadID%3D34468%26messageID%3D326456%23326456
    Note:please reward points if solution found helpfull
    Regards
    Chandrakanth.k

  • Reg SUPRESS in message mapping

    Hi Guys,
    i have a requirement to map the items from source based on a condition(if flag = X).if the condition is true then need to create item segment in target else donot create segment.
    here iam facing a problem with SUPRESS.
    if the first item is false and the second item is true and third is false and fourth is true....then, iam getting the SUPRESS for false values and not able to create the segment even for true values.
    how to remove the supress... suggest me.....this is an tough requirement for me.
    Thanks,
    M.S

    Assume if Queue have
    123
    234
    SUPPRESS
    233
    SUPPRESS
    100
    345
    SUPPRESS
    Your target should have
    123
    234
    233
    100
    345
    If so
    Please follow
    Function myudf(String a[], ResultList result, Container container)
    for(int k =0 ; k< a.length; k++)
    if (!a[k].equalsIgnoreCase(ResultList.SUPPRESS))
    result.addValues(a[k]);

  • Reg: copying standard message mapping

    Hi,
    Can any one suggest me how to copy the standard message mapping..
    I have copied it in normal way...but its not getting executed...
    when i try to execute the standard map with same payload, then it works fine...
    << Moderator message - Please do not offer points >>
    << Moderator message - Everyone's problem is important >>
    Thanks,
    Madhu
    Edited by: Rob Burbank on Oct 22, 2010 3:51 PM

    Stefan,
    I have copied the SAP delivered standared map. When I run the copied map, it is giving an error "Cannot create target element /ns1:envelope/catalogueRequest/document. Values missing in queue context". But same getting executed in original SAP map.
    here in the map node "document" occurences is 1 to 1000. For mapping logic it has been duplicated 3 times.
    ex : document(1..1000), document[1](1....1000), document[2](1...1000).
    according to payload we can able to create node document(first occurence) once. That should be enough. and standard SAP map is working accordingly.
    But when we ran copied map its giving an error. we desabled the occurence 2 then its working.
    no look ups are used.
    Any suggestion....

  • Running an Unix Script from Informatica Mapping/workflow.

    Hi ,
    I have a need to start /stop service to a remote server (away from the DAC server ,which is on UNIX os)
    how could this be done in a best way?
    I am planning to do this through Informatica Workflow,and need help (technical steps Please) to do this.
    1) workflow will call the Unix Script
    2) Unix Script logins to the remote server
    3) Unix scripts navigates to particular folder
    4) Unix script executes a command to Start/Stop the server.
    5) Unix Script logs out.
    6) if everything done success ,Workflow will success else fail with the error code.
    Regards

    Hi ,
    I have a need to start /stop service to a remote server (away from the DAC server ,which is on UNIX os)
    how could this be done in a best way?
    I am planning to do this through Informatica Workflow,and need help (technical steps Please) to do this.
    1) workflow will call the Unix Script
    2) Unix Script logins to the remote server
    3) Unix scripts navigates to particular folder
    4) Unix script executes a command to Start/Stop the server.
    5) Unix Script logs out.
    6) if everything done success ,Workflow will success else fail with the error code.
    Regards

  • HR Analytics - DAC error -Informatica mapping parameter values getting drop

    I am running HR Analytic full load against People HCM 9.1 but DAC load is failing. When I looked Informatica log files, I saw $$HR_WRKFC_EXTRACT_DATE values getting dropped. Like TO_DATE is showing O_DATE in the log file and due to that Informatica workflow is failing.
    Anyone face this kind of issue?
    Thanks

    Hi
    This is a known issue with 9.0.1 Informatica (CR 256495). Ref: ID 1325631.1
    Workaround to this issue, is to override INFA_CODEPAGENAME variable in the Integration Services to Latin1 and change the Integration Service code page to ISO 8859-1 Western European.
    1. In Informatica admin console, go to Integration Service -> Processes tab
    2. Edit the Codepage and change it to 'ISO 8859-1 Western European'
    3. Edit Environment variables and add:
    LANG = C
    INFA_CODEPAGENAME = Latin1
    4. Restart integration service.
    Let me know if this helped.
    Regards,
    Jay

  • Informatica parent child mapping

      down votefavorite I have a scenario where suppose saycountry      province      city      zip
    ind           ts           hyd       xyz
    ind          maha          mum       abc
    Desired output:id   name         parent-id
    1    india         1
    2    telengana     1
    3    hyderabad     2
    4    xyz           3
    5    mumbai        1
    6    abc           5
    I have to do it with informatica mapping any ideas how do I show this parent child relationship. I am not able to get the logic for thisThanksKumar

    Thanks for the thoroughness. There was a mistake in moving the code over for the forum. The field names are correct throughout the original source code.
    BASE_OBJECT_ID is used throughout.
    I suspect the problem lies in the one-to-many sampleItem(s) relationship that is based upon the subclassed item class. (The relationship is actually "sampleItems" in the real code and somehow got changed in the move over.)
    The problem may lie in the mapping of the attribute override in the child class to the referencing of the item class from the parent side of the relationship in the Sample class.
    I further suspect this may be specific to Eclipselink based upon other postings I've seen on the web that have similar problems...
    Any thoughts?
    Edited by: Chris-R on Mar 3, 2010 9:56 AM

  • Export all tables' name under user by informatica

    Hi,
    I have a requirement about informatica mapping,
    we want to user one mapping to export all tables' name under user,
    like sql :
    SELECT t.OWNER || '.' || t.TABLE_NAME ||
    FROM all_tables t
    WHERE t.OWNER = 'BIEBS';
    How to realize it?
    Thank you!

    Get the table all_table in source analyzer save it. While building the map choose this as ur source, modify source qualifier as select table_name,table_name from all_tables where table_name='BIEBS'.
    Once u have this, use an expression transformation to build ur expression and load it into a flat file.
    Please mark this answered or helpful if it solves your purpose.
    Rgds.

  • Informatica 9.0.1:- Table or view does not exist, error occured at recurive

    Hello every one,
    I am new to this field, I am getting error when I run my workflow in other folder, It is giving table or view does not exist, And I understood that the target table is not present,
    Actual table present is ws_twfs_olb_project_ds but in the session log it is showing wc_twfs_olb_project_ds1. I don't know why in the session log it is showing like that.
    Might be the problem occured during the importing session. If so, Can you please provide me how to resolve this issue.
    Thanks in advance!

    wc_twfs_olb_project_ds1 is an alias table and used in Informatica mapping to do self join(specifically in order to query different column multiple times).
    Ideally you sholud be checking the wc_twfs_olb_project_ds in the OLAP schema.Try connecting to OLAP scheam by toad or SQL*PLUS and let me know if you are able to view this table or not?..
    Regards,
    Veeresh Rayan

Maybe you are looking for

  • Desktop does not update when I change networks

    I'm having this odd behavior that I hope someone can shed some light on. My desktop is set to keep my icons arranged by name. All of my volumes appear on the desktop, as well. I've noticed in recent weeks that whenever I switch networks, my desktop s

  • What is the best anti phishing add on on for windows vista, i am a noice and want it understandable

    i am new to firefox, it has been suggested that i add an anti phishing add on, however even reading the reviews left me in the dark. i don't care to go thru a process just to get on the net.

  • Shrinking a Locally Managed Temporary Tablespace

    So, even thoguh the documentation is pretty clear about how to use this feature, I cannot get it to do what I expect it to do for me. And that would be shrinking the tempfile ;) Now lets face it, I have a large tempfile and want to resize it without

  • Mandatory fields to replicate vendor from ECC to SRM

    Hi all, I created a vendor in ECC with all mandatory fields. But when I tried to replicate the vendors to SRM from ECC, I am getting errors like mandatory fields are missing. Can you please tell me what are the mandatory fields in SRM when we replica

  • Display poplist from top of list when clicked?

    I have a list item, a poplist, with a number of items which cannot all be viewed at the same time so the user has to scroll through the list to find the required item. My problem is, when displaying the list it always shows the last n entries so the