Automatic Fields Mapping in Transformation - BI7

Greeting,
Getting some development procedures done automatically will definitely boost up productivity and efficiency. Examples include programs to:  create mass infoObjects, delete mass infoObjects, and activate mass InfoObjects.
Again, that are just examples.
In line with this, have you ever come across any program for automatic mass mapping in transformation?
Meaning instead of doing it manually, you just fill the source fields and the target field in the program then you press the button and have it ready on the spot.
Your quick response will be highly appreciated.

Hi,
I have not tried this out. But I guess you can use the class CL_RSTRAN_BUILD and generate transformations via an ABAP program.
Bye
Dinesh

Similar Messages

  • Problem with childs nodes and automatic key mapping in a Data Object

    Hi experts!
    I'm doing the service order tutorial from the mobile help at [this link|http://help.sap.com/saphelp_nwmobile71/helpdata/en/21/9b5b924c3b434fba4767731794b029/frameset.htm] and I have a problem...
    In the topic "Modeling the Equipment Data Object", says you have to mark the "Automatic Key Mapping" checkbox. So when I had to create a third child node ( the location node ) the system raised an exception with the message "Deselect automatic key mapping flag for more than two-level nodes". I'm trying deselecting the flag and creating the location node, but when I want mark again the automatic key mapping flag, this is disabled.
    What can I do to solve this and create the three child nodes with the flag marked? It's a configuration thing?
    Any help it's very welcome. Thanks in advance.
    Best regards,
    Simon.

    The thing is: Its not allowed to use automatic keymapping if you have more than two levels. This is why the message showed up, and this is why its been disabled.
    What automatic keymapping does: Figures out automatically which child node belongs to which parent (by guessing from the field name and type, which fields in the child correspond to which key fields of the parent).
    On three levels, this becomes more complicated => Its disabled.
    How to do keymapping yourself instead of having the DOE do it automatically: Do 'Explicit keymapping' from each child to its parent. Explicit keymapping is done by clicking on the corresponding menu button in the child node. Here you need to associate child node fields (they need not be key fields of the child, but they are allowed to be that as well) to each of its parent nodes key fields (so that each child can be associated to its parent).
    Cheers

  • How to change field mapping in IDOC

    Hi Experts,
    My requirement is to create a purchase requisition automatically through an interface using the IDOC sent by a *source client*.
    For this i have used PREQCR as the message type, PREQCR02 as basic type.
    E1BPEBANC & E1BPEBANTX are the segment type i have to use.
    I have created a Z process code in which the function module that i have used is IDOC_INPUT_PREQCR.
    With the help of this i can map all the field values to PR.
    Now there is a field in this idoc called vendor which comes from the source client.
    Now the value that is maintained in this field(vendor) should go to the Text Id that is maintained at the header level. ( i.e. the header text. i.e. In transaction ME51N, the text IDs which we can see at header level, the vendor value should go in one of this text Id ).
    Now how do i do this, because I am using standard funciton module IDOC_INPUT_PREQCR in which there is already field mapping that this vaue should go in this field and so on...
    Do I have to use an exit for the same ?? If yes then which is the exit ??
    Please suggest.
    Regards,
    Priti.

    Hi,
    I think there is no userexits available during the processing using function module "IDOC_INPUT_PREQCR".  You have to copy this function module to a Z function module.  Change as per your requirement and attach the new function module with the Z Process code created by you.
    Regards
    Vinod

  • Mapping in Transformations

    Hi Experts,
    Need help in mapping of transformations: for csv file i want to know how to map a field from source to target. this is besides the deault mappings. one field from source can be mapped to more than one fields in target?
    i get a short dump as part of creating DTP. your guidance needed.

    BI 2004S - Assuming that you already defined your datasource, have infopacakge to load your data into PSA from csv.
    You can map one source field to more than one target field. Tcode RSA1 - Modeling Infoprovider - Expand your infoprovider (cube, ODS, ...) to reveal Transformation (i.e. Update rules in 3.x) - Rightclick - choose Change, a new window will open on the right side - Position the cursor on the right edge of the source field, left click & hold while dragging the cursor to the target field, then release the mouse (you can then reapeat the same steps for the same source field to another target field)
    There are more than one place that you can create DTP. One of it is go to Modelling Datasources  - Rightclick on your datasource - Choose create Dta Transfer Process - Field in the fields - choose the green check mark (another screen willl open) - under Extraction tab, I recommend choose Extraction Mode DELTA, leave the rest with default values - Choose the icon near the hat to activate the DTP - To execute the DTP, choose the EXECUTE button under the Execute tab (this will load your data from PSA to the infoprovider only once, with the DELTA mentioned earlier). To reload again, run your infopackage FULL mode, and run your DTP DELTA mode.
    Hope this helps.

  • Field mapping and value mapping - Basic

    Hi,
    Assuming I have a repository containing a main table with 3 fields
    1)Product
    2)Description
    3)Manufacturer (lookup field)
    When i map the source file to the destination repository
    1) i map fields product and description of the source file and destination repository.
    2) map the manufacturer field between the source file and destination repository.
    3)Value map the contents of the manufacturer field between the source and destination.
    If i have a  source file like this.
    Product                                         Description                                Manufacturer
    a                                                         a                                              Hindustan lever
    b                                                         b                                              Hindustan lever
    a                                                         a                                              P&G
    b                                                         b                                              P&G
    Since i have made the manufacturer field a lookup field, it can hold the lookup values Hindustan lever and P&G.
    what happens to prduct ? i have made only a field mapping. Will the values a & b be stored in the destination repository this way?
    Is a value mapping required for product as well to hold both the values?
    Please help
    Thanks,
    Vignesh

    Hi Vignesh,
    You do not have to value map it. Import managerwill automatically assign both the values to product a and b. Basically it is set in the configuration options of the import manager. There is a setting "merge source records based on matching field". If that option is set to yes,  both the values will be assigned to product a and b respectively.
    Best Regards,
    Dheeraj

  • LSMW Source and Field Mapping

    Hi,
      I'm trying to use the LSMW.  What is the fastest way to maintain the source field and relationship?? Is it possible to define the source field and length in a text file and load it in the LSMW??
    Regards,
    Kit

    Hi Kit,
            u can create source fields in faster way.
    menupath
    In the 3rd step i.e., Maintain source fields
    source fields -> table maintenance
    here u can fill all the fields at a time.
    u can also maintain the fields in a file and copy and paste them in the Table maintenance .
    In 5th step i.e.,  Maintain Field Mapping and Conversion Rules
    Extras -> Auto field Mapping
    Now u get the auto field mapping setting pop-up.
    click ok.
    Now u Get the auto field mapping proposel pop-up.
    Here check Target field and Source field and click on Accept proposel push button for all the fields.
    with this the source and target field mapping are done automatically.
    reward if helpfull
    raam

  • CUP - field mapping from request to SU01

    Hi,
    I have configured field mapping in order for all my fields to provision to SU01 when a user is created in the system. However I am struggling with the accounting number which is a standard delivered field in CUP -> Customisation -> Request Form Screen. Should all fields used from the Request Form Screen not provision automatically?
    I tried to add Accounting Number in CUP -> Field Mapping -> Provisioning but with no success. When I add a record I pick up the accounting number under 'AC field' list but not under the 'Application field' list. I am however unsure if this step is required. I thought that all fields on the Request Form Screen should provision automatically and only custom fields should be mapped in CUP -> Field Mapping -> Provisioning.
    Any advise would be highly appreciated.
    Thanks
    Mo

    Hi,
    Does anyone know whether the fields used from CUP -> Customization -> Request Form Customization on a CUP request is automatically provisioned to SU01? Or should the provisioning of these fields be set up somewhere in configuration.
    Thanks
    Mo

  • Lazy loading or indirection for direct to field mapping striing object

    Hi,
    I need to achieve indirection for String instance variable, currently mapped using direct to field mapping.
    This field is a text column in sql server storing large data which is rarely read.
    Recently we have been noticing that it is causing lot of IO on our SAN system.
    Is there a way indirection can be achieved for this which is currently mapped as direct to field.
    I have tried using indirection in transformation mapping. But profiling db indicates above mentioned text column is still part of the select query.
    Thanks
    Nagesh

    Hi Nagesh,
    There is no support for lazy loading of DirectToField mappings in TopLink 10.1.3 POJO persistence. Typically, people who have large CLOBs or BLOBs will put them in a separate table to avoid loading them with their owner. If you can't refactor your data model to accomodate this things get more complicated. Do you ever update these large structures are they read only in your application? If they are read only then a very simple approach is to not map the large data column and instead lazily load the data with a report query.
    Ideally you would upgrade to TopLink JPA in 11g where lazy loading of Basic mappings is supported. :-)
    --Shaun                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • Technical Design Review Question: Data target Mapping and transformation

    I got my hands on technical design documentation for a project on COPA budget. I came up with a few questions but I will post them separately for fast closing and awards:
    In the discussions of Data target Mapping and transformation, there was a table of characteristics, showing dimensions, BW filed, Source field, data type, etc.
    1. What is the technique in deciding which characteristics get grouped together into a particular dimension?
    2. Why do some dimensions only have one characteristic and what is its significance?
    3. I saw one BW field, OVAL_TYPE (description= valuation type) included in three dimensions: Customer, Material and Valuation Type(only field=OVAL_TYPE). What is the significance of this repetition?
    Thanks

    Morning,
    To define Dimension means, to group all characters together, which do have a relationship "1 to n" and not "n to m" to reduce data volume (cardinality). If you reduce cardinality (which means to define 1:m relationship whenever it is possible and group them within one common dimension) you increase performance (make performance better) because of less data volume. It is essential to understand that this can not be changed easily after definition in PROD. So data modeling is also from this point of view very important.
    Example:
    An accounting object has a "m:n" relation ship to the accounting partner object. that's the reason way accounting object and accounting partner object do not belong to the same dimension.
    Ni hao
    Eckhard Lewin

  • How to add a new field in the Field Mapping

    Dear Freinds,
                   I have two un used fields in the  Source fields to the Target Fields ,
    i have to add the 3   fields which r there in the my source fields to that
    of the target fields ( as per the change in the requirement) in the 5th Step
    i.e Maintain Field Mapping and Conversion Rules
    the Target strucutre & the Source fields( by clicking the button
    Source field and selcting the fields from the popup menu) in the mapping and conversion step  ,i want to  add  3 fields which are availabe in the list of th source list  which i want to map to the Target strucutre  .
    When i tried it is adding in the othe target field
    presently i have mapping (5th sept ) as below.
       ZPA30_08                       LSMW
           Fields
               TABNAME                      Table Name
                                   Source:  ZP0008_STRC-PERNR (Personnel Number)
                                   Rule :   Default Settings
                                   Code:    ZPA30_08-TABNAME = 'ZPA30_08'.
               TCODE                        Transaction Code
                                   Rule :   Default Settings
                                   Code:    ZPA30_08-TCODE = 'PA30'.
    now i want to add 3 fields Pernr,wage type and Amount  , but when iam trying it is
    getting and saved in the TABNAME .
    How can i map please let me know .
    Regards
    syamla

    Hi,
    So you need that this new field have data in old records?
    1.- If you are in BI 7.0 and the logic or data for that New field are in the same Dimension, you can use a Remodeling to fill it. I mean if you want if you want to load from a Master Data from other InfoObject in the same Dim.
    2.- If condition "1" is not yours.
    First add the new field, then create a Backup Cube (both cubes with the new field) and make a full update with all information in the original Cube. The new field willl be empty in both cubes.
    Create an UR from BackUp_Cube to Original_Cube with all direct mapping and create a logic in the Start Routine of the UR (modiying the data_package) you can look for the data in the DSO that you often use to load.
    To do that both cubes have to be Datasources ( right click on Cube-> aditional function-> and I think is "Extract Datasource")
    Hope it helps. Regards, Federico

  • In XI Mapping multiple fields mapping to single target field.

    Hi Friends,
    In XI Mapping multiple fields mapping to single target field.
    For example my requirement is :
    Source Fields:(This RFC BAPI Structure)
    Empno                0-1
    EmpName           0-1
    Address             0-1
    Taget Field is:
    Details               0-1
    The above three fields passed to the Details Field. Here i am using Concat function
    But i have one query on that on,Every field having "line Break" required.
    Can you please help me out above this requirement.
    Thanks in Advance,
    Sateesh N.

    If you want a line break between the three fields, then try
    passing a,b,c to the udf and in the udf you would have
    return a+"\n"+b+"\n"+c;

  • ABAP code which will use function module during field mapping in LSMW

    Hi,
    During field mapping in LSMW, for one of the field called field, I want to use a function module 'QF05_RANDOM_INTEGER', which will generate random numbers and map that random number to that field.
    How do I write ABAP code which uses function module, during filed mapping for a particluar field in LSMW ?
    Please help...
    Regards,
    Tushar.

    Hi Tushar,
    You can write in FORM ROUTINE and LSMW.
    I hope you know how to write ABAP code at field level.
    If not ...
    Come to change mode->select the field->Rule->ABAP Code
    write a perform statement in ABAP code at field level.
    <b>PERFORM test</b> using val
                 changing valu.
    Go to Utilities(at menu)->switch layout-> extras ->select Layout->check Form routine-> it will display a FORM Routine... at bottom
    this is the place where you can write some code for
    <b> FORM test</b> using value
              changing value1
      FM call....
    .ENDFORM  
    Get the value1 returned and use for the Field population.
    Reward points if this Helps.
    Manish
    Message was edited by:
            Manish Kumar

  • How to create Rules with Flex Field mapping in the bpm worklist

    I Have created a flex field label and was able to map to the flex field attributes .
    But when i try to create a rules , I don't see the label or the flex attributes in the task payload .
    Can someone please help is understanding how to create Rules with Flex Field mapping in the bpm worklist .
    Even I am also searching for any scripts which will take the flex fields prompts and can directly create a label in the bpm worklist .
    Any pointers or suggestion is highly appreciated .

    Hi,
    SE38 -> Enter program
    Select Variants button and display. In the next screen, enter a variant name, (If not existing , press Create to create new one), else click on Change.
    Now the selection screen will display with a button "Variant Attributes" at the top.
    Click on that button.
    In the next screen, go to the selection variable column of the date field. Press F4 or drop down and select 'D' for date maintenance.
    In the column "Name of Variable (Input Only Using F4)" press F4 or drop down, select whichever kind of date calculation you want and save the variant.
    Now whenever you run the prgrm with this variant, date will be displayed by default.
    Regards,
    Subramanian

  • LDAP Field Mapping in 4.6C - Using WebAS 6.10+ as an LDAP Gateway

    Dear All,
      We have a need to enable CUP Functionality (we use GRC AC 5.3) for one of our oldest R/3 systems - on 4.6C. All other R/3 backends are on 4.7+ releases, so it's a multiple backend configuration for GRC AC.
      However, LDAP Field Mapping functionality is missing in 4.6C. It was enabled through LDAPMAP in the higher releases only.
      At the same time, I discovered in one SAP HR document a diagram, which shows that indeed 4.7+ can map and post data directly to LDAP, but for 4.6C and below you can use WebAS 6.10+ as an LDAP Gateway. Meaning that 4.6C calls through RFC some functions in the higher release R/3 system to use its functions for Field Mapping and further transfer of user data to the target LDAP server.
      But... I can not find anywhere how to configure 4.7 / 6.0 servers to act as an LDAP Gateway for the older 4.6C server to bypass its limitation - absence of built-in LDAP Field Mapping functionality.
      Advice on how to realise this concept will be highly appreciated.
    Thanks,
    Laziz

    Hi,
    In order to migrate users, groups and password you have to use the command ldapaddent as you did with this sintax:
    # ldapaddent -D "cn=Directory Manager" -w secret -f /etc/group group
    # ldapaddent -D "cn=Directory Manager" -w secret -f /etc/passwd passwd
    # ldapaddent -D "cn=Directory Manager" -w secret -f /etc/shadow shadowNote that you must use passwd instead of people container.
    I suggest you to check this article from BigAdmin http://www.sun.com/bigadmin/features/articles/nis_ldap_part1.jsp
    G.

  • Example of a custom field mapping?

    Ok, I admit it I am struggling here. I have simplified my example from
    what I actually have.
    I have a table that models a flat hierarchy
    ID | START_DATE | END_DATE | CLASSNAME | FIELD1 | FIELD2 | ...
    one of the objects in my hiearchy (CashFlow) has a field that is in fact
    another object called DatePeriod that contains two fields startDate and
    endDate.
    I understand that what I am trying to do is embed the DatePeriod object
    inside of the larger object when it get's persisted.
    I have the following metadata set-up
    <class name="CashFlow" persistence-capable-superclass="InstrumentFlow">
    <extension vendor-name="kodo" key="table" value="INSTRUMENT_FLOW"/>
    <extension vendor-name="kodo" key="pk-column" value="ID"/>
    <extension vendor-name="kodo" key="class-column" value="none"/>
    <extension vendor-name="kodo" key="lock-column" value="none"/>
    <field name="accrualPeriod" embedded="true"/>
    </class>
    and for my DatePeriod object
    <class name="DatePeriod">
    <extension vendor-name="kodo" key="table" value="INSTRUMENT_FLOW"/>
    <extension vendor-name="kodo" key="pk-column" value="ID"/>
    <extension vendor-name="kodo" key="lock-column" value="none"/>
    <extension vendor-name="kodo" key="class-column" value="none"/>
    <field name="startDate">
    <extension vendor-name="kodo" key="data-column" value="START_DATE"/>
    </field>
    <field name="endDate">
    <extension vendor-name="kodo" key="data-column" value="END_DATE"/>
    </field>
    </class>
    Every time I try to fetch a CashFlow object I get an error as KODO is
    trying to select the column 'ACCRUALPERIODX.'
    What am I doing wrong? Do I need to use a custom field mapping? If so
    where is the documentation to help me write a custom field mapping?
    A second question would be what happens if the DatePeriod object is used
    in a couple of places, I don't want to tie it's persistence to the
    INSTRUMENT_FLOW table.
    All help gratefully received
    Cheers
    Matt

    As you suspect, Kodo 2.x does not support embedded class mappings. Kodo
    3.0 will support embedded mappings.
    In the meantime, you can create a custom mapping, but unfortunately our
    documentation for custom mappings is lacking right now. Given how simple
    your DatePeriod object is, you're probably better off with something
    simpler (and as a bonus, less bound to Kodo):
    Just make your DatePeriod class and the field that holds the DatePeriod
    instance non-persistent. In the class that has the (now non-persistent)
    DatePeriod field, add two new persistent fields for the startDate and
    endDate. Then just use internal logic to construct the DatePeriod from
    the startDate and endDate. You can do this using the
    javax.jdo.InstanceCallbacks methods, or just do the logic in your setter
    and getter methods for the DatePeriod.

Maybe you are looking for

  • Adding more RAM to iMac 24"

    Hey Everyone, I'm considering buying a 24" iMac. I've been told that a way to save some money, is to get the RAM from elsewhere and install it myself. I'm just checking whether this is actually possible and how hard it is to actually do it. Please he

  • Insert Date into Indesign CS3 template

    I had a script that worked great in CS2 but when I upgraded to CS3 it doesn't work. I posted on here to find the problem with no help so I"m starting from scratch. Its it possible to insert the date into a template with a script in CS3 Example: SUNDA

  • Human Workflow Notification email format

    Hello, I've configured email notification in Human Workflow and able to receive email in plain text format in one line. How do i get email message with html format with text in seperate lines. I even added &#x0D; and xpath command for new line inside

  • Installing problem on HP notebook computer!

    I recently purchased Solaris 8 4/01 media kit Intel platform edition. After installing those all 8 setup CDs successfully, I got a blank screen when I reboot it. The error message says it cannot find a suitable video card driver. Unfortunately, I can

  • Printed cheque - comma and decimal point

    Hi, My company use pre-printed cheque to make payment. However, there is some error with the printed cheque. When printed, the decimal point and the comma is somehow mixed out.   For example, amount of $1,300.50   -   will appears as $1.300,50 Why is