How to use crm xif to finish batch creation of BP in crm

Hi,
   we need to upload BP data in CRM from flat file and many fields to be filled. And seems CRM provides XIF to do this. But not many materials listed in SAP help. could anyone give some suggestion ?
ths

Hi,
I've looked into your reply.
For your first query, you have to provide identification numbers. It is like an identification for the external systems when you create the same.
You do not have to pass Guid, since inside the code logic, if it is fresh insert there will be GUIDs generated also, which will cause the confusion. Let the system generate the GUID for fresh insert. You provide the identification numbers.
For your second query you need to provide the sales area data in the following segment
E101RMXIF_PR_MATERIAL_SALES of Idoc CRMXIF_PRODUCT_MATERIAL_SAVE01. Also kindly make sure that you have done the necessary customising for sales area of products. Hoping the 2nd error comes during the update of product data.
I hope this helps.
Regards,
Venkat

Similar Messages

  • How to use CRM authorization object.

    Hi All,
    I have a specific requirement to restrict user while he/she tries to save a record. It appears that if that restrictions are implemented the save logic for an entity has to be changed because there are some validation regarding relationship management in SAP system. SO I need to bypass that validation to allow some users of specific(Marketting) role to save the entity record bypassing that validation. here I am planning to use the CRM authorization objects. But dont know how to use these and which authorization object to refer.
    Please let me know if you guys have any idea.
    Regards,
    Bikramjit.

    Hi Bikramjit.,
    You might need to create a Custom authorization object and then use it. Else you can create one Z table and maintain the User ID of all users. The mainatin one field with flag and set it to X for the user that are aloowed to save the transaction.
    Also once you maintain the table, generate the table maintenance so that it becomes easier for future use.
    Hope this helps

  • How to use BAPI_MATERIAL_AVAILABILITY Function Module at batch level?

    How to use BAPI_MATERIAL_AVAILABILITY FM to check material available at batch level?
    And another question is what's the meaning of 'CHECK_RULE' in this FM. Thanks!

    Field name : PRREG
    fcheck more details on  f1 help...
    check_group defines the checking procedure to be used for  availability check in individual applications.
    along with the checkg group, checking rule  specifies the final procedure for chkg..
    eg:
    ST  for  stock transport order
    AQ for SD order project srock
    also chk this help doc.
    http://help.sap.com/saphelp_470/helpdata/en/cf/70124adf2d11d1b55e0000e82de178/content.htm
    search the forum for sample code..
    How to use BAPI_MATERIAL_AVAILABILITY Function Module....

  • How to use crm blueprint

    Hi, I'm new to using blueprint in creating the UI/portal page.
    Can anybody direct me on how to use blueprints to create buttons, tables, textfields, etc, and how to process its events.
    Thanks.

    Arnold,
    U need to go through PCUI Cook Book....and more clarifications will get in CRM forums. Here is the link where u can find the link to PCUI cook book.
    PCUI COOK BOOK
    Regards,
    Arun.

  • How to use own UOM conversion for PO creation...

    Hello Experts,
    We have a requirement wherein we need to use our OWN UOM conversion(UMREN and UMREZ)
    instead of getting it in the inforecord or the material master. Currently we are using BAPI
    BAPI_PO_CREATE1 to create PO.
    Is there anything that I can do to make the PO creation bypass the inforecord and
    the material master UOM conversion. For example in ECC we have 1 CV = 10 EA. We want it to be 1 CV = 15 EA.

    Hi Guys,
    Any help? Any exits/BADIs/enhancement spots that I can use?

  • How to use reflection package for dynamic object creation

    I want to convert a "DataObject (SDO) " into "Pure java object".
    For this i want to create a java object with the fields in Dataobject, and this should be in generic way.

    Use Java reflection if you have the java class already created.
    Otherwise use Java IO API for creating the .java file and compile it dynamically by calling javac exe from java code and executing using reflection

  • How to use LSMW on CRM for bulk master data creation?

    Hello Experts,
    I need your help to able to use LSMW on CRM for bulk master data creation. Below is described my detailed requirement; -
    We have CRM 2007 integrated with SAP ISU (ECC6.0). As per business process all Business Master Data (BP details, Business Agreement details etc) and Technical Master Data (Connection Object, PoD etc.) is created in CRM and then it is replicated in ISU. We are going to start Assembly Test soon. Now we need to have all master data in place. Tool we want to use is LSMW.
    Now can you help me right from scratch on how to use LSMW for bulk master data creation in CRM?
    Please give me documents / links etc that is relevant.
    Regards,
    Rohit

    Hi,
    I have looked into your query.
    These are some information which might be useful to you.
    What is LSMW?
    Legacy System Migration Workbench or LSMW is SAP standard program that can be used to easily load / change master data or transactions in SAP without any programming at all. It can be used to upload data at cutover, such as: open orders, contracts, materials, vendors, stock on hand. It can also be used in a production system to make changes to existing data. Very limited authorization checks are available on the transaction, so the use of it must be very well controlled.
    One input file can be used to create all the views of the material master. For all the material types. Even if different views and fields are used for the different materials.
    Steps to take
    1. Create text file with data to be loaded
    2. Set up and run LSMW
    The following link will provide good information on LSMW Idoc method:
    [http://help.sap.com/saphelp_nw04/helpdata/en/ad/2d54a41d7011d2b42e006094b944c8/frameset.htm]
    I hope this helps.
    Regards,
    Venkat

  • How to use batch files to send the load from ODC to UCM?

    Hello all.
    We're trying to do this along this week and, until now, we couldn't figure out how to use batch files containing metada on ODC and commit the output files to UCM.
    So these are our specs:
    VM 1
    ODC - 10gR3
    OS - Windows XP 32 bits SP3
    DB - Oracle DB 11.2
    VM 2
    No ODC
    UCM 11.1.1.6
    OS - Enterprise Linux Enterprise Linux Server release 5.5 (Carthage)
    We've opened a ticket with Oracle Support to know how we should do this. The number is SR 3-6108348211.
    Let me try to explain what we want to do:
    1 - we have a batch file with all metadata of many files from UCM
    2 - we want to use this file in ODC, so the application will convert the images and commit them to UCM
    3 - we should see all these files, their metadata, thumbnails and related content in UCM
    4 - this is a batch file example we want to send to ODC:
    @Properties LocalData
    IdcService=CHECKIN_UNIVERSAL
    dRevLabel = 1
    primaryFile=/tra/Relatorio_UHPF-03-04-2008-M-RL-EA.pdf
    dDocName=TRA_33
    dDocTitle=UHPF-03-04-2008-M-RL-EA
    dSecurityGroup=MAM_TRA
    dDocType=Document
    dDocAuthor=luis.neotropical
    dInDate=14/05/08
    xC_CLASSIFICACAO=EA - Educação Ambiental
    xC_DATA=01/04/08
    xC_DATAVALIDADE=30/04/08
    xC_LOCALIZACAOFISICA=N/D
    xC_USUARIO_DEP=LUIS FERNANDO DA CAMARA
    xC_TIPO=RL - Relatório
    xC_USINA=UHPF - USINA HIDRELÉTRICA PASSO FUNDO
    xC_FICHA=MAM - Relatorios Ambientais
    xC_IDDOCMAN=TRADocID=157
    xC_IDANEXODOCMAN=TRAFileID=33
    xC_DESTINOINICIAL=Tractebel
    xC_ORIGEM=Neotropical
    xC_FORMATODOCUMENTO=Digital
    xC_GENERODOCUMENTO=
    xC_DOCPRINCIPAL=Sim
    @end
    <<EOD>>
    5 - So, when we use this batch file on ODC, the file (or image) will be converted, all metadata attached and then sent to UCM.
    But we don't know how to do the link between ODC and UCM. The Oracle Support says it is necessary to do some customization. In the books I could not find any info about how to do, just generic information or how to use ODC.
    Is there a way (without customization) to do that? Can anyone give us a step-by-step guide?
    Thanks for all help and guidance.
    Fernando

    What is the one-sentence requirement or issue you're trying to solve?
    If you only want content to be converted to pdf, then you should already have that with your items checked into UCM (if you have IBR enabled).
    if you really want to feed content into ODC using a batchfile, my first thought would be to create a transformation utility/app that will convert your batchfiles to the expected ODC import server folder job format. then you'll have to move or map the content to match.
    Again, without knowing the exact issue/business requirement you're trying to work through, I'd say a redesign of your ingestion process may be in order.
    If you feed content directly into ODC through any supported ODC method (watch folder, email, scanner, fax, etc), then you'll be able to use everything ootb as you'll be able to auto-commit directly into UCM or add an indexing step and manual commit to UCM.
    ODC is basically a front end for indexing and extra ingestion options to your content repo.
    IPM is basically the same, but it includes some extra process-related details (process metadata, process integrations/bpel/bpm, other system integrations/ebs/crm, etc). I don't feel that IPM would help with the transformation-specific question you've put out there.
    -ryan

  • How to use Batch operation for two xsodata services?

    Hi All,
    I have two xsodata services. How to use submit batch for two xsodata services
    Thanks,
         Mj

    Gateway Batch Calls from SAPUI5

  • How to use batch mode in cq dam - new questions added

    Hi I am doing DAM migration. I am using custom code to pick the image and create asset by calling Assetmanager.createAsset which creates dam assets.
    This code I am calling through the scheduler which will run for every 0.5 hr. Around 1000 images we are trying to push in each cycle. I came acrosss batchMode for saving all the images per batch so that it will reduce the migration time But I am not sure how to use it.
    .setBatchMode(true); isBatchMode are the methods available in Asset api.
    Please share your thoughts on this.
    Some more Point I want to ask.
    1.What is the best appraoch when we are pushing images to DAM
         Stopping the workflows push all the images and run the work flows. (Is it possible to do this?)
         Do the both simultaniously
         Do it in the bath (i.e push bunch of images and then run the workflows)
    2.When i am migrating images (from file system which is residing in the same box where cq is running )it is almost taking 0.5 hr for 1000 images (avg size of image is 100kb). Is there any possibility to      improve this process.
    3. When I am adding images I am adding subassets to some of the images so when ever I add subasset cq is creating new version of the asset which is not required for me. What is the best approach to avoid      creating too many versions for images in DAM. By doing this is there any performance improve in the system.
    Thanks,
    Maruthi

    Hi Jorg,
    Thank you.
    I will try out batch saving option.
    I have few more doubts.
    1. Today I tried for moving aroung 5k images through the scheduler in my local machine. It went smoothly with out much issues but as usual some of the renditions are failing for some of the sub assets. This is about 3 for every 10 images. Renditions are generating properly for all assets it is failing in case of sub assets only. Along with this I am getting version exceptions. Below are errors which I am facing very frequently during migration dry run. This seems to be beacuse of the non synchronized execute methods in WF steps. What is your call on these issues how can we avoid these kind of errors.
    20.02.2013 19:16:00.082 *ERROR* [JobHandler: /etc/workflow/instances/2013-02-20/model_1361367956960012000:/content/dam/<IMAGE PATH>/subassets/<Image Name>/jcr:content/renditions/original] com.day.cq.dam.core.impl.AssetImpl setRendition: cannot set new rendition [cq5dam.thumbnail.140.100.png] for asset [<IMAGE PATH>/subassets/<Image Name>]:  javax.jcr.version.VersionException: Unable to perform operation. Node is checked-in.
    20.02.2013 19:17:01.660 *ERROR* [JobHandler: /etc/workflow/instances/2013-02-20/model_1361368020884400000:/content/dam/<IMAGE PATH>/jcr:content/metadata] com.day.cq.dam.core.impl.handler.xmp.NCommXMPHandler Stack Trace: java.lang.Exception: Unable to create revision.
    Caused by: javax.jcr.RepositoryException: Unable to update item: item.save()
    Caused by: javax.jcr.InvalidItemStateException: Item cannot be saved because it has been deleted externally: item.save()
    20.02.2013 19:19:08.666 *WARN* [JobHandler: /etc/workflow/instances/2013-02-20/model_1361368144852589000:/content/dam/<IMAGE PATH>/subassets/<Image Name>/jcr:content/renditions/original] com.day.cq.dam.core.process.ExtractMetadataProcess unexpected error occurred during metadata extraction. Cause: Unable to perform operation. Node is checked-in. javax.jcr.version.VersionException: Unable to perform operation. Node is checked-in.
    2. Can you share about experience in DAM migrations. What is the avg time it takes for migrating about 1 to 1.2 lac images which are aroung 60    gig. We thought of running the scheduler (this.scheduler.addPeriodicJob(jobName, job, config, period, canRunConcurrently);)for every 0.5 hr in  that case it will go on for 3 days.

  • How to use BAPI_INB_DELIVERY_CONFIRM_DEC do batch split

    Hi, everyone:
    I want to do batch split in inbound or outbound delivery by function.
    I found BAPI_INB_DELIVERY_CONFIRM_DEC, but our system is not a Decentralized System and I wandered in so many parameters.
    How to use this fuction?
    Thanks.

    Hi Junwen,
    Did you call the BAPI_TRANSACTION_COMMIT afterwards?
    Before this the data is not saved. After that it is saved.
    Regards,
    Rob.

  • How to Use BAPI_OUTB_DELIVERY_CHANGE for Batch Split

    Hi Experts, I have the follow delivery:
    Item        Material   Deliv.Qty    Un
    10     739911     2     PQT
    And I want to obtain this:
    Item        Material   Deliv.Qty    Un                            Batch
    10     739911     0     PQT
    900001     739911     1     PQT          ZZ00060
    900002     739911     1     PQT          ZZ00061
    But, after using the Bapi I obtained this:
    Item        Material   Deliv.Qty    Un                            Batch
    10     739911     1,372     PQT
    900001     739911     0,314     PQT          ZZ00060
    900002     739911     0,314     PQT          ZZ00061
    So as you can see the problem is with the "Deliv.Qty". I expected:
    Item        Deliv.Qty
    10            0
    900001     1     
    900002     1     
    But I obtained:
    Item        Deliv.Qty
    10     1,372
    900001     0,314
    900002     0,314
    The parameters that I am using in the Bapi are:
    HEADER_DATA                                   
    DELIV_NUMB = 801174646
    HEADER_CONTROL                                
    DELIV_NUMB = 801174646
    DELIVERY = 801174646
    TECHN_CONTROL
    UPD_IND = U
    ITEM_DATA (three records):   
    DELIV_NUMB = 801174646
    DELIV_ITEM = 000010  
    MATERIAL = 739911  
    FACT_UNIT_NOM = 1   
    FACT_UNIT_DENOM = 1   
    DELIV_NUMB = 801174646  
    DELIV_ITEM = 900001     
    MATERIAL = 739911     
    BATCH = ZZ00060    
    HIERARITEM = 000010     
    USEHIERITM = 1          
    DLV_QTY = 1
    DLV_QTY_IMUNIT = 1
    FACT_UNIT_NOM = 1      
    FACT_UNIT_DENOM = 1      
    DELIV_NUMB = 801174646  
    DELIV_ITEM = 900002     
    MATERIAL = 739911     
    BATCH = ZZ00061    
    HIERARITEM = 000010     
    USEHIERITM = 1          
    DLV_QTY = 1
    DLV_QTY_IMUNIT = 1
    FACT_UNIT_NOM = 1      
    FACT_UNIT_DENOM = 1
    ITEM_CONTROL (three records):
    DELIV_NUMB = 801174646
    DELIV_ITEM = 000010  
    CHG_DELQTY = X
    DELIV_NUMB = 801174646
    DELIV_ITEM = 900001  
    CHG_DELQTY = X
    DELIV_NUMB = 801174646
    DELIV_ITEM = 900002  
    CHG_DELQTY = X
    So I am missing something but I don't know what.
    Can any one help me with this please?
    PD: I have looked the following forums in order to use the Bapi:
    How to Use BAPI_OUTB_DELIVERY_CHANGE for Batch Split
    batch split using BAPI_OUTB_DELIVERY_CHANGE
    Re: Help in BAPI_OUTB_DELIVERY_CHANGE(batch-split)
    BAPI_OUTB_DELIVERY_CHANGE - Batch Splits don't work
    Help for BAPI_OUTB_DELIVERY_CHANGE

    Hey,
    My code is for the inbound delivery but it should work the same.
    Give this a try.
    REPORT  z_delivery_batch_split.
    DATA:header_data  LIKE  bapiibdlvhdrchg,
    header_control  LIKE  bapiibdlvhdrctrlchg,
    delivery  LIKE  bapiibdlvhdrchg-deliv_numb,
    ls_return LIKE bapiret2,
    item_data  TYPE TABLE OF  bapiibdlvitemchg,
    item_control  TYPE TABLE OF bapiibdlvitemctrlchg,
    ls_item_data  LIKE  bapiibdlvitemchg,
    ls_item_control LIKE  bapiibdlvitemctrlchg,
    return TYPE TABLE OF bapiret2 WITH NON-UNIQUE KEY type.
    header_data-deliv_numb = '1800005005'.
    header_control-deliv_numb = '1800005005'.
    delivery = '1800005005'.
    ls_item_data-deliv_numb = '1800005005'.
    ls_item_data-deliv_item = '900001'.
    ls_item_data-material = '000000000020067722'.
    PERFORM create_batch CHANGING ls_item_data-batch.
    ls_item_data-hieraritem = '000010'.
    ls_item_data-usehieritm = '1'.
    ls_item_data-dlv_qty = 80.
    ls_item_data-dlv_qty_imunit = 80.
    ls_item_data-fact_unit_nom = '1'.
    ls_item_data-fact_unit_denom = '1'.
    ls_item_data-sales_unit = 'EA'.
    ls_item_data-base_uom = 'EA'.
    APPEND ls_item_data TO item_data.
    ls_item_data-deliv_numb = '1800005005'.
    ls_item_data-deliv_item = '900002'.
    ls_item_data-material = '000000000020067722'.
    PERFORM create_batch CHANGING ls_item_data-batch.
    ls_item_data-hieraritem = '000010'.
    ls_item_data-usehieritm = '1'.
    ls_item_data-dlv_qty = 60.
    ls_item_data-dlv_qty_imunit = 60.
    ls_item_data-fact_unit_nom = '1'.
    ls_item_data-fact_unit_denom = '1'.
    ls_item_data-sales_unit = 'EA'.
    ls_item_data-base_uom = 'EA'.
    APPEND ls_item_data TO item_data.
    ls_item_control-deliv_numb = '1800005005'.
    ls_item_control-deliv_item = '900001'.
    ls_item_control-chg_delqty = 'X'.
    APPEND ls_item_control TO item_control.
    ls_item_control-deliv_numb = '1800005005'.
    ls_item_control-deliv_item = '900002'.
    ls_item_control-chg_delqty = 'X'.
    APPEND ls_item_control TO item_control.
    break cgavin.
    CALL FUNCTION 'BAPI_INB_DELIVERY_CHANGE'
       EXPORTING
         header_data          = header_data
         header_control       = header_control
         delivery             = delivery
    *   TECHN_CONTROL        = TECHN_CONTROL
       TABLES
    *   HEADER_PARTNER       = HEADER_PARTNER
    *   HEADER_PARTNER_ADDR  = HEADER_PARTNER_ADDR
    *   HEADER_DEADLINES     = HEADER_DEADLINES
         item_data            = item_data
         item_control         = item_control
    *   ITEM_SERIAL_NO       = ITEM_SERIAL_NO
    *   EXTENSION1           = EXTENSION1
    *   EXTENSION2           = EXTENSION2
         return               = return
    *   TOKENREFERENCE       = TOKENREFERENCE
    *   HANDLING_UNIT_HEADER = HANDLING_UNIT_HEADER
    *   HANDLING_UNIT_ITEM   = HANDLING_UNIT_ITEM
    *   PARTIAL_GR_OBJECTS   =          PARTIAL_GR_OBJECTS
    READ TABLE return
    INTO ls_return
    WITH TABLE KEY type = 'E'.
    IF sy-subrc = 0.
       MESSAGE ID ls_return-id TYPE ls_return-type NUMBER ls_return-number WITH ls_return-message_v1 ls_return-message_v2
       ls_return-message_v3 ls_return-message_v4.
    ENDIF.
    CALL FUNCTION 'BAPI_TRANSACTION_COMMIT'
      EXPORTING
        wait          = 'X'.
    break cgavin.
    *&      Form  CREATE_BATCH
    *       text
    *      <--P_LS_ITEM_DATA_BATCH  text
    FORM create_batch  CHANGING p_ls_item_data_batch.
       DATA: ls_bncom TYPE bncom.
       ls_bncom-matnr = ls_item_data-material.
       ls_bncom-werks = 'C333'.
       ls_bncom-lgort = '3000'.
       SELECT SINGLE mtart
         FROM mara
         INTO ls_bncom-mtart
         WHERE matnr = ls_bncom-matnr.
       CALL FUNCTION 'VB_NEXT_BATCH_NUMBER'
         EXPORTING
           matnr                          = ls_bncom-matnr
           werks                          = ls_bncom-werks
           check_batch                    = 'X'
           check_material                 = ''
           message_when_auto              = ' '
           x_bncom                        = ls_bncom
           lock_batch                     = ' '
         IMPORTING
           new_charg                      = p_ls_item_data_batch
         EXCEPTIONS
           no_material                    = 1
           no_plant                       = 2
           material_not_found             = 3
           plant_not_found                = 4
           no_batch_handling              = 5
           batch_exist                    = 6
           no_number                      = 7
           illegal_batch_number           = 8
           interval_not_found             = 9
           object_not_found               = 10
           interval_overflow              = 11
           error_automatic_batch_number   = 12
           cancelled                      = 13
           automatic_batch_num_not_active = 14
           OTHERS                         = 15.
       IF sy-subrc = 0.
         CALL FUNCTION 'BAPI_TRANSACTION_COMMIT'
           EXPORTING
             wait = 'X'.
       ENDIF.
    ENDFORM.                    " CREATE_BATCH

  • How to use batch processing with jpg pictures taken with iphone cameras?

    How to use batch processing with jpg pictures taken with iphone cameras?

    Open the editor . Go to file>process multiple files
    Sent from my iPad

  • How to use WS_DOWNLOAD funciton module in back ground (Batch Job)

    Hi all,
        Can any one tell,
    How to use WS_DOWNLOAD funciton module in back ground (Batch Job)
    Thanks,
    Ravi Kumar.

    As disussed above you can use DATASET techniques to write the file to the application server. Then you could do a bdc call to Transaction CG3Y to transfer the file to the presentation server.
    Just F1 help on keyword DATASET in SE80 for more information.
    BDC Call Info found here: <a href="http://www.sap-img.com/bdc.htm">http://www.sap-img.com/bdc.htm</a>
    Regards,
    Philip Johannesen

  • How to use custom Finish for dependency?

    Hi,
    in my project hi have tasks with sub-tasks duration depends on the number of pieces.
    I have a custom number field for number of pieces and a custom duration field for the duration of the subtask for the single piece.
    I then created a custom duration field to calculate the duration of the total time required for the single subtask for the all samples.
    Then I created a custom Finish field that uses this formula to calculate when each subtask should finish is ProjDateAdd([Start],[Duration1],[Project Calendar])
    The problem I have is that the calendar uses default Duration and Finish to schedule the tasks so Finish is always different from Finish1.
    Is there a way to update the default Duration and Finish as function of custom duration and finish?
    Michele

    michele_1,
    Well, no and yes. On the "no" side, there is no way to have custom fields update the built-in fields directly. However, it could be done with some VBA. Actually you could also use paste links but that is highly NOT recommended because you would
    set yourself up for high probability of file corruption.
    So, if you are willing to use VBA then this simple macro will do it. Note, it assumes the Number1 field is the quantity and the Duration1 field is the time for a quantity of one.
    Sub SamplesXDur()
    Dim t As Task
    For Each t In ActiveProject.Tasks
        t.Duration = t.Number1 * t.Duration1
    Next t
    End Sub
    This is a sample file prior to running the macro.
    And this is what it looks like after the macro is run.
    But just on the side, why do you need a custom finish field, or even a second custom duration field? Given that you know the start date, all you need to set the Duration field, and hence the Finish field, is the result from the custom number field, which
    you state is the quantity of pieces, times the time (duration) for a single piece.
    Hope this helps.
    John

Maybe you are looking for