How to use crm xif to finish batch creation of BP in crm
Hi,
we need to upload BP data in CRM from flat file and many fields to be filled. And seems CRM provides XIF to do this. But not many materials listed in SAP help. could anyone give some suggestion ?
ths
Hi,
I've looked into your reply.
For your first query, you have to provide identification numbers. It is like an identification for the external systems when you create the same.
You do not have to pass Guid, since inside the code logic, if it is fresh insert there will be GUIDs generated also, which will cause the confusion. Let the system generate the GUID for fresh insert. You provide the identification numbers.
For your second query you need to provide the sales area data in the following segment
E101RMXIF_PR_MATERIAL_SALES of Idoc CRMXIF_PRODUCT_MATERIAL_SAVE01. Also kindly make sure that you have done the necessary customising for sales area of products. Hoping the 2nd error comes during the update of product data.
I hope this helps.
Regards,
Venkat
Similar Messages
-
How to use CRM authorization object.
Hi All,
I have a specific requirement to restrict user while he/she tries to save a record. It appears that if that restrictions are implemented the save logic for an entity has to be changed because there are some validation regarding relationship management in SAP system. SO I need to bypass that validation to allow some users of specific(Marketting) role to save the entity record bypassing that validation. here I am planning to use the CRM authorization objects. But dont know how to use these and which authorization object to refer.
Please let me know if you guys have any idea.
Regards,
Bikramjit.Hi Bikramjit.,
You might need to create a Custom authorization object and then use it. Else you can create one Z table and maintain the User ID of all users. The mainatin one field with flag and set it to X for the user that are aloowed to save the transaction.
Also once you maintain the table, generate the table maintenance so that it becomes easier for future use.
Hope this helps -
How to use BAPI_MATERIAL_AVAILABILITY Function Module at batch level?
How to use BAPI_MATERIAL_AVAILABILITY FM to check material available at batch level?
And another question is what's the meaning of 'CHECK_RULE' in this FM. Thanks!Field name : PRREG
fcheck more details on f1 help...
check_group defines the checking procedure to be used for availability check in individual applications.
along with the checkg group, checking rule specifies the final procedure for chkg..
eg:
ST for stock transport order
AQ for SD order project srock
also chk this help doc.
http://help.sap.com/saphelp_470/helpdata/en/cf/70124adf2d11d1b55e0000e82de178/content.htm
search the forum for sample code..
How to use BAPI_MATERIAL_AVAILABILITY Function Module.... -
Hi, I'm new to using blueprint in creating the UI/portal page.
Can anybody direct me on how to use blueprints to create buttons, tables, textfields, etc, and how to process its events.
Thanks.Arnold,
U need to go through PCUI Cook Book....and more clarifications will get in CRM forums. Here is the link where u can find the link to PCUI cook book.
PCUI COOK BOOK
Regards,
Arun. -
How to use own UOM conversion for PO creation...
Hello Experts,
We have a requirement wherein we need to use our OWN UOM conversion(UMREN and UMREZ)
instead of getting it in the inforecord or the material master. Currently we are using BAPI
BAPI_PO_CREATE1 to create PO.
Is there anything that I can do to make the PO creation bypass the inforecord and
the material master UOM conversion. For example in ECC we have 1 CV = 10 EA. We want it to be 1 CV = 15 EA.Hi Guys,
Any help? Any exits/BADIs/enhancement spots that I can use? -
How to use reflection package for dynamic object creation
I want to convert a "DataObject (SDO) " into "Pure java object".
For this i want to create a java object with the fields in Dataobject, and this should be in generic way.Use Java reflection if you have the java class already created.
Otherwise use Java IO API for creating the .java file and compile it dynamically by calling javac exe from java code and executing using reflection -
How to use LSMW on CRM for bulk master data creation?
Hello Experts,
I need your help to able to use LSMW on CRM for bulk master data creation. Below is described my detailed requirement; -
We have CRM 2007 integrated with SAP ISU (ECC6.0). As per business process all Business Master Data (BP details, Business Agreement details etc) and Technical Master Data (Connection Object, PoD etc.) is created in CRM and then it is replicated in ISU. We are going to start Assembly Test soon. Now we need to have all master data in place. Tool we want to use is LSMW.
Now can you help me right from scratch on how to use LSMW for bulk master data creation in CRM?
Please give me documents / links etc that is relevant.
Regards,
RohitHi,
I have looked into your query.
These are some information which might be useful to you.
What is LSMW?
Legacy System Migration Workbench or LSMW is SAP standard program that can be used to easily load / change master data or transactions in SAP without any programming at all. It can be used to upload data at cutover, such as: open orders, contracts, materials, vendors, stock on hand. It can also be used in a production system to make changes to existing data. Very limited authorization checks are available on the transaction, so the use of it must be very well controlled.
One input file can be used to create all the views of the material master. For all the material types. Even if different views and fields are used for the different materials.
Steps to take
1. Create text file with data to be loaded
2. Set up and run LSMW
The following link will provide good information on LSMW Idoc method:
[http://help.sap.com/saphelp_nw04/helpdata/en/ad/2d54a41d7011d2b42e006094b944c8/frameset.htm]
I hope this helps.
Regards,
Venkat -
How to use batch files to send the load from ODC to UCM?
Hello all.
We're trying to do this along this week and, until now, we couldn't figure out how to use batch files containing metada on ODC and commit the output files to UCM.
So these are our specs:
VM 1
ODC - 10gR3
OS - Windows XP 32 bits SP3
DB - Oracle DB 11.2
VM 2
No ODC
UCM 11.1.1.6
OS - Enterprise Linux Enterprise Linux Server release 5.5 (Carthage)
We've opened a ticket with Oracle Support to know how we should do this. The number is SR 3-6108348211.
Let me try to explain what we want to do:
1 - we have a batch file with all metadata of many files from UCM
2 - we want to use this file in ODC, so the application will convert the images and commit them to UCM
3 - we should see all these files, their metadata, thumbnails and related content in UCM
4 - this is a batch file example we want to send to ODC:
@Properties LocalData
IdcService=CHECKIN_UNIVERSAL
dRevLabel = 1
primaryFile=/tra/Relatorio_UHPF-03-04-2008-M-RL-EA.pdf
dDocName=TRA_33
dDocTitle=UHPF-03-04-2008-M-RL-EA
dSecurityGroup=MAM_TRA
dDocType=Document
dDocAuthor=luis.neotropical
dInDate=14/05/08
xC_CLASSIFICACAO=EA - Educação Ambiental
xC_DATA=01/04/08
xC_DATAVALIDADE=30/04/08
xC_LOCALIZACAOFISICA=N/D
xC_USUARIO_DEP=LUIS FERNANDO DA CAMARA
xC_TIPO=RL - Relatório
xC_USINA=UHPF - USINA HIDRELÉTRICA PASSO FUNDO
xC_FICHA=MAM - Relatorios Ambientais
xC_IDDOCMAN=TRADocID=157
xC_IDANEXODOCMAN=TRAFileID=33
xC_DESTINOINICIAL=Tractebel
xC_ORIGEM=Neotropical
xC_FORMATODOCUMENTO=Digital
xC_GENERODOCUMENTO=
xC_DOCPRINCIPAL=Sim
@end
<<EOD>>
5 - So, when we use this batch file on ODC, the file (or image) will be converted, all metadata attached and then sent to UCM.
But we don't know how to do the link between ODC and UCM. The Oracle Support says it is necessary to do some customization. In the books I could not find any info about how to do, just generic information or how to use ODC.
Is there a way (without customization) to do that? Can anyone give us a step-by-step guide?
Thanks for all help and guidance.
FernandoWhat is the one-sentence requirement or issue you're trying to solve?
If you only want content to be converted to pdf, then you should already have that with your items checked into UCM (if you have IBR enabled).
if you really want to feed content into ODC using a batchfile, my first thought would be to create a transformation utility/app that will convert your batchfiles to the expected ODC import server folder job format. then you'll have to move or map the content to match.
Again, without knowing the exact issue/business requirement you're trying to work through, I'd say a redesign of your ingestion process may be in order.
If you feed content directly into ODC through any supported ODC method (watch folder, email, scanner, fax, etc), then you'll be able to use everything ootb as you'll be able to auto-commit directly into UCM or add an indexing step and manual commit to UCM.
ODC is basically a front end for indexing and extra ingestion options to your content repo.
IPM is basically the same, but it includes some extra process-related details (process metadata, process integrations/bpel/bpm, other system integrations/ebs/crm, etc). I don't feel that IPM would help with the transformation-specific question you've put out there.
-ryan -
How to use Batch operation for two xsodata services?
Hi All,
I have two xsodata services. How to use submit batch for two xsodata services
Thanks,
MjGateway Batch Calls from SAPUI5
-
How to use batch mode in cq dam - new questions added
Hi I am doing DAM migration. I am using custom code to pick the image and create asset by calling Assetmanager.createAsset which creates dam assets.
This code I am calling through the scheduler which will run for every 0.5 hr. Around 1000 images we are trying to push in each cycle. I came acrosss batchMode for saving all the images per batch so that it will reduce the migration time But I am not sure how to use it.
.setBatchMode(true); isBatchMode are the methods available in Asset api.
Please share your thoughts on this.
Some more Point I want to ask.
1.What is the best appraoch when we are pushing images to DAM
Stopping the workflows push all the images and run the work flows. (Is it possible to do this?)
Do the both simultaniously
Do it in the bath (i.e push bunch of images and then run the workflows)
2.When i am migrating images (from file system which is residing in the same box where cq is running )it is almost taking 0.5 hr for 1000 images (avg size of image is 100kb). Is there any possibility to improve this process.
3. When I am adding images I am adding subassets to some of the images so when ever I add subasset cq is creating new version of the asset which is not required for me. What is the best approach to avoid creating too many versions for images in DAM. By doing this is there any performance improve in the system.
Thanks,
MaruthiHi Jorg,
Thank you.
I will try out batch saving option.
I have few more doubts.
1. Today I tried for moving aroung 5k images through the scheduler in my local machine. It went smoothly with out much issues but as usual some of the renditions are failing for some of the sub assets. This is about 3 for every 10 images. Renditions are generating properly for all assets it is failing in case of sub assets only. Along with this I am getting version exceptions. Below are errors which I am facing very frequently during migration dry run. This seems to be beacuse of the non synchronized execute methods in WF steps. What is your call on these issues how can we avoid these kind of errors.
20.02.2013 19:16:00.082 *ERROR* [JobHandler: /etc/workflow/instances/2013-02-20/model_1361367956960012000:/content/dam/<IMAGE PATH>/subassets/<Image Name>/jcr:content/renditions/original] com.day.cq.dam.core.impl.AssetImpl setRendition: cannot set new rendition [cq5dam.thumbnail.140.100.png] for asset [<IMAGE PATH>/subassets/<Image Name>]: javax.jcr.version.VersionException: Unable to perform operation. Node is checked-in.
20.02.2013 19:17:01.660 *ERROR* [JobHandler: /etc/workflow/instances/2013-02-20/model_1361368020884400000:/content/dam/<IMAGE PATH>/jcr:content/metadata] com.day.cq.dam.core.impl.handler.xmp.NCommXMPHandler Stack Trace: java.lang.Exception: Unable to create revision.
Caused by: javax.jcr.RepositoryException: Unable to update item: item.save()
Caused by: javax.jcr.InvalidItemStateException: Item cannot be saved because it has been deleted externally: item.save()
20.02.2013 19:19:08.666 *WARN* [JobHandler: /etc/workflow/instances/2013-02-20/model_1361368144852589000:/content/dam/<IMAGE PATH>/subassets/<Image Name>/jcr:content/renditions/original] com.day.cq.dam.core.process.ExtractMetadataProcess unexpected error occurred during metadata extraction. Cause: Unable to perform operation. Node is checked-in. javax.jcr.version.VersionException: Unable to perform operation. Node is checked-in.
2. Can you share about experience in DAM migrations. What is the avg time it takes for migrating about 1 to 1.2 lac images which are aroung 60 gig. We thought of running the scheduler (this.scheduler.addPeriodicJob(jobName, job, config, period, canRunConcurrently);)for every 0.5 hr in that case it will go on for 3 days. -
How to use BAPI_INB_DELIVERY_CONFIRM_DEC do batch split
Hi, everyone:
I want to do batch split in inbound or outbound delivery by function.
I found BAPI_INB_DELIVERY_CONFIRM_DEC, but our system is not a Decentralized System and I wandered in so many parameters.
How to use this fuction?
Thanks.Hi Junwen,
Did you call the BAPI_TRANSACTION_COMMIT afterwards?
Before this the data is not saved. After that it is saved.
Regards,
Rob. -
How to Use BAPI_OUTB_DELIVERY_CHANGE for Batch Split
Hi Experts, I have the follow delivery:
Item Material Deliv.Qty Un
10 739911 2 PQT
And I want to obtain this:
Item Material Deliv.Qty Un Batch
10 739911 0 PQT
900001 739911 1 PQT ZZ00060
900002 739911 1 PQT ZZ00061
But, after using the Bapi I obtained this:
Item Material Deliv.Qty Un Batch
10 739911 1,372 PQT
900001 739911 0,314 PQT ZZ00060
900002 739911 0,314 PQT ZZ00061
So as you can see the problem is with the "Deliv.Qty". I expected:
Item Deliv.Qty
10 0
900001 1
900002 1
But I obtained:
Item Deliv.Qty
10 1,372
900001 0,314
900002 0,314
The parameters that I am using in the Bapi are:
HEADER_DATA
DELIV_NUMB = 801174646
HEADER_CONTROL
DELIV_NUMB = 801174646
DELIVERY = 801174646
TECHN_CONTROL
UPD_IND = U
ITEM_DATA (three records):
DELIV_NUMB = 801174646
DELIV_ITEM = 000010
MATERIAL = 739911
FACT_UNIT_NOM = 1
FACT_UNIT_DENOM = 1
DELIV_NUMB = 801174646
DELIV_ITEM = 900001
MATERIAL = 739911
BATCH = ZZ00060
HIERARITEM = 000010
USEHIERITM = 1
DLV_QTY = 1
DLV_QTY_IMUNIT = 1
FACT_UNIT_NOM = 1
FACT_UNIT_DENOM = 1
DELIV_NUMB = 801174646
DELIV_ITEM = 900002
MATERIAL = 739911
BATCH = ZZ00061
HIERARITEM = 000010
USEHIERITM = 1
DLV_QTY = 1
DLV_QTY_IMUNIT = 1
FACT_UNIT_NOM = 1
FACT_UNIT_DENOM = 1
ITEM_CONTROL (three records):
DELIV_NUMB = 801174646
DELIV_ITEM = 000010
CHG_DELQTY = X
DELIV_NUMB = 801174646
DELIV_ITEM = 900001
CHG_DELQTY = X
DELIV_NUMB = 801174646
DELIV_ITEM = 900002
CHG_DELQTY = X
So I am missing something but I don't know what.
Can any one help me with this please?
PD: I have looked the following forums in order to use the Bapi:
How to Use BAPI_OUTB_DELIVERY_CHANGE for Batch Split
batch split using BAPI_OUTB_DELIVERY_CHANGE
Re: Help in BAPI_OUTB_DELIVERY_CHANGE(batch-split)
BAPI_OUTB_DELIVERY_CHANGE - Batch Splits don't work
Help for BAPI_OUTB_DELIVERY_CHANGEHey,
My code is for the inbound delivery but it should work the same.
Give this a try.
REPORT z_delivery_batch_split.
DATA:header_data LIKE bapiibdlvhdrchg,
header_control LIKE bapiibdlvhdrctrlchg,
delivery LIKE bapiibdlvhdrchg-deliv_numb,
ls_return LIKE bapiret2,
item_data TYPE TABLE OF bapiibdlvitemchg,
item_control TYPE TABLE OF bapiibdlvitemctrlchg,
ls_item_data LIKE bapiibdlvitemchg,
ls_item_control LIKE bapiibdlvitemctrlchg,
return TYPE TABLE OF bapiret2 WITH NON-UNIQUE KEY type.
header_data-deliv_numb = '1800005005'.
header_control-deliv_numb = '1800005005'.
delivery = '1800005005'.
ls_item_data-deliv_numb = '1800005005'.
ls_item_data-deliv_item = '900001'.
ls_item_data-material = '000000000020067722'.
PERFORM create_batch CHANGING ls_item_data-batch.
ls_item_data-hieraritem = '000010'.
ls_item_data-usehieritm = '1'.
ls_item_data-dlv_qty = 80.
ls_item_data-dlv_qty_imunit = 80.
ls_item_data-fact_unit_nom = '1'.
ls_item_data-fact_unit_denom = '1'.
ls_item_data-sales_unit = 'EA'.
ls_item_data-base_uom = 'EA'.
APPEND ls_item_data TO item_data.
ls_item_data-deliv_numb = '1800005005'.
ls_item_data-deliv_item = '900002'.
ls_item_data-material = '000000000020067722'.
PERFORM create_batch CHANGING ls_item_data-batch.
ls_item_data-hieraritem = '000010'.
ls_item_data-usehieritm = '1'.
ls_item_data-dlv_qty = 60.
ls_item_data-dlv_qty_imunit = 60.
ls_item_data-fact_unit_nom = '1'.
ls_item_data-fact_unit_denom = '1'.
ls_item_data-sales_unit = 'EA'.
ls_item_data-base_uom = 'EA'.
APPEND ls_item_data TO item_data.
ls_item_control-deliv_numb = '1800005005'.
ls_item_control-deliv_item = '900001'.
ls_item_control-chg_delqty = 'X'.
APPEND ls_item_control TO item_control.
ls_item_control-deliv_numb = '1800005005'.
ls_item_control-deliv_item = '900002'.
ls_item_control-chg_delqty = 'X'.
APPEND ls_item_control TO item_control.
break cgavin.
CALL FUNCTION 'BAPI_INB_DELIVERY_CHANGE'
EXPORTING
header_data = header_data
header_control = header_control
delivery = delivery
* TECHN_CONTROL = TECHN_CONTROL
TABLES
* HEADER_PARTNER = HEADER_PARTNER
* HEADER_PARTNER_ADDR = HEADER_PARTNER_ADDR
* HEADER_DEADLINES = HEADER_DEADLINES
item_data = item_data
item_control = item_control
* ITEM_SERIAL_NO = ITEM_SERIAL_NO
* EXTENSION1 = EXTENSION1
* EXTENSION2 = EXTENSION2
return = return
* TOKENREFERENCE = TOKENREFERENCE
* HANDLING_UNIT_HEADER = HANDLING_UNIT_HEADER
* HANDLING_UNIT_ITEM = HANDLING_UNIT_ITEM
* PARTIAL_GR_OBJECTS = PARTIAL_GR_OBJECTS
READ TABLE return
INTO ls_return
WITH TABLE KEY type = 'E'.
IF sy-subrc = 0.
MESSAGE ID ls_return-id TYPE ls_return-type NUMBER ls_return-number WITH ls_return-message_v1 ls_return-message_v2
ls_return-message_v3 ls_return-message_v4.
ENDIF.
CALL FUNCTION 'BAPI_TRANSACTION_COMMIT'
EXPORTING
wait = 'X'.
break cgavin.
*& Form CREATE_BATCH
* text
* <--P_LS_ITEM_DATA_BATCH text
FORM create_batch CHANGING p_ls_item_data_batch.
DATA: ls_bncom TYPE bncom.
ls_bncom-matnr = ls_item_data-material.
ls_bncom-werks = 'C333'.
ls_bncom-lgort = '3000'.
SELECT SINGLE mtart
FROM mara
INTO ls_bncom-mtart
WHERE matnr = ls_bncom-matnr.
CALL FUNCTION 'VB_NEXT_BATCH_NUMBER'
EXPORTING
matnr = ls_bncom-matnr
werks = ls_bncom-werks
check_batch = 'X'
check_material = ''
message_when_auto = ' '
x_bncom = ls_bncom
lock_batch = ' '
IMPORTING
new_charg = p_ls_item_data_batch
EXCEPTIONS
no_material = 1
no_plant = 2
material_not_found = 3
plant_not_found = 4
no_batch_handling = 5
batch_exist = 6
no_number = 7
illegal_batch_number = 8
interval_not_found = 9
object_not_found = 10
interval_overflow = 11
error_automatic_batch_number = 12
cancelled = 13
automatic_batch_num_not_active = 14
OTHERS = 15.
IF sy-subrc = 0.
CALL FUNCTION 'BAPI_TRANSACTION_COMMIT'
EXPORTING
wait = 'X'.
ENDIF.
ENDFORM. " CREATE_BATCH -
How to use batch processing with jpg pictures taken with iphone cameras?
How to use batch processing with jpg pictures taken with iphone cameras?
Open the editor . Go to file>process multiple files
Sent from my iPad -
How to use WS_DOWNLOAD funciton module in back ground (Batch Job)
Hi all,
Can any one tell,
How to use WS_DOWNLOAD funciton module in back ground (Batch Job)
Thanks,
Ravi Kumar.As disussed above you can use DATASET techniques to write the file to the application server. Then you could do a bdc call to Transaction CG3Y to transfer the file to the presentation server.
Just F1 help on keyword DATASET in SE80 for more information.
BDC Call Info found here: <a href="http://www.sap-img.com/bdc.htm">http://www.sap-img.com/bdc.htm</a>
Regards,
Philip Johannesen -
How to use custom Finish for dependency?
Hi,
in my project hi have tasks with sub-tasks duration depends on the number of pieces.
I have a custom number field for number of pieces and a custom duration field for the duration of the subtask for the single piece.
I then created a custom duration field to calculate the duration of the total time required for the single subtask for the all samples.
Then I created a custom Finish field that uses this formula to calculate when each subtask should finish is ProjDateAdd([Start],[Duration1],[Project Calendar])
The problem I have is that the calendar uses default Duration and Finish to schedule the tasks so Finish is always different from Finish1.
Is there a way to update the default Duration and Finish as function of custom duration and finish?
Michelemichele_1,
Well, no and yes. On the "no" side, there is no way to have custom fields update the built-in fields directly. However, it could be done with some VBA. Actually you could also use paste links but that is highly NOT recommended because you would
set yourself up for high probability of file corruption.
So, if you are willing to use VBA then this simple macro will do it. Note, it assumes the Number1 field is the quantity and the Duration1 field is the time for a quantity of one.
Sub SamplesXDur()
Dim t As Task
For Each t In ActiveProject.Tasks
t.Duration = t.Number1 * t.Duration1
Next t
End Sub
This is a sample file prior to running the macro.
And this is what it looks like after the macro is run.
But just on the side, why do you need a custom finish field, or even a second custom duration field? Given that you know the start date, all you need to set the Duration field, and hence the Finish field, is the result from the custom number field, which
you state is the quantity of pieces, times the time (duration) for a single piece.
Hope this helps.
John
Maybe you are looking for
-
Tolerance limit for GR/IR clearing
Dear All, As we aware that while doing F.13, system will creates clearng document only the corresponding PO balances is zero, but one of my client scenario is he had posted GR for Rs.100/- and IR is done for for Rs.99/- and when he runs F.13, system
-
I keep getting an error message "your apple Id has been disabled". I can sign in with my apple Id but when I try to update I still get that message.
-
Just migrated using time machine from MacBook Pro to older Macbook. iPhoto 9.4.3 won't open. Tried reinstall - no change. (Mac OS X 10.7.5)
-
Reporting Options in Oracel ADF BC Application
Hi everyone, I need to understand what are various options I could use to provide some reports in ADF BC application. Integrating Crystal Reports in one of the options which sounds complicated, but is there any other easier way to generate simple rep
-
Trying to debug my application
Hi, When I'm looking in my log file, I see a line like this: 07/03/19 16:32:25 [2897] Reusing a cached session application module instance Sometimes I see this line 4 times I'm looking to find which class log this line in the debug file. Any one has