OIC_C03 dataflow...
Hi gurus,
Iam implementing in BI7.
Im confused with the dataflow of 0IC_C03. As far as my reading goes, previously they used 2lis_03_BF and 278 datasources which is now to be replaced by
2lis_03_BF and 2lis_03_BX. Thats fine..
But is there BI7 dataflow already delivered for this infocube?? I got this doubt because when i installed the cube with dataflow before i got some transformations installed. Is it that SAP delivered 7x dataflow to this infocube from
7x datasources and (7x infosources. even).
The following got installed
Transformation for 2lis_03_BX ---> OIC_C03 infocube
Transformation for 2lis_03_BF ---> OIC_C03 infocube
Transformation for 2lis_03_UM ---> OIC_C03 infocube
Update rules for 2lis_40_S278 ---> OIC_C03 infocube
The above infsources (2lis_03_BX,2lis_03_BF,2lis_03_UM) seems to be 7x infosources and 2lis_40_S278 is the only 3.x infosource.
Actually my expectation is that we will get transfer rules and update rules installed.that is i expected the a 3.x dataflow will be installed which i have to change into transformation(manually writting routines using ABAP OO).But
to my surprise i could see the transformations installed.. Does these transformation contains mappings for all the keyfigures/charecterstics?? Or is there any coding that we have to do??
Can i take this 7x dataflow ( 7x infosource -> transformations-> infocube) OR
should i install the 3x flow and then migrate it... ( i mean create transformations manually )
i wanted to confirm if my understanding that SAP has itself given the 7x dataflow(transformations) for filling this infocube.we are using Service pack 6 for BI content.
Search the How-to documents in SDN, there'S a How-to on Inventory management. You can ignore and delete everything related to 2lis_40. The 2lis_03... datasources are the ones you want, this is content that existed also in 3.x, it is not 7.0 original content. Really read that How-to before you start anything.
Similar Messages
-
BODS 3.1 : SAP R/3 data extraction -What is the difference in 2 dataflows?
Hi.
Can anyone advise as to what is the difference in using the data extraction flow for extracting Data from SAP R/3 ?
1)DF1 >> SAPR/3 (R3/Table -query transformation-dat file) >>query transformation >> target
This ABAP flow generates a ABAP program and a dat file.
We can also upload this program and run jobs as execute preloaded option on datastore.
This works fine.
2) We also can pull the SAP R/3 table directly.
DF2>>SAPR/3 table (this has a red arrow like in OHD) >> Query transformation >> target
THIS ALSO Works fine. And we are able to see the data directly into oracle.
Which can also be scheduled on a job.
BUT am unable to understand the purpose of using the different types of data extraction flows.
When to use which type of flow for data extraction.
Advantage / disadvantage - over the 2 data flows.
What we are not understanding is that :
if we can directly pull data from R/3 table directly thro a query transformation into the target table,
why use the Flow of creating a R/3 data flow,
and then do a query transformation again
and then populate the target database?
There might be some practical reasons for using these 2 different types of flows in doing the data extraction. Which I would like to understand. Can anyone advise please.
Many thanks
indu
Edited by: Indumathy Narayanan on Aug 22, 2011 3:25 PMHi Jeff.
Greetings. And many thanks for your response.
Generally we pull the entire SAP R/3 table thro query transformation into oracle.
For which we use R/3 data flow and the ABAP program, which we upload on the R/3 system
so as to be able to use the option of Execute preloaded - and run the jobs.
Since we do not have any control on our R/3 servers nor we have anyone on ABAP programming,
we do not do anything at the SAP R/3 level
I was doing this trial and error testing on our Worflows for our new requirement
WF 1 : which has some 15 R/3 TABLES.
For each table we have created a separate Dataflow.
And finally in between in some dataflows, wherein, the SAP tables which had lot of rows, i decided to pull it directly,
by-passing the ABAP flow.
And still the entire work flow and data extraction happens ok.
In fact i tried creating a new sample data flow and tested.
Using direct download and - and also execute preloaded.
I did not see any major difference in time taken for data extraction;
Because anyhow we pull the entire Table, then choose whatever we want to bring into oracle thro a view for our BO reporting or aggregate and then bring data as a table for Universe consumption.
Actually, I was looking at other options to avoid this ABAP generation - and the R/3 data flow because we are having problems on our dev and qa environments - giving delimiter errors. Whereas in production it works fine. Production environment is a old set up of BODS 3.1. QA and Dev are relatively new enviornments of BODS. Which is having this delimiter error.
I did not understand how to resolve it as per this post : https://cw.sdn.sap.com/cw/ideas/2596
And trying to resolve this problem, I ended up with the option of trying to pull directly the R/3 table. Without using ABAP workflow. Just by trial and error of each and every drag and drop option. Because we had to urgently do a POC and deliver the data for the entire e recruiting module of SAP.
I dont know whether i could do this direct pulling of data - for the new job which i have created,
which has 2 workflows with 15 Dataflows in each worflow.
And and push this job into production.
And also whether i could by-pass this ABAP flow and do a direct pulling of R/3 data, in all the Dataflows in the future for ANY of our SAP R/3 data extraction requirement. And this technical understanding is not clear to us as regards the difference between the 2 flows. And being new to this whole of ETL - I just wanted to know the pros and cons of this particular data extraction.
As advised I shall check the schedules for a week, and then we shall move it probably into production.
Thanks again.
Kind Regards
Indu
Edited by: Indumathy Narayanan on Aug 22, 2011 7:02 PM -
Issues installing 0IC_C03 dataflow in BW 7.0
Hi experts,
We are trying to install the dataflow for stocks infocube (0IC_C03) in BW (release 700 level 0015) from Business Content (release 703 level 0008) and we get the following error message when installing the transformation from Infosource 2LIS_03_BF_TR to the cube:
Start Routine: Syntax error in routine
We have checked the code in the start routine and we get the following error message:
E:In PERFORM or CALL FUNCTION "ROUTINE_9998", the actual parameter
"SOURCE_PACKAGE" is incompatible with the formal parameter
"DATA_PACKAGE".
Has anybody faced with this issue before?
Any help will be really appreciated.
Regards,
DavidHi,
Pls add the following code in the start Routine.
**************Added as per note 1052648****************************************
InfoObject: 0RMA_XCC RMA Cross-Company Code Stock Transfer.
RMA_XCC TYPE /BI0/OIRMA_XCC,
InfoObject: 0RMA_LNK RMA Stock Transfer Reference Item.
RMA_LNK TYPE /BI0/OIRMA_LNK,
InfoObject: 0RMA_RFLG RMA Relevance Flag.
RMA_RFLG TYPE /BI0/OIRMA_RFLG,
InfoObject: 0RTHFEES Stock Transfer Fees.
RTHFEES TYPE /BI0/OIRTHFEES,
*******End Added as per note*****1052648*************************************
After adding the above code no syntactical error in start routine activate the Transformation.
The Problem is Solved.
In my system Start routine code:
PROGRAM trans_routine.
CLASS routine DEFINITION
CLASS lcl_transform DEFINITION.
PUBLIC SECTION.
Attributs
DATA:
p_check_master_data_exist
TYPE RSODSOCHECKONLY READ-ONLY,
*- Instance for getting request runtime attributs;
Available information: Refer to methods of
interface 'if_rsbk_request_admintab_view'
p_r_request
TYPE REF TO if_rsbk_request_admintab_view READ-ONLY.
PRIVATE SECTION.
TYPE-POOLS: rsd, rstr.
Rule specific types
TYPES:
BEGIN OF tys_SC_1,
InfoObject: 0STORNO Reversal indicator.
STORNO TYPE /BI0/OISTORNO,
InfoObject: 0RT_PROMO Promotion.
RT_PROMO TYPE /BI0/OIRT_PROMO,
InfoObject: 0VAL_CLASS Valuation class.
VAL_CLASS TYPE /BI0/OIVAL_CLASS,
InfoObject: 0DOC_DATE Document Date.
DOC_DATE TYPE /BI0/OIDOC_DATE,
InfoObject: 0STOCKTYPE Stock type.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0STOCKCAT Stock Categories.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0PSTNG_DATE Posting date in the document.
PSTNG_DATE TYPE /BI0/OIPSTNG_DATE,
InfoObject: 0COMP_CODE Company code.
COMP_CODE TYPE /BI0/OICOMP_CODE,
InfoObject: 0BWAPPLNM Application component.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0MOVETYPE Movement Type (Inventory Management).
MOVETYPE TYPE /BI0/OIMOVETYPE,
InfoObject: 0STOCKRELEV BW: Relevance to Stock.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0CPPVLC BW: Purchase Value in Local Currency.
CPPVLC TYPE /BI0/OICPPVLC,
InfoObject: 0CPSVLC BW: Sales Value in Local Currency.
CPSVLC TYPE /BI0/OICPSVLC,
InfoObject: 0CPSTLC BW: Sales Value in Local Currency.
CPSTLC TYPE /BI0/OICPSTLC,
InfoObject: 0CPQUABU BW: Amount in base unit of measure.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0VAL_TYPE Valuation type.
VAL_TYPE TYPE /BI0/OIVAL_TYPE,
InfoObject: 0PROCESSKEY BW: Transaction Key.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0BATCH Batch number.
BATCH TYPE /BI0/OIBATCH,
InfoObject: 0MATMREA Reason for Goods Movement.
MATMREA TYPE /BI0/OIMATMREA,
InfoObject: 0BUS_AREA Business area.
BUS_AREA TYPE /BI0/OIBUS_AREA,
InfoObject: 0COSTCENTER Cost Center.
COSTCENTER TYPE /BI0/OICOSTCENTER,
InfoObject: 0SOLD_TO Sold-to party.
SOLD_TO TYPE /BI0/OISOLD_TO,
InfoObject: 0WHSE_NUM Warehouse number / warehouse complex.
WHSE_NUM TYPE /BI0/OIWHSE_NUM,
InfoObject: 0STOR_LOC Storage location.
STOR_LOC TYPE /BI0/OISTOR_LOC,
InfoObject: 0STRGE_BIN Storage bin.
STRGE_BIN TYPE /BI0/OISTRGE_BIN,
InfoObject: 0STRGE_TYPE Storage type.
STRGE_TYPE TYPE /BI0/OISTRGE_TYPE,
InfoObject: 0VENDOR Vendor.
VENDOR TYPE /BI0/OIVENDOR,
InfoObject: 0MATERIAL Material.
MATERIAL TYPE /BI0/OIMATERIAL,
InfoObject: 0DOC_NUM BW: Document Number.
DOC_NUM TYPE /BI0/OIDOC_NUM,
InfoObject: 0BASE_UOM Base Unit of Measure.
BASE_UOM TYPE /BI0/OIBASE_UOM,
InfoObject: 0DOC_YEAR BW: Document Year.
DOC_YEAR TYPE /BI0/OIDOC_YEAR,
InfoObject: 0PROFIT_CTR Profit Center.
PROFIT_CTR TYPE /BI0/OIPROFIT_CTR,
InfoObject: 0DCINDIC Debit/credit indicator.
DCINDIC TYPE /BI0/OIDCINDIC,
InfoObject: 0LOC_CURRCY Local currency.
LOC_CURRCY TYPE /BI0/OILOC_CURRCY,
InfoObject: 0PLANT Plant.
PLANT TYPE /BI0/OIPLANT,
InfoObject: 0FISCVARNT Fiscal year variant.
FISCVARNT TYPE /BI0/OIFISCVARNT,
InfoObject: 0CPNOITEMS BW: Number of Activities.
CPNOITEMS TYPE /BI0/OICPNOITEMS,
InfoObject: 0CO_AREA Controlling area.
CO_AREA TYPE /BI0/OICO_AREA,
InfoObject: 0DOC_ITEM BW: Document Line Number.
DOC_ITEM TYPE /BI0/OIDOC_ITEM,
InfoObject: 0VALUE_LC Amount in local currency.
VALUE_LC TYPE /BI0/OIVALUE_LC,
InfoObject: 0COORDER Order number.
COORDER TYPE /BI0/OICOORDER,
InfoObject: 0QUANT_B Quantity in base units of measure.
QUANT_B TYPE /BI0/OIQUANT_B,
InfoObject: 0MOVE_PLANT Receiving Plant/Issuing Plant.
MOVE_PLANT TYPE /BI0/OIMOVE_PLANT,
InfoObject: 0RECORDMODE BW Delta Process: Update Mode.
RECORDMODE TYPE RODMUPDMOD,
InfoObject: 0BWCOUNTER Additional Key Field Revaluation Document
*Record.
BWCOUNTER TYPE /BI0/OIBWCOUNTER,
InfoObject: 0INDSPECSTK Indicator: Valuation of Special Stock.
INDSPECSTK TYPE /BI0/OIINDSPECSTK,
InfoObject: 0GN_R3_SSY Source System for R/3 Entity.
GN_R3_SSY TYPE /BI0/OIGN_R3_SSY,
InfoObject: 0RT_MOVINV Movement Type is Physical Inventory.
RT_MOVINV TYPE /BI0/OIRT_MOVINV,
InfoObject: 0RT_MOVADJ Movement Type is Stock Adjustment.
RT_MOVADJ TYPE /BI0/OIRT_MOVADJ,
InfoObject: 0RSL_STMAT Structured Article.
RSL_STMAT TYPE /BI0/OIRSL_STMAT,
InfoObject: 0RT_MOVRET Movement Type is Return.
RT_MOVRET TYPE /BI0/OIRT_MOVRET,
InfoObject: 0RSL_STAUTO Automatically Completed Component.
RSL_STAUTO TYPE /BI0/OIRSL_STAUTO,
InfoObject: 0RT_MOVTRAN Movement Type is Stock Transfer.
RT_MOVTRAN TYPE /BI0/OIRT_MOVTRAN,
InfoObject: 0RT_MATPOST Transfer Posting from Article to Article.
RT_MATPOST TYPE /BI0/OIRT_MATPOST,
InfoObject: 0RT_SASTSV Share of Business Volume for Set Sales Val
*ue in Local Crcy.
RT_SASTSV TYPE /BI0/OIRT_SASTSV,
InfoObject: 0RT_SASTST Share of Business Volume for Set with Tax
*in Local Currency.
RT_SASTST TYPE /BI0/OIRT_SASTST,
InfoObject: 0RMA_XCC RMA Cross-Company Code Stock Transfer.
RMA_XCC TYPE /BI0/OIRMA_XCC,
InfoObject: 0RMA_LNK RMA Stock Transfer Reference Item.
RMA_LNK TYPE /BI0/OIRMA_LNK,
InfoObject: 0RMA_RFLG RMA Relevance Flag.
RMA_RFLG TYPE /BI0/OIRMA_RFLG,
InfoObject: 0RTHFEES Stock Transfer Fees.
RTHFEES TYPE /BI0/OIRTHFEES,
Field: RECORD.
RECORD TYPE RSARECORD,
END OF tys_SC_1.
TYPES:
tyt_SC_1 TYPE STANDARD TABLE OF tys_SC_1
WITH NON-UNIQUE DEFAULT KEY.
$$ begin of global - insert your declaration only below this line -
... "insert your code here
$$ end of global - insert your declaration only before this line -
METHODS
start_routine
IMPORTING
request type rsrequest
datapackid type rsdatapid
EXPORTING
monitor type rstr_ty_t_monitors
CHANGING
SOURCE_PACKAGE type tyt_SC_1
RAISING
cx_rsrout_abort.
METHODS
inverse_start_routine
IMPORTING
i_th_fields_outbound TYPE rstran_t_field_inv
i_r_selset_outbound TYPE REF TO cl_rsmds_set
i_is_main_selection TYPE rs_bool
i_r_selset_outbound_complete TYPE REF TO cl_rsmds_set
i_r_universe_inbound TYPE REF TO cl_rsmds_universe
CHANGING
c_th_fields_inbound TYPE rstran_t_field_inv
c_r_selset_inbound TYPE REF TO cl_rsmds_set
c_exact TYPE rs_bool.
ENDCLASS. "routine DEFINITION
$$ begin of 2nd part global - insert your code only below this line *
TYPES:
BEGIN OF tys_SC_1_full,
InfoObject: 0STORNO Stornokennzeichen.
STORNO TYPE /BI0/OISTORNO,
InfoObject: 0RT_PROMO Aktion.
RT_PROMO TYPE /BI0/OIRT_PROMO,
InfoObject: 0VAL_CLASS Bewertungsklasse.
VAL_CLASS TYPE /BI0/OIVAL_CLASS,
InfoObject: 0DOC_DATE Belegdatum.
DOC_DATE TYPE /BI0/OIDOC_DATE,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0PSTNG_DATE Buchungsdatum im Beleg.
PSTNG_DATE TYPE /BI0/OIPSTNG_DATE,
InfoObject: 0COMP_CODE Buchungskreis.
COMP_CODE TYPE /BI0/OICOMP_CODE,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0MOVETYPE Bewegungsart (Bestandsführung).
MOVETYPE TYPE /BI0/OIMOVETYPE,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0CPPVLC BW: Einkaufswert in Hauswährung.
CPPVLC TYPE /BI0/OICPPVLC,
InfoObject: 0CPSVLC BW: Verkaufswert in Hauswährung.
CPSVLC TYPE /BI0/OICPSVLC,
InfoObject: 0CPSTLC BW: VerkWmS Hauswährung.
CPSTLC TYPE /BI0/OICPSTLC,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0VAL_TYPE Bewertungsart.
VAL_TYPE TYPE /BI0/OIVAL_TYPE,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0BATCH Chargennummer.
BATCH TYPE /BI0/OIBATCH,
InfoObject: 0MATMREA Grund für die Warenbewegung.
MATMREA TYPE /BI0/OIMATMREA,
InfoObject: 0BUS_AREA Geschäftsbereich.
BUS_AREA TYPE /BI0/OIBUS_AREA,
InfoObject: 0COSTCENTER Kostenstelle.
COSTCENTER TYPE /BI0/OICOSTCENTER,
InfoObject: 0SOLD_TO Auftraggeber.
SOLD_TO TYPE /BI0/OISOLD_TO,
InfoObject: 0WHSE_NUM Lagernummer/Lagerkomplex.
WHSE_NUM TYPE /BI0/OIWHSE_NUM,
InfoObject: 0STOR_LOC Lagerort.
STOR_LOC TYPE /BI0/OISTOR_LOC,
InfoObject: 0STRGE_BIN Lagerplatz.
STRGE_BIN TYPE /BI0/OISTRGE_BIN,
InfoObject: 0STRGE_TYPE Lagertyp.
STRGE_TYPE TYPE /BI0/OISTRGE_TYPE,
InfoObject: 0VENDOR Lieferant.
VENDOR TYPE /BI0/OIVENDOR,
InfoObject: 0MATERIAL Material.
MATERIAL TYPE /BI0/OIMATERIAL,
InfoObject: 0DOC_NUM BW: Belegnummer.
DOC_NUM TYPE /BI0/OIDOC_NUM,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
InfoObject: 0DOC_YEAR BW: Belegjahr.
DOC_YEAR TYPE /BI0/OIDOC_YEAR,
InfoObject: 0PROFIT_CTR Profit Center.
PROFIT_CTR TYPE /BI0/OIPROFIT_CTR,
InfoObject: 0DCINDIC Soll-/Haben-Kennzeichen.
DCINDIC TYPE /BI0/OIDCINDIC,
InfoObject: 0LOC_CURRCY Hauswährung.
LOC_CURRCY TYPE /BI0/OILOC_CURRCY,
InfoObject: 0PLANT Werk.
PLANT TYPE /BI0/OIPLANT,
InfoObject: 0FISCVARNT Geschäftsjahresvariante.
FISCVARNT TYPE /BI0/OIFISCVARNT,
InfoObject: 0CPNOITEMS BW: Anzahl Vorgänge.
CPNOITEMS TYPE /BI0/OICPNOITEMS,
InfoObject: 0CO_AREA Kostenrechnungskreis.
CO_AREA TYPE /BI0/OICO_AREA,
InfoObject: 0DOC_ITEM BW: Belegpositionsnummer.
DOC_ITEM TYPE /BI0/OIDOC_ITEM,
InfoObject: 0VALUE_LC Betrag in Hauswährung.
VALUE_LC TYPE /BI0/OIVALUE_LC,
InfoObject: 0COORDER Auftragsnummer.
COORDER TYPE /BI0/OICOORDER,
InfoObject: 0QUANT_B Menge in Basismengeneinheiten.
QUANT_B TYPE /BI0/OIQUANT_B,
InfoObject: 0MOVE_PLANT Empfangendes/Abgebendes Werk.
MOVE_PLANT TYPE /BI0/OIMOVE_PLANT,
InfoObject: 0RECORDMODE BW Deltaverfahren: Update Modus.
RECORDMODE TYPE RODMUPDMOD,
*******************Commented based on oss note1052648*****************************
InfoObject: 0RT_RMAPIDA RMA Inventurdatum.
RT_RMAPIDA TYPE /BI0/OIRT_RMAPIDA,
End Commented based on oss note1052648************************
InfoObject: 0BWCOUNTER Zusätzliches Schlüsselfeld Umbewertungsbel
*egsatz.
BWCOUNTER TYPE /BI0/OIBWCOUNTER,
InfoObject: 0INDSPECSTK Kennzeichen Bewertung Sonderbestand.
INDSPECSTK TYPE /BI0/OIINDSPECSTK,
InfoObject: 0GN_R3_SSY Quellsystem der R/3 Entität.
GN_R3_SSY TYPE /BI0/OIGN_R3_SSY,
InfoObject: 0RT_MOVINV Bewegungsart ist Inventur.
RT_MOVINV TYPE /BI0/OIRT_MOVINV,
InfoObject: 0RT_MOVADJ Bewegungsart ist Bestandskorrektur.
RT_MOVADJ TYPE /BI0/OIRT_MOVADJ,
InfoObject: 0RSL_STMAT Strukturierter Artikel.
RSL_STMAT TYPE /BI0/OIRSL_STMAT,
InfoObject: 0RT_MOVRET Bewegungsart ist Retoure.
RT_MOVRET TYPE /BI0/OIRT_MOVRET,
InfoObject: 0RSL_STAUTO Automatisch ergänzte Komponente.
RSL_STAUTO TYPE /BI0/OIRSL_STAUTO,
InfoObject: 0RT_MOVTRAN Bewegungsart ist Umlagerung.
RT_MOVTRAN TYPE /BI0/OIRT_MOVTRAN,
InfoObject: 0RT_MATPOST Umbuchung Artikel an Artikel.
RT_MATPOST TYPE /BI0/OIRT_MATPOST,
InfoObject: 0RT_SASTSV Umsatzanteil am Set Verkaufswert in Hauswä
*hrung.
RT_SASTSV TYPE /BI0/OIRT_SASTSV,
InfoObject: 0RT_SASTST Umsatzanteil am Set Verkaufswert mit Steue
*r in Hauswährung.
RT_SASTST TYPE /BI0/OIRT_SASTST,
**********************Added as per note1052648*****************************************
InfoObject: 0RMA_XCC RMA Cross-Company Code Stock Transfer.
RMA_XCC TYPE /BI0/OIRMA_XCC,
InfoObject: 0RMA_LNK RMA Stock Transfer Reference Item.
RMA_LNK TYPE /BI0/OIRMA_LNK,
InfoObject: 0RMA_RFLG RMA Relevance Flag.
RMA_RFLG TYPE /BI0/OIRMA_RFLG,
InfoObject: 0RTHFEES Stock Transfer Fees.
RTHFEES TYPE /BI0/OIRTHFEES,
**********************End Added as per note1052648**************************************
Field: RECORD.
RECORD TYPE RSARECORD,
END OF tys_SC_1_full.
TYPES:
tyt_SC_1_full TYPE STANDARD TABLE OF tys_SC_1_full
WITH NON-UNIQUE DEFAULT KEY.
TYPES:
BEGIN OF tys_TG_1_full,
InfoObject: 0CHNGID Änderungslauf ID.
CHNGID TYPE /BI0/OICHNGID,
InfoObject: 0RECORDTP Satztyp.
RECORDTP TYPE /BI0/OIRECORDTP,
InfoObject: 0REQUID Request ID.
REQUID TYPE /BI0/OIREQUID,
InfoObject: 0CALDAY Kalendertag.
CALDAY TYPE /BI0/OICALDAY,
InfoObject: 0CALMONTH Kalenderjahr / Monat.
CALMONTH TYPE /BI0/OICALMONTH,
InfoObject: 0CALWEEK Kalenderjahr / Woche.
CALWEEK TYPE /BI0/OICALWEEK,
InfoObject: 0CALYEAR Kalenderjahr.
CALYEAR TYPE /BI0/OICALYEAR,
InfoObject: 0MATERIAL Material.
MATERIAL TYPE /BI0/OIMATERIAL,
InfoObject: 0PLANT Werk.
PLANT TYPE /BI0/OIPLANT,
InfoObject: 0STOR_LOC Lagerort.
STOR_LOC TYPE /BI0/OISTOR_LOC,
InfoObject: 0BATCH Chargennummer.
BATCH TYPE /BI0/OIBATCH,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0GN_VENDOR Lieferant.
GN_VENDOR TYPE /BI0/OIGN_VENDOR,
InfoObject: 0RECVS_VAL Zugangswert Bewerteter Bestand.
RECVS_VAL TYPE /BI0/OIRECVS_VAL,
InfoObject: 0ISSVS_VAL Abgangswert Bewerteter Bestand.
ISSVS_VAL TYPE /BI0/OIISSVS_VAL,
InfoObject: 0ISSBLOSTCK Abgangsmenge Gesperrter Bestand.
ISSBLOSTCK TYPE /BI0/OIISSBLOSTCK,
InfoObject: 0ISSCNSSTCK Abgangsmenge Konsignationsbestand.
ISSCNSSTCK TYPE /BI0/OIISSCNSSTCK,
InfoObject: 0ISSQMSTCK Abgangsmenge Qualitätsbestand.
ISSQMSTCK TYPE /BI0/OIISSQMSTCK,
InfoObject: 0ISSTRANSST Abgangsmenge Transitbestand.
ISSTRANSST TYPE /BI0/OIISSTRANSST,
InfoObject: 0RECBLOSTCK Zugangsmenge Gesperrter Bestand.
RECBLOSTCK TYPE /BI0/OIRECBLOSTCK,
InfoObject: 0RECCNSSTCK Zugangsmenge Konsignationsbestand.
RECCNSSTCK TYPE /BI0/OIRECCNSSTCK,
InfoObject: 0RECQMSTCK Zugangsmenge Qualitätsbestand.
RECQMSTCK TYPE /BI0/OIRECQMSTCK,
InfoObject: 0RECTRANSST Zugangsmenge Transitbestand.
RECTRANSST TYPE /BI0/OIRECTRANSST,
InfoObject: 0ISSSCRP Abgangsmenge Ausschuss.
ISSSCRP TYPE /BI0/OIISSSCRP,
InfoObject: 0ISSVALSCRP Abgangswert Ausschuss.
ISSVALSCRP TYPE /BI0/OIISSVALSCRP,
InfoObject: 0RECTOTSTCK Zugangsmenge Gesamt Bestand.
RECTOTSTCK TYPE /BI0/OIRECTOTSTCK,
InfoObject: 0ISSTOTSTCK Abgangsmenge Gesamtbestand.
ISSTOTSTCK TYPE /BI0/OIISSTOTSTCK,
InfoObject: 0ISSVALSTCK Abgangsmenge Bewerteter Bestand.
ISSVALSTCK TYPE /BI0/OIISSVALSTCK,
InfoObject: 0RECVALSTCK Zugangsmenge Bewerteter Bestand.
RECVALSTCK TYPE /BI0/OIRECVALSTCK,
InfoObject: 0VENCONCON Verbrauchswerte Lieferantenkonsignationsbe
*stand.
VENCONCON TYPE /BI0/OIVENCONCON,
InfoObject: 0LOC_CURRCY Hauswährung.
LOC_CURRCY TYPE /BI0/OILOC_CURRCY,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
Field: RECORD Nummer des Datensatzes.
RECORD TYPE RSARECORD,
END OF tys_TG_1_full.
TYPES:
BEGIN OF tys_SC_1__RULE_9,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_9.
TYPES:
BEGIN OF tys_SC_1__RULE_10,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_10.
TYPES:
BEGIN OF tys_SC_1__RULE_11,
InfoObject: 0GN_R3_SSY Quellsystem der R/3 Entität.
GN_R3_SSY TYPE /BI0/OIGN_R3_SSY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0VENDOR Lieferant.
VENDOR TYPE /BI0/OIVENDOR,
END OF tys_SC_1__RULE_11.
TYPES:
BEGIN OF tys_SC_1__RULE_12,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPPVLC BW: Einkaufswert in Hauswährung.
CPPVLC TYPE /BI0/OICPPVLC,
InfoObject: 0INDSPECSTK Kennzeichen Bewertung Sonderbestand.
INDSPECSTK TYPE /BI0/OIINDSPECSTK,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0LOC_CURRCY Hauswährung.
LOC_CURRCY TYPE /BI0/OILOC_CURRCY,
END OF tys_SC_1__RULE_12.
TYPES:
BEGIN OF tys_SC_1__RULE_13,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPPVLC BW: Einkaufswert in Hauswährung.
CPPVLC TYPE /BI0/OICPPVLC,
InfoObject: 0INDSPECSTK Kennzeichen Bewertung Sonderbestand.
INDSPECSTK TYPE /BI0/OIINDSPECSTK,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0LOC_CURRCY Hauswährung.
LOC_CURRCY TYPE /BI0/OILOC_CURRCY,
END OF tys_SC_1__RULE_13.
TYPES:
BEGIN OF tys_SC_1__RULE_15,
InfoObject: 0GN_R3_SSY Quellsystem der R/3 Entität.
GN_R3_SSY TYPE /BI0/OIGN_R3_SSY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0VENDOR Lieferant.
VENDOR TYPE /BI0/OIVENDOR,
END OF tys_SC_1__RULE_15.
TYPES:
BEGIN OF tys_SC_1__RULE_19,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_19.
TYPES:
BEGIN OF tys_SC_1__RULE_20,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_20.
TYPES:
BEGIN OF tys_SC_1__RULE_21,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0INDSPECSTK Kennzeichen Bewertung Sonderbestand.
INDSPECSTK TYPE /BI0/OIINDSPECSTK,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_21.
TYPES:
BEGIN OF tys_SC_1__RULE_22,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0INDSPECSTK Kennzeichen Bewertung Sonderbestand.
INDSPECSTK TYPE /BI0/OIINDSPECSTK,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_22.
TYPES:
BEGIN OF tys_SC_1__RULE_25,
InfoObject: 0GN_R3_SSY Quellsystem der R/3 Entität.
GN_R3_SSY TYPE /BI0/OIGN_R3_SSY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0VENDOR Lieferant.
VENDOR TYPE /BI0/OIVENDOR,
END OF tys_SC_1__RULE_25.
TYPES:
BEGIN OF tys_SC_1__RULE_31,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPPVLC BW: Einkaufswert in Hauswährung.
CPPVLC TYPE /BI0/OICPPVLC,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0LOC_CURRCY Hauswährung.
LOC_CURRCY TYPE /BI0/OILOC_CURRCY,
END OF tys_SC_1__RULE_31.
TYPES:
BEGIN OF tys_SC_1__RULE_37,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_37.
TYPES:
BEGIN OF tys_SC_1__RULE_38,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_38.
TYPES:
BEGIN OF tys_SC_1__RULE_39,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_39.
TYPES:
BEGIN OF tys_SC_1__RULE_40,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_40.
TYPES:
BEGIN OF tys_SC_1__RULE_41,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKCAT Bestandstypen.
STOCKCAT TYPE /BI0/OISTOCKCAT,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_41.
TYPES:
BEGIN OF tys_SC_1__RULE_42,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0PROCESSKEY BW:Vorgangsschlüssel.
PROCESSKEY TYPE /BI0/OIPROCESSKEY,
InfoObject: 0STOCKRELEV BW: Best.Relevanz.
STOCKRELEV TYPE /BI0/OISTOCKRELEV,
InfoObject: 0STOCKTYPE Bestandsausprägung.
STOCKTYPE TYPE /BI0/OISTOCKTYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_42.
TYPES:
BEGIN OF tys_SC_1__RULE_43,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPQUABU BW: Menge in Basismengeneinheit.
CPQUABU TYPE /BI0/OICPQUABU,
InfoObject: 0MOVETYPE Bewegungsart (Bestandsführung).
MOVETYPE TYPE /BI0/OIMOVETYPE,
InfoObject: 0BASE_UOM Basismengeneinheit.
BASE_UOM TYPE /BI0/OIBASE_UOM,
END OF tys_SC_1__RULE_43.
TYPES:
BEGIN OF tys_SC_1__RULE_51,
InfoObject: 0BWAPPLNM Anwendungskomponente.
BWAPPLNM TYPE /BI0/OIBWAPPLNM,
InfoObject: 0CPPVLC BW: Einkaufswert in Hauswährung.
CPPVLC TYPE /BI0/OICPPVLC,
InfoObject: 0MOVETYPE Bewegungsart (Bestandsführung).
MOVETYPE TYPE /BI0/OIMOVETYPE,
InfoObject: 0LOC_CURRCY Hauswährung.
LOC_CURRCY TYPE /BI0/OILOC_CURRCY,
END OF tys_SC_1__RULE_51.
Additional types for start routine interface
TYPES:
data_package_structure type tys_SC_1_full.
Additional declaration for update rule interface
DATA:
MONITOR type standard table of rsmonitor WITH HEADER LINE,
MONITOR_RECNO type standard table of rsmonitors WITH HEADER LINE,
RECORD_NO LIKE SY-TABIX,
RECORD_ALL LIKE SY-TABIX,
SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS.
global definitions from update rules
INCLUDE rsbctgn_top.
INCLUDE rsbctgn_update_rules.
FORM routine_9998
TABLES DATA_PACKAGE TYPE tyt_SC_1_full
CHANGING
ABORT LIKE sy-subrc
RAISING
cx_sy_arithmetic_error
cx_sy_conversion_error.
init variables
fill the internal table "MONITOR", to make monitor entries
see OSS note 571669
LOOP AT DATA_PACKAGE.
IF DATA_PACKAGE-stockcat EQ 'V' OR
DATA_PACKAGE-stocktype EQ 'V'.
DELETE DATA_PACKAGE.
ENDIF.
ENDLOOP.
if abort is not equal zero, the update process will be canceled
ABORT = 0.
ENDFORM. "routine_9998
FORM routine_0012
USING
COMM_STRUCTURE type tys_SC_1__RULE_9
CHANGING
RESULT TYPE tys_TG_1_full-ISSCNSSTCK
RETURNCODE LIKE sy-subrc
ABORT LIKE sy-subrc
RAISING
cx_sy_arithmetic_error
cx_sy_conversion_error.
init variables
fill the internal table "MONITOR", to make monitor entries
only goods issue is considered
IF ( COMM_STRUCTURE-processkey EQ '100' "Other Issues
OR COMM_STRUCTURE-processkey EQ '101' "Returns / Vendor
OR COMM_STRUCTURE-processkey EQ '104' "Material Transfer
OR COMM_STRUCTURE-processkey EQ '105' "Stock Adjustment InvD
OR COMM_STRUCTURE-processkey EQ '106' "Stock Adjustment Other
OR COMM_STRUCTURE-processkey EQ '110' ) "Issue from Stock Transfer
AND COMM_STRUCTURE-bwapplnm EQ 'MM'
only movements which are relevant for stock control
AND COMM_STRUCTURE-stockrelev EQ '1'
AND COMM_STRUCTURE-cpquabu <> 0.
only consignment stock is considered
CASE COMM_STRUCTURE-stockcat.
WHEN 'K'.
RESULT = COMM_STRUCTURE-cpquabu.
RETURNCODE = 0.
WHEN space.
IF COMM_STRUCTURE-stocktype CA 'KLM'.
RESULT = COMM_STRUCTURE-cpquabu.
RETURNCODE = 0.
ELSE.
RETURNCODE = 4.
ENDIF.
WHEN OTHERS.
RETURNCODE = 4.
ENDCASE.
ELSE.
if the returncode is not equal zero, the result will not be updated
RETURNCODE = 4.
ENDIF.
if abort is not equal zero, the update process will be canceled
ABORT = 0.
ENDFORM. "routine_0012
FORM routine_0014
USING
COMM_STRUCTURE type tys_SC_1__RULE_10
CHANGING
RESULT TYPE tys_TG_1_full-RECCNSSTCK
RETURNCODE LIKE sy-subrc
ABORT LIKE sy-subrc
RAISING
cx_sy_arithmetic_error
cx_sy_conversion_error.
init variables
fill the internal table "MONITOR", to make monitor entries
only goods receipt is considered
IF ( COMM_STRUCTURE-processkey EQ '000' "Other Receipts
OR COMM_STRUCTURE-processkey EQ '001' "Goods Receipt / Vendor
OR COMM_STRUCTURE-processkey EQ '004' "Material Transfer
OR COMM_STRUCTURE-processkey EQ '005' "Stock Adjustment InvD
OR COMM_STRUCTURE-processkey EQ '006' "Stock Adjustment Other
OR COMM_STRUCTURE-processkey EQ '010' ) "Receipt from Stock Transfer
AND COMM_STRUCTURE-bwapplnm EQ 'MM'
only movements which are relevant for stock control
AND COMM_STRUCTURE-stockrelev EQ '1'
AND COMM_STRUCTURE-cpquabu <> 0.
only consignment stock is considered
CASE COMM_STRUCTURE-stockcat.
WHEN 'K'.
RESULT = COMM_STRUCTURE-cpquabu.
RETURNCODE = 0.
WHEN space.
IF COMM_STRUCTURE-stocktype CA 'KLM'.
RESULT = COMM_STRUCTURE-cpquabu.
RETURNCODE = 0.
ELSE.
RETURNCODE = 4.
ENDIF.
WHEN OTHERS.
RETURNCODE = 4.
ENDCASE.
ELSE.
if the returncode is not equal zero, the result will not be updated
RETURNCODE = 4.
ENDIF. -
Using the Execute Preloaded Option for ABAP Dataflows in SAP BODS
Hello All,
This is regarding the use of Advanced Option in SAP Application Type (ECC) Datastore settings in SAP BODS 4.2, where there are two options to select from for ABAP Execution Option property: (1)Generate and Execute and (2) Execute Preloaded. Since our ECC client is often locked by BASIS team even on DEV environment, we would like to make use of the second option 'Execute Preloaded' so that we could extract the data from ECC tables without having to ask the BASIS team to unlock the ECC client every time before extraction.
The problem is that we are getting an error upon generating and uploading the ABAP program to ECC client. I have searched the blogs and so far I have only found that there are certain ABAP programs or function modules that come with SAP BODS which need to be installed by BASIS team on ECC side to allow the ABAP dataflows to be generated and uploaded to ECC server. I would appreciate it if anyone could provide a list of which function modules BASIS needs to install on ECC or a blog that provides details around using this option.
So far, from the SAP BODS designer, we are performing below steps but getting an error upon generating and uploading the ABAP dataflow program:
1. Create a test ABAP Dataflow using SAP ECC datastore. Provide the ABAP program options.
2. Right click, select Generate ABAP Program.
3. Once the ABAP Program Generation Dialog box appears, check the box "Upload Generated Program".
4. Upon clicking OK, we are getting the following error:
The ABAP program <ZRTEST01> for ABAP data flow <RT_TEST_R3> (datastore <R3_DS>) was not uploaded: < RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_MESSAGE- Exception condition "NOT_SUPPORTED_BY_GUI" trigger[SAP NWRFC 720][SAP Partner ### ][clientname][servername][accountname][4103]>. >.
Any help would be greatly appreciated.
Thanks,
RizwanAll,
The BASIS team reviewed steps provided in the BODS document and attempted to install the Function Modules but now none of the BODS jobs would work. All BODS jobs are being terminted with the Syntax Error error when extracting data from ECC using ABAP dataflows:
17740 16404 R3C-150412 06/17/2014 9:38:46 AM |Data flow RT_DF_TEST_PC207
17740 16404 R3C-150412 06/17/2014 9:38:46 AM RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_RUNTIME_FAILURE-(Exception_Key: SYNTAX_ERROR)- Syntax
17740 16404 R3C-150412 06/17/2014 9:38:46 AM error in program /BODS/SAPLBODS .[SAP NWRFC 720][SAP Partner 740 ][DEV1][Server][account][4103]>.
22052 20572 R3C-150412 06/17/2014 9:38:46 AM |Data flow RT_DF_TEST_PC207
22052 20572 R3C-150412 06/17/2014 9:38:46 AM RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_RUNTIME_FAILURE-(Exception_Key: SYNTAX_ERROR)- Syntax
22052 20572 R3C-150412 06/17/2014 9:38:46 AM error in program /BODS/SAPLBODS .[SAP NWRFC 720][SAP Partner 740 ][DEV1][Server][account][4103]>. -
We have an SSIS package that runs on clustered MSSQL 2012 Enterprise Nodes that is failing. We use a job to executer the package.
Environmental information:
Product - Microsoft SQL Server Enterprise: Core-based Licensing (64-bit)
Operating System - Microsoft Windows NT 6.1 (7601)
Patform - NT x64
Version - MSSQL Version 11.0.3349.0
Package is set to 32 -bit. All permissions verified. Runs in lower environments, same MSSQL version. All environments are clustered. In the failing environment, all nodes are at the same service pack. I have not verified if all
nodes in the failing environment have SSIS installed. Data access is installed. We have other simpler packages that run in this environment, just not this one. Time to ask the community for help!
Error:
Source: Data Flow Task - Data Flow Task (SSIS.Pipeline) Description: The version of Lookup is not compatible with this version of the DataFlow. End Error Error: Code: 0xC0048020
Description: Component "Conditional Split, clsid {7F88F654-4E20-4D14-84F4-AF9C925D3087}" could not be created and returned error code 0x80070005 "Access is denied.". Make sure that the component is registered correctly. End Error
Description: The component is missing, not registered, not upgradeable, or missing required interfaces. The contact information for this component is "Conditional Split;Microsoft Corporation; Microsoft SQL Server; (C) Microsoft Corporation; All Rights
Reserved; http://www.microsoft.com/sql/support;0". End Error
(Left out shop specific information. This is the first error in the errors returns by the job history for this package. )
Thanks in advance.Hi DeveloperMax,
According to your description, the error occurs when you execute the package with Agent job on clustered MSSQL 2012 Enterprise Nodes.
As per my understanding, I think this issue can be caused by you use SQL Server Agent to schedule a SQL Server Integration Services package in a 64-bit environment. And the SSIS package is referencing some 32-Bit DLL or 32-Bit drivers which are available
only in 32-bit versions, so the job failed.
To fix this issue, we should use the 32-bit version of the DTExec.exe utility to schedule the 64-bit SQL Server Agent to run a package. To run a package in 32-bit mode from a 64-bit version of SQL Server Agent, we can go to the Job Step dialog box, then
select “32 bit runtime” in the Advanced tab.
Besides, we should make sure that SQL Server Integration Services is installed on the failing environment.
If there are any other questions, please feel free to ask.
Thanks,
Katherine Xiong
Katherine Xiong
TechNet Community Support -
How to work around dataflow order in front panel simulation?
I'm on the steep part of the Labview learning curve. The attached vi is my attempt to simulate the user interface of the product I'm working on. I'm posting because every workaround I dream up to deal with Labview's dataflow architecture results in more complexity. I think I'm putting bandaids on top of bandaids!
If you run the vi you should see what I'm trying to accomplish. The 5 Mode Select buttons send the machine into one of 5 available modes, implemented with a case structure. Delay blocks within the cases simulate how the machine will react. Random number generators simulate occurrence of fault conditions. RS flip flops simulate momentary contact buttons.
Some rules:
- pushing a mode button should instantly switch the machine to the selected mode (my vi doesn't work that way).
- probe test mode should execute just once. My vi simulates this one-time operation by latching the test result - kludge!
- if a fault occurs, it can only be cleared by "power cycling" (stopping and restarting the vi)
Questions:
1) is there an simpler way to simulate momentary contact buttons?
2) is there a simpler, better way to select the cases? I'd like to automatically go from Test to Stop after a single execution of the test case.
3) what is an efficient way to initialize the mode indicators? They 'remember' their last state through a power cycle.
Thanks for any and all insight!
Jeff
Jeff
Climbing the Labview learning curve!
Sanarus Medical
Pleasanton, CAAn alternative method for the UI is to use the radio buttons control. You can customize the standard one to use the square buttons, and I believe this would give you the equivalent effect that you're doing with the lights. See attached VI along with customized control. For the customized control I took a standard radio button control, opened it for editing, and replaced the circular booleans with square buttons. The radio button control is an enumeration, with the enumeration items being the labels of the individual controls (not the text inside the buttons). The control is set so that the labels on the individual square buttons is not visible.
Attachments:
Event Structure Example2.vi 45 KB
Control 1.ctl 6 KB -
ABAP Dataflow using FTP data transfer method
Hi,
I am facing this error while executing the ABAP dataflow, the datastore configuration has FTP as data transfer method.
I wanted to know if the ABAP program is generated succesfully and the error is while executing the ABAP program or there is an issue with the generation of ABAP program itself... Please help me understand the issue.
Since I am unable to attach the doc, please see the error message below.
Job name: Job_Module4_ECC_To_BW
(12.2) 02-28-11 11:14:45 (E) (16396:17152) R3C-150607: |Data flow DF_Module4_ECC_To_BW
Execute ABAP program <C:/TEMP/ZECC_To_BW> error < Open File Error --
madap8004\tmp/SAPECC_To_BW_ZCUST_NAME.txt>.
(12.2) 02-28-11 11:14:45 (E) (16772:15444) R3C-150607: |Data flow DF_Module4_ECC_To_BW
Execute ABAP program <C:/TEMP/ZECC_To_BW> error < Open File Error --
madap8004\tmp/SAPECC_To_BW_ZCUST_NAME.txt>.
Regards,
Sri
Edited by: srilakshmi.ponnada on Feb 28, 2011 11:58 AMHi,
I am really not clear with your question. This is the first time i am executing the ABAP interface with FTP data transfer method. The madap8004 is the FTP server name. and the path is where the file whould be generated I belive. Is it the problem with any read/write permissions on the server?
-Sri -
SQL server 2014 and VS 2013 - Dataflow task, read CSV file and insert data to SQL table
Hello everyone,
I was assigned a work item wherein, I've a dataflow task on For Each Loop container at control flow of SSIS package. This For Each Loop container reads the CSV files from the specified location one by one, and populates a variable with current
file name. Note, the tables where I would like to push the data from each CSV file are also having the same names as CSV file names.
On the dataflow task, I've Flat File component as a source, this component uses the above variable to read the data of a particular file. Now, here my question comes, how can I move the data to destination, SQL table, using the same variable name?
I've tried to setup the OLE DB destination component dynamically but it executes well only for first time. It does not change the mappings as per the columns of the second CSV file. There're around 50 CSV files, each has different set off columns
in it. These files needs to be migrated to SQL tables using the optimum way.
Does anybody know which is the best way to setup the Dataflow task for this requirement?
Also, I cannot use Bulk insert task here as we would like to keep a log of corrupted rows.
Any help would be much appreciated. It's very urgent.
Thanks, <b>Ankit Shah</b> <hr> Inkey Solutions, India. <hr> Microsoft Certified Business Management Solutions Professionals <hr> http://ankit.inkeysolutions.comThe standard Data Flow Task supports only static metadata defined during design time. I would recommend you check the commercial COZYROC
Data Flow Task Plus. It is an extension of the standard Data Flow Task and it supports dynamic metadata at runtime. You can process all your input CSV files using a single Data Flow Task
Plus. No programming skills are required.
SSIS Tasks Components Scripts Services | http://www.cozyroc.com/ -
How to pass a parameter into execute sql task and later use it into dataflow task?
i am in a situation, where i have a logging table in which i have a primary key called ETL_log_ID which is an identity column and acts as a foreign key for various fact table and dimension tables which are populated using SSIS packages. Now i wanna use the
ETL_log_ID as a parameter in the execute sql task which populates the log table and pass the same value in the data flow task which populates the facts and dimension. Can you let me know how to pass the parameter in a step by step procedure.
Thanks,
Nikhil
Nikhil,
You can check the following :
http://www.programmersedge.com/post/2013/03/05/ssis-execute-sql-task-mapping-parameters-and-result-sets.aspx
http://stackoverflow.com/questions/7610491/how-to-pass-variable-as-a-parameter-in-execute-sql-task-ssis
Regarding the usage in Dataflow task, Can you elaborate on that a little?
Thanks,
Jay
<If the post was helpful mark as 'Helpful' and if the post answered your query, mark as 'Answered'> -
Merge data, dataflows from 2 BW systems into one to reduce cost and one way of working
Hello BW gurus,
I have a question here in my current project we are asked to merge data from 2 BW system to one.
This basis for this is there each BW system has their own R3 systems.
First they have now started to migrate data from ECC1 system to ECC2 system using DSM tooling and mapping tables to align to one way of working.
Only open items are migrated from ECC1 to ECC2.
BW1 system is connected to ECC1 and BW2 with ECC2.
Now they want to merge BW1 to BW2.
In BW2 historical data from BW1 is needed.Data for open items from BW1 will come from ECC2 as they are being migrated.
What should be the approach to migrate the historical data from BW1 to BW2.
There can be several scenarios :
1)BW Objects migration
2)Data migration
3)How to deal with common object names used in both systems for DSO and cubes where the data flows in both systems are same?
4)How to deal with common MD info objects (eg:0customer,0vendor) and also if the cutomer numbers in both system are same but with different description and vice versa.
5)How to deal with same kind of transaction data?
I went through many other discussions on the same topic but there it is mentioned to concatenate the 0logsys or sourcesytem in the dataflows where common master is used but creating z objects a copy of 0 objects like 0customer etc.
In these cases if I need to do this to existing 0customer object in that case it would mean deleting data from all flows where 0customer is used if I want to add sourcesytem the compounding which is not favourable.or in case of creating z objects of existing MD IO then add sourcesytem compounding in those cases also it would mean I have to use the new zobjects in all those DSO/Cubes/MP/Queries in place of original object which is also a lot of effort and also load data into it.
This can be done for DSO and cube flows but not MD IO.
Please suggest some good approaches /ways to the above criterias I have mentioned.
System speccification: Currently we are on BW 7.4 version but with the merge we cannot go on Hana due to cost control so we need to use the BW system only.Ritu,
There is no way you can create single model for 2 different R/3 systems (unless SAP decides to implement this in later versions).
If you still want to stick with a single model approach, I would suggest to use <b>Webservice</b> model where the R/3 interaction part would be taken care by WebService.
Ashutosh -
Problems with BC Dataflow for PO History &Confirmations
Hello everyone,
I´m fighting to combine Purhase Order History from SRM (DataSource 0SRM_TD_PO) with Confirmations (Goods receipts) and Invoices from ECC in BI in the DSO 0SRPO_D1.
I will load the PO´s from SRMi do not get the Confirmations from ECC (TR: ME23N)
How is the correct Dataflow for this? When i have a look into the Business Content then it looks like that i have to use the following:
DataSources/InfoSource/DSO/CUBE:
0SRM_TD_PO - PO History -> IS 0SRM_TD_PO -> DSO 0SRPO_D1 --> Cube 0SR_PO_C1
2LIS_02_HDR - ECC/MM -> IS 2LIS_02_HDR -> DSO 0SRPO_D1 --> Cube 0SR_PO_C1
2LIS_02_ITM - ECC/MM -> IS 2LIS_02_ITM -> DSO 0SRPO_D1 --> Cube 0SR_PO_C1
2LIS_02_SCL ?? -> IS 2LIS_02_SCL -> DSO 0SRPO_D1 --> Cube 0SR_PO_C1
2LIS_06_IV Invoice Verifictation. -> DSO 0SRIV_D3 -> Cube: 0SR_IV_C1
When I load the 2LIS_02_ITM into the PSA then i will find the confirmation Quantity. But when I load it then into the DSO 0SRPO_D1 then he will overwrite the records and the confirmation Quantity is empty.
Does anyone of you if there are some additional changes to the BC are necessary? I have also not the Material Document....
many Thanks,
AndreasHi Enzo
When I tried in my system, value is shown correctly.
It must be a bug. Pl report in OSS message.
Best regards
Ramki -
Dataflow in HLS doesn't work in real hardware implementations (13 benchmarks)
Hi all,
I use Vivaod HLS to generate a hardware IP core using streaming interface (hls::stream) and implement it using Vivado. (ARM controls DMA to send the streaming data to hardware IP and after the hardware processing the data, use DMA to transfer the result back to memory from hardware).
The C code in HLS is applied with Dataflow pragma, and the initialtion interval (II) is less than the latency reported by HLS (This is correct, because we are utilized the task level parallelism). Theoretically, the execution time should be like the following graph (Providing 4 frames data and invoking this hardware IP for 4 times) , which means that the execution cycle should be 16386*2 + 95811 * 4
However, in reality, when I implement it on real hardware and measure the execution cycle via SCUtimer on ARM, the execution time is (16386*2+ 95811) * 4, which means that we do not have any benefit from this "dataflow" task level parallelism. The whole hardware is correct and all the results sending back from device to memory are verified. The way I invoke this hardware in the host code is similar to the following pattern:
for (i=0; i<4; i++) {
start timer st;
1. Send i-th frame data from DMA to device;
2. Send i-th frame result from device to DMA;
end timer et;
time += et-st;
Any steps or things I miss here? Does anyone make the dataflow work on the real implementation? Or it just works on the HLS and can not support in Vivado.
Thanks,
Henryhenry.c wrote:
for (i=0; i<4; i++) {
start timer st;
1. Send i-th frame data from DMA to device;
2. Send i-th frame result from device to DMA;
end timer et;
time += et-st;
Hello Henry,
The way you test this does not reflect your flow / dataflow diagram.
In your test, the VHLS IP runs between steps 1 and 2 but you are still serially executing everything : DMA in , buffer in , compute , buffer out, DMA out, DMA in, buffer in, compute, buffer out, DMA out , DMA in, buffer in etc so the total time is the sum of the individual execution.
In your flow diagram (which is the same as the UG description for dataflow) you have the input buffering for frame N happening before the end (and even before the start) of output buffering of frame N-1; so you need to have your embedded code processor doing exactly that.
I'm not so familiar with the way the DMA works and if you can stack up transactions beforehand; that would be the easiest, otherwise you'll have to make it work some other way.
Also your timing measurements should be outside the for loop so that you measure the total execution. (<- that sentence is the EDIT)
I hope this helps. -
7.0 data flow to 3.x dataflow :(
I'm curious to see if there are others who are in my situation and want to know how they are coping with it,
I have been on 7.0 dataflow from 2005 as part of a ramp up customer and have been on it since then. Now I moved to a new project and this customer still uses 3.x dataflow. I feel so lost and I feel as if I'm trying to relearn the old stuff again - and I feel I downgraded myself. 3,x dataflow looks so prehistoric with export datasources and infosources. I cant believe what I got myself into - it took me more than 15minutes to figure out the dataflow for a cube.
I'm looking forward to hear others experiences
back to 3.x days folks.I recently got trained on BI and my trainer insists on first getting it right with the 3.X flow before progressing to the 7.X flow. He also tells me that the 7.X flow is still not completely "stabilized" and SAP releases frequent corrections in that area.
So now I would like to know if there's any iota of truth in it.
pk
ps: My trainer, I gather is well-versed with the BI system and has been a recent Top Contributor here in that area. -
Count of Work flow and dataflow in a single job
Hi,
We would like to know the count of all the worlflow,dataflow which is present in a single job. Basically we want to know what are all the things are called in side the job and how many?
Is there any way to acheive this by seeing the repositary table or any other options are available?
So far we tried to count manually. If single job contains multiple workflow,dataflow,scripts and datastore. It is a very tough process to calculate this or find this. So if some one help us ton this ,It would be helpful for us to calculate the metrix.
Thanks,
VinoYou should be able to get this information by querying the repository. Look for the AL_USAGE table (or ALVW_PARENT_CHILD). The Reference Guide has a chapter on "Metadata in Repository Tables and Views" that explains how to use this table (with some sample SQL).
-
Hi Experts. I have seen the following dataflow for the new GL
0FI_GL_10 --> 0FI_GL_10 --> 0FIGL_C10 (for Totals)
0FI_GL_14 --> 0FIGL_O14 (for Line item)
Question1: The DSO 0FIGL_O14 does not feed the cube. Any particular reason for that?
So If I want line item reporting my best solution will be RRI? Correct!!
Question2: SAP recommended dataflow for new GL uses virtual cube 0FIGL_V10. Why do we need it or what reporting needs it satisfies?
Thanks.Question1: The DSO 0FIGL_O14 does not feed the cube. Any particular reason for that?
So If I want line item reporting my best solution will be RRI? Correct!!
U already have GL totals available at 0FIGL_C10 and then no point in loading line items (O14) again to cube.
Yes.. Best approach for line item reports is RRI i.e. from 0FIGL_C10 to O14.
Question2: SAP recommended dataflow for new GL uses virtual cube 0FIGL_V10. Why do we need it or what reporting needs it satisfies?
SAP delivered function module RS_BCT_FIGL_DATA_GET_VC10 pulls data from 0FIGL_C10 to load 0FIGL_V10... Logic for cumulative balances considering carry forward balances from previous year for Balance sheet & P/L reports, is maintained in that function module.
(So, make sure this FM got installed while u installed standard content for GL model
Do the following, to see weather FM got installed.
Goto change mode of virtual provider (ZFIGL_V10). click information icon (blue 'i')
click 'Type/ Namespace'.. Then click 'details' button
If there is no FM, u won't get any data to 0FIGL_V10 and related queries.
And also 0FIGL_V10 uses Financial Statement Version object (0GLACCEXT), which is not available in 0FIGL_C10.
For Balance sheet & P/L reports, 0GLACCEXT is key object, so 0FIGL_VC10 is must.
U should install & load master (attr, hier, text) data for 0GLACCEXT.
http://help.sap.com/bp_bw370/BBLibrary/documentation/B70_BB_ConfigGuide_EN_DE.doc
Maybe you are looking for
-
How to get the number of minutes or hours from date arithmetic ?
Thanks in advance, I have two date fields that I would like to derive the number of minutes or hours via the following: select start_date - end_date from dual; How do I get HH:MM results ? Thanks
-
Ran DiskUtility on Bootcamp partition, now windows cant boot
I installed Windows 7 Ultimate on my MBPr and it was booting fine several times. I noticed yesterday the bootcamp partition was no longer showing System Preferences -> Startup Disk I could still get Windows 7 to boot if I did a restart on Mac OSX 10.
-
i just made a movie in iMovie and exported in MP4 format, now it won't open, error message says quicktime can't open the file ***? Anyone?
-
everytime I open ffox I have to go into Advance option and un-select the "warn me when sites try to redirect or reload pages" and re-select "make ffox my browser". I also tried selecting 'allow third party cookies' with "from visited' or Always but i
-
How & where to get Javadoc for CRM ISA 5.0?
Hi, SDN Fellows. I read a forum blog, knowing that to get the javadocs of CRM ISA applications, I should perform this step: when building the internet-sales-ear-file, the sources are included as zip-file within the ear-file. after unzipping the sourc