Documentation of data flow and processes in BW
Hi,
Is anyone aware of the possibility to document data flow and job-flows from BW (source systems, datasources etc all the way to queries) and export it from BW.
I am looking for a SAP certified tool like ie. Erwin that can collect this documentation from BW. I have not been able to find out if Erwin supports BW.
My hope is that - now that these things are done in BW - I can extract/export this documentation and use it in our general documentation environment.
The alternative is to manually document it.
Regards,
Mick
Denmark
Hi Mick,
I don't think that there is a tool like the one you are talking about...
however, if you select your infoprovider and then right-click->display data flow...in this way, you can obtain a picture of your data flow from the datasource to your provider with all the interdependencies !
Hope it helps !
Bye,
Roberto
Similar Messages
-
Analysing Task Audit, Data Audit and Process Flow History
Hi,
Internal Audit dept has requested a bunch of information, that we need to compile from Task Audit, Data Audit and Process Flow History logs. We do have all the info available, however not in a format that allows proper "reporting" of log information. What is the best way to handle HFM logs so that we can quickly filter and export required audit information?
We do have housekeeping in place, so the logs are partial "live" db tables, and partial purged tables that were exported to Excel to archive the historical log info.
Many Thanks.I thought I posted this Friday, but I just noticed I never hit the 'Post Message Button', ha ha.
This info below will help you translate some of the information in the tables, etc. You could report on it from the Audit tables directly or move them to another appropriate data table for analysis later. The concensus, though I disagree, is that you will suffer performance issues if your audit tables get too big, so you want to move them periodically. You can do this using a scheduled Task, manual process, etc.
I personally just dump it to another table and report on it from there. As mentioned above, you'll need to translate some of the information as it is not 'human readable' in the database.
For instance, if I wanted to pull Metadata Load, Rules Load, Member List load, you could run a query like this. (NOTE: strAppName should be equal to the name of your application .... )
The main tricks to know at least for task audit table are figuring out how to convert times and determing which activity code corresponds to the user friendly name.
-- Declare working variables --
declare @dtStartDate as nvarchar(20)
declare @dtEndDate as nvarchar(20)
declare @strAppName as nvarchar(20)
declare @strSQL as nvarchar(4000)
-- Initialize working variables --
set @dtStartDate = '1/1/2012'
set @dtEndDate = '8/31/2012'
set @strAppName = 'YourAppNameHere'
--Get Rules Load, Metadata, Member List
set @strSQL = '
select sUserName as "User", ''Rules Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
from ' + @strAppName + '_task_audit ta, hsv_activity_users au
where au.lUserID = ta.ActivityUserID and activitycode in (1)
and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + '''
union all
select sUserName as "User", ''Metadata Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
from ' + @strAppName + '_task_audit ta, hsv_activity_users au
where au.lUserID = ta.ActivityUserID and activitycode in (21)
and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + '''
union all
select sUserName as "User", ''Memberlist Load'' as Activity, cast(StartTime-2 as smalldatetime) as "Time Start",
cast(EndTime-2 as smalldatetime) as ''Time End'', ServerName, strDescription, strModuleName
from ' + @strAppName + '_task_audit ta, hsv_activity_users au
where au.lUserID = ta.ActivityUserID and activitycode in (23)
and cast(StartTime-2 as smalldatetime) between ''' + @dtStartDate + ''' and ''' + @dtEndDate + ''''
exec sp_executesql @strSQLIn regards to activity codes, here's a quick breakdown on those ....
ActivityID ActivityName
0 Idle
1 Rules Load
2 Rules Scan
3 Rules Extract
4 Consolidation
5 Chart Logic
6 Translation
7 Custom Logic
8 Allocate
9 Data Load
10 Data Extract
11 Data Extract via HAL
12 Data Entry
13 Data Retrieval
14 Data Clear
15 Data Copy
16 Journal Entry
17 Journal Retrieval
18 Journal Posting
19 Journal Unposting
20 Journal Template Entry
21 Metadata Load
22 Metadata Extract
23 Member List Load
24 Member List Scan
25 Member List Extract
26 Security Load
27 Security Scan
28 Security Extract
29 Logon
30 Logon Failure
31 Logoff
32 External
33 Metadata Scan
34 Data Scan
35 Extended Analytics Export
36 Extended Analytics Schema Delete
37 Transactions Load
38 Transactions Extract
39 Document Attachments
40 Document Detachments
41 Create Transactions
42 Edit Transactions
43 Delete Transactions
44 Post Transactions
45 Unpost Transactions
46 Delete Invalid Records
47 Data Audit Purged
48 Task Audit Purged
49 Post All Transactions
50 Unpost All Transactions
51 Delete All Transactions
52 Unmatch All Transactions
53 Auto Match by ID
54 Auto Match by Account
55 Intercompany Matching Report by ID
56 Intercompany Matching Report by Acct
57 Intercompany Transaction Report
58 Manual Match
59 Unmatch Selected
60 Manage IC Periods
61 Lock/Unlock IC Entities
62 Manage IC Reason Codes
63 Null -
Losing source-target mapping from a data flow diagram process
Hi,
When I reload a model design from an xml repository, I am losing source-target mappings from a data flow diagram process. I wonder why.
The way it goes is:
- I define a data flow diagram with information stores, processes, and flows.
- Information structures are defined from the flows and the information trickles down to processes and information stores.
- In a process, I define Source-Target Mapping by selecting a target element and then selecting the source element(s).
- I save the whole model.
- So far, the Source-Target Mapping keeps the association between target and source elements.
- If I exit from Data Modeler, the next time I reload the saved model, the Source-Target Mapping target to source elements associations are lost!?
Does anyone have the same issue or would know how to resolve it?
Thanks,
/MarioI am still having problems getting my source to target mappings to save after I apply the fix described in this thread. Any thoughts?
From my TargetElement.xml
<?xml version = '1.0' encoding = 'UTF-8'?>
<oracle.dbtools.crest.model.design.process.TargetElement parentClass="java.lang.Object">
<property name="objectID" dataType="java.lang.String" readOnly="true" defaultValue="" xmlName="id" xmlType="attribute" getter="getObjectID" setter="setObjectID" reffered="false"/>
<property name="target" dataType="oracle.dbtools.crest.model.design.DesignObject" defaultValue="" xmlName="target" xmlType="element" getter="getTarget" setter="setTarget" reffered="true"/>
<property name="description" dataType="java.lang.String" defaultValue="" xmlName="description" xmlType="element" getter="getDescription" setter="setDescription" reffered="false"/>
<property name="name" dataType="java.lang.String" readOnly="true" defaultValue="" xmlName="name" xmlType="attribute" getter="getName" setter="setName" reffered="false"/>
<property name="transformationType" dataType="java.lang.String" defaultValue="" xmlName="transformationType" xmlType="element" getter="getTransformationType" setter="setTransformationType" reffered="false"/>
<collection name="sources" dataType="java.util.List" xmlName="sources" itemXmlName="sourceRef" getter="getSources" itemIsRef="true" refsXMLName="sourceRefs" addItem="addSource" getItemByID="getSourceByID" nativeList="true" orderTheList="true"/>
<roproperty name="sources" dataType="interface java.util.List" getter="getSources"/>
</oracle.dbtools.crest.model.design.process.TargetElement>
Process.xml:
<?xml version="1.0" encoding="UTF-8" ?>
- <oracle.dbtools.crest.model.design.process.Process parentClass="oracle.dbtools.crest.model.design.process.Transformation">
<property name="type" dataType="int" defaultValue="0" xmlName="type" xmlType="element" getter="getType" setter="setType" reffered="false" />
<property name="batchMinimumTransactions" dataType="int" defaultValue="" xmlName="batchMinimumTransactions" xmlType="element" getter="getBatchMinimumTransactions" setter="setBatchMinimumTransactions" reffered="false" />
<property name="batchTimeUnit" dataType="java.lang.String" defaultValue="" xmlName="batchTimeUnit" xmlType="element" getter="getBatchTimeUnit" setter="setBatchTimeUnit" reffered="false" />
<property name="diagram" dataType="oracle.dbtools.crest.model.design.process.DataFlowDesign" defaultValue="" xmlName="diagram" xmlType="element" getter="getDiagram" setter="setDiagram" reffered="true" />
<property name="footNote" dataType="java.lang.String" defaultValue="" xmlName="footNote" xmlType="element" getter="getFootNote" setter="setFootNote" reffered="false" />
<property name="frequencyTimeUnit" dataType="java.lang.String" defaultValue="" xmlName="frequencyTimeUnit" xmlType="element" getter="getFrequencyTimeUnit" setter="setFrequencyTimeUnit" reffered="false" />
<property name="frequencyTimes" dataType="int" defaultValue="" xmlName="frequencyTimes" xmlType="element" getter="getFrequencyTimes" setter="setFrequencyTimes" reffered="false" />
<property name="interactiveLongestResponseTime" dataType="int" defaultValue="" xmlName="interactiveLongestResponseTime" xmlType="element" getter="getInteractiveLongestResponseTime" setter="setInteractiveLongestResponseTime" reffered="false" />
<property name="interactiveResponseTimeUnit" dataType="java.lang.String" defaultValue="" xmlName="interactiveResponseTimeUnit" xmlType="element" getter="getInteractiveResponseTimeUnit" setter="setInteractiveResponseTimeUnit" reffered="false" />
<property name="mode" dataType="java.lang.String" defaultValue="" xmlName="mode" xmlType="element" getter="getMode" setter="setMode" reffered="false" />
<property name="paramsWrappersString" dataType="java.lang.String" defaultValue="" xmlName="paramsWrappersString" xmlType="element" getter="getParamsWrappersString" setter="setParamsWrappersString" reffered="false" />
<property name="priority" dataType="java.lang.String" defaultValue="" xmlName="priority" xmlType="element" getter="getPriority" setter="setPriority" reffered="false" />
<property name="transformationtask" dataType="oracle.dbtools.crest.model.design.process.TransformationTask" defaultValue="" xmlName="transformationtask" xmlType="element" getter="getTransformationTask" setter="setTransformationTask" reffered="true" />
<property name="peakPeriodsString" dataType="java.lang.String" defaultValue="" xmlName="peakPeriodsString" xmlType="element" getter="getPeakPeriodsString" setter="setPeakPeriodsString" reffered="false" />
<collection name="events" dataType="java.util.List" xmlName="events" itemXmlName="eventRef" getter="getEvents" itemIsRef="true" refsXMLName="eventRefs" removeItem="removeEvent" addItem="addEvent" getItemByID="getPartyByID" nativeList="true" orderTheList="true" />
<collection name="targetMappings" dataType="java.util.List" xmlName="targetMappings" itemXmlName="targetMapping" getter="getTargetElements" createItem="createTargetElement" removeItem="removeEvent" addItem="addTargetElement" getItemByID="getPartyByID" nativeList="true" orderTheList="true" />
<roproperty name="componentFlowsFor" dataType="interface java.util.Collection" getter="getComponentFlowsFor" />
<roproperty name="events" dataType="interface java.util.List" getter="getEvents" />
<roproperty name="flowEvents" dataType="interface java.util.List" getter="getFlowEvents" />
<roproperty name="paramMapping" dataType="oracle.dbtools.crest.model.design.DesignObject" getter="getParamMapping" />
<roproperty name="paramsWrappers" dataType="interface java.util.List" getter="getParamsWrappers" />
<roproperty name="parentprocess" dataType="oracle.dbtools.crest.model.design.process.Process" getter="getParentProcess" />
<roproperty name="parentprocessInDFD" dataType="oracle.dbtools.crest.model.design.process.Process" getter="getParentProcessInDFD" />
<roproperty name="peakperiods" dataType="[Z" getter="getPeakPeriods" />
<roproperty name="processedAttributesAndFlows" dataType="interface java.util.List" getter="getProcessedAttributesAndFlows" />
<roproperty name="processedEntitiesAndFlows" dataType="interface java.util.List" getter="getProcessedEntitiesAndFlows" />
</oracle.dbtools.crest.model.design.process.Process> -
How to solve delay in a program with data aquisition and processing
Hello, I am a starter in Labview programing. I am working on a system which contains a roller, a piston and a A/D cart which is from Data Translation Inc (DT304). I am using labview to get speed data of the roller (which is voltage first then be converted into speed) from Analog Input Channels, and then using Analog Output Channels to control the friction force being appylied to the roller by using a piston.
I am using a Waveform Chart to show the speed of roller. However, as I am adding more components to the program, I always get delay in the speed display and also the output control of the piston. Also, as I spent more time using the program eg. from morning to night, it became slower too.
My question is, how does labview store data and how are the data been stored and released in the buffer.Is it because I am loading too much data in the buffer so that it became slower? And also how to solve these delays?
I am using the "save to file" function to write data to a ".lvm" file, but I did some changes to the saving by using a button to enable and disable saving. So I don't think the problem is because I am always saving the data into a file.
Thank you very much.
JessieI'm betting it's two things:
1. Program Architecture
and 2. Dynamically resizing arrays.
With the program architecture... if you have one do-while loop that acquires data from teh DAQ, processes it, opens/indexes/writes/closes a file, then comes back around... as your arrays and files increase in size, the loop is going to take longer to execute. A Producer-Consumer loop (a good tutorial on them can be found here) has one loop that acquires data and stuffs the data into a queue. This will buffer the data while it's in transition between acquisition and processing. A second loop in parallel takes data in from the queue, processes it, then comes back around. The two loops operate independently of each other, so even if the consumer loop takes longer as the files or math gets more complex, the producer loop continues to run full speed.
Second is the arrays. Every time you append data into an array, LabVIEW has to make a copy of the data that's in the array. If you append small amounts of data to an array over and over and over again, eventually LabVIEW is going to be copying very large amounts of data over and over and over again. The producer-consumer architecture can alleviate this problem. -
If I have 3 data flow tasks A > B > C
Then a SQL task that executes a procedure.
Can the procedure refer to records that have been added in one of the previous data flow tasks?
All the tasks are in a container that will commit when its finished, but if its not finished yet, will the SQL task and the procedure be able to refer to any of the records that were loaded?That depends on how you handle connections. Changes made in a transaction are visible to only the connection that made the changes until the transaction commits or rolls back.
David
David http://blogs.msdn.com/b/dbrowne/ -
Difference between Service data objects and process data object
Hi
Can anybody tell me what is the difference between SDO (Service Data Objects) & PDO (Process Data Objects).I am using 2 port PCI-CAN series 2 card.
If anybody knows the answer then please reply.
Thanking YouHi,
See the online help file from our NI CANopen Library for LabVIEW for some basic informations about SDOs and PDOs.
More information should be available from the web.
DirkW
Attachments:
lvcanopenvhelp.zip 154 KB -
Data flows and technical terms
Hi all,
Anyone please give me SD data flow i.e what is sales order and when it will be generated and how it is linked with billing and delivery and what is invoice, cost centre, profit center, business area, wbs element etc.
And also the purchase data flow ie Purchase order purchase requisition, material document etc.
Provide me what exactly these mean in simple words.
I am new to BI and i am very confused with these technical terms.
Regards,
Nishuv.Hi,
Cost centre & profit Centre Hierarchy:
you have to look at the Implementation Guide (IMG) for Profit Center Accounting, Master Data -> Create/Change/Display Hierarchy or something similar..with Transaction Code:KCH3 .
But if you want specfic one , need to build a specific hierarchy in R3 and load it in BW.
You can also build it directly in BW.
If it is product hierachy u have to generate the levels in R/3 side (using same transaction in SBIW), which activates PRODH1, PRODH2...etc.
SBIW or SAP reference IMG under General Logistics -> Master Material -> Material or Settings for Central Fields -> Sales-Relevant Data -> Define Product Hierarchies
Please check this link for the detailed procedure to create hiearchy;
http://help.sap.com/saphelp_nw04/helpdata/en/0e/fd4e3c97f6bb3ee10000000a114084/frameset.htm.
Create Hierarchy in R/3 -Sample
http://www.bwexpertonline.com/go/sample/
Pls chk this links;
Re: Hierarchy Loading from R/3 to BW
Product Hirarchies for 0MATERIAL (from R3)
R/3 Hierarchy Text upload issue
http://help.sap.com/saphelp_nw70/helpdata/en/b3/41cc373de44040e10000009b38f8cf/frameset.htm
http://help.sap.com/saphelp_nw70/helpdata/en/c6/f3dc3709e9ea12e10000009b38f8cf/frameset.htm
Hope this helps,
Regards
CSM Reddy -
How to create process chain for this data flow
Hi experts,
My data flow is:
======================================
2lis_11_vahdr->Transfer rules->2lis_11_vahdr->Update rules->DSO1->Transformation->DTP->Cube1.
2lis_11_vaitm->Transfer rules->2lis_11_vaitm->Update rules->DSO1->Transformation->DTP->Cube1.
======================================
2lis_12_vchdr->Transfer rules->2lis_12_vchdr->Update rules->DSO2->Transformation->DTP->Cube1.
2lis_12_vcitm->Transfer rules->2lis_12_vcitm->Update rules->DSO2->Transformation->DTP->Cube1.
======================================
2lis_13_vdhdr->Transfer rules->2lis_13_vdhdr->Update rules->DSO3->Transformation->DTP->Cube1.
2lis_13_vditm->Transfer rules->2lis_13_vditm->Update rules->DSO3->Transformation->DTP->Cube1.
======================================
Here for each datasource info package brings data upto dso and then from dso dtp fetches the data.
For deltas i want to keep this data flow in process chain.
Anyone please guide me how to keep this in a process chain.
Full points will be assigned.
Regards,
Bhadri m.
Edited by: Bhadri M on Sep 2, 2008 7:20 PM
Edited by: Bhadri M on Sep 2, 2008 7:21 PMHello,
Sure it is possible to maintain that dataflow.
Start here:
This is a very good document about process chains:
https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/8da0cd90-0201-0010-2d9a-abab69f10045
Let me know if you have any doubts.
Regards,
Jorge Diogo -
SSIS Control and Data Flow Items not available in Customize Toolbox
I cannot add additional Data Flow and Control Flow items to the SSIS Toolbox. I am using SQL Server 2014 and VS 2013 (BIDS). The standard SSIS items appear - but they are all of them. How do I add the additional ones that are not loaded by default. I didn't
have this issue with SQL Server 2012.
Thank you,
Johhn
JohnJohn, which ones you do not see? AFAIK nothing has changed regarding how the items get added to SSIS toolbox in SSDT or VS 2013 with SSDT installed (no BIDS in SSIS 2012 and up).
Arthur My Blog -
I am using a Data Flow and an OLE DB Source to read my staged 3rd party external data. I need to do various Lookups to try and determine if I can find the external person in our database...by SSN...By Name and DOB...etc...
Now I need to do some more data verification based on the Lookup that is successful. Can I do those data edits against our SQL server application database by utilizing an OLE DB Command? Using a Stored Procedure or can I sue straight SQL to perform my edit
against every staging row by using a parameter driven query? I'm thinking a Stored Procedure is the way to go here since I have multiple edits against the database. Can I pass back the result of those edits via a variable and then continue my SSIS Data Flow
by analyzing the result of my Stored Procedure? And how would I do that.
I am new to the SSIS game here so please be kind and as explicit as possible. If you know of any good web sites that walk through how to perform SQL server database edits against external data in SSIS or even a YouTube, please let me know.
Thanks!Thanks for that...but can I do multiple edits in my Stored Procedure Vaibhav and pass back something that I can then utilize in my SSIS? For example...
One and Only one Member Span...so I'd be doing a SELECT COUNT(*) based on my match criteria or handle the count accordingly in my Stored Procedure and passing something back via the OLE DB Command and handling it appropriately in SSIS
Are there "Diabetes" claims...again probably by analyzing a SELECT COUNT(*)
Am I expecting too much from the SSIS...should I be doing all of this in a Stored Procedure? I was hoping to use the SSIS GUI for everything but maybe that's just not possible. Rather use the Stored Procedure to analyze my stged data, edit accordingly, do
data stores accordingly...especially the data anomalies...and then use the SSIS to control navigation
Your thoughts........
Could you maybe clarify the difference between an OLE DB Command on the Data Flow and the Execute SQL Task on the Control Flow...
You can get return values from oledb comand if you want to pipeline.
see this link for more details
http://josef-richberg.squarespace.com/journal/2011/6/30/ssis-oledb-command-and-procedure-output-params.html
The procedure should have an output parameter defined for that
I belive if you've flexibility of using stored procedure you may be better off doing this in execute sql task in control flow. Calling sp in data flow will cause it to execute sp once for each row in dataset whereas in controlflow it will go for set based
processing
Please Mark This As Answer if it solved your issue
Please Vote This As Helpful if it helps to solve your issue
Visakh
My Wiki User Page
My MSDN Page
My Personal Blog
My Facebook Page -
Data flow not visible while creating remote cube.
Hi SDN,
I am working on a remote cube. I need to link the infosource to remote cube. I have selected the source system and assigned.
I have done, like--Remote cube --> Context Menu --> Show Data flow , and I wanted to see the source system , Info source and the remote cube, but could not find them .
Guide me what I've missed /went wrong.
Thanks in Advance
AnkitHi,
Remote cube is technology where you report on data which is stored in the source system. While creating the Infocube you have to select Remote Infocube Radiobutton and in the next screen it will ask you for the infosource.
In that infosource you can assign the datasource in the infosource menu.
Now you see the data flow.
Hope it works,
Regards,
Sasi -
Interleave​d data saving and the TDMS Viewer VI.
Hello,
I'm using LabVIEW 2009 Service Pack 1 ( ver 9.0.1 32-bit ).
I was trying to take a piece of data and save it in interleaved mode.
The save operation seems to be completed correctly, but when the TDMS Viewer VI is launched it crashes, and I'm unable to browse the data.
I'm attaching the example I used.
What I want to know is:
1) am I using the write properly? What I'm trying to acomplish is to push down an array if I16 to a TDMS write block, recieve 30 channels of data this way.
2) I managed to use this VI and it crashes, on two PC's, one is a win7 32bit, the second one was an XP. But let me know if this runs properly for you.
Maciej
Attachments:
InterleavedDataErr.zip 16 KBI have a buffer that can accumulate up to 65536 I16 values.
In that buffer I get for example 6x 10000 values.
Every "chunk" of the 10000 values is my interleaved data. One point is one channel.
My data is shaped like in the opening post:
http://forums.ni.com/t5/LabVIEW/TDMS-Streaming-in-interleaved-mode/m-p/1189201#M513943
Since I'm running a windows machine, sometimes I'll get 5x 10000 values, sometimes it will be 3x , sometimes 6x during one buffer swipe.
The data rate and the channel length is variable.
But let's talk about one case:
10k channels and 10 kHz rate of I16 vals.
That gives me roughly 200 MB/s of data that I need to get out of the buffer.
What I want to do on my PC is take that data, and save it interleaved to a TDMS file, while being able to safely stay above the 200 MB/s.
I would be intersted in aiming for 300 MB/s so that if we push our card further down the line, I'm not stuck with the streaming bandwitch.
I have noticed that unfortunately if you use an ammount of channles that is 10000 the writing performance drops dramatically.
What I figured out is that I'm not conducting 1 write on 10000 channels on a small ammount of data. But rather than that a large accumulated amount of data (like 8000 x 10000 of I16)
And I split the save of 10000 channels into 10x 1000 channels save.
First channs from 0-999 , than 1000-1999 etc.
All is done in interleaved mode.
This is Very Very fast indeed. I got rates of 360 MB/s. That's smoking fast!
Now the trick is... to handle the appropriate data shaping in the RAM, from my buffer, to the large shaped RAM chunk, that will be cut in to pieces for Streaming. LV is designed for data flow and copies data a lot, hence I get easlity in the out of memory problem, I'm going to have to use some ticks like storing the data structure via a VI Server call, or perhaps queues don't know yet. But I'm going to do some reading on that and it seems that this should be achieavable.
Thank you both for your input,
I will post if I get stuck.
Maciej -
Data source and infocube in BC
Hi all,
I have some basic questions related to BC,
How to identify the IC for the given Data source?/How to identify the Data source for the given IC?
Can we assign more than one data source to one info cube?Hi,
The best input would be from SP help for each application component. Alernatively goto metadata repository and search there.
If you already have some extractions for that particular component, you can right click and select show data flow and it'll give you the entire picture if you have mapped everything.
The other way is if you know for ex a cube, you can goto business content and drag it to the right hand side window. Give the method as data flow before and afterwards and it'll pull everything. That way you can get to know all the components.
You can assign multiple datasources to a cube.
Cheers,
Kedar -
Hi
can we replicate ABAP data flow and do modifications for history data upload?It is a copy and whatever changes you will make is not going to impact the other ABAP dataflows.
-
Required, SD DATA FLOW with MM Integration.
Dear Experts,
I have used the search functionality.. but got links..
Please Explain in your words, 1.SD Data Flow.
and also please mention its Integration with MM.
Points Assured.
ThanksLink Between SAP SD, MM & FI
The link between SD and MM :-
1. When you create sales order in SD, all the details of the items are copied from Material master of MM.
2. MRP and availibility check related data is also taken from MM although you control this data in SD also.
3. While you create inbound/outbound delivery with reference to a sales order,the shipping point determination takes place with the help of the loading group, plant data, shipping conditions etc. This also refers to Material Master.
4. The material which you are entering in a sales order must be extended to the sales area of your sales order/customer otherwise you cannot transact with this material.
There are many such links between SD and MM.
Now the link between SD and FI :-
1. Whenever you create a delivery with reference to a sales order, goods movement takes place in the bacgground. eg. In case of standard sales order, you create an outbound goods delivery to the customer.
Here movement 601 takes place. This movement is configured in MM. Also, this movement hits some G/L account in FI. Every such movement of good s hits some G/L account.
2. The accounts posting in FI is done with reference to the billing documents (invoice, debit note, credit note etc) created in SD. Thus this is a link between SD and FI
3. Tax determination: In case of a tax determination also, there is a direct link between SD and MM
SD Integration points with other modules
SD module is highly integrated with the other modules in SAP.
Sales Order
Integration Points Module
Availability Check - MM
Credit Check - FI
Costing - CO/ MM
Tax Determination - FI
Transfer of Requirements - PP/ MM
Delivery & Goods Issue
Integration Points Module
Availability Check - MM
Credit Check - FI
Reduces stock - MM
Reduces Inventory $ - FI/ CO
Requirement Eliminated - PP/ MM
Billing -
Integration Points Module
Debit A/R - FI/ CO
Credit Revenue - FI/ CO
Updates G/ L - FI/ CO
(Tax, discounts, surcharges, etc.)
Milestone Billing - PS
Return Delivery & Credit Memo -
Integration Points Module
Increases Inventory - MM
Updates G/ L - FI
Credit Memo - FI
Adjustment to A/R - FI
Reduces Revenue - FI
SD Transaction Code Flow:
Inquiry / Document type IN
Tcode for creation VA11,VA12,VA13. tables VBAK,VBAP
Quotation / QT
Tcode for creation VA21,VA22,VA23. tables VBAK,VBAP
Sales Order OR
Tcode for creation VA01,VA02,VA03. tables VBAK,VBAP
Delivery LF
Tcode for creation VL01,VL02,VL03. tables LIKP,LIPS
Billing F2
Tcode for creation VF01,VF02,VF03. tables VBRK,VBRP
Regards,
Rajesh Banka
Reward suitable points.
Maybe you are looking for
-
How do I remove the Verizon instructions from end of my customized greeting?
Hello everyone. I've called Verizon repeatedly about this one and they say they can't help. I want to remove the Verizon system message from the end of my customized greeting and reduce the number of rings (currently 5) before my callers can leave
-
New to ABAP Objects. Need help using events
Hi all, I got a program from one of the text books which has the output as pursur helps pilot stewardess helps passenger on seat 11 stewardess helps passenger on seat 17 I am trying to modify this program to get an output:
-
White Balance: can you apply setting from one image to a series of other images?
I like to make a grey card reference image whenever the light changes and use the image of the grey card to set the white balance. In Bridge, it was possible to set the white balance on any number of other images by applying the setting made on the r
-
Creating a Report Using a SharePoint List
I am using a SharePoint List that Named as 'Service Requests'. It has some fields with Date, Time, Status. The Date & Time will be entered. The Status as three values & those will be selected using a Drop Down List Box. The Statuses are 'In Progress,
-
US AT&T iPhone4 (iOS5) -- use in England with local SIM card?
I will be travelling to England for several weeks. Can I use my US AT&T iPhone4 (iOS5) by purchasing a SIM card from a U.K. based company? Are there any resources on how to do this, and what is the best network (London, Somerset, Cornwall, Kent) Th