Data flows and technical terms

Hi all,
Anyone please give me SD data flow i.e what is sales order and when it will be generated and how it is linked with billing and delivery and what is invoice, cost centre, profit center, business area, wbs element etc.
And also the purchase data flow ie Purchase order purchase requisition, material document etc.
Provide me what exactly these mean in simple words.
I am new to BI and i am very confused with these technical  terms.
Regards,
Nishuv.

Hi,
Cost centre & profit Centre Hierarchy:
you have to look at the Implementation Guide (IMG) for Profit Center Accounting, Master Data -> Create/Change/Display Hierarchy or something similar..with Transaction Code:KCH3 .
But if you want specfic one , need to build a specific hierarchy in R3 and load it in BW.
You can also build it directly in BW.
If it is product hierachy u have to generate the levels in R/3 side (using same transaction in SBIW), which activates PRODH1, PRODH2...etc.
SBIW or SAP reference IMG under General Logistics -> Master Material -> Material or Settings for Central Fields -> Sales-Relevant Data -> Define Product Hierarchies
Please check this link for the detailed procedure to create hiearchy;
http://help.sap.com/saphelp_nw04/helpdata/en/0e/fd4e3c97f6bb3ee10000000a114084/frameset.htm.
Create Hierarchy in R/3 -Sample
http://www.bwexpertonline.com/go/sample/
Pls chk this links;
Re: Hierarchy Loading from R/3 to BW
Product Hirarchies for 0MATERIAL (from R3)
R/3 Hierarchy Text upload issue
http://help.sap.com/saphelp_nw70/helpdata/en/b3/41cc373de44040e10000009b38f8cf/frameset.htm
http://help.sap.com/saphelp_nw70/helpdata/en/c6/f3dc3709e9ea12e10000009b38f8cf/frameset.htm
Hope this helps,
Regards
CSM Reddy

Similar Messages

  • Documentation of data flow and processes in BW

    Hi,
    Is anyone aware of the possibility to document data flow and job-flows from BW (source systems, datasources etc all the way to queries) and export it from BW.
    I am looking for a SAP certified tool like ie. Erwin that can collect this documentation from BW. I have not been able to find out if Erwin supports BW.
    My hope is that - now that these things are done in BW - I can extract/export this documentation and use it in our general documentation environment.
    The alternative is to manually document it.
    Regards,
    Mick
    Denmark

    Hi Mick,
    I don't think that there is a tool like the one you are talking about...
    however, if you select your infoprovider and then right-click->display data flow...in this way, you can obtain a picture of your data flow from the datasource to your provider with all the interdependencies !
    Hope it helps !
    Bye,
    Roberto

  • Multiple data flow and sql tasks in a container - are records available in other tasks before transaction commits?

    If I have 3 data flow tasks A > B > C
    Then a SQL task that executes a procedure.
    Can the procedure refer to records that have been added in one of the previous data flow tasks?
    All the tasks are in a container that will commit when its finished, but if its not finished yet, will the SQL task and the procedure be able to refer to any of the records that were loaded?

    That depends on how you handle connections.  Changes made in a transaction are visible to only the connection that made the changes until the transaction commits or rolls back.
    David
    David http://blogs.msdn.com/b/dbrowne/

  • Large data sets and key terms

    Hello, I'm looking for some guidance on how BI can help me. I am a business analyst in a health solutions firm, but not proficient in SQL. However, I have to work with large data sets that just exceed the capabilities of Excel.
    Basically, I'm having to use Excel to manaully search for key terms and apply a values to those results. For instance, I have a medical claims file, with Provider Names, Tax ID, Charges, etc. It's 300,000 records long and 15-25 columsn wide. I need to search for key terms in the provider name like Ambulance, Fire Dept, Rescue, EMT, EMS, etc. Anything that resembles an ambulance service. Also, need to include abbreviations of them such as AMB, FD, or variations like EMT, E M T, EMS, E M S, etc. Each time I do a search, I have filter and apply an "N/A" flag.
    That's just one key term. I also have things like Dentists or DDS, Vision, Optomemtry and a dozen other Provider Types that need to be flagged as "N/A".
    Is this something that can be handled using BI? I have access to a BI group, but I need to understand more about the capabilities of what can be done. As an analyst, I'm having to deal with poor data inegrity. So, just cleaning up the file can be extremely taxing and cumbersome.
    Some insight would be very helpful. Thanks.

    I am not sure if you are looking for an explanation about different BI products? If so, may be this forum is not the place to get a straight answer.
    But, Information Discovery product suite might be useful in your case. Regarding the "large date set" you mentioned, searching and analyzing 300,000 records may not be considered a large data set at least in Endeca standards :).
    All your other requests, could also be very easily implemented using Endeca's product suite. Please reach out to Oracle's Endeca product team and they might guide you on how this product suite would help you.

  • Where is the "search" feature in ibook?  How do I create a vocabulary list of words and technical terms?

    Is there a "search" feature in iBook on my ipad; and how can I create a "vocalbulary List" of terms or words while reading or studying with ipad?

    Tap the page of the iBook then tap the magnifying glass for the search feature. Use Notes for your vocabulary list.

  • SSIS Control and Data Flow Items not available in Customize Toolbox

    I cannot add additional Data Flow and Control Flow items to the SSIS Toolbox. I am using SQL Server 2014 and VS 2013 (BIDS). The standard SSIS items appear - but they are all of them. How do I add the additional ones that are not loaded by default. I didn't
    have this issue with SQL Server 2012.
    Thank you,
    Johhn
    John

    John, which ones you do not see? AFAIK nothing has changed regarding how the items get added to SSIS toolbox in SSDT or VS 2013 with SSDT installed (no BIDS in SSIS 2012 and up).
    Arthur My Blog

  • Data flow not visible while creating remote cube.

    Hi SDN,
    I am working on a remote cube. I need to link the infosource to remote cube. I have  selected the source system and assigned.
    I have done,  like--Remote cube --> Context Menu --> Show Data flow , and I wanted to see the source system , Info source and the remote cube, but could not find them .
    Guide me what I've missed /went wrong.
    Thanks in Advance
    Ankit

    Hi,
    Remote cube is technology where you report on data which is stored in the source system. While creating the Infocube you have to select Remote Infocube Radiobutton and in the next screen it will ask you for the infosource.
    In that infosource you can assign the datasource in the infosource menu.
    Now you see the data flow.
    Hope it works,
    Regards,
    Sasi

  • Interleave​d data saving and the TDMS Viewer VI.

    Hello,
    I'm using LabVIEW 2009 Service Pack 1 ( ver 9.0.1 32-bit ).
    I was trying to take a piece of data and save it in interleaved mode.
    The save operation seems to be completed correctly, but when the TDMS Viewer VI is launched it crashes, and I'm unable to browse the data.
    I'm attaching the example I used.
    What I want to know is:
    1) am I using the write properly? What I'm trying to acomplish is to push down an array if I16 to a TDMS write block, recieve 30 channels of data this way.
    2) I managed to use this VI and it crashes, on two PC's, one is a win7 32bit, the second one was an XP. But let me know if this runs properly for you.
    Maciej
    Attachments:
    InterleavedDataErr.zip ‏16 KB

    I have a buffer that can accumulate up to 65536 I16 values.
    In that buffer I get for example 6x 10000 values.
    Every "chunk" of the 10000 values is my interleaved data. One point is one channel.
    My data is shaped like in the opening post:
    http://forums.ni.com/t5/LabVIEW/TDMS-Streaming-in-​interleaved-mode/m-p/1189201#M513943
    Since I'm running a windows machine, sometimes I'll get 5x 10000 values, sometimes it will be 3x , sometimes 6x during one buffer swipe.
    The data rate and the channel length is variable.
    But let's talk about one case:
    10k channels and 10 kHz rate of I16 vals.
    That gives me roughly 200 MB/s of data that I need to get out of the buffer.
    What I want to do on my PC is take that data, and save it interleaved to a TDMS file, while being able to safely stay above the 200 MB/s.
    I would be intersted in aiming for 300 MB/s so that if we push our card further down the line, I'm not stuck with the streaming bandwitch.
    I have noticed that unfortunately if you use an ammount of channles that is 10000 the writing performance drops dramatically.
    What I figured out is that I'm not conducting 1 write on 10000 channels on a small ammount of data. But rather than that a large accumulated amount of data (like 8000 x 10000 of I16)
    And I split the save of 10000 channels into 10x 1000 channels save.
    First channs from 0-999 , than 1000-1999 etc.
    All is done in interleaved mode.
    This is Very Very fast indeed. I got rates of 360 MB/s. That's smoking fast!
    Now the trick is... to handle the appropriate data shaping in the RAM, from my buffer, to the large shaped RAM chunk, that will be cut in to pieces for Streaming. LV is designed for data flow and copies data a lot, hence I get easlity in the out of memory problem, I'm going to have to use some ticks like storing the data structure via a VI Server call, or perhaps queues don't know yet. But I'm going to do some reading on that and it seems that this should be achieavable.
    Thank you both for your input,
    I will post if I get stuck.
    Maciej

  • Data source and infocube in BC

    Hi all,
    I have some basic questions related to BC,
    How to identify the IC for the given Data source?/How to identify the Data source for the given IC?
    Can we assign more than one data source to one info cube?

    Hi,
    The best input would be from SP help for each application component. Alernatively goto metadata repository and search there.
    If you already have some extractions for that particular component, you can right click and select show data flow and it'll give you the entire picture if you have mapped everything.
    The other way is if you know for ex a cube, you can goto business content and drag it to the right hand side window. Give the method as data flow before and afterwards and it'll pull everything. That way you can get to know all the components.
    You can assign multiple datasources to a cube.
    Cheers,
    Kedar

  • ABAP Data flow

    Hi
    can we replicate ABAP data flow and do modifications for history data upload?

    It is  a copy and whatever changes you will make is not going to impact the other ABAP dataflows.

  • Required, SD DATA FLOW  with MM Integration.

    Dear Experts,
    I have used the search functionality.. but got links..
    Please Explain in your words, 1.SD Data Flow.
    and also please mention its Integration with MM.
    Points Assured.
    Thanks

    Link Between SAP SD, MM & FI
    The link between SD and MM :-
    1. When you create sales order in SD, all the details of the items are copied from Material master of MM.
    2. MRP and availibility check related data is also taken from MM although you control this data in SD also.
    3. While you create inbound/outbound delivery with reference to a sales order,the shipping point determination takes place with the help of the loading group, plant data, shipping conditions etc. This also refers to Material Master.
    4. The material which you are entering in a sales order must be extended to the sales area of your sales order/customer otherwise you cannot transact with this material.
    There are many such links between SD and MM.
    Now the link between SD and FI :-
    1. Whenever you create a delivery with reference to a sales order, goods movement takes place in the bacgground. eg. In case of standard sales order, you create an outbound goods delivery to the customer.
    Here movement 601 takes place. This movement is configured in MM. Also, this movement hits some G/L account in FI. Every such movement of good s hits some G/L account.
    2. The accounts posting in FI is done with reference to the billing documents (invoice, debit note, credit note etc) created in SD. Thus this is a link between SD and FI
    3. Tax determination: In case of a tax determination also, there is a direct link between SD and MM
    SD Integration points with other modules
    SD module is highly integrated with the other modules in SAP.
    Sales Order –
    Integration Points                      Module
    •Availability Check             -       MM
    •Credit Check                   -       FI
    •Costing                        -       CO/ MM
    •Tax Determination              -       FI
    •Transfer of Requirements       -       PP/ MM
    Delivery & Goods Issue –
    Integration Points                      Module
    •Availability Check             -       MM
    •Credit Check                   -       FI
    •Reduces stock                  -       MM
    •Reduces Inventory $            -       FI/ CO
    •Requirement Eliminated         -       PP/ MM
    Billing -
    Integration Points                      Module
    •Debit A/R                      -       FI/ CO
    •Credit Revenue                 -       FI/ CO
    •Updates G/ L                   -       FI/ CO
      (Tax, discounts, surcharges, etc.)
    •Milestone Billing              -       PS
    Return Delivery & Credit Memo -
    Integration Points                      Module
    •Increases Inventory            -       MM
    •Updates G/ L                   -       FI
    •Credit Memo                    -       FI
    •Adjustment to A/R              -       FI
    •Reduces Revenue                -       FI
    SD Transaction Code Flow: 
    Inquiry / Document type IN
    Tcode for creation VA11,VA12,VA13. tables VBAK,VBAP
    Quotation / QT
    Tcode for creation VA21,VA22,VA23. tables VBAK,VBAP
    Sales Order OR
    Tcode for creation VA01,VA02,VA03. tables VBAK,VBAP
    Delivery LF
    Tcode for creation VL01,VL02,VL03. tables LIKP,LIPS
    Billing F2
    Tcode for creation VF01,VF02,VF03. tables VBRK,VBRP
    Regards,
    Rajesh Banka
    Reward suitable points.

  • Can I use an OLE DB Command Task to call a parameterized stored procedure, perform some data editing and pass variables back to the SSIS for handling?

    I am using a Data Flow and an OLE DB Source to read my staged 3rd party external data. I need to do various Lookups to try and determine if I can find the external person in our database...by SSN...By Name and DOB...etc...
    Now I need to do some more data verification based on the Lookup that is successful. Can I do those data edits against our SQL server application database by utilizing an OLE DB Command? Using a Stored Procedure or can I sue straight SQL to perform my edit
    against every staging row by using a parameter driven query? I'm thinking a Stored Procedure is the way to go here since I have multiple edits against the database. Can I pass back the result of those edits via a variable and then continue my SSIS Data Flow
    by analyzing the result of my Stored Procedure? And how would I do that.
    I am new to the SSIS game here so please be kind and as explicit as possible. If you know of any good web sites that walk through how to perform SQL server database edits against external data in SSIS or even a YouTube, please let me know.
    Thanks!

    Thanks for that...but can I do multiple edits in my Stored Procedure Vaibhav and pass back something that I can then utilize in my SSIS? For example...
    One and Only one Member Span...so I'd be doing a SELECT COUNT(*) based on my match criteria or handle the count accordingly in my Stored Procedure and passing something back via the OLE DB Command and handling it appropriately in SSIS
    Are there "Diabetes" claims...again probably by analyzing a SELECT COUNT(*)
    Am I expecting too much from the SSIS...should I be doing all of this in a Stored Procedure? I was hoping to use the SSIS GUI for everything but maybe that's just not possible. Rather use the Stored Procedure to analyze my stged data, edit accordingly, do
    data stores accordingly...especially the data anomalies...and then use the SSIS to control navigation
    Your thoughts........
    Could you maybe clarify the difference between an OLE DB Command on the Data Flow and the Execute SQL Task on the Control Flow...
    You can get return values from oledb comand if you want to pipeline.
    see this link for more details
    http://josef-richberg.squarespace.com/journal/2011/6/30/ssis-oledb-command-and-procedure-output-params.html
    The procedure should have an output parameter defined for that
    I belive if you've flexibility of using stored procedure you may be better off doing this in execute sql task in control flow. Calling sp in data flow will cause it to execute sp once for each row in dataset whereas in controlflow it will go for set based
    processing
    Please Mark This As Answer if it solved your issue
    Please Vote This As Helpful if it helps to solve your issue
    Visakh
    My Wiki User Page
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • LO cockpit extraction - data flow question

    This question is related with the LO cockpit extraction.
    Are update tables are meant to be the same  as set tables or
    are they different?
    can anybody provide a highlevel data flow for the logistics data
    using LO cockpit from R3 to BW
    please correct / enhance the below data flow for different methods of LO delta updates as i assume
    and inculude the set up table where it fits.
    (for direct update)
    R3 db talbes -> delta queue -> extract structure -> transfer structure -> infosource -> data target
    (for queued delta)
    R3 db talbes -> extraction queue -> delta queue -> extract structure -> transfer structure -> infosource -> data target
    (for unserialized V3 update)
    R3 db talbes -> postings to update tables -> delta queue -> extract structure -> transfer structure -> infosource -> data target
    I also want to know where does the delta queue reside - in R3 or in BW? Can the content of delta queue be viewed?
    Thanks for ur help - i will give full points for your input.
    ak

    Check the diagram in links below. It clearly identifies the data flow and whether the component is in R3 or BW.
    /people/sap.user72/blog/2005/01/19/logistic-cockpit-delta-mechanism--episode-three-the-new-update-methods
    /people/sap.user72/blog/2004/12/16/logistic-cockpit-delta-mechanism--episode-one-v3-update-the-145serializer146
    /people/sap.user72/blog/2004/12/23/logistic-cockpit-delta-mechanism--episode-two-v3-update-when-some-problems-can-occur

  • Data Models and Data Flow diagrams.

    Hi  Gurus,
        Can anybody brief me the concept of Data Models and Data Flow Diagrams and their development with illustrations. And is it a responsibility of a Technical or a Functional consultant..i.e to translate Business requirements and functional specifications into technical specifications, data flow diagrams and data models.
    Your valuable answers will be rewarded.
    Thanks in advance.

    Hi,
    Concept of Data Models
    Data model or Data modelling is basically how you define or design your BW Architecture based on Business requirements. It deals with designing and creating a effcient BW architecture sticking to standard practices.
    Multi-Dimensional Modeling with SAP NetWeaver BI
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/6ce7b0a4-0b01-0010-52ac-a6e813c35a84
    /people/githen.ronney3/blog/2008/02/13/modeling-strategies
    Modeling the Data Warehouse Layer with BI
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/3668618d-0c01-0010-1ab5-aa75c3a4dfc2
    /people/gilad.weinbach2/blog/2007/02/23/a-beginners-guide-to-your-first-bi-model-in-nw2004s
    Data Flow Diagrams
    This show the path of data flow for each individual object in BW. How data dets loaded into that object and how it is going out of the object etc.
    Right click on the data target > show data flow .
    It shows all the intermdeiate layer through which data comes into that particular object.
    Responsibility of a Technical or a Functional consultant
    This is done genrally in the designing phase itself by a Senior Technical Consultant with the help of a Functional consultant or a Techno=Functional consultant interacting with Business.
    Hope this helps.
    Thanks,
    JituK

  • SSIS 2005 and SSIS 2012 Have Different Data Flow Output

    Hi Guys,
    I'm currently converting our SSIS 2005 packages to 2012 and I'm finding an odd difference on the Data Flow Output when it imports a txt file to a table.
    We have a pipe delimited file ( contains a totals on the bottom row).
    CATNO|Title|Total Stk Avail|Stk on Order|Rank|QOH|Allocated|Reserved|BackOrd|WIP|Cleared|Held
    CATNUM000067|Fast and Furious|9804|0||9804|0|0|0|0|0|0
    CATNUM000068|Frozen|0|0||0|0|0|0|0|0|0
    CATNUM000029|Brave|0|0||0|0|0|0|0|0|0
    CATNUM000029|Brave with Kinect|0|0||0|0|0|0|0|0|0
    CATNUM000029|Big Hero|0|0||0|0|0|0|0|0|0
    CATNUM000057|Transformers|27376|0||27376|0|0|0|0|0|0
    CATNUM000013|James Bond|121|0||121|0|0|0|0|0|0
    ,Total Available Stock,524259
    And these are the settings for the flat file connection
    In SSIS 2005, the total rows are being omitted automatically by the data flow. Basically only 7 rows are inserted whereas in SSIS 2012. 8 rows were inserted including the totals.
    if we think about it, SSIS 2012 is correct since we just feed what's ever in the file.
    But on the conversion and testing perspective, it's now harder as there's discrepancies between those 2 version.
    I just like to check if there's any other discrepancies in terms on results/output between SSIS 2005 and SSIS 2012
    Many thanks. 

    Hi Vinxster,
    After testing the issue that Flat File Connection Manager cannot handle file with total row in my SQL Server 2005 environment, I can reproduce it.
    Based on my research, the issue is caused by the Column Delimiter gets first preference and then Row delimiter. This is by design in SQL Server 2005, SQL Server 2008 or SQL Server 2008 R2.
    Good news is that the issue is fixed in SQL Server Data Tools which comes with SQL Server 2012. In SQL Server 2012, by default, the Flat File Connection Manager always checks for a row delimiter in unquoted data, and starts a new row when a row delimiter
    is found. This enables the connection manager to correctly parse files with rows that are missing column fields.
    If you want to fix the issue in SQL Server 2005, there is a sample component posted to the CodePlex that might help you with this:
    http://ssisdfs.codeplex.com/
    References:
    http://blogs.msdn.com/b/dataaccesstechnologies/archive/2013/03/13/flat-file-source-cannot-handle-file-with-uneven-number-of-columns-in-each-row.aspx
    https://connect.microsoft.com/SQLServer/feedback/details/293193/ssis-import-of-flat-file-with-uneven-number-of-columns
    Thanks,
    Katherine Xiong
    If you have any feedback on our support, please click
    here.
    Katherine Xiong
    TechNet Community Support

Maybe you are looking for

  • Multiple windows open up in one browser. I drag and dropped websites to the taskbar and now can't get rid of it.

    I thought it would be convenient for 3 windows to open automatically in one browser. It slows my laptop down and I want to be able to click on Firefox and have just one window open. Any ideas how to fix?

  • Payment to  Customers via automatic payment run (Just like suppliers)

    Hi all, I do have a client requirement which is described below.  Thank you for reading my post and suggesting any solution. Customers made various payments in the past and those payments were cleared against invoices.  During this process which cont

  • UCCX script and abandoned calls

    Hi Guys I got question about that UCCX script and abandoned calls , so currently we had 4 different script for  support the call follow . So the Main script is just AA let the people select menu 1 , 2 ,3 then go to different queue , after select 1 (f

  • Maximum SQL statement on ODBC 8176 or higher ?

    Hello. I am developing a DB reporting program using ODBC. My customer complained that the program shut down when query the SQL statement. But, my computer didn't shut down when I queried the same SQL statement. The difference is my customer used 8177

  • Cross system MRP

    Hi, I am trying to integrate a ECC and a 4.0 system. with some components in ECC being sourced from plants on the 4.0 system and vice versa. I know we can setup dummy plants for 4.0 in ECC and generate the PO/SO via ALE interface tp get these compone