Data flow into data source

Hello,
Is there any nice documentation which describes exactly how the data flows from the source R/3 table all the way to the data source? Is this very different from ds to ds or is there a common logic?
As I understand it, a function module fetches the data from the data source when being called upon from info package in BW... But how do I trace the flow all the way back to the source tables in R/3?
Best regards,
Fredrik

Hi Fredrik,
Yes. The data flow for the datasources would be different. You can refer to this blog by Roberto:
/people/sap.user72/blog/2005/09/05/sap-bw-and-business-content-datasources-in-pursuit-of-the-origins
Bye
Dinesh

Similar Messages

  • Error: 0xC020902A at Data Flow Task, XML Source [24515]: The "component "XML Source" (24515)" failed because truncation occurred, and the truncation row disposition on "output column "MsgLev1" (26196)" specifies failure on truncation. A truncation error o

    When I was Importing data from XML to SqlServer using SSIS , I am getting this error. The import is working if i use small file and not working if I use large XMl file. Can any one of you guys help me out with the issue?
    Error: 0xC020902A at Data Flow Task, XML Source [24515]: The "component "XML Source" (24515)" failed because truncation occurred, and the truncation row disposition on "output column "MsgLev1" (26196)" specifies failure on truncation. A truncation error occurred on the specified object of the specified component.
    Error: 0xC02092AF at Data Flow Task, XML Source [24515]: The component "XML Source" (24515) was unable to process the XML data. Pipeline component has returned HRESULT error code 0xC020902A from a method call.
    Error: 0xC0047038 at Data Flow Task: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on component "XML Source" (24515) returned error code 0xC02092AF.  The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "SourceThread0" has exited with error code 0xC0047038.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread0" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread1" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread3" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread4" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047039 at Data Flow Task: SSIS Error Code DTS_E_THREADCANCELLED.  Thread "WorkThread2" received a shutdown signal and is terminating. The user requested a shutdown, or an error in another thread is causing the pipeline to shutdown.  There may be error messages posted before this with more information on why the thread was cancelled.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread0" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread1" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread2" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread3" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.
    Error: 0xC0047021 at Data Flow Task: SSIS Error Code DTS_E_THREADFAILED.  Thread "WorkThread4" has exited with error code 0xC0047039.  There may be error messages posted before this with more information on why the thread has exited.

    The reason is in the first line of the error.  It doesn't have anything to do with the size of your XML file - it has to do with the contents of the "MsgLev1" column.  You (or the XSD) has indicated that the values in this column not exceed a certain size - but one of the values in the file that's failing is larger than that.
    In order to fix the problem, you're going to need to increase the allocated space for that column.  In order to do that, you're going to need to find out what the required size of that data is - either from someone who ought to know, or by direct examination of the file.  If you want to know which entity has this overly large data element in it, you need to configure the error handling of the XML Source to "redirect truncation errors".  Then you can hook up the XML Source's error output to another destination where you can see which rows are problematic.
    Talk to me on

  • Message mapping (data flow editor, data dependcies  buttons not visible)

    Hi,
    In my SAP PI...IR ...in message mapping screen some buttons are not enable/visible like
    data flow editor, data dependcies,clear mapping  buttons not visible)
    any oss notes/help will be appreciated.
    Thanks

    Hi,
    This will help you
    /people/shabarish.vijayakumar/blog/2006/02/13/unable-to-open-iresrid-xipipi-71-updated-for-pi-71-support
    Go for 1.4.X version than 1.6.X The Problem will be resolved
    Regards
    Sangeetha

  • Data Flow from SAP Source (ECC) system to SAP BI system

    Hi All,
    I wanted to know how data will be flown from SAP Source system to SAP BI system.Data flow should include
    1) Data will be flown by using the IDOCs?
    2) What all are the interfaces involved while data is transferring?
    3) What will happen exactly, if you execute the PSA?.
    If you have any info on this, could you please post here....I
    Regards,
    K.Krishna Chaitanya.

    Hi Krishna,
    Please go through  this article :
    "http://www.trinay.com/C6747810-561C-4ED6-B85C-8F32CF901602/FinalDownload/DownloadId-C2EB7035A229BFC0BB16C09174241DC8/C6747810-561C-4ED6-B85C-8F32CF901602/SAP%20BW%20Extraction.pdf".
    Hope this answers all the mentioned questions.
    Regards,
    Sarika

  • Mixing of 3.x data flow into 7.0 flow

    There is an Info Object with Master data and already used in different cubes.
    The whole for extracting Master data into this object and also loading to the cubes is also in 3.x flow.
    There is a requirement for adding new attributes to this Info Object, and for these attributes I am loading text data from R/3 table using Generic extraction for texts.
    If I load the text data to these attributes in 7.0 flow and when these objects are added to already existing InfoObject whose data flow is in 3.x, will it affect performance or generate any errors.
    Will any compatibility errors occurs during Process chains schedule or attribute change runs?

    Hi Akshara,
    There shouldn't be anyproblem, you can use 3.x and 7.0 flow for a same object.
    for example:- you can load a master data info object with 3.x flow and also 7.0 flow.
    i did this in one of my previous assignments....we had a masterdata object (created in 3.x)which has been used across the system..like, in cubes, DSOs etc.. 
    Later on, the system has been upgarded to 7.0 and we had to add few new attributes to add to the same master data infoobject. we added those new attributes and created a 7.0 flow(transformations and DTPs) inorder to load the newly added attributes.
    There has been absoluetly no problem till date....this has been implemented more than an year ago.
    your case is pretty simple when comapred to mine.....you are just loading the text data for the new attributes using a 7.0 flow.
    Bottom line:  you can use 3.x and 7.0 flow for a same object simultaneously.
    Regards,
    Sudheer.

  • How can i convert this data(00000000) into date format(yyyy-MM-dd)

    Hi,
    i am using this method to convert date format from string.
    public static string FormatDate(string inputDate, string inputFormat)
                System.DateTime date = DateTime.ParseExact(inputDate, inputFormat, null);
                string datepresent = DateTime.Now.Date.ToString("yyyy-MM-dd");
                return datepresent;
    its working properly,
    but when input data is like 00000000 i am getting error this is not valid.
    how can i convert this into date format any help..!

    Have you tried the above code:
    I can see it is working with both Date and DateTime destination nodes.
    Map:
    Functoid Parameters:
    Functoid Script:
    public static string FormatDate(string inputDate, string inputFormat)
    string datepresent;
    if (inputDate == "00000000")
    datepresent = "0000-00-00";
    else
    System.DateTime date = DateTime.ParseExact(inputDate, inputFormat, null);
    datepresent = date.ToString("yyyy-MM-dd");
    return datepresent;
    Input:
    <ns0:InputDateString xmlns:ns0="http://DateFormat.SourceSchema">
    <DateString>00000000</DateString>
    </ns0:InputDateString>
    Output:
    <ns0:OutputDate xmlns:ns0="http://DateFormat.DestinationSchema">
    <Date>0000-00-00</Date>
    </ns0:OutputDate>
    If this answers your question please mark as answer. If this post is helpful, please vote as helpful.

  • Importing data tables into data tablespace and indexes into tablespaces

    Hi
    I want to import data into new schema and i want to store tables into data tablespaces and index into index tablespace ...can anyone tell me how it will possible...

    I want to import data into new schema and i want to store tables into data tablespaces and index into index tablespace ...can anyone tell me how it will possible...
    imp userid=/user/passwd show=y indexfile=import.sql indexes=n full=y
    imp userid=/user/passwd show=y indexfile=import2.sql full=y
    Edit the import.sql and import2.sql to modify the tables' tablespace and indexes tablespace.
    execute import.sql the script in the database. this will create the tables in their respective tablespace.
    imp userid=/user/passwd full=y ignore=y indexes=n constraints=y - to import just the data since the tables have already been created.
    imp userid=/user/passwd full=y ignore=y rows=n  - to import just the indexes since the tables and data have already been imported.

  • Need to know how to check the outbound queue in SAP for data flow into CRM

    Hi Specialist's,
    I have changed an entry in Customer Master Data in Sap r/3 and this change has come fine to CRM system.I checked the inbound BDOC in CRM ( Trn:SMW01) and it has come fine into CRM system.I am also able to see the affected field in CRM with new value.
    I need to know how to check this from SAP side i.e. I have checked the outbound IDOC list (WE02) but could not find the record.
    Is there any other way in SAP that i can check?
    Also please let me know if there is any other way except IDOC in SAP to send data from R/3 to CRM?
    Usefull answers will definitely be rewarded with points .
    Thanks,
    Abhinav.

    Hi Abhinav,
    There are no IDOCs generated in R/3 for replication. R/3 system calls BAPI and Function modules remotely on CRM system which are part of the Adaptors provided for middleware.
    Try to deregister the outbound queues in R/3 (Transaction SMQS) and see if you get an entry in the Outbound Queue (Transaction SMQ1). I have not tried this.
    If you don't get any entry in the outbound queue then definitely there would be some FM at R/3 end which would be calling CRM system remotely.
    <b>Reward points if it helps.</b>
    Regards,
    Amit Mishra

  • Data flowing into incorrect CC when triggered from a SAP - RFC

    hi Experts,
    I have a RFC to SOAP scenario. Data is triggered from SAP through RFC. Initially, this was done from client 200. Now for the same system there is one more client 210 from which data needs to be triggered.
    I completed the configuration. But when data was triggered from client 210, the communication channel of the client 200 is showing the following error.
    Error in processing caused by: com.sap.aii.adapter.rfc.afcommunication.RfcAFWException: alternativeServiceIdentifier: party/service from channel configuration are not equal to party/service from lookup of alternativeServiceIdentifier (CPA party/service/channel ''/'ECQCLNT200'/'CC_SAP_RFCEproPOSender', lookup of alternativeServiceIdentifier for party '', schema 'TechnicalSystem' with 'ECQ#210' returned party/service of ''/'ECQCLNT210')
    The client 210 communication channel is not showing any message. Also no messages are seen in moni.
    Waiting for your replies.

    Thank you for your reply everyone.
    I cleared the cache and now am getting the following error in the communication channel of client 210.
    Error in processing caused by: com.sap.aii.adapter.rfc.afcommunication.RfcAFWException: alternativeServiceIdentifier: party/service from channel configuration are not equal to party/service from lookup of alternativeServiceIdentifier (CPA party/service/channel ''/'ECQCLNT200'/'CC_SAP_RFCEproPOSender', lookup of alternativeServiceIdentifier for party '', schema 'TechnicalSystem' with 'ECQ#210' returned party/service of ''/'ECQCLNT210')
    I have created the communication channel under the bussiness system of client 210 and also have created the receiver agreements and sender agreement specific for the client 210. Still getting the error.
    Need your help in solving this.

  • Problem in data flow into other pages in TDS SAP Script

    Hi Gurus,
    I have solved all my problems which i had faced through the creation of the TDS Certificate Layout & programs. But blocked with one problem, which is when there are more line items to be printed and which are moved to the second page. Now for Example in my first page if there are 10 line items printed & then moved to the second page only 11 items are printed and again in third page 11 items are printed, where in my page's are wasted when we take a printout. what could be the problem where i am notable to fix. Please Suggest me what can i do to solve this problem.
    Regards
    Syama

    Hi,
    Use PROTECT -
    ENDPROTECT statements.
    Regards,
    Jyothi CH.

  • Data loading into Data warehouse Tables

    Hi all,
    I am having scenario like this. I am having table with lacks of records. Every day, new data will come with some records DELETE ed, some new records
    INSERT ed, some records UPDATE ed. I have to load this new data into parent table.
    Currently, we are DELETING all parent data and inserting new data. It is taking a lot of time. There is NO PRIMARY KEY on this parent
    table.
    What are the best options to load the new data. Actually, there are several tables are there. I have mentioned one table for example.
    Oracle version is 9i.
    Thanks,
    Pal

    I need to know which is a better way of loading flat
    files data into Warehouse tables. SQL*Loader or
    Oracle DataMart Suite. Doesn't make much odds. The advantage of using the
    Delimited or Fixed Length transform is that it keeps
    everything within Data Mart Suite.
    1. What all types of databases are supported by DataMart
    ODBC. DataMart Suite will support most RDBMS's as an
    ODBC source. Sybase, SQL Server, Informix, DB2, Redbrick,
    Teradata. Note this is only a source for the
    data, not a store for the DataMart
    2. Can DataMart suite be programmed for auto operation of
    loading data.There is some limited scheduling functionality based on
    the NT scheduler service. Where you want to run one plan
    following another, you may need to get a bit clever, by
    including some preprocessing to check that the previous
    plan has completed or use a third party scheduler which
    would not start the plan until the previous plan had
    completed.
    3. What is the extend of errors handling in loading operation,
    supported by this suite. Pretty good. All the stuff from SQL Loader, plus each of the
    transforms has it's own error handling.
    ===
    Manish Jain (guest) wrote:
    : Hello,
    : I need to know which is a better way of loading flat
    : files data into
    : Warehouse tables. SQL*Loader or Oracle DataMart Suite.
    : 1. What all types of databases are supported by DataMart ODBC.
    : 2. Can DataMart suite be programmed for auto operation of
    : loading data.
    : 3. What is the extend of errors handling in loading operation,
    : supported by this suite.
    : regards,
    : Manish.
    Oracle Technology Network
    http://technet.oracle.com
    null

  • Data Flow Migration without Infosource

    Hello All...
    I have installed a BI Content cube with the data flow. The data flow is in 3.5 ie, with transfer rules and update rules. I want to migrate the data flow into transformation and DTP. Is there any way I can convert the entire data flow into 7.0 dataflow without the info source. ie, I want only one transformation directly from Datasource to my cube. Is that possible using any tool, or i shud do that manually???
    Regards,
    BIJESH

    Hi,
    Have you migrated the update rule ?
    Once you have migrated the update rule, you will find the option of copying the transformation
    when u right click on the transformation.
    At that time you provide the source and target for the transformation.
    You can find the steps for migration in the link mentioned below.
    [Business Content Activation Errors - Transformations;
    I hope this helps.
    Thanks,
    Kartik

  • Error on Data Flow Task MSSQL 2012 Clustered "Description: The version of Lookup is not compatible with this version of the DataFlow. "

    We have an SSIS package that runs on clustered MSSQL 2012 Enterprise Nodes that is failing.  We use a job to executer the package.
    Environmental information:
    Product - Microsoft SQL Server Enterprise: Core-based Licensing (64-bit)
    Operating System - Microsoft Windows NT 6.1 (7601)
    Patform - NT x64
    Version - MSSQL Version 11.0.3349.0
    Package is set to 32 -bit.  All permissions verified.  Runs in lower environments, same MSSQL version.  All environments are clustered.  In the failing environment, all nodes are at the same service pack.  I have not verified if all
    nodes in the failing environment have SSIS installed.  Data access is installed.  We have other simpler packages that run in this environment, just not this one.  Time to ask the community for help!
    Error:
    Source: Data Flow Task - Data Flow Task (SSIS.Pipeline)     Description: The version of Lookup is not compatible with this version of the DataFlow.  End Error  Error:  Code: 0xC0048020    
    Description: Component "Conditional Split, clsid {7F88F654-4E20-4D14-84F4-AF9C925D3087}" could not be created and returned error code 0x80070005 "Access is denied.". Make sure that the component is registered correctly.  End Error 
    Description: The component is missing, not registered, not upgradeable, or missing required interfaces. The contact information for this component is "Conditional Split;Microsoft Corporation; Microsoft SQL Server; (C) Microsoft Corporation; All Rights
    Reserved; http://www.microsoft.com/sql/support;0".  End Error 
    (Left out shop specific information.  This is the first error in the errors returns by the job history for this package. )
    Thanks in advance.

    Hi DeveloperMax,
    According to your description, the error occurs when you execute the package with Agent job on clustered MSSQL 2012 Enterprise Nodes.
    As per my understanding, I think this issue can be caused by you use SQL Server Agent to schedule a SQL Server Integration Services package in a 64-bit environment. And the SSIS package is referencing some 32-Bit DLL or 32-Bit drivers which are available
    only in 32-bit versions, so the job failed.
    To fix this issue, we should use the 32-bit version of the DTExec.exe utility to schedule the 64-bit SQL Server Agent to run a package. To run a package in 32-bit mode from a 64-bit version of SQL Server Agent, we can go to the Job Step dialog box, then
    select “32 bit runtime” in the Advanced tab.
    Besides, we should make sure that SQL Server Integration Services is installed on the failing environment.
    If there are any other questions, please feel free to ask.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • XML data inserting into Master, not Page

    First, I fully admit that I jumped into the deep end of the pool.  This is my first InDesign project and I'm trying to set up a 2-page spread master to support loading my content from XML.
    Thanks to the great online help, these forums/community resources, and my growing bookshelf, I've successfully (or so it seems) created my Master layout, with frames appropriately tagged for my XML (as confirmed by the Structure).  Since I'm very novice at InDesign, my testing isn't always unambiguous, but I HAVE successfully loaded my external XML (including graphics and anchored text, plus the story) into my document.
    The problem is that the data seems to be flowing into the Master, and not the Page.  So if I create all the pages I need ahead of time (right now just 2 spreads for my test), both pages are populated with the first record.
    If I only create the first spread, then the data flows and I get the magic "+" for overflow, but I can't click on it in the Page.  If I change to the Master I can click on it, which seems the major clue as to why this isn't working.  I just have no idea how to fix it.
    From everything I've read it seems that I should be able to tag my frames in the Master (which I want to reuse for other documents) and have the data flow into the Pages.
    Am I missing something obvious? 
    Or is what I'm trying to do just not possible?  (if so, does that mean I have to keep re-tagging the frames every time I make a new Document?  ugh)
    I've spent hours trying to resolve this, and would really appreciate a pointer to get me moving again.
    Thanks!
    julie

    Hm, that seems contradictory to everything I've read and the examples I've seen  (although I may have misunderstood the examples - none are quite as complicated as my layout).  My text frame has several anchors for XML data that doesn't just flow with the story.  How do I set up the tagged anchor frames I need if the story text frame is in the Master but not tagged?
    I thought that the story text frame and the anchor frames (some graphics and some text) all needed to be in the same context.  If I put them in the Master I get the XML flowing into the Master.  If I put them into the Page then the new pages created don't have the proper frames.  If I just put the story text frame in the Master, how do I associate it with the anchor frames (I only know how to do this by using the Story editor on the tagged frame as described in A Designer's Guide To Adobe InDesign and XML)?
    I'll go back and work through the samples again, but I definitely don't understand how it all fits together.
    Thanks!
    j

  • How data flow when SSIS packages are run on a different server than the DB server

    The scenario is that i have a dedicated SQL SErver 2014 SSIS machine that executes the packages.
    The database server is a separate machine with SQL Server 2008R2.
    1) Running SSIS packages that transfer data within SQL Server 2008R2 (same machine)
    2) Running SSIS packages that transfer data between 2 separate SQL Server servers.
    How the data flow in these two cases and what resource is being used where? (cpu,disk,ram,network)
    Elias

    When you have a dedicated SSIS server, all data read flows to that server, is processed using the resources of that ETL server and then sent back over the network to the destination server.
    It doesn't matter if source and destination are the same server. If you use a data flow, all data flows over the network twice.
    The only exception is when you don't use a data flow, but only SQL statements. In that case, data flows only between source and destination.
    MCSE SQL Server 2012 - Please mark posts as answered where appropriate.

Maybe you are looking for