Mapping data flow tracking?

We are using OWB repository 10.2.0.1.0 and OWB client 10.2.0.1.31. The Oracle version is 10 G (10.2.0.3.0). OWB is installed on Linux.
We have an existing mapping. Do we able track column from source to target or target to source in Mapping?
Suppose we have employee_salary in source. It goed through number of transforms before reaching the target column sal_measure. If we want to track the complete dataflow of employee_salary to sal_measure or sal_measure to employee_salary. Are there any way in OWB mapping editor?
Thanks in advance.
RI

Dear David,
Thanks for your reply. I have issue with Ipact piece as well.
Here is the posted link.
Impact Analysis results error ??
Thanks in advance.
RI

Similar Messages

  • Mapping data flow from R/3 to BW

    Hello,
    I am pretty new to BW and I have been tasked with creating a detailed map of the data flow from R/3 into BW. 
    I need to record where the data originates from in R/3 (field names/tables) and literally track the flow of that data all the way including any info objects along the way to any cubes that it may be sitting in.
    How do I track this flow ? And how can I identify what a characteristic in BW is in R/3 ?
    Has anybody had to create a similar data flow ? If so how did you approach this ?
    Many Thanks,
    Matt

    Hi Matthew,
    From the R/3 side:
    BW treats all the data from R/3 as Datasources.
    From the Datasource the upload of data to the cube is done as..        
    <b>Datasource->Transfer Rule->psa/infosource->communication structure->cube</b>
    (for a 3.5 system)
    in case of 7.0 system... data flow is as follows...
    <b>Datasource->infopackage->psa->transformation/DTP-> Data target(cube)</b>
    -> Go to transaction <b>RSA5</b>( for Business Content datasources ) and <b>RSA6</b>( for all the active Datasources ) found in the system.
    -> There you can find all the data that you want...(For your mapping purpose this will do..)
    -> You can as well check from the BI side in the transaction RSA1 -> click on the Monitor button on the left ( for custom objects ) or Business Content button -> choose the object from the tree... right click and replicate to find if all of them were used.
    Hope this helps!!
    <b>*</b><i>Reward Pts if useful</i><b>*</b>
    regards,
    Naveenan.

  • Message mapping (data flow editor, data dependcies  buttons not visible)

    Hi,
    In my SAP PI...IR ...in message mapping screen some buttons are not enable/visible like
    data flow editor, data dependcies,clear mapping  buttons not visible)
    any oss notes/help will be appreciated.
    Thanks

    Hi,
    This will help you
    /people/shabarish.vijayakumar/blog/2006/02/13/unable-to-open-iresrid-xipipi-71-updated-for-pi-71-support
    Go for 1.4.X version than 1.6.X The Problem will be resolved
    Regards
    Sangeetha

  • Problem with context mapping and data flow in a FPM application

    Hi All,
    I am trying to develop an ESS application using FPM. For the same, the requirement is to see the history of an employee in the second view.
    The first view has got just the overview information and the second one has got the detail. So, the records or the fields are the same on both the views.
    As per the FPM guidelines, the Model is residing in the Fc component and the respective Vc components are using the model data accordingly.
    I am executing the model in the Fc component calling the executable method in the interfaceController of the first view and then trying to display the output data of the BAPI in the first view which provides the overview information.This is working fine.
    But when i am trying to map the same output node to the Table UI for the second view, the record size is coming zero and thus no information is available.
    For the above issue, I am again executing the RFC in the InterfaceController of the second view to populate the records, which is incorrect as it is already executed and the data is available for the first view.
    I request you to let me know the correct approach to Context mapping and data flow when using FPM-roadmap. Is their any standard method or approach available to deal with such requirements? Please let me know.
    Thanks in advance.
    Regards
    DK

    Hi Idhaya,
    I model node is available in Fc and the Fc interface controller is being used in the first Vc and the second Vc.
    So the idea is, as the executable method is generated in the Fc, so i have created a custom method to call the executable method in Fc, where the input parameter is getting passed and this custom method is finally getting called is the first Vc.
    So , now my first Vc is ready to call the custom method in Fc and execute the RFC. Once the RFC is executed, the nodes in the Fc should get populated which is the ideal case.
    And as the Fc is used as a component in the second Vc, the same node is available to the UI elements.
    But, when I check the record size for the output node, it is always zero, for the second Vc.
    Regards
    DK

  • Owb3i data flow connectors problem in mapping editor

    Hi,
    To define mappings between source and target, the manual says:
    steps 1-3 and then,
    "Repeat steps one through three until you have created all the data flow
    connection appropriate for your situation."
    The manual also says:
    "To connect Mapping Operators, you draw lines from output attributes or output
    attribute groups to input attributes or groups between the operators."
    The question:
    When I draw a lines for individual connections, it is fine. But I have a source/target with 201 columns. So I dragged my mouse between the groups. This created 201 additional attributes in the target. Am I doing it wrong or is it a 'feature'? How else I can make all the 201 connections less painfully?
    TIA.
    --Rao.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

    sunil kumar wrote:
    Hi,
    >
    > As said above convert transfer rules to transformations (right click on transfer rules -> additional prop -> create transformation)
    >
    > Then right click on datasource -> Migrate
    >
    > After doing this, your datasource will be linked to Infocube directly with one transformations automatically.
    >
    > Regards,
    > Sunil
    Hi Sunil & Sravan,
    THanks for ur quick reply.As you said click on transfer rules and create transformation.As i can see in my system i have a transformation from data source to infosource .i.e. starting with RSDS_  and then another transformation from info source to info cube.i.e. starting with TRCS_  .  what i did was created a transformation by right click on info cube and by giving data source name as 2LIS_03_BF .this is a manual transformation i am mapping manually.Is this a correct way? Please tell me how to do it.
    as i have to do this immediately.
    Thanks in advance

  • Losing source-target mapping from a data flow diagram process

    Hi,
    When I reload a model design from an xml repository, I am losing source-target mappings from a data flow diagram process. I wonder why.
    The way it goes is:
    - I define a data flow diagram with information stores, processes, and flows.
    - Information structures are defined from the flows and the information trickles down to processes and information stores.
    - In a process, I define Source-Target Mapping by selecting a target element and then selecting the source element(s).
    - I save the whole model.
    - So far, the Source-Target Mapping keeps the association between target and source elements.
    - If I exit from Data Modeler, the next time I reload the saved model, the Source-Target Mapping target to source elements associations are lost!?
    Does anyone have the same issue or would know how to resolve it?
    Thanks,
    /Mario

    I am still having problems getting my source to target mappings to save after I apply the fix described in this thread. Any thoughts?
    From my TargetElement.xml
    <?xml version = '1.0' encoding = 'UTF-8'?>
    <oracle.dbtools.crest.model.design.process.TargetElement parentClass="java.lang.Object">
    <property name="objectID" dataType="java.lang.String" readOnly="true" defaultValue="" xmlName="id" xmlType="attribute" getter="getObjectID" setter="setObjectID" reffered="false"/>
    <property name="target" dataType="oracle.dbtools.crest.model.design.DesignObject" defaultValue="" xmlName="target" xmlType="element" getter="getTarget" setter="setTarget" reffered="true"/>
    <property name="description" dataType="java.lang.String" defaultValue="" xmlName="description" xmlType="element" getter="getDescription" setter="setDescription" reffered="false"/>
    <property name="name" dataType="java.lang.String" readOnly="true" defaultValue="" xmlName="name" xmlType="attribute" getter="getName" setter="setName" reffered="false"/>
    <property name="transformationType" dataType="java.lang.String" defaultValue="" xmlName="transformationType" xmlType="element" getter="getTransformationType" setter="setTransformationType" reffered="false"/>
    <collection name="sources" dataType="java.util.List" xmlName="sources" itemXmlName="sourceRef" getter="getSources" itemIsRef="true" refsXMLName="sourceRefs" addItem="addSource" getItemByID="getSourceByID" nativeList="true" orderTheList="true"/>
    <roproperty name="sources" dataType="interface java.util.List" getter="getSources"/>
    </oracle.dbtools.crest.model.design.process.TargetElement>
    Process.xml:
    <?xml version="1.0" encoding="UTF-8" ?>
    - <oracle.dbtools.crest.model.design.process.Process parentClass="oracle.dbtools.crest.model.design.process.Transformation">
    <property name="type" dataType="int" defaultValue="0" xmlName="type" xmlType="element" getter="getType" setter="setType" reffered="false" />
    <property name="batchMinimumTransactions" dataType="int" defaultValue="" xmlName="batchMinimumTransactions" xmlType="element" getter="getBatchMinimumTransactions" setter="setBatchMinimumTransactions" reffered="false" />
    <property name="batchTimeUnit" dataType="java.lang.String" defaultValue="" xmlName="batchTimeUnit" xmlType="element" getter="getBatchTimeUnit" setter="setBatchTimeUnit" reffered="false" />
    <property name="diagram" dataType="oracle.dbtools.crest.model.design.process.DataFlowDesign" defaultValue="" xmlName="diagram" xmlType="element" getter="getDiagram" setter="setDiagram" reffered="true" />
    <property name="footNote" dataType="java.lang.String" defaultValue="" xmlName="footNote" xmlType="element" getter="getFootNote" setter="setFootNote" reffered="false" />
    <property name="frequencyTimeUnit" dataType="java.lang.String" defaultValue="" xmlName="frequencyTimeUnit" xmlType="element" getter="getFrequencyTimeUnit" setter="setFrequencyTimeUnit" reffered="false" />
    <property name="frequencyTimes" dataType="int" defaultValue="" xmlName="frequencyTimes" xmlType="element" getter="getFrequencyTimes" setter="setFrequencyTimes" reffered="false" />
    <property name="interactiveLongestResponseTime" dataType="int" defaultValue="" xmlName="interactiveLongestResponseTime" xmlType="element" getter="getInteractiveLongestResponseTime" setter="setInteractiveLongestResponseTime" reffered="false" />
    <property name="interactiveResponseTimeUnit" dataType="java.lang.String" defaultValue="" xmlName="interactiveResponseTimeUnit" xmlType="element" getter="getInteractiveResponseTimeUnit" setter="setInteractiveResponseTimeUnit" reffered="false" />
    <property name="mode" dataType="java.lang.String" defaultValue="" xmlName="mode" xmlType="element" getter="getMode" setter="setMode" reffered="false" />
    <property name="paramsWrappersString" dataType="java.lang.String" defaultValue="" xmlName="paramsWrappersString" xmlType="element" getter="getParamsWrappersString" setter="setParamsWrappersString" reffered="false" />
    <property name="priority" dataType="java.lang.String" defaultValue="" xmlName="priority" xmlType="element" getter="getPriority" setter="setPriority" reffered="false" />
    <property name="transformationtask" dataType="oracle.dbtools.crest.model.design.process.TransformationTask" defaultValue="" xmlName="transformationtask" xmlType="element" getter="getTransformationTask" setter="setTransformationTask" reffered="true" />
    <property name="peakPeriodsString" dataType="java.lang.String" defaultValue="" xmlName="peakPeriodsString" xmlType="element" getter="getPeakPeriodsString" setter="setPeakPeriodsString" reffered="false" />
    <collection name="events" dataType="java.util.List" xmlName="events" itemXmlName="eventRef" getter="getEvents" itemIsRef="true" refsXMLName="eventRefs" removeItem="removeEvent" addItem="addEvent" getItemByID="getPartyByID" nativeList="true" orderTheList="true" />
    <collection name="targetMappings" dataType="java.util.List" xmlName="targetMappings" itemXmlName="targetMapping" getter="getTargetElements" createItem="createTargetElement" removeItem="removeEvent" addItem="addTargetElement" getItemByID="getPartyByID" nativeList="true" orderTheList="true" />
    <roproperty name="componentFlowsFor" dataType="interface java.util.Collection" getter="getComponentFlowsFor" />
    <roproperty name="events" dataType="interface java.util.List" getter="getEvents" />
    <roproperty name="flowEvents" dataType="interface java.util.List" getter="getFlowEvents" />
    <roproperty name="paramMapping" dataType="oracle.dbtools.crest.model.design.DesignObject" getter="getParamMapping" />
    <roproperty name="paramsWrappers" dataType="interface java.util.List" getter="getParamsWrappers" />
    <roproperty name="parentprocess" dataType="oracle.dbtools.crest.model.design.process.Process" getter="getParentProcess" />
    <roproperty name="parentprocessInDFD" dataType="oracle.dbtools.crest.model.design.process.Process" getter="getParentProcessInDFD" />
    <roproperty name="peakperiods" dataType="[Z" getter="getPeakPeriods" />
      <roproperty name="processedAttributesAndFlows" dataType="interface java.util.List" getter="getProcessedAttributesAndFlows" />
      <roproperty name="processedEntitiesAndFlows" dataType="interface java.util.List" getter="getProcessedEntitiesAndFlows" />
      </oracle.dbtools.crest.model.design.process.Process>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Data flow fails on packed decimal field moving iSeries DB2 data from one iSeries DB to another

    I' trying to use SSIS to move table content from one iSeries DB2 database to another.  I'm using the .Net providers for OleDb\IBM DB2 for i5/OS IBMDA400 OLE DB Provider in the connection managers for the source and destination and the test connection
    works fine.  When I try to run the data flow task however it fails on the first packed decimal field it encounters with the exceptions ...
    [select from hydro520 hydroweb2 blpmstr [16]] Error: SSIS Error Code DTS_E_INDUCEDTRANSFORMFAILUREONERROR.  The "component "select from hydro520 hydroweb2 blpmstr" (16)" failed because error code 0x80004002 occurred, and the error
    row disposition on "output column "MSPRIB" (55)" specifies failure on error. An error occurred on the specified object of the specified component.  There may be error messages posted before this with more information about the failure.
    [select from hydro520 hydroweb2 blpmstr [16]] Error: The component "select from hydro520 hydroweb2 blpmstr" (16) was unable to process the data. Pipeline component has returned HRESULT error code 0xC0209029 from a method call.
    [SSIS.Pipeline] Error: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on component "select from hydro520 hydroweb2 blpmstr" (16) returned error code 0xC02090F5.  The component returned a failure code when the pipeline
    engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    ...in the progress tab.  Can someone kindly tell me what I need to do to get the connection manager to work with DB2 packed decimal fields?  Or is it a different issue all together?  Thanks tonnes for any help, Roscoe

    Hi rpfinn,
    From the Data Types mapping rules between SSIS and DB2, we can see that both the NUMERIC and DECIMAL data types in DB2 are mapped to DT_NUMBERIC data type in SSIS. Now that the source data in your DB2 database is NUMERIC data type, changing the DT_NUMBRIC
    data type to DT_DECIMAL is invalid. Besides, if we check the data types of the target External column and Output column from the Advanced Editor for ADO NET Source adapter, the data type should be defined as DT_NUMERIC with Precision as 9 and Scale as 2. I
    am not clear where you see the DT_NUMBERIC(9,0) e.g. DT_NUMERIC with Precision as 9 and Scale as 0, but it may be the cause of the issue. You need to make sure the DT_NUMERIC data type also has Scale 2 instead of 0.
    If you don’t know how to modify the data type, please elaborate the Data Flow Task of the package so that we can make further analysis. Besides, the error messages you posted are incomplete, it will be helpful if you post the complete error message.
    Regards,
    Mike Yin
    If you have any feedback on our support, please click
    here
    Mike Yin
    TechNet Community Support

  • How to derive month/year from date in SAP BW 3.5 data flow

    Hi
    How we can derive cal year/month and fiscal month/year from date in SAP BW 3.5 data flow (we're using transfer and update rule)..
    Thanks,
    PK

    Hi,
    if you have any date filed in source side you can just map to any time char system will automatically convert to target objects.
    please look at the screen shot for understanding. (not 3.x it is 7.x)
    Thanks,
    Phani.

  • Read from sql task and send to data flow task - [OLE DB Source [1]] Error: A rowset based on the SQL command was not returned by the OLE DB provider.

    I have created a execut sql task -
    In that, i have a created a 'empidvar' variable of string type and put sqlstatement = 'select distinct empid from emp'
    Resultset=resultname=0 and variablename=empidvar
    I have added data flow task of ole db type and I put this sql statement under sql command - exec emp_sp @empidvar=?
    I am getting an error.
    [OLE DB Source [1]] Error: A rowset based on the SQL command was not returned by the OLE DB provider.
    [SSIS.Pipeline] Error: component "OLE DB Source" (1) failed the pre-execute phase and returned error code 0xC02092B4.

    shouldnt setting be Result
    Set=Full Resultset as your query returns a resultset? also i think variable to be mapped should be of object type.
    Then for data flow task also you need to put it inside a ForEachLoop based on ADO.NET recordset and map your earlier variable inside it so as to iterate for every value the sql task returns.
    Also if using SP in oledb source make sure you read this
    http://consultingblogs.emc.com/jamiethomson/archive/2006/12/20/SSIS_3A00_-Using-stored-procedures-inside-an-OLE-DB-Source-component.aspx
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • SQL Query using a Variable in Data Flow Task

    I have a Data Flow task that I created. THe source query is in this file "LPSreason.sql" and is stored in a shared drive such as
    \\servername\scripts\LPSreason.sql
    How can I use this .sql file as a SOURCE in my Data Flow task? I guess I can use SQL Command as Access Mode. But not sure how to do that?

    Hi Desigal59,
    You can use a Flat File Source adapter to get the query statement from the .sql file. When creating the Flat File Connection Manager, set the Row delimiter to a character that won’t be in the SQL statement such as “Vertical Bar {|}”. In this way, the Flat
    File Source outputs only one row with one column. If necessary, you can set the data type of the column from DT_STR to DT_TEXT so that the Flat File Source can handle SQL statement which has more than 8000 characters.
    After that, connect the Flat File Source to a Recordset Destination, so that we store the column to a SSIS object variable (supposing the variable name is varQuery).
    In the Control Flow, we can use one of the following two methods to pass the value of the Object type variable varQuery to a String type variable QueryStr which can be used in an OLE DB Source directly.
    Method 1: via Script Task
    Add a Script Task under the Data Flow Task and connect them.
    Add User::varQuery as ReadOnlyVariables, User::QueryStr as ReadWriteVariables
    Edit the script as follows:
    public void Main()
    // TODO: Add your code here
    System.Data.OleDb.OleDbDataAdapter da = new System.Data.OleDb.OleDbDataAdapter();
    DataTable dt = new DataTable();
    da.Fill(dt, Dts.Variables["User::varQuery"].Value);
    Dts.Variables["QueryStr2"].Value = dt.Rows[0].ItemArray[0];
    Dts.TaskResult = (int)ScriptResults.Success;
    4. Add another Data Folw Task under the Script Task, and join them. In the Data Flow Task, add an OLE DB Source, set its Data access mode to “SQL command from variable”, and select the variable User::QueryStr.
    Method 2: via Foreach Loop Container
    Add a Foreach Loop Container under the Data Flow Task, and join them.
    Set the enumerator of the Foreach Loop Container to Foreach ADO Enumerator, and select the ADO object source variable as User::varQuery.
    In the Variable Mappings tab, map the collection value of the Script Task to User::QueryStr, and Index to 0.
    Inside the Foreach Loop Container, add a Data Flow Task like step 4 in method 1.
    Regards,
    Mike Yin
    TechNet Community Support

  • Automatic creation of BW data flow documentation

    Dear Gurus,
    I need to write documentation of the data flow of a huge project which I haven't implemented by myself.
    The documentation should contain a mapping of the objects in the dataprovider, towards objects in the source system(s).
    Eventually with the info in which dataproviders the objects are included, e.g. between the multiprovider and the source system.
    Details of transformations can be ignored; eventually mentioning there's a routine involved, but that's the maximum.
    With the data repository, I can have the content of cubes in a graphical overview, but it doesn't really provide me useful information.
    You can imagine I prefer an automatic way to create this documentation.
    Anybody who knows a solution, even if it only provides part of the purpose?
    Any solution via query, standard SAP or customized program, ...
    Recommendations would be very highly appreciated!
    Thx & Rgds, sam

    Worldwide documentation is made on SAP BW projects, but no reply on automatic documentation.
    A lot of time must be lost by manually creating documentation on mapping objects to source system fields.
    ==> SAP, please, work out a solution.
    I didn't find a satisfying solution, but I've done it the following way:
    List all objects for a multiprovider via the meta data repository, and paste in excel document.
    Then listing all objects for the underlying dataproviders, and paste in separate sheets of this excel.
    Compare the objects of the MP with the objects on the other sheets using excel functions, and sign when a dataprovider contains a certain object.
    For the datasources, I checked if an object is present, and if yes, give the original source field.
    This in summary as a not optimal and not complete solution, but it prevents making mistakes.
    Rgds. sam

  • Foreach Loop Container with a Data Flow Task looking for file from Connection Manager

    So I have a Data Flow Task within a Foreach Loop Container. The Foreach Loop Container has a Variable Mapping of User:FileName to pass to the Data Flow Task.
    The Data Flow Task has a Flat File Source since we're looking to process .csv Files. And the Flat File Source has a Flat File Connection Manager where I specified the File name when I created it. I thought you needed to do this even though it won't really
    use it since it should be getting its File name from the Foreach Loop Container. But when attempting to execute, it is blowing up because it seems to be looking for my test file name that I indicated in the Flat File Connection Manager rather than the file
    it should be trying to process in User:FileName from the Foreach Loop Container.
    What am I doing wrong here??? I thought you needed to indicate a File name within the Flat File Connection Manager even though it really won't be using it.
    Thanks for your review...I hope I've been clear...and am hopeful for a reply.
    PSULionRP

    The Flat File Connection manager's Connection String needs to be set to reference the variable used in the ForEach Loop:
    Arthur My Blog

  • Error  in mapping data fields using BODS Designer

    HI All
    I am trying to map the data from excel sheet to SAP ECC 5.0 using Business Objects data services Designer .
    However after mapping the data Fields when i try to do the Validation It is giving me the error message :
    u201D(Data Flow:New_data Flow) (Query:Query) FROM: The input schemas are not specified for <Query.Queryname >u201D
    If anybody has experienced the same error in the past Please let me know how to resolve this error.
    Thanks in Anticipation

    Hello,
    Please check your "From" tab whether it is having the input table name is there or not. also check the mapping tab it should gives you "Schema.field"  format.
    Kindly revert with your findings.
    thx
    Deep

  • Oracle to SAP BI with BCS u2013 Best Data flow Design.

    Hi,
      We are a SAP implementation team. We are @ Blue Print stage. My client is a RETAIL Business Gaint. Client has 50 % of Transaction data and Master data in Oracle data base. Now we are moving to BI 7.0 and also has plans to use SAP-BCS.
      We would like to map all the existing Oracle tables to BI. Provide any clue regarding the best Data flow (From Oracle 10G to BI 7.0).
    Your quick and valuable suggestion/ links are highly appreciated.
    Warm Regards,
    Bab

    Hi Ashok,
    You have mentioned that you have a Oracle 10g  system as a data inflow which perfectly sets the platform to extract the data from the Oracle  system to SAP BW system.
    This Can be done through the DB connect ,through where you could select the necessary tables and form them as a datasource in BI system,further creating the usual BI objects on top of the DS.
    Once the Data is in BI ,we could pull the BI cube or form a replica of the cube in BCS application format to use it in the BCS environment.
    Hope this helps,
    Regards,
    Rajesh.

  • BI 7.0 Data Flow in 7.0

    Hi SDN Guru's,
    I am working BI 7.0 version, I have Activated BI Content for SD & MM Modules, Actualy I am taking at only necessary object in Grouping (tap).
    Actualy I am Activating BI Content for BI 3.5 version, Now, I want to do in Data flow in BI 7.0 for SD & MM Modules.
    How to do data flow BI 7.0? Please provide me solutions ASAP.......
    Thanks & REgards,
    Kumar.

    Hi,
    Kindly follow below steps to convert in to BI 7 dataflow.
    If the Dataflow is from Data source to DSO and DSO to Cube
    -->First Step u2013 Select DSO to Cube update rule and right click on update rule -> additional function -> create transformation
    This will copy and map all the objects with routine if exist and do the syntax check once before activation
    -->Second Step u2013 Then select the transfer rule from Data source to DSO to do the same thing as above.
    -->Third step - Then migrate the Data source , in   this always select the u2018with exportu2019 option which  will enable to revert back to 3.x version .. If needed
    If from one data source many targets are getting updated then use infosource
    -->Create DTP for all
    In addition to this I will suggest you to use LSA architecture if it is fresh implementation.
    Regards,
    Ashish

Maybe you are looking for

  • How many Sql licenses do i need

    hi , With one sql license how many users can connect concorrently to SAP Business One.If there are more than 50 SBO users how many sql licenses do i need to take? Thanks OmPrakash.

  • What are Smartforms and SAPscripts?

    Hey Everyone, Where do we use Smartforms and SAPscripts...how to we create these...if u have step step procedures that will be helpful... Thanks in Advance... Cheers: Jim

  • Can't use font functions

    I was using a downloaded trial version of Contribute 3 and later purchased my own copy. But with the new (purchased) copy, I can no longer use the font functions from the menu - when I go to them it says "disabled". It's VERY frustrating, particularl

  • Error -1100 when updating system software?

    "The operation couldn't be completed. (NSURLErrorDomain error -1100.)" I get this error from the Mac App Store whenever I try to find new system software.

  • Application sever file open

    hi to all,           I have one file in application server. Path let say D:\usr\sap\CCMS\TEST. I displayed path in report. my requirement is when ever i click on that  path ,need to open that file data with (word,notepad..etc). do you have any idea .