Data Reconciliation between PSA and DSO.

Hi Experts,
The data records which is showing in PSA is 1000 and when we ran to load the data  till DSO the records were only 600 and we dont understand why the records are not fetching.
DTP Settings:
Extraction mode is :FUll
If anyone having the good documentation on How to design DSO please farward some good links.
Regards.

Hi,
This is not issue with the DSo.
There might be filters at various points due to which you are getting some of the records filtered.
Check the following.
Check the PSA for number of records if the records is not equal o 1000 then the records are filtered while fetching from source
1)Check in PSA if there are any filters.
If the records in PSA equal to 1000 then
1)Check in DTP also for some filters
2)Check the routines in transformation also if there are any records that are filtering in the transformation.
Hope this helps,
Sri...

Similar Messages

  • Data Reconciliation between BI and R/3 systems

    Hi Gurus,
    I want to know how to reconciliation the data between BI and R/3 systems.
    is there any easy mentods to reconcile the data.i have also gone through the HOW TO GUIDE document, that document dosen't help me out.
    Regards
    Sreenivas.Y

    1) Either we can reconcile using standard R3 tables eg VBAK , VBAP for sales .
    2) You can go for automatic reconcillation where you will have to create a reconcillation data source and extract it to BW using virtual cube (Virtual cube would extract real time data from R3).
    Include original cube and this virtual cube in a multiprovider .
    Make a report taking keyfigures from both the infoproviders and add a column for difference .If there is 0 difference it means BW data matches with R3 .

  • Data reconciliation between R/3 and the BW Systems

    How do we do data reconciliation between R/3 and the BW system for the following areas?
    Purchasing
    Controlling:
    Project System:
    COPA
    SD
    AP
    Trgards,
    Tony G

    Tony
    Have you looked these documents??
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/biw/how%20to%20validate%20infocube%20data%20by%20comparing%20it%20with%20psa%20data
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/7a5ee147-0501-0010-0a9d-f7abcba36b14
    Re: BW v. R/3 data reconciliation
    Hope this helps
    Thanks
    Sat

  • Link Between PSA and Data Targets

    Hi Experts,
    Could you please help me is there any table which will have PSA & data target (Cube or ODS) details.
    I know the table which will have data target manage screen details and other table wil give the details like data loaded into till PSA.
    But i want to know any SAP table will have link between PSA & data target.
    I want to know if the request is loaded till PSA or it is loaded into data target.
    Any table will give details like PSA request no, if target is there then some kind of flag like....
    Please let me know if any one know the table which will have PSA as well as Data target details.
    Thaks in Advance
    Shaik

    Hi,
    I clearly mentioned im aware of Manage screen details table, i want link between PSA and Data target table which will return if requst loaded till PSA or PSA & target.
    Im using BI 7.0 .
    Regards
    Shaik

  • Data Replication Between Sqlserver and Oracle11g using materialized view.

    I have Sqlserver 2005 as my source and oracle11g as my target.I need to populate the target daily with change data from source.
    for that we have created a dblink between sqlserver and oracle and replicated that table as a Materialized view in Oracle.
    problem we are getting here is Fast refresh option is not available.each day it will pick full data from the source.
    is there any way to use Fast refresh in this scenario??
    Thanks in advance.
    Regards,
    Balaram.

    Pl do not post duplicates - Data Replication Between Sqlserver and Oracle11g using materialized view.

  • Data streaming between server and client does not complete

    Using an ad-hoc app, data streaming between server
    and client does not complete as it supposed to be.
    The process runs well in solaris 5.8, however under 5.9
    we have found the characters stream buffer length limitation
    is around 900 to 950 characters (by default we are using 3072
    characters).
    Example:
    - We are transfering HTML file, which will be displayed
    in the App client, with buffer=3072, the HTML only displayed / transfered
    as xxxxxxxx characters, but with buffer=900 the HTML is displayed properly,
    in this case, the only problem that we have is the file transfer will
    eventually longer than usual.
    - There is another case, where we have to transfer information (data) as stream
    to the client. A long data stream will not appear at all in the client.
    Any ideas why the change between 5.8 and 5.9 woudl cause problems?
    The current app-driver that we are using is compiled using Solaris 5.6,
    if possible we would like to have use of a later version, which is compiled using Solaris 5.9, do you think this will probably solve our problem?
    Thanks
    Paul

    Does this have anything to do with Java RMI? or with Java come to think of it?

  • Secure the file/data transfer between XI and any third-party system

    Hi All,,
    I would like to use to "secure" SSH on OS Level the file/data transfer between XI and any third-party system Run OS Command before processing and OS command After processing. right now my XI server installed on iSeries OS.
    with ISeries we can't call the Unix commands hope we need to go for AS400 (CL) Programming. If we created the AS400 programm how i can call that in XI.
    If any one have idea pls let me know weather it will work or not.
    Thanks in adavance.
    Venkat

    Hi,
    Thanks for your reply.
    I have red some blogs like /people/krishna.moorthyp/blog/2007/07/31/sftp-vs-ftps-in-sap-pi to call the Unix Shell script in XI.
    But as i know in iSeries OS we can write the shell script we need to go for AS400 programe. If we go with AS400 how we need to call that programe and it will work or not i am not sure there i need some help please.
    Thanks,
    Venkat

  • Data mismatch between 10g and 11g.

    Hi
    We recently upgraded OBIEE to 11.1.1.6.0 from 10.1.3.4.0. While testing we found data mismatch between 10g and 11g in-case of few reports which are including a front end calculated column with division included in it, say for example ("- Paycheck"."Earnings" / COUNT(DISTINCT "- Pay"."Check Date")) / 25.
    The data is matching for the below scenarios.
    1) When the column is removed from both 10g and 11g.
    2) When the aggregation rule is set to either "Sum or Count" in both 10g and 11g.
    It would be very helpful and greatly appreciated if any workaround/pointers to solve this issue is provided.
    Thanks

    jfedynic wrote:
    The 10g and 11.1.0.7 Databases are currently set to AL32UTF8.
    In each database there is a VARCHAR2 field used to store data, but not specifically AL32UTF8 data but encrypted data.
    Using the 10g Client to connect to either the 10g database or 11g database it works fine.
    Using the 11.1.0.7 Client to go against either the 10g or 11g database and it produces the error: ORA-29275: partial multibyte character
    What has changed?
    Was it considered a Bug in 10g because it allowed this behavior and now 11g is operating correctly?
    29275, 00000, "partial multibyte character"
    // *Cause:  The requested read operation could not complete because a partial
    //          multibyte character was found at the end of the input.
    // *Action: Ensure that the complete multibyte character is sent from the
    //          remote server and retry the operation. Or read the partial
    //          multibyte character as RAW.It appears to me a bug got fixed.

  • Data sync between oracle and sql server

    Greetings Everyone,
    Your expert views are highly appreciable regarding the following.
    We at work are evaluation different solutions to achieve data synchronization between oracle and sql server data bases. Data sync i mentioned here is for live applications. We are runnign oracle EBS 11i with custom applications and intending to implement a custom software based on .NET and SQL Server. Now the whole research is to see updates and data changes whenever happens between these systems.
    I googled and found Oracle Golden Gate, Microsoft SSIS, Wisdom Force from Informatica....
    If you can pour in more knowledge then it's great.
    Thank You.

    Most of the work involved has to be done through scripts and there is no effective GUI to implement OGG.However using commands is not vey togh and they are very intutive.
    These are the steps, from a high level:
    1.Get the appropriate GG Software for your source and target OS.
    2.Install GG on source and target systems.
    3.Create Manager and extract processes on source system
    4.Create Manager and replicat processes on target system
    5.Start these processes.
    First try to achieve uni-directional replication. Then Bi-directional is easy.I have implemented bi-directional active active replication using Oracle DBs as source and target. Refer to Oracle installation and admin guides for more details.
    Here is a good article that may be handy in your case.
    http://www.oracle.com/technetwork/articles/datawarehouse/oracle-sqlserver-goldengate-460262.html
    Edited by: satrap on Nov 30, 2012 8:33 AM

  • Please write the difference between ODS and DSO...

    Hi all,
             Please write me the differences between ODS and DSO.. I think both are same in structure and have same update types (Over write, Addition) and also with the tables (changelog, Activation queue, Active table).. Is there still a difference between ODs and DSO...I can assign the points..
    thanks
    arya

    Hi,
    You can check the below document for new Features in BI:
    http://help.sap.com/saphelp_nw04s/helpdata/en/a4/1be541f321c717e10000000a155106/frameset.htm
    /people/martin.mouilpadeti/blog/2007/08/24/sap-netweaver-70-bi-new-datastore-write-optimized-dso
    Reg
    Pra

  • Relation between psa and data source

    Hi everyone,
    What is the relation between DataSource and PSA.

    PSA tables get created when u activate datasource in BW...replication is just required for purpose of Meta data loading for that data source....PSA will store data temporarily when u will load data from R/3....just u can manage PSA request datasource wise....by right clicking datasource...selecting manage tab u can see all the PSA req corresponding to that particular datasource.

  • Data difference between PSA/Query and InfoCube

    Hi All!
    I have a problem during the data extraction from R/3 to a basic Cube. When the currency is different from EUR, the key figure 0CS_YTD_TC data it’s different from the others (0CS_YTD_GC and 0CS_YTD_LC). In the cube content the value is correct but in the PSA and in queries is not.
    What can I do?

    Ricardo,
    Do you mean that one of your KF value (for example, in the cube) differs from the value shown in a query 10 times?
    If so, look at the table TCURX. In my system PTE has 0 decimals. It means that cube/query will differ 100 times.
    Look here at the problem description:
    Re: Local Currency COP, KRW values wrong in report
    and here for workaround on a query level:
    Re: is conditional calculation in the report possible?
    Best regards,
    Eugene

  • Doesn't load from source system to PSA and DSO

    Hi Gurus,
    We have launched a Delta load from Source System (POSDM) to DSO. In the source system there are 5000000 rows but none was loaded into DSO or PSA. The delta load was ok, but any data was loaded.
    If the delta  load was lunch again, would be loaded those 5000000 rows. Or those 5000000 rows wouldn't be able to load again.
    Any idea abut this issue? Any feedback will be really appreciated.
    Thanks in advance.

    Hi David,
    Are you sure this 5 million records are delta records and should be pulled into BW as delta??
    did you count the number of records in the underlying tables??
    which data source you are using??
    Delta loads are suppose to bring the new records or changed records since the last data load and not all the records.
    Since the request is green as you said and still it shows 0 records then it is possible that nothing has changed since last delta.try to see the details of the data load... if it failed...may be thats the reason you are not able to see any records.
    If you schedule the delta again it will bring the records changed or created after last delta.
    If the delta was unsuccessfull then turn the QM status and overall status of request to red manually in the monitor of the request and delete it from PSA and all the targets in which its loaded and schedule the delta again and it should bring the delta.
    Thanks
    Ajeet
    Thanks
    Ajeet

  • Transformation between DS and DSO missing Non-cumulative KYF

    Hi SDN,
    I installed the business content for CML, and am working with 0CML_DELTA_CAP datasource.  I would like to create a transformation between this datasource and DSO 0CML_O03.
    By default, when you install the CML business content, the datasources use the 3.x dataflow concept (transfer rules, infosources, and update rules).
    I would like to use the 7.x dataflow concept and created a transformation between the 0CML_DELTA_CAP datasource and the 0CML_O03 DSO.  In the transformation, it is missing the fields 0CML_B330, 0CML_B340, 0CML_B360, 0CML_B380 in the datatarget.  These key figures are non-cumulative with NCum Value change (0CML_B330 uses 0CML_B331 as value change).  The value change key figures show up in the transformation, but the non-culmulative key figures do not. 
    Does anyone have any ideas why the non-cuml. kyf are not showing up in the transformation?
    Thanks,

    Hi Brendon,
    The non- cumulative key figures are not mapped in the transformation. Only the 'Inflow' and 'Outflow' key figures for the non- cumulative key figure are mapped.
    You can check the property for the non- *** KF in RSD1, where you would find the corresponding inflow and outflow kf. If both of these are mapped in the transformation, the non- *** kf would calculate the value in the report in which it is included as:
    Non-*** KF value = Inflow value - Outflow value.
    Hope this helps.
    Regards,
    Abhishek.

  • Report for date variances between delivery and goods receipt date

    dear guru ,
    i search a standard report for date variances between delivery date in the purchase order and the goods receipt date.
    Do you help me ?
    Thanks

    Hi,
    Use Report ME80FN and here check the delivery schedule option in output screen.

Maybe you are looking for

  • How do i delete video playlists?

    I was looking through my iPod one day and I noticed that I had a "John Mayer" playlist in my "video playlist" section, even though they were songs, not videos. I have no idea how it got there in the first place, but I have been looking around iTunes

  • PSB - Create GL Budget Journals Error

    We tried to transfer a PSB worksheet but are now getting an error: "Budget worksheet 7841 has already been posted for GL period Sep-10." I look at PSB_WORKSHEETS but the worksheet is no longer there - so I'm not sure what to do. Thanks, Rob

  • Browser object in a desktop application

    i wantto add a browser object in my APP beacuse iwant to d/l site to and parse some data from them! how can i do this? thnaks in advabnce peleg

  • How to copy cell content without formula

    In numbers, how do I copy and paste the cell content without transferring the underlying formula?

  • Tandberg C20 with ASA

    I have a Tandberg C20 connected to the DMZ interface of my ASA, when I  make outbound calls I can hear the people, the can see and hear me but I cannot see any incoming video. We moved the unit outside the firewall and it worked correctly. I am not s