Issue at PSA Level data

Hi,
I am trying to extract data from the 0CRM_COMPLAINTS_I into BI. I have to enhance this DS with two Z fields. I can clearly see data populated for these Z fields in RSA3 with the enhancement code.
When I tried to get the same into BI, the data is not  being populated even at the PSA level. I tried deleting the old requests at PSA level and target level and then extracting but of no use. Instead it is showing some junk values for the Z fields at PSA.
I am  working on BI 7.0 and CRM 5.0.
Please let me know ASAP.
Thanks in advance,
Rama Murthy.

Hi,
you need to replicate the datasource as well as to maintain the transfer structure. In there you need to put the new fields from the right hand tab (datasource) to the left hand tab (transfer structure). Activate the transfer rules. Reload the data and you will see the new fields.
regards
Siggi

Similar Messages

  • PSA  level Errors are not displayed

    Hi friends,
        I'm using BI 7.0....up to PSA  i'm getting data  perfectly.. Even though some errors are there  i mean.. data mis match in my flatfile(for NUMC fields  I was enterd Charecter data... ... that time also.. psa level data was  showing green only.... if i upload into ODS  then only i'm getting  problem..
          but   in previous versions  it was showing  Red  (for error reacords..) in PSA level itself...
         for what reason  I'm not getting  like this.... is there  any  problem  with my procedure.. or   is this BI 7.0 feature..
    plz clarrify..
    Thanks
    babu

    Hi,
        Using the Error Handling function on the Update Parameters tab page in the scheduler, you can control the behavior of the BI system if data records with errors occur in the data flow with 3.x objects (use of transfer and update rules) when you are loading data using the PSA.
    The system checks for data records with errors:
          in the transfer rules
          in the update rules
          when data is updated to the master data tables, text tables, or hierarchy tables
          when data is updated to the InfoCube
          when checking for referential integrity within an InfoObject against the master data tables or DataStore objects on the communication structure level.
    Features
    Incorrect records arising in transfer rules, update rules, when updating to a data target, or when checking for referential integrity, are highlighted and returned to the error handler. The error handler locates the original PSA records and writes the corresponding log. This function is not yet available for updating to DataStore objects.
    Using the Error Handling function on the Update Parameters tab page in the scheduler, you can control how the system responds to incorrect data records.
    Activities
           1.      Define how the system should react to incorrect data records
    a.       No update, no reporting (default)
    If errors occur, the system terminates the update of the entire data package. The request is not released for reporting. However, the system continues to check the records.
    b.       Update valid records, no reporting (request red)
    This option allows you to update valid data. This data is only released for reporting after the administrator checks the incorrect records that have not been updated and manually releases the request (using a QM action, that is, by setting the Overall status on the Status tab page in the monitor).
    c.       Update valid records, reporting possible
    Valid records can be reported immediately. Automatic follow-up actions, such as adjusting the aggregates, are also carried out.
    With option a), the incorrect records are marked in red in the PSA maintenance. You can display the relevant error messages amodally, edit the records, and update the request manually. If it was not possible to write to the PSA (hierarchy or transfer method IDoc), an application log is written instead.
    With options b) and c), a new request that is only updated to the PSA is formed from the incorrect records. Here you can edit the records of the new request in the PSA and start the update manually.
    If errors arise when an InfoCube is updated, new requests are generated for incorrect records for each InfoCube combination. Among these requests, the incorrect records are updated to the PSA. These new requests have entered as data targets the InfoCubes for which there were incorrect records. If, for example, a record was updated to two InfoCubes by mistake, a request is generated for this record that contains both InfoCubes as data targets. In the PSA tree, the requests appear as successfully updated in the PSA. These new requests can then be triggered again for updating to the InfoCubes.
           2.      Specify the maximum number of incorrect data records that are allowed before the system terminates the load process. In the PSA, you can display the errors that have arisen up to that point. If you do not specify the maximum number of incorrect records after which termination should occur, the handling of incorrect data records is inactive and the update terminates after the first error.
           3.      Define how the BI system should react if the number of data records received by the BI system is not the same as the number of data records updated in BI.
    A difference between the number of received and updated records can occur:
             in the transfer rules
             in the update rules
             when updating
    Records are sorted, aggregated, or added.
    If you set the No Aggregation Permitted indicator, the request is deemed incorrect if the number of received records is not the same as the number of updated records.
    Regardless of this indicator, an error occurs if the number of selected records is not the same as the number of records received in the BI system.
    Result
    The monitor directs the user to the relevant error case and enables a direct navigation into the PSA or the respective error request.
    If you want to continue processing the incorrect data records in a program that you have written yourself, you can use the method GET_ERRORS of class CL_RSSM_ERROR_HANDLER. The use of this method is documented in the RS_ERRORLOG_EXAMPLE.
    More Information:
    Checking and Changing Data Using the PSA APIs
    To perform complex checks on data records, or to carry out specific changes to data records regularly, you can use delivered function modules (PSA-APIs) to program against a PSA table.
    If you want to execute data validation with program support, select Tools ® ABAP Workbench ® Development ® ABAP Editor and create a program.
    If you use transfer routines or update routines it may be necessary to read data in the PSA table afterwards.
    Employee bonuses are loaded into an InfoCube and sales figures for employees are loaded into a PSA table. If an employee’s bonus is to be calculated in a routine in the transformation, in accordance with his/her sales, the sales must be read from the PSA table.
    Procedure
           1.      Call up the function module RSSM_API_REQUEST_GET to get a list of requests with request ID for a particular InfoSource of a particular type. You have the option of restricting request output using a time restriction and/or the transfer method.
    You must know the request ID, as the request ID is the key that makes managing data records in the PSA possible.
           2.      With the request information received so far, and with the help of the function module, you can
             read RSAR_ODS_API_GET data records from the PSA table
             write RSAR_ODS_API_PUT changed data records in the PSA table.
    RSAR_ODS_API_GET
    You can call up the function module RSAR_ODS_API_GET with the list of request IDs given by the function module RSSM_API_REQUEST_GET. The function module RSAR_ODS_API_GET no longer recognizes InfoSources on the interface, rather it recognizes the request IDs instead. With the parameter I_T_SELECTIONS, you can restrict reading data records in the PSA table with reference to the fields of the transfer structure. In your program, the selections are filled and transferred to the parameter I_T_SELECTIONS.
    The import parameter causes the function module to output the data records in the parameter E_T_DATA. Data output is unstructured, since the function module RSAR_ODS_API_GET works generically, and therefore does not recognize the specific structure of the PSA. You can find information on the field in the PSA table using the parameter E_T_RSFIELDTXT.
    RSAR_ODS_API_PUT
    After merging or checking and subsequently changing the data, you can write the altered data records into the PSA table with the function module RSAR_ODS_API_PUT. To be able to write request data into the table with the help of this function module, you have to enter the corresponding request ID. The parameter E_T_DATA contains the changed data records. 
    Result
    The corrected data is now available for continued updates.
    if it helps plz assign points

  • Error in psa level - Request not updated to any data target using delta

    Hi,
    This is pavan and am facing this issue any one know for below issue to solve.
    Data Flow is BW Version and Implemented in BI  7.20 version.The issue was data has come
    to psa level but from here it is not updated to infocube so for that i have tried update manually
    it is showing information that 1334 SID IS SMALLER THAN / EQUAL TO 47287 CANCELLED.
    In Process Chain DTP is not included so how can i update data to infocube with out disturbing process chain
    Regards,
    Pavan Kumar

    Hi Pavan Kumar ,
    Please log on Tcode RSRQ -then give old request number (I think in your case 1334 is old request number),
    and you will get monitor for that request then change the QM status to red after that load new request .
    Thanks
    Hanumanthu

  • InfoPack processing (PSA and data targets in parallel) data and lock issues

    Good morning,
    We are on BW 3.5 and contemplating changing our InfoPackages from option 1 (PSA then data targets) to option 2 (PSA and data targets in parallel) for performance reasons (test have shown a gain of 50% in load time compared to using option 1).
    Issues have been raised that using option 2 may lead to data consistency and lock issues especially for SD infosources where we are dealing with multiple daily status changes for same document that need to be read in a specific order.
    From what I have read, such issues were possible in BW 2.x e.g. OSS Note 384023 and 364577. However, I am under the impression that the issues may have been resolved in BW 3.x and that it is "safe" to use option 2 (PSA and data targets in parallel).
    Would really appreciate if you could please share your experience when using processing option 2 (PSA // DT) especially in a BW 3.xx system.
    Thank you in advance,
    Catherine

    Hi Amit,
    The help text (vers. 3.5) for this method of loading says simply -
    "A process is started to write the data from this data packet into the PSA for each data package. If the data is successfully updated in the PSA, a second parallel process is started. In this process, the transfer rules are used for the package data records, data is adopted by the communication structure, and it is finally written to the data targets. Posting of the data occurs in parallel by packet.
    This method is used to update data into the PSA and the data targets with a high level of performance. The BW system receives the data from the source system, writes it to the PSA, and starts the update immediately and in parallel into the corresponding data target"
    Not sure where you got your description from, but is not explained well and confusing.
    Regards,
    Ken

  • PSA level issue USD is coming as USD,  for some records .

    Hello Experts ,
                           Very Good evening , I am facing one issue regarding the data load in the cube , actually I am loading data from a flat file having some 51 records ,till my PSA data is coming correct but with one modification i.e. in some records near about 16 records out of 51 one field data USD is coming like USD, means one comma (,) is automatically coming in the PSA level ,thats Y my rest of the records are not getting loded , can u please help me in this case to resolve it , waiting for the reply , and Thanks in advance .

    Open the CSV flat file in Notepad (right click on file and open with notepad) and see if there is any comma after USD.You can make correctionsin notepad version and save it.Do not open in CSV and make corrections.
    Regards,
    JB

  • AUTHORIZATION ISSUE: cube level data restriction in BI

    Hi all,
    I have few cubes and ODS which are containing data. The requirement is to restrict the cube level data.
    Eg : we have option to see the cube data in RSA1( ADMIN WORKBENCH).Right click on cube manage data.
    the requirement is to restrict to see the data company code = 111(only)
    Likewise for few users only company code = 222.
    If they try to see other than 111, they should get u201Cno authorization messageu201D.
    Cube data               
    company code         distribution channel         account     amount
    111                         10                                   10002     100
    222                        20                                      10002     200
    333                        30                                       10002     300
    444                        10                                      10002     400
    1111                20                                       10002     500
    1111                30                                      10002     600
    2222               30                                           10002     700
    Thanks in advance.
    Jo

    Hi MaikI,
    Thank for the inputs.
    Actually i want restrict the data based on the ANALYSIS AUTHORIZATION - 0TCAIPROV.
    I want to give s_rs_comp-provider value *, and i want to control the query (data) access through analysis authorization. i want to create zanalysisauth which contains 0TCAIPROV = $variable.
    variable is populated with one/two provider values at runtime .
    based on the runtime variable population user should get access.
    But with this implementation user is able to open all queries.where i am going wrong? how can i do this?
    Regards,
    Joseph

  • ERROR IN PSA LEVEL

    *Hi Experts,*
    *This is kalyan and i am facing a issue in BI 7.20 VER from past 1 week am not able to solve.*
    *this issue is process chain scheduled daily basis and process chain getting failed failed from 01-01-2011 to till date*
    *but records available in psa level but not reached target then when i try to do schedule manullay but*
    *in data target tab it is not showing target.i have tried with process manually this option available ipsa under header tab*
    *when i choose that option it is showing this error SID 1334 is smaller than /equal to update SID of DTA ZSD_BLCO3/47287;Canceled please help me in this issue if one knows and what will be solution.*
    *Thanks In Advance,*
    *Kalyan*

    Hi
    As you told this is running in Process chain. Do you have "Update from PSA" step after this IP is completed.
    check the job log of IP load in SM37, check the error log in Job.
    If you don't have Update from PSA step in PC, include it in your PC or make the changes to IP ( change the option to  PSA and then data target in series in update tab)
    Regards,
    Venkatesh

  • Issue when uploading Sales data from DSO to Cube.

    Dear All,
    I have an issue when I am uploading Sales data from DSO to Cube.  I am using BI 7.0 and I have uploaded all sales document level data to my DSO.  Then I use a transformation rule to calculate the Sales Value when I am doing DTP to cube.  Cube has customer wise aggregation data.
    In DSO I have NetPrice(KF) and Delivered_QTY(KF).  I do a simple multiplication routine in the transformation from DSO to Cube.
    RESULT =   SOURCE_FIELDS-NET_PRICE * SOURCE_FIELDS-DLV_QTY .
    At the moment I use Active Table(With out Archive) on the DSO to get the data since this is my first load.
    Issue is that the figure (Sales Value) in the cube is incorrect. I am getting a very large values which is impossible.
    Can someone please help me.
    Shanka

    Hi,
    are you sure that cube has customer wise aggregated data..since it will always aggregate the values of the key figures for the same set of characters...
    did you checked the values for the key figures as well and if even they are inflated or is the problem with this key figure only.
    During the data load the records may aggregates first for the same values of characters and then multiplication will be happening....if that is the case then you may have to multiply the values before storing it in the data package and then let it aggregate....can be achived through start routine...
    but first verify if other key figures are also having the same issue.
    Thanks
    Ajeet

  • Item level data in sap script

    Hello Friends,
    I have a report which prints Normal Tax invoice & Returns Tax, I modified this report just by adding one extra "delivery channel". The report call one more program which dose the printing job based on the delivery no.
    The issue, I can able to see the header data in th the print preview of the tax invoice, and not able to see item level data.
    Please help me how to resolve this issue.
    Regards,
    Shreekant

    Hi,
    You need to pass the line item data to the smartform/script. Please check in debug mode whether the data is passed correctly.
    Best regards,
    Prashant

  • Dont allow to change item level data in sales order.

    Hi all,
    I have a requirement in which, users should not be allowed to change  item level data or not allowed to add any new items in sales order on a certain condition. But they should be allowed to change the header level data.
    How can i achieve this.
    Can anyone help me?

    Hi,
    Check below exit.
    MV45AFZZ and in form USEREXIT_MOVE_FIELD_TO_VBAP.
    Here check for ur validation, If passes then CHECK variable SVBAP-TABIX. If it is 0 then item is created. If it is GT 0 then item is changed. Other way could be...
    select data from VBAP for each sales document and item in xvbpa internal table.
    If for any item u don't have data in VBAP table that means u r adding that item. So issue error message.
    *       FORM USEREXIT_MOVE_FIELD_TO_VBAP                              *
    *       This userexit can be used to move some fields into the sales  *
    *       dokument item workaerea VBAP                                  *
    *       SVBAP-TABIX = 0:  Create item                                 *
    *       SVBAP-TABIX > 0:  Change item                                 *
    *       This form is called at the end of form VBAP_FUELLEN.          *
    Thanks,
    Vinod.

  • Converted Crystal Statement Report not showing line level data properly

    Hi all,
    We have an issue with a Cystral-converted PLD Customer Statement layout behaving strangely when printing for a customer with multiple pages.
    When the statements are printed for a customer that has line level data which fills about an A4 page, this print absolutely fine; customer header data, line data etc. However, when we print for a customer that has more lines that can fit onto the one page, Crystal does the following:
    Page 1 - Customer Header info and column headers (OK); Line level info until bottom of page (OK)
    Page 2 - Customer Header info and column headers (MISSING); Line level info (ONLY ONE LINE APPEARS INDIVIDUALLY ON THIS PAGE BUT IS THE NEXT LINE CONTINUING FROM FROM THE PREVIOUS PAGE)
    Page 3 - Customer Header info and column headers (OK); Line level info until bottom of page (OK - CONTINUING FROM THE INDIVIDUAL LINE FROM THE PREVIOUS PAGE)
    Page 4 - Customer Header info and column headers (MISSING); Line level info (ONLY ONE LINE APPEARS INDIVIDUALLY ON THIS PAGE BUT IS THE NEXT LINE CONTINUING FROM FROM THE PREVIOUS PAGE)
    etc...
    So what basically seems to be happening is that every other page, Crystal just sticks in an individual line on it's own without any headers or anything OR other lines.
    Help please! Quite likely a 'Section Expert' issue but we don't know what setting could be causing this.

    Sounds like the report is trying to print multiple pages with blank pages in between, is the normal report ie details fit onto an A4 page printing with a blank back page?
    It will be page number resets and suppression of details on pagenumber = 2
    YOu need to go to section expert and go through each section noting what conditions have been suppressed and if any formula boxes are red then open and see what they do.
    Try turn conditions on and off and commenting out any formula conditiona and see what happens. By a process of elimination you should be able to work it out. However, you may have to stop the introduction of a blank page some how.
    Ian

  • Issues while generating Schema DAT files

    We are facing two type of issues when generating Schema ".dat" files from Informix Database on Solaris OS using the
    "IDS9_DSML_SCRIPT.sh " file.
    We are executing the command on SOLARIS pormpt as follows..
    "IDS9_DSML_SCRIPT.sh <DBName> <DB Server Name> ".
    The first issue is ,after the command is excuted ,while generating the ".dat" files the following error is occuring .This error is occuring for many tables
    19834: Error in unload due to invalid data : row number 1.
    Error in line 1
    Near character position 54
    Database closed.
    This happens randomly for some schemas .So we again shift the script to a different folder in Unix and execute it.
    Can we get the solution for avoiding this error.
    2. The second issue is as follows..
    When the ".dat" files are generated without any errors using the script ,these .dat files are provided to the OMWB tool to load the Source Model.
    The issue here is sometimes OMWB is not able to complete the process of creating the Source Model from the .dat files and gets stuck.
    Sometimes the tables are loaded ,but with wrong names.
    For example the Dat files is having the table name as s/ysmenus for the sysmenus table name.
    and when loaded to oracle the table is created with the name s_ysmenus.
    Based on the analysis and understanding this error is occuring due to the "Delimiter".
    For example this is the snippet from a .dat file generated from the IDS9_DSML_SCRIPT.sh script.The table name sysprocauthy is generated as s\ysprocauthy.
    In Oracle this table is created with the name s_ysprocauthy.
    s\ysprocauthy║yinformixy║y4194387y║y19y║y69y║y4y║y2y║y0y║y2005-03-31y║y65537y║yT
    y║yRy║yy║y16y║y16y║y0y║yy║yy║y╤y
    Thanks & Regards
    Ramanathan KrishnaMurthy

    Hello Rajesh,
    Thanks for your prompt reply. Please find my findings below:
    *) Have there been any changes in the extractor logic causing it to fail before the write out to file, since the last time you executed it successfully? - I am executing only the standard extractors out of the extractor kit so assumbly this shouldnt be a issue.
    *) Can this be an issue with changed authorizations? - I will check this today, bt again this does not seem to be possible as the same object for a different test project i created executed fine and a file was created.
    *) Has the export folder been locked or write protected at the OS level? Have the network settings (if its a virtual directory) changed? - Does not seem so because of the above reason.
    I will do some analysis today and revert back for your help.
    Regards
    Gundeep

  • PO Form Issue with Header Level Changes

    Issue:
    It has been observed that whenever changes are done in PO Header (e.g. Header Text), system prints Billing Address and Communications Details as blank. If the changes are made to Line Item level along with header then system prints the Billing address and Communication Details correctly.
    Analyse:
    we are using the SELECT statement below to fetch the field for
    Billing Address and Communications Details.
    SELECT SINGLE
    mandt z_plant z_company_code z_toll_free_no z_local_no z_fax_no
    z_email z_name1 z_name2 z_street z_house_no z_district z_post_code
    z_city z_country z_vat_no z_loc_term_cond z_langu
      FROM zptp_poformdata
      INTO w_poformdata
      WHERE z_plant = ls_ekpo-werks
    AND   z_company_code = ls_ekpo-bukrs .  
    And when we only change the header level  data no value is passing for comany code and plant . as a result Select statement gets failed and we are getting blank window.
    NOTE: we are getting blank box only when we fire a print or fax.The preview is coming fine.

    by asking so many questions in a single post,you are violating the rule of engagement given below
    Only One Question Per Posting
    For each question that you have, make a separate posting. Don't ask multiple questions in one posting, this is confusing for people and might not get answered, since the subject line does not reflect all your problems. If a new and unrelated question comes up during a thread, start a new thread.

  • User exit to change item level data in purchase order

    Hi,
    Can anyone let me know the user exit to change item level data in purchase order . there is a badi ME_PROCESS_PO_CUST for this but the issue is its method process_item gets triggered only when the item is changed. My requirment is
    For purchase order document types u201CZSOu201D and u201CZCOu201D, where the purchase order is a u201CLimits Orderu201D only i.e. no materials or services on the purchase order, the print price indicator field should be set to u201Cblanku201D (unchecked).   now i cant use ME_PROCESS_PO_CUST  because process_item wont get triggered if there is no change in itemlevel data.
    Regards,
    Rahul

    Hi Rahul,
    Probably EXIT_SAPLMEKO_002.
    hope it helps,
    Edgar

  • Help - User exit to change item level data in Purchase Order

    Hi,
    Can anyone let me know the user exit to change item level data in purchase order . there is a badi ME_PROCESS_PO_CUST for this but the issue is its method process_item gets triggered only when the item is changed. My requirement is For purchase order document types u201CZSOu201D and u201CZCOu201D, where the purchase order is a u201CLimits Orderu201D only i.e. no materials or services on the purchase order, the print price indicator field should be set to u201Cblanku201D (unchecked).   now i cant use ME_PROCESS_PO_CUST  because process_item wont get triggered if there is no change in item level data.
    Thanks,
    Rahul

    Hi Rahul,
    Probably EXIT_SAPLMEKO_002.
    hope it helps,
    Edgar

Maybe you are looking for

  • Problem receiving from sender, showing in SmtpReceive log but not Message Tracking

    Our Exchange 2007 server (SBS2k8 really) has had problems receiving mail from a certain sender the last few weeks.  With the SMTPReceive set to verbose, I can see the mail coming in but it never shows up in the message tracking logs.   Looking at sim

  • Can not print text file.

    I have a problem since I upgrade to 10.6.7. All text file, including docx, pdf, and so on, can not be printed. Only the image file can be printed correctly. Here is a report: Process: cupstomcdufr2 [7642] Path: /Library/Printers/Canon/UFR2/Cores/cups

  • Refering text from Shipping doc to billing doc

    Hi dear Friends, how do I copy text from delivery to invoice in invoice the document of reference is always the sales order in structure REFERENZ there is only the order the problem is the same for VBBK and VBBP can you help me??? thank you very much

  • Exporting from swf

    i have a large swf (900 x 600) and need to export a printscreen image when the user clicks a button. it is possible to export the printscreen as another swf? and how would one do this? any tutorial recommendations? i have read some tutorials about ho

  • Identification of OS user with DAD authentication scheme

    I have following situation: Authentication scheme : no authentication (I'm using dads.conf which has PlsqlDatabaseUsername APEX_PUBLIC_USER) APEX/database is on a Unix server Apex application is run from Windows desktop browsers. I want to capture de