APD Process for Cube -  ODS data transfer.

hi,
Two data sources (cube and ODS) are feeding another ODS. I want to delete the data in Data target (ODS)belongs to that particular infosource for every data load if it comes from the same infosource without using process chain or events. (I want the similar option found in cube). I also want to write some complex functionality (ABAP CODE) in start routine of update rules.
Bhanu suggested me to use APD Process. Can any one (including BHANU) explain me in detail about APD Process. Please also give me navigations.
Regards,
Avneet

Hi Avneet,
You can read details about the APD here:
http://help.sap.com/saphelp_nw04/helpdata/en/49/7e960481916448b20134d471d36a6b/content.htm
Basically you can get to the APD from transaction RSANWB. Start creating a new process using a query as the source of data. You can provide a transactional ODS as a data target and you also have the opportunity of changing the data using predefined functions, or also ABAP code before loading into the ODS.
Hope this helps...

Similar Messages

  • Missing Data Target in Infopackage for Update ODS Data in Data Target Cube

    Hello & Best Wishes for the New Year to all of you,
    I have 3 ODS (1 on Full Update and 2 with Delta Updates). All these 3 ODS update data to a single CUBE. In my development system this works correctly. Data load from PSA to ODS to Cube.
    Now I transported this to my QA and Production System. In QA and Production System, I am able to load data upto all the 3 ODS and ACTIVATE Data in all these 3 ODS.
    When I am trying for "Update ODS Data in Data Target" to load data from ODS to Cube in QA /PD, the system created Infopackage (ODS to CUBE), doesnot get the Data Target details. (Initial Upload / Full Upload). The Data Target Tab is Blank (expected Cube details).
    I have tried to transport again after deleting the update rules.
    Can you suggest what could be the problem ?
    regards - Rajesh Sarin

    Thanks Dinesh,
    I have the ODS to CUBE Update Rules ACTIVE in the QA and PD system. Still the problem exists only in QA and PD. In DV the ODS to CUBE Data Target is available in the Infopackage and loading the data correctly to Cube.
    Listing all the trials I have done :
    1) Originally transported with all the Collected Objects. Inspite of having Active Update Rules, Data Target Cube was missing in QA and PD.
    2) After this problem, I again transported only the Update Rules through the Transport Connection, still the problem didnot get solved in QA and PD.
    3) Again, I sent a transport to delete the Update Rules which deleted the ODS to CUBE update rules in (DV), QA, PD. After that I sent another request to CREATE the ODS To CUBE UPDATE Rules in (DV), QA and PD. Still the Data Target is missing, inspite of having Active Update Rules in QA and PD.
    In DV the ODS to CUBE Data Target is available and loading the data correctly to Cube, even now.
    4) I have also tried "Generate Export Data Source" for the 3 ODS in QA and PD. Still it doesnot help.
    Can you please suggest ?
    regards - Rajesh Sarin

  • Processing for request without data

    Hi Experts,
    I need to save additional information in my custom table for each request loaded into my infoprovider (cubes und dso).
    I thought I can make it with some ABAP in the start routine. It works fine, but not for requests without data. If there is a request without data, start routine is not executed at. all. Do you know a pleace where I can input my processing for request without data records.
    thanks
    Wojciech

    Hi,
    I have to save information about the open posting period in R/3 when the request is loaded. I dont want to have this information in the data, becouse I want to have the possibility to delete requests belonging to an posting period.
    I don't think, that this information is needed to understand my problem.
    Greetings
    Wojciech

  • Modifying Cube/ODS data for few entries

    Hi friends,
    Is there any way to modify few record entries of couple of chars. in either
    cube/ODS without actually distrubing the settings.
    regards.

    Hi Nasiroddin,
        Create a generic datasource based on ODS Active table. Create update rules from this ODS into this ODS only(Loop back).
        Pull data till PSA and edit manually(if records are less) or write some routines to change data accordingly and load except key figures(or map only required fields in overwrite mode).
        Push delta from this ODS to subsequent target, data will be corrected automatically.
    Hope it Helps
    Srini

  • Delta for cubes & master data

    Bonjour,
    I am a bit confused about some concepts of BW. Hope to get clarified here in good ole SDN.
    1. Infocube is said to be ADD only. Yet, its data deletable eg. by request. How do I interpret this? Is it deletable only by requests? How about selectively as detail as single records? How about field level deletion (instead of a whole record)?
    2. Is a 'new' image (rec mode "N") = 'after' image (rec mode " ")?
    3. How is Repair Request compared to Pseudo Delta? Are they both only and must selectively delete by month / time char only? Can either be deleted by single records?
    4. Master data is said to be usually overwritten. How is it so? What Delta type(s) are we talking about by this? How is Master data aggregated? Can it be? What does serialization and duplicate master data processing mean? How can master data prevent duplicates?
    What's the usual and possible practices to manage delta for Master data attr,txt,hier?
    Awaiting your reply.
    best regards,
    Pascal Gabin

    1. Infocube is said to be add only because it is different from ODS or master data where you can 'merge or overwrite' data. When 2 records are loaded into an Infocube, if they are different in any one field, they will be stored as 2 seperate records (in the fact table of the cube). Whereas if 2 records are loaded into an ODS, and the records are similar in terms of the key of the ODS, only 1 record will be created in the ODS with the second records details overwriting the first if there is a clash.
    From a cube you can delete data either selectively by specifying ranges or delete an entire request. You cannot do field level deletions.
    2. No they are different. Let us say there is a record with 2 fields Customer and Amount being loaded into an ODS. First record loaded is (Pascal, 10.00). Then there will be a New-image record (Pascal, 10.00,N). After this a second record (Pascal, 15.00) is loaded. After this there will be 3 records.
    Pascal, 10.00, New Image
    Pascal, -10.00, Before Image
    Pascal, 15.00, After Image.
    All 3 add to 15.00.
    3. Not able to understand your question. Deletion of all requests are I think the same. Selective deletion can be done only for a Infocube not for an ODS. See answer 1.
    4. Master data usually changes very little. For eg, a customer 100013 is always 100013. Once in a while his phone number might change. When it does, it is picked up by the delta and the existing record in BW is overwritten. There is only one type of delta for master data. Not sure of what is the exact name. Master data is aggregated by specifying the key to the master data table.
    The best practice is to maintain delta only for large master data objects (Customers, Materials etc). For all of these SAP has provided delta-enabled datasources which can be used. For the rest, full-uploads suffice.

  • For Init without data transfer showing records 0 from 0

    Hi all,
    Iam doing Init without data transfer to ODS. In the ODS manage it is showing 1 from 1 which is correct. This ODS data is automatically updated to a cube. There in the cube manage it is showing 0 from 0. My mappings in update rules also correct , but why it is not giving 1 from 1.
      In the cube manage the status is showing green but in the monitor of cube always it is yellow.
    Help me please
    Ram

    Hi Ram
    If you load data with INIT without data transfer then system shows the defualt record in the monitor.
    If check that record in the PSA then you dont find it and on the cube as well.
    If the load is green with 1 record then extractor working fine.It's not actaully a data reocrd.
    Hope it helps...
    Regards,
    Chama.

  • ALE/iDoc for Work Center data transfer

    Hello, community.
    I am working on several interfaces for the PP and PM modules using iDocs/ALE.
    One of the requirements is that only created and modified data must be distributed from R/3.
    For Work Centers, we are using basic type LOIWCS02 and message type LOIWCS. We are generating iDocs through transaction POIM, which has two transfer modes: Complete (data for all Work Centers etc.) and Change (only new and updated data).
    The Change Transfer Mode assumes you have set both Change Pointers and Change Documents for the objects to be transfered. Otherwise, it'll work just like the Complete Transfer Mode.
    Our PP consultant told me there is no configuration to generate Change Documents for Work Center data modifications.
    Can anybody give me other opinions? I'll resort to Z-table solutions only if I have no other options.

    Hi Danilo,
    Do you know what the process codes are for
    LOIWCS and LOIROU ?
    Appreciate your help.
    Thanks,
    Gopi.

  • Creat Job for LSMW (periodical data transfer)- idoc BLAORD

    Hi,
    We are trying to create job for LSMW (thru frame program for periodic data transfer) for idoc BLAORD (Purchasing contract)
    When I am running the job that I created, the job is cancelled and the error is:
    "Upload from front-end is not possible in batch mode.
    Message no. /SAPDMC/LSMW_OBJ_070 020"
    How can I correct that? so I will be able to run a job (the job mast run background?)
    Thanks,
    Tal Sasson

    Hello,
    I was also receiving the same message /SAPDMC/LSMW_OBJ_070 020 even though I specified in /SAPDMC/SAP_LSMW_INTERFACE the server path. The problem was that in LSMW->Specify Files you have to delete the definition of front-end files and define application server files.
    That Flag File (Path and Name) (field /SAPDMC/LSOINP-FILENAME in program /SAPDMC/SAP_LSMW_INTERFACE) is kind of useless since it is ignored anyway and files defined in LSMW object are imported and they have to be read anyway if you use more than one file.
    Regards,
    Peter

  • Reg lsmw for customer master data transfer

    Hello All,
    I want to know while transfering data by lsmw for customer master data .
    Scenario : There are more then one ship-to-party and bill to party so how to upload the data in this case?????
    Thanks,
    Sunny

    hi
    just refer to the link below
    http://www.sapmaterial.com/?gclid=CN322K28t4sCFQ-WbgodSGbK2g   
    <b>step by step procedure with screen shots</b>          
    regards
    vijay
    <b>plz dont forget to reward points if helpful</b>

  • Source for Init with data transfer and full load

    Hi Experts,
    1. Please tell me where the data comes when we do the
        following actions :-
        a) Init with data transfer
        b) Full load
      i want to know the source of data i.e (setup table or main table or any other table)
    Helpful if you can provide the data-flow.
    kindly tell which is prefrable and why?
    Regards,
    RG

    Hi.......
    When you do init with data transfer it will read data from the set up table only..........but beside this it will also set the init flag..........due to this from next time any new records will go to the delta queue...........and next time when you run delta load it will pick records from delta queue...........
    Now suppose data are alrady available in the BW side...........then you can run init without data transfer.......this will only set the init flag.........it will not pick any records.........
    Extraction structure gets data from set up table in case of full upload ............and in case of delta load from delta queue.......
    Regards,
    Debjani...........
    Edited by: Debjani  Mukherjee on Sep 20, 2008 11:45 AM

  • ODS TO ODS  DATA TRANSFER

    hi GURUS,
    can anyone pls suggest the steps to transfer data from one ODS to another ODS.
    regards
    ravi kiran

    dear ravi
    to pass data from one ODS to another ODS just follow the aboave process.
    1.go to source ODS- right click- create export data source-make.
    2.go to RSA1- source system- choose ur system name- replicate datasource.
       in this step u might get some problem regarding the myself source system defination. If this step dont work properly, then contant to basis person and check the myself source system defination.
    3.go to the target ODS and create update rule from ur source ODS. if u need any modification to the date then write teh update routine or start routine as per ur requirement.
    i think this way u can pass data from One ODS to Another ODS.
    Kindly assign the points if it helps..

  • Custome Process for update the date

    Hello Every one,
    I have a page with the check-box where the user can select a record and update the record by pressing on the button.
    I have three processes, and each of these process update one field ( which is a date field) of the page by pressing on the specific button. here are the three processes:
    declare
    begin
    for i in 1..apex_application.g_f01.count loop
    UPDATE  REC_RET_ADD_RECORD
    SET DATE_ADMIN_APPROVED=:P8_DATE_ADMIN_APPROVED
    WHERE REC_RET_ID = APEX_APPLICATION.G_F01(I);
    end loop;
    commit;
    end;
    Condition: When Admin button pressed (submit)
    The second process:
    declare
    begin
    for i in 1..apex_application.g_f01.count loop
    UPDATE REC_RET_ADD_RECORD
    SET DATE_COMMITTEE_APPROVED=:P8_DATE_COMMITTEE_APPROVED
    WHERE  REC_RET_ID = APEX_APPLICATION.G_F01(I);
    end loop;
    commit;
    end;
    Condition: When Committee button pressed (submit)
    The third process:
    declare
    begin
    for i in 1..apex_application.g_f01.count loop
    UPDATE REC_RET_ADD_RECORD
    SET DATE_OHS_APPROVED=:P8_DATE_OHS_APPROVED
    WHERE  REC_RET_ID = APEX_APPLICATION.G_F01(I);
    end loop;
    commit;
    end;
    Condition: When OHS button pressed (submit)
    Now the problem is, when the user selects the row and selects the date for Admin field and presses the Admin button, the system inserts the date but after that if the user wants to selects the same row and update the OHS date, the system will remove the date for Admin field and inserts the OHS date in OHS field and visa verse.
    I don't know why this is happening. Could you please help me. I appreciate it.

    Here is the report query:
    SELECT  apex_item.checkbox (1,
                               REC_RET_ID,
                               'onchange="spCheckChange(this);"',
                               :f_REC_RET_ID_list,
                              ) checkbox,
    "REC_RET_ID",
    "RET_SCHEDULE_NUM",
    "RET_COST_CENTER",
    "RET_TITLE",
    "RET_TITLE_DESC",
    "RET_PERIOD",
    "DATE_USER_INSERTED",
    "DATE_ADMIN_APPROVED",
    "DATE_COMMITTEE_APPROVED",
    "RC3_FORM_REQUIRE",
    "DATE_OHS_APPROVED",
    "DESTRUCTION_SUSPENDED",
    "DESTRUCTION_SUSPENDED_NOTE",
    "INACTIVE_DATE",
    "INACTIVE_DATE_APPROVED",
    "CREATION_DATE",
    "CREATED_BY",
    "LAST_UPDATE_DATE",
    "LAST_UPDATE_BY"
    from REC_RET_ADD_RECORD

  • How to Search BPM Processes for a Given Date Range through a VC Report

    Hi All,
    We have a VC component build to view the BPM process information. Here we are supposed to provide a Date range option for searching the processes. But the Range search is not working.
    I tried in below ways.
    1. On Action Submit, I set the Occurance_D value as "=DSTR(@DATE_FROM,"M/D/YY")&" - "&DSTR(DADD(@DATE_TO,1,"D"),"M/D/YY")" it is working only if there are processes existing either on DATE_FROM or on DATE_TO, otherwise its not working.
    Like, I have 5 processes triggered on Sep-5-2011, then if i give the Date Range as 9/5/11 - 9/11/11  or 9/1/11 - 9/5/11  it is working. But if I give 9/1/11 - 9/11/11 it is not working.
    I tried this way as well.
    =DSTR(@DATE_FROM,"M/D/YY")&" : "&DSTR(DADD(@DATE_TO,1,"D"),"M/D/YY") . Even this is not working.
    Can some one please help.
    Regards,
    SrinivaS

    Dear SrinivaS,
    The main things you need to consider are:
    - store your date value from the UI in a store
    - the store gets used to submit a filter string in format (MM/DD/YY) to
    BI_START_TIME_D.
    If you follow http://help.sap.com/saphelp_nw72/helpdata/en/81/541300d147
    41b39057c8ab3be61f7a/frameset.htm
    you find that the process data source can be filtered in format  "9/9/11
    - 9/12/11" for field BI_START_TIME_D and BI_END_TIME_D.
    The trick is now at submission time to concatenate the strings
    accordingly and submit to BI_START_TIME_D. Use the following dynamic
    expression
    =DSTR(store@START_DATE_FROM,"MM/DD/YY")&" - "
    &DSTR(store@START_DATE_TO,"MM/DD/YY")
    to format the string during submission. I was able to verify this on my
    machine. I hope that also your locale settings work with this
    implementation.
    Find attached a link to a small sample application [(test_date.zip) |please find attached a small sample application (test_date.zip) which]which shows you the proper usage of a date picker in conjunction with the
    process data source. Please import the zip file into VC and deploy it.
    Best regards,
    Ralf.

  • Looking for a better data-transfer mechanism

    We are currently developing a rich client desktop application.
    The client can talk to back-end server via the web service.
    The web service is responsible for querying data from the back-end database (oracle).
    And for function-like requests from clients, the web service simply submit a job to the oracle db.
    The business logic resides in oracle stored procedure.
    And the data is in the same database.
    When the client request for the data that could make the database to retrieve 1,000 records out of, say 1,000,000 records.
    what we do now is making multiple web service calls, each time for 100 records (in order to make the response data package small enough),
    until no more could be retrieved.
    In detail, we make all the queries appended with an order by clause,
    and we send over the starting index along with the number of records retrieving,
    and use JDBC result set (in the web service method) to re-position and get the needed data,
    then use OracleCachedRowSet to hold the data and then send back to the client.
    The problem here seemed to be significant.
    Multiple web service calls are being wasted.
    Each web service call will make the db run the query (in-efficiently) again, and wasted whole bunch of data.
    In this querying scenario,
    we do not want to separate the data in to different pages (like the table in traditional web-app),
    and we want to know the meta data for the query result (our client application have to make use of the metas).
    Due to our in-experienced background, we can not figure out an efficient mechanism for this querying scenario.
    We thought socket-programming might work for us since once it's opened, the connection will be always be established,
    and then we would not have to waste multiple db queries.
    But socket can not by-pass the firewalls, and we would lost much the benefits introduced by the web service.
    Is there an efficient way of doing this in the web service world?
    Please enlighten us.
    Any suggestion / criticism welcomed.
    Thanks in advance.

    Yes, agree with SAB, zalman 7000 is a good one, however I´ve got 2 very simular K8N Neo2-systems but with 2 different cpu coolers.
    One with zalman 7000alcu and one with Thermalright XP-90 and a zalman 92mm fan ( you can use what ever 80/92mm fan you like high speed, low speed, led colored, etc).
    Both are equal regarding cooling to my experience, but I prefered the XP-90 because I did not have to remove the backplate ( I think zalman recommends mounting it´s own backplate to get the right distance/pressure between the cooler and cpu ).
    With the XP-90 you just switch the retension mechanism which is easy ( even with the MoBo in the case ).

  • Table for Cube & ODS count

    Hi Gurus,
    Could you please tell me the name of the tables to find the Total count of
    1) Cubes
    2) ODS
    3) Queries (Report)
    4) No of total users,
    in BW system.
    Thanks
    Nisha

    Hi,
                                 Table name
    No of total users      USR02
    Cubes                     RSRREPDIR
    Queries                   RSRREPDIR
    Hope this helps...

Maybe you are looking for

  • [CS4] Crossdomain security settings when exporting to SWF

    I'm using CS4 (and CS4 server) to export single pages to SWF files, then I need to embed those SWF to a parent SWF binary (written with Flex) and publish them on a web server. Everything works fine as long as both the parent and children SWF share th

  • Retrieve Image from SQL Server and Store in Oracle

    Hi All I want to retrieve the image stored in sql server as a image type and store that image to oracle as a BLOB type. I tried to retrive the image from sqlserver using getBytes and when storing to Oracle i have to type cast the byte array to BLOB t

  • Safari always remembers wrong email address

    Once I typed my email address wrong (with a '.' in place of a '-') in a form and since then Safari always auto-completes with the wrong address regardless of how many times I've corrected it. I've tried clearing the cache and also resetting(?) Safari

  • Use of Video Camcorder on Skype

    I'm on Tiger 10.4 - and want to use my JVC camcorder through the USB port as the video input over skype. I'm having big problems getting skype to acknowledge my camcorder exists ! Audio is working fine. Have tried macams driver but still not working.

  • RMI on a single computer

    Can I develop and test rmi applications on a single computer without network card?