Data Load Requirement through Process Chain

Hello All,
I have implemented an area through a custom data source and following is my data load requirement through Process Chain. Request you to kindly help me for the same.
1. The design is an InfoCube and updated using a Transformation through the custom data source.
2. For the entire year 2008 I want a single request to be loaded. So it gets loaded into the PSA and then into the Infocube through a (Delta) DTP.
3. Now, I have created an InfoPackage (Full Update) with year as 2009 in the selection. Tht is I want tht hencforth the data load should be for the year 2009.
4. Hence, the Infopackage will run to bring the 2009 data into PSA and I run the (Delta) DTP to update the same in the Cube.
5. Now, what i require is everyday the InfoPackage (Full Update) with 2009 should run and bring data into PSA. and the same should be updated in the InfoCube after deleting the 2009 request already present keeping 2008 request loaded previously intact.
I hope the above is nt confusing.
Please let me know if i can elaborate the same.
Thank you.
Regards,
Kunal Gandhi

Hi,
Please go through the links.
http://help.sap.com/saphelp_nw04/Helpdata/EN/21/15843b74f7be0fe10000000a114084/frameset.htm
http://help.sap.com/saphelp_nw04/Helpdata/EN/f8/e5603801be792de10000009b38f842/frameset.htm
http://help.sap.com/saphelp_nw04/Helpdata/EN/b0/078f3b0e8d4762e10000000a11402f/frameset.htm
These may help you in designing the process chain as required.
Regards,
Sunil
Edited by: sunil kumar on Apr 17, 2009 6:20 PM

Similar Messages

  • Steps for master data loading by using process chain

    Hi
    I want to load the Master data by using the process chain . Can anyone tell me the step-by-step method for loading the master data .
    I means what are the processes should i need to include in the process chain . Pls tell me in sequenece .
    i ll assing the points
    kumar

    Hi,
    The process os loading master data is similar to transaction data in terms of using the infopackages.
    First you have to load Attr followed by texts and hierarchy.For master data you have to include the process type called as Attribute change run.This can be included as a separate process type or using the AND process.
    Hope this helps.
    Assign points if useful.
    Regards,
    Venkat

  • Master data loading by using process chain

    hi
    I want to load the master data by using the process chain in BI7.0 . How can i do this one .
    Can anyone tell me the step-by-step method for loading the master data.
    kumar

    Hi Kumar,
    check this link to load the master data by using the process chain
    Re: Process Chain : Master Data
    cheers
    Sunil

  • Flat file data loading error using process chain

    Hi SAP Experts,
    I am having a problem loading the flat file data to the cube using process chain. The issue is that when i run the process chain it fails giving the message " Date format error, please enter the date in the format  _.yyyy" . I am using " 0calmonth in the datasource" . Strange is that when i manually execute the infopackage, i dont get any errors and am able to load the same file successfully to the dso and the cube. Is there any special setting for the process chain that i am missing?
    The date format in the flat file is mm/yyyy. I have tried all the options i could including recreating the datasource and invain dont see any success so far. please help me solving this problem as we r in the middle of testing cycle.
    Thanking you all for your quick response and support all the time.
    Kind Regards,
    Sanjeev

    Hi Sanjeev,
    I believe you are opening the .csv file again after saving it. In this case the initial 0 in single digit month (say 02/2010) is getting changed to 2/2010 and the resulting file is not readable to the system. Just do not open the file after you have entered data in the .csv file and saved and closed it. If required, open the file in notepad, but not in excel in case you want to re-check the data.
    Hope this helps.
    regards,
    biplab

  • BPC:: Master data load from BI Process chain

    Hi,
    we are trying to automatize the master data load from BI.
    Now we are using a package with:
    PROMPT(INFILES,,"Import file:",)
    PROMPT(TRANSFORMATION,%TRANSFORMATION%,"Transformation file:",,,Import.xls)
    PROMPT(DIMENSIONNAME,%DIMNAME%,"Dimension name:",,,%DIMS%)
    PROMPT(RADIOBUTTON,%WRITEMODE%,"Write Mode",2,{"Overwirte","Update"},{"1","2"})
    INFO(%TEMPNO1%,%INCREASENO%)
    INFO(%TEMPNO2%,%INCREASENO%)
    TASK(/CPMB/MASTER_CONVERT,OUTPUTNO,%TEMPNO1%)
    TASK(/CPMB/MASTER_CONVERT,FORMULA_FILE_NO,%TEMPNO2%)
    TASK(/CPMB/MASTER_CONVERT,TRANSFORMATIONFILEPATH,%TRANSFORMATION%)
    TASK(/CPMB/MASTER_CONVERT,SUSER,%USER%)
    TASK(/CPMB/MASTER_CONVERT,SAPPSET,%APPSET%)
    TASK(/CPMB/MASTER_CONVERT,SAPP,%APP%)
    TASK(/CPMB/MASTER_CONVERT,FILE,%FILE%)
    TASK(/CPMB/MASTER_CONVERT,DIMNAME,%DIMNAME%)
    TASK(/CPMB/MASTER_LOAD,INPUTNO,%TEMPNO1%)
    TASK(/CPMB/MASTER_LOAD,FORMULA_FILE_NO,%TEMPNO2%)
    TASK(/CPMB/MASTER_LOAD,DIMNAME,%DIMNAME%)
    TASK(/CPMB/MASTER_LOAD,WRITEMODE,%WRITEMODE%)
    But we need to include these tasks into a BI process chain.
    How can we add the INFO statement into a process chain?
    And how can we declare the variables?
    Regards,
    EZ.

    Hi,
    i have followed your recomendation, but when i try to use the process /CPMB/MASTER_CONVERT, with the parameter TRANSFORMATIONFILEPATH and the root of the transformation file as value, i have a new problem. The value only have 60 char, and my root is longer:
    \ROOT\WEBFOLDERS\APPXX\PLANNING\DATAMANAGER\TRANSFORMATIONFILES\trans.xls
    How can we put this root???
    Regards,
    EZ.

  • Data load Error using Process chain

    Hi,
    In NW2004s when I schedule delta loads for 2LIS_11_VAITM using process chains, when there is zero delta, the manage screens of the infocubes show green light with zero records correctly whereas the process chain log display shows red. The details of the process monitor show zero records. Is this right? I assume that even if there is zero delta, the log must show green and all the subsequent processes like activation of ODS, deleting of PSA, construction of indexes must be carried out. Please confirm. Is this a bug?
    Thanks.
    Param

    Hi Victor,
    Can you please advice how you solved issue as we are having similar issue where it works fine when loaded manually but throws communication error when process chain runs, following is error message..
    Communication error: call FM RSSDK_DATA_REMOTE_GET
    Thanks in advance for your help!!
    Sandeep

  • Regarding Master Data Loading by using Process Chain

    hai
    can anyone tell me 'step-by-step process and what are processes , process types' are used for loading the Master Data , Transaction into ODS and into Infocube.
    i ll assing maximum points
    bye
    rizwan

    HI Mohammand,
    1. Master Data loading:
    http://help.sap.com/saphelp_nw04/helpdata/en/3d/320e3d89195c59e10000000a114084/content.htm
    2. Transactional data to ODS:
    http://help.sap.com/saphelp_nw04/helpdata/en/ce/dc87c1d711f846b34e0e42ede5ebb7/content.htm
    3. Transactional data to CUBE:
    http://help.sap.com/saphelp_nw04/helpdata/en/d5/e80d3dbd82f72ce10000000a114084/content.htm
    Hope it Helps
    Srini

  • Unable to load deltas through Process chain(PC)

    Hello Experts ,
    I have scheduled the PC , deltas are not getting loaded from PC.
    Load type   : Immediate , periodic updates : for every 30 min.
    (ii) Stopped the PC and tried to run deltas manually
    Status 0 from 0 records in ODS
    also 0 from 0 records in PSA
    RSA3  contains records.
    1.Can I load all the deltas without PC. (or)
    2.What changes should I take to run deltas through PC.
    Kindly help
    Points will be assigned.

    Hi Vijay,
    Your not getting the records into Bw Right. on Which Functional module ur workin in BW ( i.e ur SD,FI,MM) why bcoz if it is other than FI
    u have to check R/3 side
    1. "Jobcontrol" has done periodically  for every 30min (why bcoz ur pulling data for every 30min right)  <b>OR</b> Not.
    If Jobcontrol has not done u have to to do periodically for every 30 min .
    2. Next u have to pull the data It will work definately.
    I am very much sure about this.
    <b>Do period Jobcontrol for every 30 min in Source system.</b>
    Thanks,
    kiran.
    IT will work surely.

  • How to stop the data loads through process chains

    hi,
    I want to stop all the data loads to BI through Process chains where load happens periodic.
    kindly suggest how can I proceed.

    Hi,
    Goto RSPC find your PC and double click on START then change the timings, i.e. give starting date is 01.01.9999 like that Save and ACtivate the PC, it won't start till 01.01.9999.
    Thanks
    Reddy

  • Loading through Process Chains 2 Delta Loads and 1 Full Load (ODS to Cube).

    Dear All,
    I am loading through Process chains with 2 Delta Loads and 1 Full load from ODS to Cube in 3.5. Am in the development process.
    My loading process is:
    Start - 2 Delta Loads - 1 Full Load - ODS Activation - Delete Index - Further Update - Delete overlapping requests from infocube - Creating Index.
    My question is:
    When am loading for the first am getting some data and for the next load i should get as Zero as there is no data for the next load but am getting same no of records for the next load. May be it is taking data from full upload, i guess. Please, guide me.
    Krishna.

    Hi,
    The reason you are getting the same no. of records is as you said (Full load), after running the delta you got all the changed records but after those two delta's again you have a full load step which will pick whole of the data all over again.
    The reason you are getting same no. of records is:
    1> You are running the chain for the first time.
    2> You ran this delta ip's for the first time, as such while initializing these deltas you might have choosen "Initialization without data transfer", as such now when you ran these deltas for the first time they picked whole of the data.Running a full load after that will also pick the same no. of records too.
    If the two delats you are talking are one after another then is say u got the data because of some changes, since you are loading for a single ods to a cube both your delta and full will pick same "For the first time " during data marting, for they have the same data source(ODS).
    Hope fully this will serve your purpose and will be expedite.
    Thax & Regards
    Vaibhave Sharma
    Edited by: Vaibhave Sharma on Sep 3, 2008 10:28 PM

  • Incorrect data after activating the request through Process chain.

    Dear SDN chaps.
    Today morning. I encountered a strange issue in DSO..
    I have DSO which is updating from the AL11(application server) flat file.
    While i am loading it to PSA there were no issues and after loading it to the DSO there is no issue and its passing through the routine and the data is populating properly in NEW data Table .But after successful activation of  the request through process i am getting the wrong records in active data table.
    Then i deleted the request and reran it manually i mean triggered the DTP and ran the manual activation surprisingly accurate records are coming through manual process..
    I am just wondering why it is not working through process chain and why it is showing incorrect records through process chain execution and how it is showing accurate records through manual uploading process..'
    Could some one please help to come out from this..By the way mine is SAP BI 7 SP20 &SP05 for BW 7.01
    Thanks
    K M R
      

    Hi Pra
    Thanks for your response..
    We are doing PSA deletion and then we are uploading the data to PSA as well as DSO.
    Now the issue is not in the part of loading we are facing the issue in Actiation of request if i am executing the activation through process chain it is sucess but the values are incorrect. If i am doing the manual activation it sucess with correct data.
    Even i tried with a new chain but still i am facing the issue.
    Surprise thing is in new data table the data is perfect in both the ways like manual upate and Process chain update only during activation i am getting incorrect record in the active data table..
    Appreciate your help on this....
    Thanks
    K M R
    Edited by: K M R on Jul 9, 2010 11:09 AM

  • Error in the Source system while data loding through Process Chain

    Hi,
    I am facing issue while data loading for certain extractors through Process chain. Load of 0BPM_WIHEAD, 0BP_LOGHIST, OBPM_OBJREL, 0BPM_DEADLINES(there are 2-3 more extractors) is getting failed daily with message Error occurred in the source system and if we repeat the same it is getting executed successfully.
    Regards,
    Javed

    Hi..
    It means that the extraction job is failing in the source system..
    Please check the Job log in the source system  --> Copy the request nimber --> In the source system go to SM37 --> Give * then the request number......execute --> Check the Job log..
    Look there may be multiple reasons behind this....
    1) Connection problem : Go to SM59 and test the connection, If you don't have the access in SM59...then go to RSA1 --> Source Systems --> Right click --> Connection Parameters --> Connection Test
    2) Or may be work processes are not available and due to which jobs are failing due to time out...
    In both the cases you can check with the Basis Team if they can suggest something....or you can change the process chain scheduling time if possible, and schedule it in such a time when less no of jobs are running..
    Regards,
    Debjani..

  • Broadcasting through process chains

    How to broadcast reports to indivdual mailboxs in the company through process chains. say current data should be broadcasted into individual email account every time a new sales order or purchase order is created or by the end of the day , it should automatically trigger the mail with the report. can anyone help on this.

    Hi,
    Here see the blog details.
    The goal of information broadcasting is to distribute the right information, in the appropriate format, to the right people, through different channels, at the right time.
    With the BEx Broadcaster, you can precalculate queries, query views, Web templates, reports and workbooks and broadcast them by e-mail or to the portal. In addition to the precalculated documents in various formats (HTML, MHTML, ZIP and so on) that contain historical data, you can also generate online links.
    Accessing the Broadcaster
    The broadcaster can also be accessed via the Portal through the delivered BI Role.
    You can determine the Scheduling for Information Broadcaster
    Based on a data change event triggered by a process chain.
    Based on a pre-defined time point.
    Freely definable scheduling.
    Steps to Schedule Information Broadcaster based on a data change event triggered by a process chain
    Create the Query in the Query Designer.
    Execute the report in the BEx Web Analyzer.
    Click on u201CSendu201D to create the settings for broadcasting.
    The Broadcast Wizard takes you through a series of prompts where you supply the key information required to develop a broadcast. At any time you can leave the Wizard and use the standard settings dialogs which offer more settings.
    Then schedule the broadcast. If you start the Broadcaster for a query (or template or workbook) that gets data from an InfoProvider that will be selected in the process chain, you can select the InfoProvider for Scheduling.
    Create the process chain and include the event data change, include process type "Trigger Event Data Change (for Broadcaster), itu2019s available under "Load Process and Post -Processing".
    The Process Chain is created including the process types: (1) Start (2) Execute InfoPackage (3) Delta Data Transfer Process (4) Activate DSO (5) Trigger Event data Change.
    When you create the Variant for the Event Data Change, using checkbox we can indicate when the Broadcast should trigger.
    As soon as that InfoProvider is affected by a Process Chain, the Broadcasting is triggered.
    After successful activation you can now schedule your chain. Press button u201CScheduleu201D or menu u201CExecution -> scheduleu201D. The chain will be scheduled as background job. You can see it in SM37. You will find a job named u201CBI_PROCESS_TRIGGERu201D. Unfortunately every process chain is scheduled with a job with this name. In the job variant you will find which process chain will be executed. During execution the steps defined in RSPCPROCESSCHAIN will be executed one after each other. The execution of the next event is triggered by events defined in the table.  You can watch SM37 for new executed jobs starting with u201CBI_u201D or look at the protocol view of the chain.
    You can monitor the Broadcaster from the SOST transaction.
    Note:
    Depending on authorizations, end-users can schedule their Broadcasting Settings.
    Only for those queries which are checked "Execution with Data Change in the Infoprovider" while you schedule, will be triggered by process chain event.
    You may wish to refer the Note on Settings for Information Broadcasting- 760775
    Hope this help you
    Regards,
    Rakesh

  • Archiving Infocube through Process Chain...

    Hi All,
    I need help in creating process chain for archiving infocube.I am able to archive infocube manually but not getting through process chain.
    Is it possible to archive infocube through Process Chain?If yes then please give steps to create process chain for archiving.
    Thanks in advance.
    Bandana.

    Hi,
    It is possible to Archive data from an infocube via process chain.
    Have a start process followed by Archive data from an infoprovider. Here the trick lies in the variants used for the archive steps used to make the chain.
    Create a process by dragging the "Archive data..." process and give name for the variant and use this variant for writing the archive file, choose your archive process (the same archive process u have created to archive data from the infocube) , as this is write phase do not check "Continue Open Archiving  requests" chk box, and choose "40 Wirte phase completed succesfully option" from the options in Continue Process Until target status.Now enter the required selection conditions. In case u want to reuse this chain give a relative value under 'Primary Time Restriction tab' and save this variant. This is your variant 1.
    Now drag the same 'Archiving process' and create a variant and in this process u need to select "Continue Open Archiving  request" and choose the option from the dropdown list '70 Deletion phase confirmed and request completed' this variant for deleting the data from the infocube. This is your variant 2.
    So now have a Process chain with Start >Archive Process with Variant 1 ( to write the archive file) > Archive Process with Variant 2 ( to delete the data from the cube).
    Thats it !!
    Regards
    Edited by: Ellora Dobbala on Apr 8, 2009 5:28 PM

  • PSA Deletion through Process Chain

    Hi Experts,
    Currently I am working on the BW3.5 version. I would like to delete the old PSA req through Process Chain. I need some clarification. Please provide me your suggestions.  Thanks in advance.
    1)In SAP-BW 3.5 version, there is only one process type u2018Deleting Requests from the PSAu2019 is available to do this activity. This process type can delete the requests from the PSA only or Will it delete PSA and Change Log tables!!!!! In the SDN threads, some people said it delete the PSA requests only and some said  it delete the PSA requests as well as Change Log.
    2)Currently we are having six process chain and each has master & transaction data load(Data load through PSA).
    Similarly I am planning to create six PSA deletion process chain which includes master & transaction data deletion. Here I am got little complication on finding Object Name(PSA Table Name).Please refer the screen shot.Is there any short cut to find full list of Object Name(PSA Table) which is currently available in the Process chain every day load.
    3)On the request selection, i prefer "Only sucessfully booked/updated req.". I did not select"Only those req with errors,that are not booked in a data target". Please share your view on this selection preference.
    http://img818.imageshack.us/img818/3963/psa1.jpg
    Thanks,
    RR

    Hi Murali,
    Thanks for the response. I do understand about the identifying the data source & PSA retention period(days). Let me elaborate my question little more. Thanks.
    1)In SAP-BW 3.5 version, there is only one process type u2018Deleting Requests from the PSAu2019 is available to do this activity. This process type can delete the requests from the PSA only or Will it delete PSA and Change Log tables!!!!! In the SDN threads, some people said it delete the PSA requests only and some said it delete the PSA requests as well as Change Log.If i am not mistaken, In BW 3.5 PSA deletion will leads through Change Log also.Am i right?
    2)Currently we are having six process chain and each has master & transaction data load(Data load through PSA).
    Similarly I am planning to create six PSA deletion process chain which includes master & transaction data deletion. Here I am got little complication on finding Object Name(PSA Table Name).Please refer the screen shot.Is there any short cut to find full list of Object Name(PSA Table) which is currently available in the Process chain every day load.In the very day process chain, we are having many stages in master & transaction load. So I can go through info pkg level and find the data source,then i can identify the PSA table name. But is there any other simplest way to find the PSA table name.
    http://img818.imageshack.us/img818/3963/psa1.jpg
    3)On the request selection, i prefer "Only sucessfully booked/updated req.". I did not select"Only those req with errors,that are not booked in a data target". Please share your view on this selection preference. May i know that which option is optimal on normal business process. If i did not select both options, then will that delete all the request in PSA!!!
    Thanks.
    RR.

Maybe you are looking for