SAP R/3 data flow

Hi all,
I am working on the Accounts Payable Rapid mart . Can i have a job that first creates all the .dat files on the SAP working directory and another job that executes the .dat file from the application shared directory without having to again run the R3 data flow
If you didnt get it ..
1st job get the data from the sap r3 table and puts in the data transport ( ie..It writes the .dat file on the working directory of the SAP server).
2nd job gets the .dat file from the application shared directory without having to do the first job again
Is the above method possible if there is a way.
I would really appreciate any comments or explanations on it.
Thanks
OJ

Imagine the following case:
You execute your regular job.
It starts a first dataflow
A first ABAP is started...runs for a while...then is finished.
Now the system knows there is a datafile on the SAP server and wants to get it
Because we configured the datastore to use a custom transfer program as download, the tool expects our bat file to download the file from the SAP server to the DI server
Our custom transfer program shall do nothing else than wait for 15 minutes because we know the file will be copied without our intervention automatically. So we wait and after 15 minutes we return with "success"
DI then assumes the file is copied and starts reading it from the local directory...
The entire trick is do use the custom transfer batch script as a way to wait for the file to be transported automatically. In the real implementation the batch script will not wait but check if the file is finally available....something along those lines.
So one job execution only, no manual intervention.
Got it? Will it work?

Similar Messages

  • Data Flow from SAP Source (ECC) system to SAP BI system

    Hi All,
    I wanted to know how data will be flown from SAP Source system to SAP BI system.Data flow should include
    1) Data will be flown by using the IDOCs?
    2) What all are the interfaces involved while data is transferring?
    3) What will happen exactly, if you execute the PSA?.
    If you have any info on this, could you please post here....I
    Regards,
    K.Krishna Chaitanya.

    Hi Krishna,
    Please go through  this article :
    "http://www.trinay.com/C6747810-561C-4ED6-B85C-8F32CF901602/FinalDownload/DownloadId-C2EB7035A229BFC0BB16C09174241DC8/C6747810-561C-4ED6-B85C-8F32CF901602/SAP%20BW%20Extraction.pdf".
    Hope this answers all the mentioned questions.
    Regards,
    Sarika

  • SRM Product / Material creation - Data flow to SAP MDM

    Dear SAP friends,
    I need your help with 2 issues. Thanks in advance for your collaboration.
    Our client needs to create products and make them available to purchase almost immediately.
    The scenario we had working for the last 2 years was: product creation in SRM (fields: short description / product category / measure unit), that would then be sent to SAP MDM for data enrichment, through BBP_CCM_TRANSFER_CATALOG (JOB running every 1 minute). This solution worked well for the last 2 years.
    Recently, we are having some problems with this job. We noticed that, sometimes SRM doesn´t send the new products to MDM, and we think it´s a report performance problem. The report doesn´t maintain is normal behavior when executed every 1 minute.
    SAP advice is to only run this report once a day. That would obviously require having a minimum time frame of 1 day between product creation and its availability (purchase), which isn´t acceptable in this client business model.
    Question 1: Did anyone have a client with similar requirement? What was the implemented solution?
    Because of report performance problems, we are running it (BBP_CCM_TRANSFER_CATALOG) every 5 minutes. It happens that sometimes, more than one catalog manager creates products in that time frame, and MDIS when importing the products to MDM, creates one unique workflow JOB making it impossible to assign it to both catalog managers for enrichment.
    Question 2: How can we ensure that SRM sends only one product per xml files sent to XI / PI ? Or, perhaps, ensure that XI / PI splits the xml file sent from SRM into several files with one product each.
    Thank you very much for your help in advance,
    Best Regards,
    João S. Carvalho

    Hi meka
    In product costing values flows as follows;
    In product cost calculation raw material prices are costed which comes from material master in MM01
    Product operations: i.e material BOM and master receipe or routings from PP module.From PP module operation costs comes in product costing.
    In CO module: In Tcode KP26 we maintain activity\cost center planning for activities like machine hours, set up hours, labour hours etc.Here we define plan quantity and plan prices.
    In Cost Sheet (KZS2) we maintain base cost elements & overheads and While order creation we calculate planned cost of that particular  material.and then check the variances after getting actual cost of product.
    Hope this will help you and if found useful assign points.
    Let me know if you need more information.
    Regards
    Nilesh.

  • How to derive month/year from date in SAP BW 3.5 data flow

    Hi
    How we can derive cal year/month and fiscal month/year from date in SAP BW 3.5 data flow (we're using transfer and update rule)..
    Thanks,
    PK

    Hi,
    if you have any date filed in source side you can just map to any time char system will automatically convert to target objects.
    please look at the screen shot for understanding. (not 3.x it is 7.x)
    Thanks,
    Phani.

  • Help Required Regarding - SAP Job names using R3 data flows

    We are calling a set of SAP Jobs using R3 data flows in data services. When ever a job fails we first kill the active SAP jobs by logging into SAP and then restarting the Jobs.
    There are about 100 odd SAP jobs that we call using these Data services Jobs so we wanted to kill the jobs using a reusable code on the SAP side by passing the Job name just before every R3 flows just incase its still in active status.
    So wanted to know if there are any short cuts to retrive the set of associated SAP job names because it will be a tedious process to hardcode the SAP job names and pass them as parameters for all the 100 + SAP job names in the custom defined resuable code.
    Any help or advice on this please !!

    The program is not meeting the expectations
    and the problem is due to reflection.Do we know this for certain?
    ... my application gets the class name, field name
    etc. from an XML file so i don't know their method names
    beforehand .
    Now since every class instance corresponds to a row
    in the database and i have to call get and set
    methods of each class instance so the performance
    keeps on degrading as the number of columns and rows increase .
    Can somebody suggest some improvement regarding this
    and regarding creating multiple instances of the same object Class.forName() will be using a hash already, so there is probably not much room for improvement.
    Class.newInstance() probably does not take significantly more processing than a simple "new Fubar();".
    Umpteen reflective method invokations (one per column) for each row/instance - Are you saying these are the problem?
    You can test this easy enough.
    If you comment out the reflective method invocations and leave the rest of your code untouched,
    does your application processing speed up significantly?

  • Oracle to SAP BI with BCS u2013 Best Data flow Design.

    Hi,
      We are a SAP implementation team. We are @ Blue Print stage. My client is a RETAIL Business Gaint. Client has 50 % of Transaction data and Master data in Oracle data base. Now we are moving to BI 7.0 and also has plans to use SAP-BCS.
      We would like to map all the existing Oracle tables to BI. Provide any clue regarding the best Data flow (From Oracle 10G to BI 7.0).
    Your quick and valuable suggestion/ links are highly appreciated.
    Warm Regards,
    Bab

    Hi Ashok,
    You have mentioned that you have a Oracle 10g  system as a data inflow which perfectly sets the platform to extract the data from the Oracle  system to SAP BW system.
    This Can be done through the DB connect ,through where you could select the necessary tables and form them as a datasource in BI system,further creating the usual BI objects on top of the DS.
    Once the Data is in BI ,we could pull the BI cube or form a replica of the cube in BCS application format to use it in the BCS environment.
    Hope this helps,
    Regards,
    Rajesh.

  • Data flow in SAP XI

    Any one there , will provide me the Data flow in SAP XI ,
    thanking you
    sridhar

    Hi sridhar
    1) Sender Adapter & Sender agreement & communication channel
    2) receiver determination
    3) interface determination
    4) message branch
    5) receiver agreement & communication channel
    6) call inbound adapter
    Regards Mario

  • Regarding data flow in SAP XI

    Hello Friends,
    In Landscape where XI connected to system(s), is there any way to find how the data flows through or how it is designed to connect other systems in SAP XI other than SLD
    thanks

    If you are looking for any specific landscape and trying to make out how the message flow, then i think you need
    to get the Technical and functional specification documents and based on that you can get a clear picture of
    how different interface are designed and which systems they are connecting.
    Regards
    Srinivas

  • SAP r3 Data flow

    Any In depth documentation of the extracting from R3 as in how RFC is used and general architecture design of how R3 connection is established( involvement of  RFC)
    And a question that i had
    We use Shared Directory Access option on SAP data store
                             SAP Working Directory - Work\Dir
                             Application Shared Directory -
    servername\foldername
    So the architecture we are currently using we have a middleware which handles all the transport
    So for ex and R3 data flow writes .dat file on the SAP working Directory once its done we have an ABAP code that we run /ose38 to move the file from the working directory to the out directory and there is a shuttle that moves the file to the application shared directory so it is not a one step process
    One step does the extract and after the shuttle moves it to the shared directory  the other job reads the flat file from the shared directory
    Now they want to change the architecture and use as a one step process.
    I want to know what kind of access is needed on the Working directory to move the files to the application shared directory? Any special access
    Does the working directory needs a write access to the shared directory?
    Does DI handle the movement from Working directory to the shared directory or is there any function that needs to be run on the SAP to make this work?

    Sounds like a custom_transfer method in case you find an exe that you can call which itself transports the file.
    https://wiki.sdn.sap.com/wiki/display/BOBJ/ChosingtheTransport+Method

  • Data Flow in SAP BI - Please Help

    Hello All,
    I have to prepare a Data flow in SAP BI for one of the Insurance system application.
    There are 4 flat files coming from the source. Each flat file has records on insurances details like policies and so on.
    Each flat file represents a line of business. Now the data model is newly being made. There will also be historic data for last three years for reporting and current year data is loaded in monthly basis.
    Also the error records will go in different target for reporting for each line of business.
    Please suggest a data flow which which would be made in best.
    Edited by: Syed786 on Oct 23, 2009 5:09 PM
    Edited by: Syed786 on Oct 26, 2009 11:31 AM

    Hi sridhar
    1) Sender Adapter & Sender agreement & communication channel
    2) receiver determination
    3) interface determination
    4) message branch
    5) receiver agreement & communication channel
    6) call inbound adapter
    Regards Mario

  • Need to know how to check the outbound queue in SAP for data flow into CRM

    Hi Specialist's,
    I have changed an entry in Customer Master Data in Sap r/3 and this change has come fine to CRM system.I checked the inbound BDOC in CRM ( Trn:SMW01) and it has come fine into CRM system.I am also able to see the affected field in CRM with new value.
    I need to know how to check this from SAP side i.e. I have checked the outbound IDOC list (WE02) but could not find the record.
    Is there any other way in SAP that i can check?
    Also please let me know if there is any other way except IDOC in SAP to send data from R/3 to CRM?
    Usefull answers will definitely be rewarded with points .
    Thanks,
    Abhinav.

    Hi Abhinav,
    There are no IDOCs generated in R/3 for replication. R/3 system calls BAPI and Function modules remotely on CRM system which are part of the Adaptors provided for middleware.
    Try to deregister the outbound queues in R/3 (Transaction SMQS) and see if you get an entry in the Outbound Queue (Transaction SMQ1). I have not tried this.
    If you don't get any entry in the outbound queue then definitely there would be some FM at R/3 end which would be calling CRM system remotely.
    <b>Reward points if it helps.</b>
    Regards,
    Amit Mishra

  • BODS 3.1 : SAP R/3 data extraction -What is the difference in 2 dataflows?

    Hi.
    Can anyone advise as to what is the difference  in using the data extraction flow for extracting Data from SAP R/3 ?
    1)DF1 >> SAPR/3 (R3/Table -query transformation-dat file) >>query transformation >> target
    This ABAP flow generates a ABAP program and a dat file.
    We can also upload this program and run jobs as execute preloaded option on datastore.
    This works fine.
    2) We also can pull the SAP R/3 table directly.
    DF2>>SAPR/3 table (this has a red arrow like in OHD) >> Query transformation >> target
    THIS ALSO Works fine. And we are able to see the data directly into oracle.
    Which can also be scheduled on a job.
    BUT am unable to understand the purpose of using the different types of data extraction flows.
    When to use which type of flow for data extraction.
    Advantage / disadvantage - over the 2 data flows.
    What we are not understanding is that :
    if we can directly pull data from R/3 table directly thro a query transformation into the target table,
    why use the Flow of creating a R/3 data flow,
    and then do a query transformation again
    and then populate the target database?
    There might be some practical reasons for using these 2 different types of flows in doing the data extraction. Which I would like to understand.  Can anyone advise please.
    Many thanks
    indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 3:25 PM

    Hi Jeff.
    Greetings. And many thanks for your response.
    Generally we pull the entire SAP R/3 table thro query transformation into oracle.
    For which we use R/3 data flow and the ABAP program, which we upload on the R/3 system
    so as to be able to use the option of Execute preloaded - and run the jobs.
    Since we do not have any control on our R/3 servers nor we have anyone on ABAP programming,
    we do not do anything at the SAP R/3 level
    I was doing this trial and error testing on our Worflows for our new requirement
    WF 1 : which has some 15 R/3 TABLES.
    For each table we have created a separate Dataflow.
    And finally in between in some dataflows, wherein, the SAP tables which had lot of rows, i decided to pull it directly,
    by-passing the ABAP flow.
    And still the entire work flow and data extraction happens ok.
    In fact i tried creating a new sample data flow and tested.
    Using direct download and - and also execute preloaded.
    I did not see any major difference in time taken for data extraction;
    Because anyhow we pull the entire Table, then choose whatever we want to bring into oracle thro a view for our BO reporting or aggregate and then bring data as a table for Universe consumption.
    Actually, I was looking at other options to avoid this ABAP generation - and the R/3 data flow because we are having problems on our dev and qa environments - giving delimiter errors.  Whereas in production it works fine. Production environment is a old set up of BODS 3.1. QA and Dev are relatively new enviornments of BODS. Which is having this delimiter error.
    I did not understand how to resolve it as per this post : https://cw.sdn.sap.com/cw/ideas/2596
    And trying to resolve this problem, I ended up with the option of trying to pull directly the R/3 table. Without using ABAP workflow.  Just by trial and error of each and every drag and drop option. Because we had to urgently do a POC and deliver the data for the entire e recruiting module of SAP. 
    I dont know whether i could do this direct pulling of data - for the new job which i have created,
    which has 2 workflows with 15 Dataflows in each worflow.
    And and push this job into production.
    And also whether i could by-pass this ABAP flow and do a direct pulling of R/3 data, in all the Dataflows in the future for ANY of our SAP R/3 data extraction requirement.  And this technical understanding is not clear to us as regards the difference between the 2 flows.  And being new to this whole of ETL - I just wanted to know the pros and cons of this particular data extraction. 
    As advised I shall check the schedules for a week, and then we shall move it probably into production.
    Thanks again.
    Kind Regards
    Indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 7:02 PM

  • Display Data Flow - Short Dump

    Hi all,
    When i select display data flow of any cube...it is going for a short dump.
    I have searched for the answer in previous Forum questions. I could find only for previous BW versions but not for for BI7.
    Could you please let me know the solution for this issue.
    Thanks & Regards,
    Eswari

    Hi All,
    Thank you very much for all of your responces.....
    I am working on Support Package 10.
    Here is the detailed description of the short dump.
    Short text
        The current application triggered a termination with a short dump.
    What happened?
        The current application program detected a situation which really
        should not occur. Therefore, a termination with a short dump was
        triggered on purpose by the key word MESSAGE (type X).
    What can you do?
        Note down which actions and inputs caused the error.
        To process the problem further, contact you SAP system
        administrator.
        Using Transaction ST22 for ABAP Dump Analysis, you can look
        at and manage termination messages, and you can also
        keep them for a long time.
    Error analysis
        Short text of error message:
        GP: Control Framework returned an error; contact system administrator
        Long text of error message:
         Diagnosis
             The Graphical Framework is based on the basis technology known as
             the Control Framework. A method in the Control Framework returned
             an error.
         Procedure
             It probably involves a programming error. You should contact your
             system administrator.
         Procedure for System Administration
             Check the programming of the graphics proxy especially for the
             parameters that were sent and, if necessary, correct your program.
        Technical information about the message:
        Message class....... "APPLG"
        Number.............. 229
        Variable 1.......... " "
        Variable 2.......... " "
        Variable 3.......... " "
        Variable 4.......... " "
    How to correct the error
        Probably the only way to eliminate the error is to correct the program.
        If the error occures in a non-modified SAP program, you may be able to
        find an interim solution in an SAP Note.
        If you have access to SAP Notes, carry out a search with the following
        keywords:
        "MESSAGE_TYPE_X" " "
        "CL_AWB_OBJECT_NET_SAPGUI======CP" or "CL_AWB_OBJECT_NET_SAPGUI======CM005"
        "PBO"
        If you cannot solve the problem yourself and want to send an error
        notification to SAP, include the following information:
        1. The description of the current problem (short dump)
           To save the description, choose "System->List->Save->Local File
        (Unconverted)".
        2. Corresponding system log
           Display the system log by calling transaction SM21.
           Restrict the time interval to 10 minutes before and five minutes
        after the short dump. Then choose "System->List->Save->Local File
        (Unconverted)".
        3. If the problem occurs in a problem of your own or a modified SAP
        program: The source code of the program
           In the editor, choose "Utilities->More
        Utilities->Upload/Download->Download".
       4. Details about the conditions under which the error occurred or which
       actions and input led to the error.
    Thanks,
    Eswari.

  • R/3 data flow is timing out in Data Services

    I have created an R/3 data flow to pull some AP data in from SAP into Data Services.  This data flow outputs to a query object to select columns and then outputs to a table in the repository.  However the connection to SAP is not working correctly.  When I try to process the data flow it just idles for an hour until the SAP timeout throws an error.  Here is the error:
    R/3 CallReceive error <Function Z_AW_RFC_ABAP_INSTALL_AND_RUN: connection closed without message (CM_NO_DATA_RECEIVED)
    I have tested authorizations by adding SAP_ALL to the service account I'm using and the problem persists.
    Also, the transports have all been loaded correctly.
    My thought is that it is related to the setting that controls the method of generating and executing the ABAP code for the data flow, but I can't find any good documentation that describes this, and my trial and error method so far has not produced results.
    Any help is greatly appreciated.
    Thanks,
    Matt

    You can't find any good documentation??? I am working my butt off just.......just kiddin'
    I'd suggest we divide the question into two parts:
    My dataflow takes a very long time, how can I prevent the timeout after an hour? Answer:
    Edit the datastore, there is a flag called "execute in background" to be enabled. With that the abap is submitted as a background spool job, hence does not have the dialog-mode timeout. Another advantage is, you can watch it running by brwosing the spool jobs from the SAP GUI.
    The other question seems to be, why does it take that long even? Answer:
    Either the ABAP takes that long because of the data volume.
    Or the ABAP is not performing well, e.g. join via ABAP loops with the wrong table as inner.
    Another typical reason is to use direct_download as transfer method. This is fine for testing but it takes a very long time to download data via the GUI_DOWNLOAD ABAP function. And the download time would be part of the ABAP execution.
    So my first set of questions would be
    a) How complex is the dataflow, is it just source - query - data_transfer or are there joins, lookups etc?
    b) What is the volume of the table(s)?
    c) What is your transfer method?
    d) Have you had a look at the generated abap? (in the R/3 dataflow open the menu Validation -> Generate ABAP)
    btw, some docs: https://wiki.sdn.sap.com:443/wiki/display/BOBJ/ConnectingtoSAP

  • Data Flow terminated due to error 120307

    Hi.
    I get this error when executed  project.
    Source system: SyBase IQ.
    Target system: SAP HANA.
    Part of tables copied successfully, but job terminated anyway.
    I attached screenshots with Progress screen and Monitoring.
    Error log is empty.
    In trace log I see errors from SUBJ.
    Also I have another one strange message in trace log:
    Cache statistics determined that data flow <SYBASE_IQ_2_HOD_DBA_FACT_FINAL> uses 0 caches with a total size of 0 bytes, which is less than (or equal to) 3757047808 bytes available for caches in virtual memory. Data flow will use IN MEMORY cache type.

    I executed this job from data services designer and get another error.
    main Bufman: An error was detected on a database page. You may have a damaged index. For additional information, please check your IQ message file or run sp_iqcheckdb
    Trying to find issue in goggle

Maybe you are looking for

  • Increasing Mobile Me Gallery on iWeb Page

    I am inserting a MobileMe Gallery into my Web Page. Unfortunately the size is to small 160 x 140. I want to at least double the size. I click on the outer edge and find the 6 small boxes that would normally allow increase/decrease and no change of cu

  • TS3276 Since upgrading to Mountain Lion, Mail does not let me send e-mails using Comcast mail server.  Why?

    Why is this?  Comcast's SMTP outgoing mail server is set for my e-mai account but I keep getting an Apple error dialog saying the server can't be used.

  • Macbook freezing on certain programmes.

    So my Macbook Pro is freezing when I play games such as "League of Legends", depending whether its full screen or windowed it freezes, so if its on full screen at the time I have to restart, but if its windowed I can close and reopen. My Mac sounds l

  • Using iphoto with iview

    I am trying to set up a simple system for archiving and backing up photos using the best combination of iview and iphoto. As far as i understand, iphoto makes copies of the original photos to store in its own folder, and iview simply mnakes pointers

  • Officejet 4500 g510n-z driver download link?

    Just wondering if there was a direct download link for Officejet 4500 g510n-z? I've been trying all morning and haven't been able to from the HP site. Keeps getting stuck at .......compiling your results.  Thanks in advance.