How do name PSA in data flow

Hi guys,
I am preparing landscape, I want know how do we mention the name for PSA,
PSA  -- 0FI_AR_03_XX,
  here in  place of  XX  what I need to mentioned .
thanks in advance
Ram

Hey Ram,
It is based on the source system.
When you go to Tools menu you have an option to see assign source system ID.
This would be the ID which comes after that underscore.
It varies between Dev/QA/Prod
Better way is to mention PSA for (XXXXXX) instead of the technical name
Mey

Similar Messages

  • Help Required Regarding - SAP Job names using R3 data flows

    We are calling a set of SAP Jobs using R3 data flows in data services. When ever a job fails we first kill the active SAP jobs by logging into SAP and then restarting the Jobs.
    There are about 100 odd SAP jobs that we call using these Data services Jobs so we wanted to kill the jobs using a reusable code on the SAP side by passing the Job name just before every R3 flows just incase its still in active status.
    So wanted to know if there are any short cuts to retrive the set of associated SAP job names because it will be a tedious process to hardcode the SAP job names and pass them as parameters for all the 100 + SAP job names in the custom defined resuable code.
    Any help or advice on this please !!

    The program is not meeting the expectations
    and the problem is due to reflection.Do we know this for certain?
    ... my application gets the class name, field name
    etc. from an XML file so i don't know their method names
    beforehand .
    Now since every class instance corresponds to a row
    in the database and i have to call get and set
    methods of each class instance so the performance
    keeps on degrading as the number of columns and rows increase .
    Can somebody suggest some improvement regarding this
    and regarding creating multiple instances of the same object Class.forName() will be using a hash already, so there is probably not much room for improvement.
    Class.newInstance() probably does not take significantly more processing than a simple "new Fubar();".
    Umpteen reflective method invokations (one per column) for each row/instance - Are you saying these are the problem?
    You can test this easy enough.
    If you comment out the reflective method invocations and leave the rest of your code untouched,
    does your application processing speed up significantly?

  • How Make a software Design Data Flow Diagrams (DFD)..?

    I want a build up software for design DFD (Data Flow Diagrams). Can you help show to me a simple script for design DFD (Data Flow Diagrams). Thanks for all..
    My mail : [email protected]

    -Also the problem is how to do various validations of objects(line, oval, process) and move these objects using mouse? Which all specific events are needed ?This will help you [http://java.sun.com/docs/books/tutorial/uiswing/painting/step3.html]
    -Another problem is how to have mutiple windows framework in applet. By clicking on the process another level corresponding to the process(i.e. its sub-process) will open?I think JInternalFrames would do.
    I general please refer to [The Swing Tutorial|http://java.sun.com/docs/books/tutorial/uiswing/], it has answers to many (all) swing related problems. Go through it.
    Thanks!

  • How to run just the data flow

    We are in the process of converting from one ETL tool to Data Services 3.0.  I have a workflow with several data flows underneath it and have to make a change to the data flow.  Is there a way to run just the dataflow without executing the whole job?  For example, I just want to run DF4 below.
    WF --> DF1 --> DF2 --> DF3 --> DF4 --> DF5
    Thanks,
    Dan

    Although this topic is solved, just a few clarifications:
    When you create a new job and drag 'n drop the workflow/dataflow into it, you are not creating a second object (class), you are calling the object a second time (instantiate). It is like a BASIC program. I created a sub-procedure and called in DF1. In my filesystem I can see that sub-procedure as a seperate file in the folder object_library -> dataflow_procedures.
    Then I have two main programs, both just calling that sub-procedure with "gosub DF1".
    In the job you see the calls, when you drill into DF1 you see its definition and what object it calls.
    Using a conditional is okay but not adviced in my opinion. You are losing too much, the handling is complex and development takes a few seconds longer. my two cents only.
    https://wiki.sdn.sap.com/wiki/display/BOBJ/Testing

  • How to Restrict the consumer data flow from CRM to R/3

    Hi Experts
    I have a scenario where my client doesn't wish to flow consumers data into R/3 which are created in CRM's B2C shop.He wants to map consumer data with R/3 One time customers so that further process such as Delivery  and all take place and this status gets updated in B2C shop via CRM
    hence i request you all could you please help me to solve my issue
    i would appreciate you help
    Thanking you in advance
    Regards
    Rao.

    Hi Mark
    thanks very much for you reply
    See my scenario is i have set up the B2C SHOP  in CRM not in ECC  as and when consumer gets  register in B2C SHOP  he creates an Order,  so standard functionality is that  CRM  Order needs to flow into R/3 along with  consumer data such as (sold_to,ship_to_bill_to---) for further process like delivery and billing.
    Now my requirement is client does not wish consumer data should be flown into R/3 for which he wants to map consumer data which is created in CRM to  R/3 with One Time customer for delivery, so that when ever consumer gets created in B2C SHOP it should map with the same R/3 One time customers
    which means that very time one time customer data should be replaced with consumer data which are created in B2C shop.and delivery should take place
    so could you please help me is there any way to resolve my issue
    Thanking you
    Regards
    Rao

  • Dynamically pick the table names in data flow task SSIS

    Hi All,
    I want to create a SSIS package which loads the data to a table on the other server every day. I have around 250 tables to load everyday and source and destination table names are available in a metadata table, table names have to read from the metadata
    table and data should be loaded. Is there a way that we can configure the source and destination table names dynamically in Data flow task?
    I am newbie to SSIS can any help with the solution for this problem.

    You can do that, not a big deal. The underlying problem is say suppose you constructed a ETL based on some x source and y destination and have put x(3 columns) as source and y destination(3 columns).. As you said if we have choice of dynamically pick the
    table names.. ETL might fail when you face below situation
    source x(3 columns)   destination y (4 columns) and there will be no mapping as it is dynamic. Sometimes mapping also fails even if the source and destination have same number of columns. If you still want to do... follow below steps:
    Create two variables:
    1.variable1 , datatype string
    2.variable datatype string  
    take one execute sql task, pick your source table names dynamically as you desired from metadata table \
    "SELECT sourcetblname as Res FROM @metadata WHERE ID=1" in sql statement and then go to name the result name as Res (I meant same as table alias) and map it to variablename1
    And in variable2 go to expression and write "Select * from "+@[User::variable].. and this will be your constructed dynamic command for oledb destination.
    And connect that execute sql task to (Data flow task)oledb source and choose data access mode as sql command with variable,and choose variable2. below is the diagram.
    - please mark correct answers

  • Customised data flow for SPM

    Hello ,
    Wanted to gain some insights on how to accomodate a customised data flow into SPM UI. If needed can we create a new flow and have a bex query on top of that which can then be accomodated into SPM UI? Also how will we manage the data upload set up in this case?

    Hi Gundeep
    Have you looked at the SPM BPx site for all relevant resources? There are some good ones that will help you get an understanding.
    Here are some new ones on exactly the subject you are referring to:
    /people/rajesh.prabhu/blog/2010/12/29/spm-application-loosely-coupled-to-bring-about-a-perfect-fit
    /people/rajesh.prabhu/blog/2010/12/23/spm-customization-series-loading-data-in-customer-defined-dimension (this talks about loading customer defined dimension or master data, I will later create another blog for customer defined master data, but a lot of conecpts are similar)
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/e0fda858-48f5-2d10-d4af-8885c4d18c0e
    /people/rajesh.prabhu/blog/2010/12/23/upload-types-in-spm
    As usual the SPM BPx site is at:
    Spend Analytics
    A consolidated list of all resources (duplicated out of the BPx site) for SPM can be found at:
    https://wiki.sdn.sap.com/wiki/display/CPM/AllSPMresourcesonSDN
    Regards
    Rajesh

  • Building a data flow task, within a foreach loop for dynamic table name, but ole db source not allowing variable

    In my control flow, I set up a variable for the table name, enumerated by SMO, following the instructions from the link here:
    http://www.bidn.com/blogs/mikedavis/ssis/156/using-a-for-each-loop-on-tables-ssis
    Now, I put a data flow task inside the foreach. I selected the OLE DB connection manger for my database, set the Data access mode to "Table name or view name variable", and selected my variable name from the drop down. So far so good. When I click on OK,
    it gives me an error 0x80040E37, basically saying it can't open the rowset for "my variable", Check that the object exists in the database.
    So, I assume I won't be able to do this "that' easily, and I will need to build a "SQL command from variable" or some such thing. Any advice on how to build this Source editor to dynamically name my columns from the variable?
    Thanks in advance!
    mpleaf

    Hi mpleaf,
    Please try to set "ValidateExternalData" to False in your OLE DB Source Properties and "DelayValidation" property to TRUE, please refer to similar threads:
    http://social.msdn.microsoft.com/Forums/en-US/sqlintegrationservices/thread/620557d9-41bc-4a40-86d5-0a8d2f910d8c/
    http://social.msdn.microsoft.com/Forums/en-US/sqlintegrationservices/thread/456f2201-447c-41b3-bf27-b3ba47e6b737
    Thanks,
    Eileen
    Eileen Zhao
    TechNet Community Support

  • How to push a perticular request from PSA to Data target using DTP in BI

    Hello all.
      I have mulitple requests in PSA & wanted to move only perticular requests (Randomly,  say first request & last request ) to data target using DTP in BI 7.0. How can we do this.. ? Is there any option in DTP (BI 7.0)for selective loading of requests from PSA to Data Target..
      Thanks in advance..
    Cheers,
    sami.

    Hi,
      It is possible by using 'zero fetch' in DTP
    If you want only the recent req to be loaded.
    1. In PSA, change the status of the recent request to red
    2. Do a zero fetch
    that is processing mode - No data transfer; delta status in source: fetched
    With this processing mode you execute a delta without transferring data.
    3. change the particular request status back to green
    4. Run a delta load
    Regards,
    Priya.D

  • How to delete the duplicate data  from PSA Table

    Dear All,
    How to delete the duplicate data  from PSA Table, I have the purchase cube and I am getting the data from Item data source.
    In PSA table, I found the some cancellation records for that particular records quantity  would be negative for the same record value would be positive.
    Due to this reason the quantity is updated to target but the values would summarized and got  the summarized value  of all normal and cancellation .
    Please let me know the solution how to delete the data while updating to the target.
    Thanks
    Regards,
    Sai

    Hi,
    in deleting the records in PSA table difficult and how many you will the delete.
    you can achieve the different ways.
    1. creating the DSO maintain the some key fields it will overwrite the based on key fields.
    2. you can write the ABAP logic deleting the duplicate records at info package level check with the your ABAPer.
    3.you can restrict the cancellation records at query level.
    Thanks,
    Phani.

  • How to derive month/year from date in SAP BW 3.5 data flow

    Hi
    How we can derive cal year/month and fiscal month/year from date in SAP BW 3.5 data flow (we're using transfer and update rule)..
    Thanks,
    PK

    Hi,
    if you have any date filed in source side you can just map to any time char system will automatically convert to target objects.
    please look at the screen shot for understanding. (not 3.x it is 7.x)
    Thanks,
    Phani.

  • How can i export the data to excel which has 2 tables with same number of columns & column names?

    Hi everyone, again landed up with a problem.
    After trying a lot to do it myself, finally decided to post here..
    I have created a form in form builder 6i, in which on clicking a button the data gets exported to excel sheet.
    It is working fine with a single table. The problem now is that i am unable to do the same with 2 tables.
    Because both the tables have same number of columns & column names.
    Below are 2 tables with column names:
    Table-1 (MONTHLY_PART_1)
    Table-2 (MONTHLY_PART_2)
    SL_NO
    SL_NO
    COMP
    COMP
    DUE_DATE
    DUE_DATE
    U-1
    U-1
    U-2
    U-2
    U-4
    U-4
    U-20
    U-20
    U-25
    U-25
    Since both the tables have same column names, I'm getting the following error :
    Error 402 at line 103, column 4
      alias required in SELECT list of cursor to avoid duplicate column names.
    So How can i export the data to excel which has 2 tables with same number of columns & column names?
    Should i paste the code? Should i post this query in 'SQL and PL/SQL' Forum?
    Help me with this please.
    Thank You.

    You'll have to *alias* your columns, not prefix it with the table names:
    $[CHE_TEST@asterix1_impl] r
      1  declare
      2    cursor cData is
      3      with data as (
      4        select 1 id, 'test1' val1, 'a' val2 from dual
      5        union all
      6        select 1 id, '1test' val1, 'b' val2 from dual
      7        union all
      8        select 2 id, 'test2' val1, 'a' val2 from dual
      9        union all
    10        select 2 id, '2test' val1, 'b' val2 from dual
    11      )
    12      select a.id, b.id, a.val1, b.val1, a.val2, b.val2
    13      from data a, data b
    14      where a.id = b.id
    15      and a.val2 = 'a'
    16      and b.val2 = 'b';
    17  begin
    18    for rData in cData loop
    19      null;
    20    end loop;
    21* end;
      for rData in cData loop
    ERROR at line 18:
    ORA-06550: line 18, column 3:
    PLS-00402: alias required in SELECT list of cursor to avoid duplicate column names
    ORA-06550: line 18, column 3:
    PL/SQL: Statement ignored
    $[CHE_TEST@asterix1_impl] r
      1  declare
      2    cursor cData is
      3      with data as (
      4        select 1 id, 'test1' val1, 'a' val2 from dual
      5        union all
      6        select 1 id, '1test' val1, 'b' val2 from dual
      7        union all
      8        select 2 id, 'test2' val1, 'a' val2 from dual
      9        union all
    10        select 2 id, '2test' val1, 'b' val2 from dual
    11      )
    12      select a.id a_id, b.id b_id, a.val1 a_val1, b.val1 b_val1, a.val2 a_val2, b.val2 b_val2
    13      from data a, data b
    14      where a.id = b.id
    15      and a.val2 = 'a'
    16      and b.val2 = 'b';
    17  begin
    18    for rData in cData loop
    19      null;
    20    end loop;
    21* end;
    PL/SQL procedure successfully completed.
    cheers

  • How to push the data from PSA to Data Target in 3.5 version

    Hi Experts,
    I'm currently working on BI 7. I have some master data load being maintained in 3.5 version.
    The infopackage has an option to load the data to psa and data target when extracted.
    I want to stage the data at PSA only.
    Later I want to trigger it to the target.
    In 7 version we use DTP but I'm not very sure what should I do here.
    Can anyone please help me to push the data to the target please using 3.5 version.
    Thanks

    Hi pathak,
    Thanks for your reply. I'm able to do it.
    Can you please tell me how do I filter the load to target.
    We have filter in DTP where i can restrict the load.
    Do we have any option in 3.5.
    Please advice.
    Thanks

  • How can I fix consistent TCP timeout and make data flow simpler?

    Hi!
    I'm acquiring data from a Scanivalve Digital Scanning Array through a TCP/IP connection. I'm having problems with the connection timing out regularly. It will run fine if you take data several times in a row, but if the VI sits for several minutes (while looking at previous data runs or adjusting test setup) the connection will timeout the next time you try to take data. The subVIs (reading the tcp connection, processing the data pakcet, etc) were provided by the manufacturer several years ago. The midlevel VI were written by another engineer, and I adapted them to run with the top level VI that I needed for this test. The data flow in the VIs are convoluted and while I can follow the ones t
    hat I wrote, I'm not sure how to troubleshoot the others.
    I posted a question on an unrelated problem and the responses mentioned that race conditions were going to be a problem b/c of all of the global variables.
    Please be patient with my lack of knowledge in some areas. Any help would be appreciated.
    Thanks so much!
    -Sarah
    Engineering intern
    Techsburg, Inc.
    Attachments:
    DSA_Acquistion_VIs.zip ‏253 KB

    I'm not positive about the best solution in this situation, but there is lots of information available regarding error 56 when using TCP/IP communication.
    You might find some of these useful:
    Error 56 Occurred at TCP Open: Windows XP Fails as TCP/IP Server with LabVIEW 6.1.
    Error 56 Occurs When Using TCP Listen.vi
    TCP/IP Error Codes and Related Time-out Issues in LabVIEW

  • How data flow when SSIS packages are run on a different server than the DB server

    The scenario is that i have a dedicated SQL SErver 2014 SSIS machine that executes the packages.
    The database server is a separate machine with SQL Server 2008R2.
    1) Running SSIS packages that transfer data within SQL Server 2008R2 (same machine)
    2) Running SSIS packages that transfer data between 2 separate SQL Server servers.
    How the data flow in these two cases and what resource is being used where? (cpu,disk,ram,network)
    Elias

    When you have a dedicated SSIS server, all data read flows to that server, is processed using the resources of that ETL server and then sent back over the network to the destination server.
    It doesn't matter if source and destination are the same server. If you use a data flow, all data flows over the network twice.
    The only exception is when you don't use a data flow, but only SQL statements. In that case, data flows only between source and destination.
    MCSE SQL Server 2012 - Please mark posts as answered where appropriate.

Maybe you are looking for