Customer Hierarchy Creation by using Flat file

Hi all,
I am trying to create Customer hierarchy (Transaction - BPH), from flat fileu2026
Here, I find like BPH transaction using FM u2018BUPA_HIERARCHY_SAVEu2019 to create hierarchyu2026
Nowu2026 I want to do same thing through Flat file and my file contains below information
First section:
1. Hierarchy Category
2. Hierarchy Tree
3. Tree description
Second section:
BP ID information likes u2013
1. ABC (Higher level BP) u2013 First node
2. XYZ (Next level BP) u2013 Second node
Here, I want to create customer hierarchy with header details (given in first sections) and need creates nodes (given in second section),
Can any one tell me in which way I need to pass data to FM to create hierarchy?
Thanks in advance!!!
Regards,
Sreenu Valeti

Try using BAPI_CUSTOMER_HIERARCHIE_INS

Similar Messages

  • Steps to load the data by using flat file for hierarchies in BI 7.0

    Hi Gurus,
    steps to load the data by using flat file for hierarchies in BI 7.0

    hi ,
    u will get the steps int he following blog by Prakash Bagali
    Hierarchy Upload from Flat files
    regards,
    Rathy

  • How to use 'flat file extraction' option in Infospoke

    Hello Gurus,
    We are trying to extract the BW data using Infospoke to a 3rd party ETL tool (Datastage). Now in the current BW 3.5 release, the customer is on SP11 and we are not able to extract any data using a 3rd party system option in Infospoke. Due to the fact, we found out from SAP, that since we are getting an error (3rd party system is not set)in infospoke, we have to upgrade the SP level from 11 to either 16 or 17. The customer that we are dealing with is not going to do that until July - 2007.
    So now we have a second option to load the 'flat file' in BW App. server and FTP to Datastage. Is this something workable option?
    When using that option within Infospoke destination Tab, we can see two options. One is to use File option and load the data within the standard DIR HOME. I tried that but I can't see the file and the entire data within the standard directory. Can anyone explain why and what will be the procedure to follow?
    The second option is to use Logical file. I never done that before. Can anyone explain step by step process or provide the detail document and/or link?
    Since we are at the critical timeframe stage, I would appreciate if I can get the faster response from the Gurus.
    I will make sure to give 'reward points' to each and every helpful answer(s).
    thanks in advance,
    Maulesh

    Hi shashi,
    first, let me clarify one thing here! I am not saying '3rd party tool' as an option in infospoke. I know Infospoke provides only two options (DB table, and file). But when you use DB table option, you check mark on 3rd party destination and that's where you define an RFC destination for 3rd party tool. So when you execute an infospoke, it will start extracting data and load into the defined destination. Since we are on SP11 on BW 3.5 release, we are getting an error while executing an Infospoke using 'DB table option'. The error is like '3rd party destination is not set'. When I went to Marketplace and found an OSS note, I found out that we need an upgrade on SP level (upto 17).
       As I mentioned earlier, the client doesn't want to upgrade until next year, and we are at the critical timeframe stage, where we have to make a decision whether we hold this project until the SP upgrade happens or found alternate approach where we can extract the data from BW (somehow) and load into Teradata database via Datastage - BW Pack (3rd party ETL).
       So we might be able to use the second option in Infospoke destination, which is to use 'Flat file' load.
       Now when I use File option with App. server checked, the data loads into default Server and directory on BW. Which is DIR_HOME. I can see those file extracted and loaded into that directory under the transaction AL11. But I can't see any data. Do you think I have to request for an increase in size limit? If I see all the data, how can I FTP from the BW server? Do I have to work with SAP Basis or create a UNIX script?
       When I use Logical file name, I guess I have to follow what you have described on your response, correct? Can i use an existing Logical path and file name? Do you have a 'How to' document on creating YEAR and PER FM? Is it possible that Basis can do this work for us? What happens after we assign a logical Path and logical file name into the destination tab on Infospoke? What is the use of logical file versus file? Overall, I am looking for a process from extracting the data from Infospoke using either file or logical file, all the way to load into 3rd party database.
      looking forward for your response!
    thanks,
    Maulesh

  • How to load hierrarchies using dtp  using flat file  in bw  ineed clear ste

    how to load hierrarchies using dtp  using flat file  in bw  ineed clear steps

    Hi,
    If you want to load InfoObjects in the form of hierarchies, you have to activate the indicator with hierarchies for each of the relevant InfoObjects in the InfoObject maintenance. If necessary, you need to specify whether the whole hierarchy or the hierarchy structure is to be time-dependent, whether the hierarchy can contain intervals, whether additional node attributes are allowed (only when loading using a PSA), and which characteristics are allowed.
    1.Defining the source system from which you want to load data:
    Choose Source System Tree ® Root (Source System) ® Context menu (Right Mouse Button) ® Create.
    For a flat file, choose: File System, Manual Metadata; Data via File Interface.
    2.Defining the InfoSource for which you want to load data:
    Optional: Choose InfoSource Tree ®Root (InfoSources) ® Context Menu (Right Mouse Button) ® Create Application Component.
    Choose InfoSource Tree ® Your Application Component ® Context Menu (Right Mouse Button) ® Create InfoSource ® Direct Update
    Choose an InfoObject from the proposal list, and specify a name and a description.
    3.Assigning the source system to the InfoSource
    Choose InfoSource tree ® Your Application Component ® One of your InfoSources ® Context Menu (Right Mouse Button) ® Assign Source System. You are taken automatically to the transfer structure maintenance.
    The system automatically generates DataSources for the three different data types to which you can load data.
          Attributes
          Texts
          Hierarchies (if the InfoObject has access to hierarchies)
    The system automatically generates the transfer structure, the transfer rules, and the communication structure (for attributes and texts).
    4.Maintaining the transfer structure / transfer rules
    Choose the DataSource to be able to upload hierarchies.
    Idoc transfer method: The system automatically generates a proposal for the DataSource and the transfer structure. This consists of an entry for the InfoObject, for which hierarchies are loaded. With this transfer method, during loading, the structure is converted to the structure of the PSA, which affects performance.
    PSA transfer method: The transfer methods and the communication structure are also generated here. 
    5.Maintaining the hierarchy:
    Choose Hierarchy Maintenance, and specify a technical name and a description of the hierarchy.
    PSA Transfer Method: You have the option here to set the Remove Leaf Value and Node InfoObjects indicator. As a result, characteristic values are not transferred into the hierarchy fields NODENAME, LEAFFROM and LEAFTO as is normally the case, but in their own transfer structure fields.  This option allows you to load characteristic values having a length greater than 32 characters.
    Characteristic values with a length > 32 can be loaded into the PSA, but they cannot be updated in characteristics that have a length >32.
    The node names for pure text nodes remain restricted to 32 characters in the hierarchy (0HIER_NODE characteristic).
    The system automatically generates a table with the following hierarchy format (for sorted hierarchies without removed leaf values and node InfoObjects):
    Description
    Field Name
    Length
    Type
    Node ID
    NODEID
    8
    NUMC
    InfoObject name
    INFOOBJECT
    30
    CHAR
    Node name
    NODENAME
    32
    CHAR
    Catalog ID
    LINK
    1
    CHAR
    Parent node
    PARENTID
    8
    NUMC
    First subnode
    CHILDID
    8
    NUMC
    Next adjacent node
    NEXTID
    8
    NUMC
    Language key
    LANGU
    1
    CHAR
    Description - short
    TXTSH
    20
    CHAR
    Description - medium
    TXTMD
    40
    CHAR
    Description- long
    TXTLG
    60
    CHAR
    The system transfers the settings for the intervals and for time-dependency from the InfoObject maintenance. Depending on which settings you have defined in the InfoObject maintenance, further table fields can be generated from the system.
    The valid from and valid to field is filled if you select Total Hierarchy Time-dependent in the InfoObject maintenance. The time-dependent indicator is activated if you select the Hierarchy Nodes Time-dependent option in the InfoObject maintenance.
    6.Save your entries.
    Depending on which settings you defined in the InfoObject maintenance, additional fields can be generated from the system. Also note the detailed description for Structure of a Flat Hierarchy File for Loading via an IDoc and for Structure of a Flat Hierarchy File for Loading via a PSA.
    Also find the below blog for your reference...
    /people/prakash.bagali/blog/2006/02/07/hierarchy-upload-from-flat-files
    You can load hierarchy using process chain...
    Find the below step by step procedure to load hierarchy using process chain...
    http://help.sap.com/saphelp_nw70/helpdata/EN/3d/320e3d89195c59e10000000a114084/frameset.htm
    Assign points if this helps u...
    Regards,
    KK.
    Edited by: koundinya karanam on Apr 8, 2008 1:08 PM
    Edited by: koundinya karanam on Apr 8, 2008 1:09 PM

  • Loading Hierarchies using flat file

    Hi experts,
    I have requirement to load hierarchies using flat file.
    I have to create hier on Person Responsible, sub nodes are Sales Office, Project Definition, Project Type and Project Definition Description. I have file, i need to load into BW.
    I dont know much about hier loading, can you tell me what are the Attributes to be inserted and how to prepare the hier loading file which contains HIENM, VERSION, HCLASS, DATEFROM and DATETO.
    I dont know these fields are enough to load my hier.
    I have created one hier, when i try to activate it's giving error like ' no data exists'.
    Can anyone tell me how to prepare file to load into hierarchy?
    Thanks in advance,
    Venky

    Hi Venkatesh,
    There are several formats to load a hierarchy via file. The following two are the commonly used formats.
    1) Default
    Node ID                     NODEID
    InfoObject Name     INFOOBJECT
    Node Name     NODENAME
    Link Name     LINK
    HghrLvlNode     PARENTID
    Language Key     LANGU
    Description - Short     TXTSH
    Description - Medium     TXTMD
    Description - Long     TXTLG
    2) Sorted
    Node ID                     NODEID
    InfoObject Name     INFOOBJECT
    Node Name     NODENAME
    Link Name     LINK
    HghrLvlNode     PARENTID
    First Subnode     CHILDID
    Next Node Along     NEXTID
    Language Key     LANGU
    Description - Short     TXTSH
    Description - Medium     TXTMD
    Description - Long     TXTLG
    If Default format is used, the sequence of node IDs appeared in the flat file determines the structure of the hierarchy. Whereas the sorted hierarchy format defines the structure via PARENTID, CHILDID, NEXTID relationship.
    You make the selection in DataSource/Tran.Structure tab of InfoSource Page by clicking on the "Hier. Structure button.
    Good luck!
    Bill

  • Object reference not set to an instance of an object error when generating a schema using flat file schema wizard.

    I have a csv file that I need to generate a schema for. I am trying to generate a schema using flat file schema wizard but I keep getting "Object reference not set to an instance of an object." error when I am clicking on the Next button after
    specifying properties of the child elements on the wizard. At the end I get schema file generated but it contains an empty root record with no child elements.
    I thought may be this is because I didn't have my project checked out from the Visual SourceSafe db first but I tried again with the project checked out and got the same error.
    I also tried creating a brand new project and generating a schema for it but got the same error.
    I am not sure what is causing Null Reference exception to be thrown and there is nothing in the Windows event log that would tell me more about the problem.
    I am using Visual Studio 2008 for my BizTalk development.
    I would appreciate if some has any insides on this issue.

    Hi,
    To test your environment, create a new BizTalk project outside of source control.
    Create a simple csv file on the file system.
    Name,City,State
    Bob,New York,NY
    Use the Flat file schema Wizard to create the flat file schema from your simple csv instance.
    Validate the schema.
    Test the schema using your csv instance.
    This will help you determine if everything is ok with you environment.
    Thanks,
    William

  • How to create the Export Data and Import Data using flat file interface

    Hi,
    Request to let me know based on the requirement below on how to export and import data using flat file interface.....
    Please provide the steps involved for the same.......
    BW/BI - Recovery Process for SNP data. 
    For each SNP InfoProvider,
    create:
    1) Export Data:
    1.a)  Create an export data source, InfoPackage, comm structure, etc. necessary to create an ASCII fixed length flat file on the XI
    ctnhsappdata\iface\SCPI063\Out folder for each SNP InfoProvider. 
    1.b)  All fields in each InfoProvider should be exported and included in the flat file. 
    1.c)  A process chain should be created for each InfoProvider with a start event. 
    1.d)  If the file exists on the target drive it should be overwritten. 
    1.e)  The exported data file name should include the InfoProvider technical name.
    1.f)  Include APO Planning Version, Date of Planning Run, APO Location, Calendar Year/Month, Material and BW Plant as selection criteria.
    2) Import Data:
    2.a) Create a flat file source system InfoPackage, comm structure, etc. necessary to import ASCII fixed length flat files from the XI
    ctnhsappdata\iface\SCPI063\Out folder for each SNP InfoProvider.
    2.b)  All fields for each InfoProvider should be mapped and imported from the flat file.
    2.c)  A process chain should be created for each InfoProvider with a start event. 
    2.d)  The file should be archived in the
    ctnhsappdata\iface\SCPI063\Archive directory.  Each file name should have the date appended in YYYYMMDD format.  Each file should be deleted from the \Out directory after it is archived. 
    Thanks in advance.
    Tyson

    Here's some info on working with plists:
    http://developer.apple.com/documentation/Cocoa/Conceptual/PropertyLists/Introduc tion/chapter1_section1.html
    They can be edited with any text editor. Xcode provides a graphical editor for them - make sure to use the .plist extension so Xcode will recognize it.

  • Deleting master data after loading transactional data using flat file

    Dear All,
    I have loaded transaction data  into an infocube using a flat file . While loading DTP i have checked the  option "load  transactional data with out master data exists" . So transactional data is loaded even if no master data is there in BW.
    While loading the flat file, I made a mistake for DIVISION Characteristic  where original master data value is '04000' , but i loaded the transactional data with value '4000' .Then i later realized after seeing the data from the infocube and deleted the request. then i reloaded data with value '04000'. Till now every thing is fine.
    But when I see the master data for DIVISION , i can see a new entry  with value '4000'.
    My question is how to delete the entry value '4000' from DIVISION. I tried deleting manually this entry from 'maintaining masterdata' , but it is not allowing me to do so .
    I have also checked if any transactional data exists for that value '4000' , as i said earlier I have deleted the transactional data with that values. even tried to delete the entries from the master data table, but i donot see a option to delete entries there.
    Please suggest me on this.
    Regards,
    Veera

    Hi,
    Goto RSA1 right click on the Info object and select Delete Master data. This will delete the master data unused existing in the table.
    If this master data is not used any where else just delete the master data completely with SID option.
    If even this doesnt work you can delete the complete table entire in SE14. But this will wipe out the entire table. Be sure if you wanna do this.
    Hope this helps
    Akhan.

  • How to Use Flat File As source in operation mapping?

    Hi dont want to use an FCC. I have written an JAVA Mapping to convert a flat file to XML.
    But in Interface Mapping How i can test this?

    /*This java mapping will give same output as input payload */
    package com.mapping;
    import java.io.InputStream;
    import java.io.OutputStream;
    import com.sap.aii.mapping.api.AbstractTransformation;
    import com.sap.aii.mapping.api.StreamTransformationException;
    import com.sap.aii.mapping.api.TransformationInput;
    import com.sap.aii.mapping.api.TransformationOutput;
    public class Java_Mapping extends AbstractTransformation {
         public void transform(TransformationInput transformationInput,
                   TransformationOutput transformationOutput)
                   throws StreamTransformationException {
              try {
                   InputStream inputstream = transformationInput.getInputPayload()
                             .getInputStream();
                   OutputStream outputstream = transformationOutput.getOutputPayload()
                             .getOutputStream();
                   byte[] b = new byte[inputstream.available()];
                   inputstream.read(b);
                   String strContent = new String(b);
                   outputstream.write(strContent.getBytes());
              } catch (Exception exception) {
                   exception.printStackTrace();
    Edited by: Raghu Vamsee on Jan 5, 2011 3:46 PM

  • MAPPING USING FLAT FILE SOURCE MODULE

    I created a mapping using a flat file as source and a dimension as target, I've already configure both modules and their locations.
    When I try to deploy this mapping in the deploy manager, it generates the script for the control file and when I hit the deploy button it says that the deploy was succesfull, but I really don�t know what it did.
    Does it put the file in my target schema?, what do I suposs to do next?

    Hi,
    So,i don't have to manually run the sql loader from the command prompt???
    All i have to do is Execute the map normally from the deployment manager and it will load the .ctl file itself.Ok great,i did so but i got the following error:
    Starting Execution BM_DEBIT_CARDS_DAT_MAP
    Starting Task BM_DEBIT_CARDS_DAT_MAP
    C:\Documents and Settings\CORAL 2000\Desktop\Flat filesBM_DEBIT_CARDS_DAT_MAP.ctl (The file access permissions do not allow the specified action.)
    RPE-01008: Recovery of this request is in progress.
    RPE-01009: Completion of this task could not be completed due to an earlier system failure.
    Completing Task BM_DEBIT_CARDS_DAT_MAP
    Completing Execution BM_DEBIT_CARDS_DAT_MAP
    So are there any permissions i should grant or what?
    Dina Nagia

  • Query using flat file in the FROM Statement

    I have a flat file that contains about 10k rows that I need to join to a table with 2million plus rows. My problem is I cannot load the flat file into a Oracle table or create a new table to contain the 10k rows because I lack of permissions to create table on my user account. I have tried using MS-Access to query the two tables using ODBC, but the query times were horribly slow and would time out. I was curious if there is a way of calling the flat file as a table in my SQL statement from Oracle?
    I know the following doesn't work but I was wondering if there is something out there that would allow me to do something like the following?
    example:
    SELECT a.COL1, a.COL2, ff.COLa, FF.COLb
    FROM apple a, c:\temp\flat_file.csv ff
    WHERE a.COLa=ff.COLa
    db version - 10g
    OS - Windows XP Pro
    Thanks in Advance!

    user9339442 wrote:
    I have a flat file that contains about 10k rows that I need to join to a table with 2million plus rows. My problem is I cannot load the flat file into a Oracle table or create a new table to contain the 10k rows because I lack of permissions to create table on my user account. I have tried using MS-Access to query the two tables using ODBC, but the query times were horribly slow and would time out. I was curious if there is a way of calling the flat file as a table in my SQL statement from Oracle?
    I know the following doesn't work but I was wondering if there is something out there that would allow me to do something like the following?
    example:
    SELECT a.COL1, a.COL2, ff.COLa, FF.COLb
    FROM apple a, c:\temp\flat_file.csv ff
    WHERE a.COLa=ff.COLa
    db version - 10g
    OS - Windows XP Pro
    Thanks in Advance!Is there some reason you cannot get the data loaded into the local database?
    If you absolutely cannot (either through creating and loading objects, or having someone else do it for you) i would likely recommend installing your own database (Oracle XE is free and will easily handle your flat file)
    http://www.oracle.com/technetwork/database/express-edition/downloads/index.html
    I would install that, load your data into a schema you create, and then create a database link to the remote database and use that in your query.
    I would highly push you to work with the administrators of the database though and get them to set you up a schema where you can create objects (with access to the objects in other schemas that you require, if only read permitted).

  • Processing records using Flat file ActiveSync resource

    Hi Folks,
    I was working on flat file activeSync resource adapter. Documentation says the following..
    The Flat File Active Sync adapter can track the timestamp of a flat file. In addition, the adapter can archive the last file processed and then compare it to the most recent version. Identity Manager will then act on the accounts that are different in the two files.
    If these features are enabled, the first time Identity Manager polls the source flat file, the system copies the file and places it in the same directory. The copied (archived) file is named FFAS_timestamp.FFAS, with the timestamp indicating the last time the original file was changed. The format of the timestamp is determined by the operating system on which the source file resides.
    On each subsequent poll, Identity Manager compares the timestamp on the original file with the most recent timestamp. If the new timestamp value is the same as the previous value, then the file has not changed, and no further processing is performed until the next poll. If the timestamp values are different, Identity Manager checks for the presence of the FFAS file. If the file does not exist, Identity Manager processes the updated source file as if it were a new file.
    If the timestamps are different and the archived FFAS file exists, Identity Manager compares the source file with the archived file. The comparison will filter any users that have not changed. If a user has changed, then it will be sent through the adapter in the normal manner, and the configured process, correlation and delete rules determine what to do with the user.
    I have set the following attributes in synchronization policy as
    Track Last Processed Timestamp : true     
    Process Differences Only : true
    Unique Key for Diff : <some attribute>
    I have the following flat file..
    PERSONID,accountID,FIRSTNAME,LASTNAME,NATID,BIRTHDATE,EXTERNALID,TYPE,COUNTRY
    10004,,test,4,123444,12/01/1804,204,A,USA
    10005,,test,5,,12/01/1805,205,M,USA
    There is no user account in IDM repository. After processing this file, my code creates the above accounts in IDM.
    I deleted the above two accounts from IDM using admin interface and the accounts are not created again. So far good.
    Now I opened the above flat file and just saved again. It creates the account in IDM.
    According to the above documentation, it should not create the account as the flat file is already processed based on diff operation. Then Why it is creating the account again in IDM? Any logic??
    Your repsonse will be appreciated.

    Hi,
    I am pretty sure this has been working for me but with IAPI.Cancel. Please try to capitalize the C.
    Regards,
    Patrick

  • Query regarding Import Dimensions using Flat File in EPMA

    Hi All,
    I am trying to import dimensions and Dimension properties using a flat text file to the master dimension library. If I try to include any properties (HFM) in the !Members Section, I am getting an error "Input Line ... does not have the expected format of 1 columns". But if I remove the properties and just build the members and hierarchies without any properties definition, I am able to succeed. Can someone guide me on what I might be missing with regards to member properties? Below is the format of the input text file.
    !Section=Dimensions
    'Name,DimensionClass,DimensionAlias,CustomDimensionID
    HFM_Entity,Entity,HFM_Entity,
    !MEMBERS=HFM_Entity
    'Name,Allow Adjustments
    10001000,Y
    10001100,Y
    !HIERARCHIES=HFM_Entity
    'Parent,Child
    Please let me know if any other information is required.
    Cheers,
    HyperionUser

    Hi,
    Have a look in directory \Hyperion\products\Planning\bin\sampleapp
    there is a sample ads file :- SampApp Source Flat File.ads
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Reverse engineering using Flat file

    Hello,
    Can someone help me please.
    My environment,
    I am new to ODI and trying to work with it.
    My source DB = FLAT FILE
    Target DB = MySQL5.1
    I created a new project and models. I have the physical and logical architecture and all the connections are working. In the physical architecture, I specified the path of the csv file as c:\textfile.csv
    I created a new model and a data store and pointed the the file to above directory and file.
    When I click on reverse in column tab, it says
    ERROR: The directory c:\textfile.csv specified in your schema, does not exist.
    I am unable to figure out the problem. Can someone help me please.
    Thank you.
    Edited by: user9946607 on Mar 17, 2010 5:52 PM

    Hi,
    Thank you.
    I was able to solve the error for the file loading, now i have used the lkm as file to sql and LKM sql control append, it says execution session started, but when i go to operator, it fails at Load and gives an error:
    7000 : null : java.sql.SQLException: Could not read heading rows from file
    java.sql.SQLException: Could not read heading rows from file
    at com.sunopsis.jdbc.driver.file.x.<init>(x.java)
    at com.sunopsis.jdbc.driver.file.a.d.d.a(d.java)
    at com.sunopsis.jdbc.driver.file.g.a(g.java)
    at com.sunopsis.jdbc.driver.file.w.executeQuery(w.java)
    at com.sunopsis.sql.SnpsQuery.executeQuery(SnpsQuery.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execCollOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlC.treatTaskTrt(SnpSessTaskSqlC.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i(e.java)
    at com.sunopsis.dwg.cmd.g.y(g.java)
    at com.sunopsis.dwg.cmd.e.run(e.java)
    at java.lang.Thread.run(Unknown Source)

  • Upload Currency Vlaues using Flat File.

    Hi All,
    I need a clarification in Loading the Currency Values through flat files. The flat file is created with the similar table structure as that of TCURR.
    But there is a problem that am facing.
    In SPRO there is an screen to maintain the Exchange rates. After doing the flat file upload in AWB under the source system tab, i checked in SPRO Exchange rates screen to see if the values for Indirect Quotation and Direct Quotations are populated. But it is always populating the exchange value only in Direct Quotation and not in indirect quotation.
    Apart from that i am not able to navigate in that screen as am getting an error message in the Exchange rate screen as :"MAnintain conversion factors for <FROM>/<TO> currency.
    Please let me know how to do a flat file upload of currency values and also on the Indirect/Direct Quot mapping.
    Sajan.M

    Hi Sajan,
    The Steps that you followed is correct but only thing is you have to maintain Exchange Rate Type with use of T-code : RRC1 .
    Maintain the Flatfile same as the table.
    Note : Exchange Rate Type should be the existing one.
    In the Source system, right click the concern source system and choose the transfer exchange rate...and specify your flat file.we can load it from the local or application server.
    But be carefull sometimes it will delete the existing data in the table.
    Regards,
    Suresh

Maybe you are looking for

  • ReaderX, WinXP SP3, Protected Mode for Admins

    Hello, We are having an issue running Reader X on Windows XP SP3 ***IF*** the Windows user is a member of the Local Admin group, it will not open in protected mode. **IF** the Windows user is a member of the local user group, Reader will open in prot

  • Every time i sync my iphone my itunes doesn't respond

    I am trying to sync my iphone, but every time it gets to step 6, it stops responding. I have the most recent itunes software, so i cant figure out what is going on... any input?

  • Configuring signature through IPS Manager in CSM3.0

    I was trying to customize the siganture with IPS manager in CSM3.0.Any changes in CSM3.0 only displays in the window but looks like it is not been applied to IDSM2 which has IPS v5.1.The same change If I make through IDM to idsm2 it works fine. This

  • Regain vector information after import with editable styles turned off

    Hi,  my colleague imported a bunch of prepped psd files into after effects cc and proceeded to animate the seperate layers.  The idea was that our character animator would then go back to these files and add much more to the animation by animating th

  • Losing emails and browsing history since upgraded to Mountain Lion

    Ever since I upgraded to Mountain Lion, I've been losing my emails and even browsed webpages.  Everytime I restarted or boot up my MacPro, I would lose almost two days of emails and web browsing/bookmarks.  Does anyone have this problem?  Thanks in a