About Importing the Data

Hello Expert
                           I have to Import the Incoming Payment For Particular Customer i have Used ORCT template for importing the record but what i need is to also clear the incoming Payment for this customer so that the Particular Invoice Status of this customer should be closed, How Should i do this which other Template to use or which field should be filled in the (ORCT Payment) template can anybody Explain with example .

Hi......
If you want to make payment against a particular Invoice then along with ORCT you can use RCT2 for adjustment of Invoices in which you have to put the document entry in DocEntry column of RCT2 Template which links invoice to the ORCT payment and the invoice will be closed.
Regards,
Rahul

Similar Messages

  • How to import the data in a .xls or .xlsx file into a oracle database table

    Hi,
    Please tell me how to import the data in a .xls or .xlsx file into a oracle database table in Oracle 10gR2 using Oracle Warehouse Builder 10gR2.

    ....can we do something through Non-Oracle->ODBC?Yes, it is possible, look at this thread
    [SQLServer access from AIX Warehouse builder|http://forums.oracle.com/forums/thread.jspa?messageID=2502982]
    If your server (with target DB and OWB runtime) is on Windows OS this configuration will be simpler - you can use single server.
    And additional link on OWB blog (with 11g transparent gateway)
    [http://blogs.oracle.com/warehousebuilder/2008/01/11g_heterogeneous_agent.html]
    (configuring nonoracle connection with 10g generic connectivity very similar to 11g gateway)
    Also look at
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:4406709207206#18830681837358
    Regards,
    Oleg

  • Can we use impdp to import the data from an normal exp dump?

    Hi All,
    I have a export dump taken from a 9i database. Can i use impdp to import the data into 10g database from that 9i exp dump?
    Please suggest
    thanks and Regards
    Arun

    Hi,
    I have a export dump taken from a 9i database. Can i use impdp to import the data into 10g database from that 9i exp dump?Yes, it can be.
    Refer:
    http://wiki.oracle.com/thread/3734722/can+a+9i+dump+file+be+imported+into+10%3F
    thanks,
    X A H E E R

  • Error while importing the data

    Hi All,
    In fdm i am trying to Import the data but i am getting the below error while importing text file.
    Error: An error occurred importing the file.
    Detail: Data access error.
    also the below is the log found ..
    2012-02-14-01:04:42
    User ID...........     admin
    Location..........     mine
    Source File.......     C:\Hyperion\products\FinancialDataQuality\Mineapp\Test2\Inbox\mine\load_demo2.txt
    Processing Codes:
    BLANK............. Line is blank or empty.
    ESD............... Excluded String Detected, SKIP Field value was found.
    NN................ Non-Numeric, Amount field contains non numeric characters.
    RFM............... Required Field Missing.
    TC................ Type Conversion, Amount field could be converted to a number.
    ZP................ Zero Suppress, Amount field contains a 0 value and zero suppress is ON.
    Create Output File Start: [2012-02-14-01:04:42]
    [TC] - [Amount=NN]     Accounts,Market,Year,Product,Scenario,Data
    [Blank] -      
    Excluded Record Count..............1
    Blank Record Count.................1
    Total Records Bypassed.............2
    Valid Records......................2
    Total Records Processed............4
    Begin SQL Server (BCP) Process (2): [2012-02-14-01:04:42]
    [RDMS Bulk Load Error Begin]
         Message:      (-2147217900) - Data access error.
         See Bulk Load File:      C:\Hyperion\products\FinancialDataQuality\Mineapp\Test2\Inbox\tWadmin56465844032.tmp
    [RDMS Bulk Load Error End]
    Please help me out i am really helpless and dont know how to solve this error.

    When you perform a bulk load, you basically tell SQL Server to execute a command local to the SQL server. What that means is that if you pass it a parameter that says load the file :
    C:\Hyperion\products\FinancialDataQuality\Mineapp\Test2\Inbox\mine\load_demo2.txt
    SQL Server is going to look on it's C drive for the file.
    Unless you have SQL Server running on the FDM server, this is going to fail obviously.
    The 'right' thing to do is setup your FDM app using a share name. (i.e. \\FDMSERVER\FDMAPPS\Mineapp instead of c:\Hyperion\products\FinancialDataQuality\Mineapp).
    If you are relatively handy with FDM, there is a config file (in XML) that has the path information for the apps and you just need to change it from the hardcoded C drive reference to the share you create.)
    Your other option is to NOT load the fiels as bulk. If you have a lot of large data files being loaded, this would be a bad idea. If you have relatively small load files every month and not a lot of users on the system, this may not impact performance. I would update the path information if it were me.
    Charles

  • Can I import the data that is dumped from Oracle 8.1.5 into 8.0.4?

    I got the following message while trying to import the data into 8.0.4:
    Connected to: Oracle8 Enterprise Edition Release 8.0.4.0.0 - Production
    With the Partitioning and Objects options
    PL/SQL Release 8.0.4.0.0 - Production
    Export file created by EXPORT:V08.01.05 via conventional path
    IMP-00008: unrecognized statement in the export file:
    est">
    IMP-00008: unrecognized statement in the export file:
    afds
    IMP-00008: unrecognized statement in the export file:
    .9
    IMP-00008: unrecognized statement in the export file:
    .9~⌂
    IMP-00009: abnormal end of export file
    IMP-00034: Warning: FromUser "SHARP" not found in export file
    Import terminated successfully with warnings.
    I appriciate if any help.
    Thanks,
    Anup.
    null

    When I want to do an export from Oracle 8.0.6 to Oracle 7.3.4, I use the exp and imp from Oracle 7.3.4. This way the whole excercise works fine. The only thing I have to do is run the script catexp7.sql on the Oracle 8.0.6 version (delivered with Oracle 8.0.6). Perhaps You can try the same thing with Oracle 8.0.4, use the exp and imp from 8.0.4 to do the work.

  • How do I import the data from Excel (.xls) file into Oracle Database table

    I have an excel file with 5 columns of 5000 rows.
    I have to import 5000 rows of data into my oracle table with
    same column names & Data types.
    What is the best advise for this ?
    Thanks.
    Srini

    You can connect to Oracle from Excel directly. I know you can read data but am not sure if you can export data.
    Other option is to import the data into Access.
    Create an ODBC connection for Oracle
    Use the to connect to Oracle from Access
    Export data to Oracle from Access
    Shakti
    http://www.impact-sol.com
    Developers of Guggi Oracle - Tool for DBAs and Developers

  • HCI error - unable to import the data because the row is duplicated in the master data

    Hi, i am working on HCI while i am mapping the data from ECC to HCI target table i am getting this error :Unable to import the data because the row is duplicated in the master data .Could you please help me how to solve this
    Message was edited by: Mariana Mihaylova

    Moved to Process Integration (PI) & SOA Middleware as per New SAP HANA Cloud Integration (HCI) content category (unfortunately, I can't edit the discussion and add the category as I'm not a moderator in this space). Perhaps Mariana Mihaylova can help.
    Susmitha Yerradhodi please read and follow The SCN Rules of Engagement - in particular, the more details you provide, the better the chances of getting your question answered. (e.g. "HCI" is not a helpful subject)

  • How to import the data to Data manager

    Hi All,
    I am new to MDM and need your help in handling the following issue
    In our setup we are sending the Customer master data from R/3->XI->MDM (R/3-> XI IDOCs,  XI->MDM XML files).
    with the help of extraction job in R/3 IDOCs r generated and XI has onverted these IDOCs to XML Files and created at MDM Server. Now my Question is how to import the Data from these XML files to Data Manager !!!!
    Note : Import Map settings are done already in the system.
    Can any one help me the steps requeired to get this data to Data manager .
    and am I need to do some activity in MDM to import the data to Data manager each and every time or is there any automatic way of doing this .
    Please guide me.
    Thanks in advance!

    Hi MDM User,
    Generally there are 2 cases while importing the data,
    1) Initial Load
    2) Delta Load
    Initial Load means importing the data into MDM at first time.This is usually done by manual importing
    without using import server.
    Delta Load means importing the data into MDM when we have changes to existing data or added new data
    to existing data.This is usually done by automatically using import server.
    To manually import the data,just create a remote system and port in console and login into import manager
    using remote system.Create ,save the map and execute the map.There is no other configurations needed.
    To automatically import the data without manual intervention,we go for import server.
    To let this happen we need to set some configuarations in mdis.ini file of import server.
    Give the repository name,username and password which are set in console for that repository and also
    give schedule time based on which import server picks the file from ready folder.
    Go to the import manager , change the map and import action depends on your requriment and save that
    map.
    and also you need to set port type as inbound and automatic and select the map which you saved earlier in
    MDM console.
    As soon as we start the import server, it picks the file from ready folder of MDM server based on
    configurations which we set in mdis.ini file and executes the map.
    Here i am giving complete path of ready folder.
    SAP MDM 5.5> Server>Distributions>bp1bocap080.bp.co_MSQL> BP3_PoC_Customer> Inbound>Siebel> SIEBEL_OB_CUS_SIEBELCUS01> Ready)
    This path has to be given in target directory of receiver communication channel of XI so that it puts XML
    file in the above path.
    For better understanding,go through below blogs.
    /people/balas.gorla/blog/2007/02/05/r3-xi-mdm-outbound-scenario
    /people/balas.gorla/blog/2006/09/27/mdm-xi-r3-integration
    Hope it helps
    Reward points,if found useful
    Thanks
    Narendra

  • NULL error while importing the data

    Hai everybody. In one of our branches there was some problem. To simulate the situation. i advised them to send the backupdata. While importing data from a dump file, i am getting the erro message
    "cannot insert NULL into (“APGB5168”. “D010014”. “LBRCODE”)’.
    Can i change the NULL constraint and import OR can i import first only the table strucuture and modify the constraint and can import the data afterwards.
    Pl can any one suggest me how to import the data from this dump file so that i can simulate the situation.
    Thanks - JayaDev(JD)

    Hi,
    It means you have NOT NULL column and you are trying to insert NULL data.
    Can i change the NULL constraint and import OR can i import first only the table strucuture and modify the constraint and can import the data afterwards.It all depends on you and your business needs. If you are sure that removing the constraint will have no harm on your application and data then you can, but think why it has NOT NULL constraint if data has NULL values. Either your table data is incorrect or your table design. Rectify the problem and take action accordingly.
    Cheers,
    Avinash

  • Import the data in CMS

    hi...Fri's
    i exported data from dev to quality  using CMS , when i am trying to import the data in CMS , the import failed.. how can i import the data in CMS..
    thanks in Advance,
    Pasi.

    Hi
    have u checked the log
    cms -> details of the import -> log
    check out this discussion Re: cms import failed
    Step by step CMS
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/b8d8f7b2-0701-0010-b09a-cda4cca2c98e
    'How ToTransport XI Content Using CMS'.
    /people/sap.india5/blog/2005/11/28/xi-software-logistics-solution-iii-cms
    To check out the actual problem of why the DC import failed go to the following link
    http://<hostname>:<port>/webdynpro/dispatcher/sap.com/tc.CBS.WebUI/WebUI
    Message was edited by:
            Anusha  Ramsiva

  • Need a script to import the data from flat file

    Hi Friends,
    Any one have any scripts to import the data from flat files into oracle database(Linux OS). I have to automate the script for every 30min to check any flat files in Incoming directory process them with out user interaction.
    Thanks.
    Srini

    Here is my init.ora file
    # $Header: init.ora 06-aug-98.10:24:40 atsukerm Exp $
    # Copyright (c) 1991, 1997, 1998 by Oracle Corporation
    # NAME
    # init.ora
    # FUNCTION
    # NOTES
    # MODIFIED
    # atsukerm 08/06/98 - fix for 8.1.
    # hpiao 06/05/97 - fix for 803
    # glavash 05/12/97 - add oracle_trace_enable comment
    # hpiao 04/22/97 - remove ifile=, events=, etc.
    # alingelb 09/19/94 - remove vms-specific stuff
    # dpawson 07/07/93 - add more comments regarded archive start
    # maporter 10/29/92 - Add vms_sga_use_gblpagfile=TRUE
    # jloaiza 03/07/92 - change ALPHA to BETA
    # danderso 02/26/92 - change db_block_cache_protect to dbblock_cache_p
    # ghallmar 02/03/92 - db_directory -> db_domain
    # maporter 01/12/92 - merge changes from branch 1.8.308.1
    # maporter 12/21/91 - bug 76493: Add control_files parameter
    # wbridge 12/03/91 - use of %c in archive format is discouraged
    # ghallmar 12/02/91 - add global_names=true, db_directory=us.acme.com
    # thayes 11/27/91 - Change default for cache_clone
    # jloaiza 08/13/91 - merge changes from branch 1.7.100.1
    # jloaiza 07/31/91 - add debug stuff
    # rlim 04/29/91 - removal of char_is_varchar2
    # Bridge 03/12/91 - log_allocation no longer exists
    # Wijaya 02/05/91 - remove obsolete parameters
    # Example INIT.ORA file
    # This file is provided by Oracle Corporation to help you customize
    # your RDBMS installation for your site. Important system parameters
    # are discussed, and example settings given.
    # Some parameter settings are generic to any size installation.
    # For parameters that require different values in different size
    # installations, three scenarios have been provided: SMALL, MEDIUM
    # and LARGE. Any parameter that needs to be tuned according to
    # installation size will have three settings, each one commented
    # according to installation size.
    # Use the following table to approximate the SGA size needed for the
    # three scenarious provided in this file:
    # -------Installation/Database Size------
    # SMALL MEDIUM LARGE
    # Block 2K 4500K 6800K 17000K
    # Size 4K 5500K 8800K 21000K
    # To set up a database that multiple instances will be using, place
    # all instance-specific parameters in one file, and then have all
    # of these files point to a master file using the IFILE command.
    # This way, when you change a public
    # parameter, it will automatically change on all instances. This is
    # necessary, since all instances must run with the same value for many
    # parameters. For example, if you choose to use private rollback segments,
    # these must be specified in different files, but since all gc_*
    # parameters must be the same on all instances, they should be in one file.
    # INSTRUCTIONS: Edit this file and the other INIT files it calls for
    # your site, either by using the values provided here or by providing
    # your own. Then place an IFILE= line into each instance-specific
    # INIT file that points at this file.
    # NOTE: Parameter values suggested in this file are based on conservative
    # estimates for computer memory availability. You should adjust values upward
    # for modern machines.
    # You may also consider using Database Configuration Assistant tool (DBCA)
    # to create INIT file and to size your initial set of tablespaces based
    # on the user input.
    # replace DEFAULT with your database name
    db_name=DEFAULT
    db_files = 80 # SMALL
    # db_files = 400 # MEDIUM
    # db_files = 1500 # LARGE
    db_file_multiblock_read_count = 8 # SMALL
    # db_file_multiblock_read_count = 16 # MEDIUM
    # db_file_multiblock_read_count = 32 # LARGE
    db_block_buffers = 100 # SMALL
    # db_block_buffers = 550 # MEDIUM
    # db_block_buffers = 3200 # LARGE
    shared_pool_size = 3500000 # SMALL
    # shared_pool_size = 5000000 # MEDIUM
    # shared_pool_size = 9000000 # LARGE
    log_checkpoint_interval = 10000
    processes = 50 # SMALL
    # processes = 100 # MEDIUM
    # processes = 200 # LARGE
    parallel_max_servers = 5 # SMALL
    # parallel_max_servers = 4 x (number of CPUs) # MEDIUM
    # parallel_max_servers = 4 x (number of CPUs) # LARGE
    log_buffer = 32768 # SMALL
    # log_buffer = 32768 # MEDIUM
    # log_buffer = 163840 # LARGE
    # audit_trail = true # if you want auditing
    # timed_statistics = true # if you want timed statistics
    max_dump_file_size = 10240 # limit trace file size to 5 Meg each
    # Uncommenting the line below will cause automatic archiving if archiving has
    # been enabled using ALTER DATABASE ARCHIVELOG.
    # log_archive_start = true
    # log_archive_dest = disk$rdbms:[oracle.archive]
    # log_archive_format = "T%TS%S.ARC"
    # If using private rollback segments, place lines of the following
    # form in each of your instance-specific init.ora files:
    # rollback_segments = (name1, name2)
    # If using public rollback segments, define how many
    # rollback segments each instance will pick up, using the formula
    # # of rollback segments = transactions / transactions_per_rollback_segment
    # In this example each instance will grab 40/5 = 8:
    # transactions = 40
    # transactions_per_rollback_segment = 5
    # Global Naming -- enforce that a dblink has same name as the db it connects to
    global_names = TRUE
    # Edit and uncomment the following line to provide the suffix that will be
    # appended to the db_name parameter (separated with a dot) and stored as the
    # global database name when a database is created. If your site uses
    # Internet Domain names for e-mail, then the part of your e-mail address after
    # the '@' is a good candidate for this parameter value.
    # db_domain = us.acme.com      # global database name is db_name.db_domain
    # FOR DEVELOPMENT ONLY, ALWAYS TRY TO USE SYSTEM BACKING STORE
    # vms_sga_use_gblpagfil = TRUE
    # FOR BETA RELEASE ONLY. Enable debugging modes. Note that these can
    # adversely affect performance. On some non-VMS ports the db_block_cache_*
    # debugging modes have a severe effect on performance.
    #_db_block_cache_protect = true # memory protect buffers
    #event = "10210 trace name context forever, level 2" # data block checking
    #event = "10211 trace name context forever, level 2" # index block checking
    #event = "10235 trace name context forever, level 1" # memory heap checking
    #event = "10049 trace name context forever, level 2" # memory protect cursors
    # define parallel server (multi-instance) parameters
    #ifile = ora_system:initps.ora
    # define two control files by default
    control_files = (ora_control1, ora_control2)
    # Uncomment the following line if you wish to enable the Oracle Trace product
    # to trace server activity. This enables scheduling of server collections
    # from the Oracle Enterprise Manager Console.
    # Also, if the oracle_trace_collection_name parameter is non-null,
    # every session will write to the named collection, as well as enabling you
    # to schedule future collections from the console.
    # oracle_trace_enable = TRUE
    # Uncomment the following line, if you want to use some of the new 8.1
    # features. Please remember that using them may require some downgrade
    # actions if you later decide to move back to 8.0.
    #compatible = 8.1.0
    Thanks.
    Srini

  • Fastest way to import the data for big tables

    Hi friends,
    I have recently joined a billing product set up team for TELECOM giant.As part of the setup process /initial testing we need to put huge data (nearly 1.5 billion records/~200 GB size of data) in some tables. Currently i am using impdp for importing .It takes a lot of time (nearly >9 hours for loading this data).
    I have tried the below ways to import the data(All these tables are partitioned):
    1. Normal impdp (single thread i.e. parallel=1): This one takes a lot of time (>24 hours).
    2. Normal impdp but partition-wise .This one is completed in relatively less amout of time compared to the 1st method.
    3. Drop drop the indexes ,do inset append from another schema in same instance which has the data and then recreate the indexes ( total process takes ~9 hours)
    My questions to all:
    1.Is there any other way/trick in the book which I can try to put the data in less amount of time
    2.What i have observed that even if I give parallel=8 or any number ,sometimes parallel workers are spawned and sometimes not. Can someone educate me why ?
    3.How can i come to know (before running the impdp ) that my parallel clause is going to spawn the parallel workers or not. I have searched on this subject but no success as to how I will come to know this.
    I dont know what strategy to follow ,because this is a involved and repetitive task for me. Due to the above task I am not able to concentrate on other DBA tasks due to this which come to me.
    Cheers,
    Kunwar

    assuming you have a table with 3 indexes, then create a script for each index (reblace index_name_n with your index name) and start them at the same time, after the import of the table is completed
    DUMPFILE=<DUMPFILE> or NETWORK_LINK=<DBLINK>
    DIRECTORY=DATA_PUMP_DIR
    LOGFILE=<IMP_INDEX_NAME_n>
    CONTENT=ALL
    PARALLEL=1
    JOB_NAME=IMP_<INDEX_NAME_n>
    INCLUDE=TABLE_EXPORT/TABLE/INDEX:"IN('<INDEX_NAME_n>')"
    TABLES=<SCHEMA>.<TABLE>change following settings during import:
    increase pga_aggregate_target
    increase db_writer_processes
    db_block_checking=false
    db_block_checksum=false
    set undo and temp ts to autoextend
    set noarchivelogmode
    create 4GB redo logs
    hth

  • How can i import the data from multiple sources into single rpd in obiee11g

    how can i import the data from multiple sources into single rpd in obiee11g

    Hi,
    to import from multiple data sources, first configure ODBC connections for respective data sources. then you can import data from multiple data sources. When you import the data, a connection pool will create automatically.
    tnx

  • Any method to Import the data from Ingres Database to Oracle Database

    Hi,
    Is there any method to import the datas from Ingres Database to Oracle Database,
    Thanks in Advance.

    Check out this link.
    HTH...

  • Getting error while importing the data using loadercli

    Hello,
    I want to copy the data using export/ import  via (loadercli).
    scenario:
    1) I have 2 server and i want to export data from old server and import in to new MAXDB server using loadercli.
    I have done this once and it went fine , but now I want to do the export import again so i will have te latest data on the new server.
    when i tried to do that it gives error that the table already exist . Can i use loadercli to import hte data again ?
    can any one help me in this ?
    Regards,
    Bhavesh

    > Do you want to add and/or update the data in the already existing tables or do you want to replace the content completely?
    >
    > so in that way :
    > bot the options are fine what ever take less time.
    Sorry mate, but YOU have to know what you want here.
    I gave you an easy to follow set of steps.
    As you don't seam to mind the outcome, just might just use them...
    > I wanted to know weathe i can use the  loadercli for thie export import or not? if yes then is there any new steps to do before i do the export import?
    We had this discussion before...
    >
    > For that the easiest option would be just to drop the tables of SAPR3 and run the import again.
    >
    > For ease of use you could also just do:
    > - logon as superdba
    > - drop user SAPR3
    > - create user SAPR3 password SOMEPW not exclusive dba
    >
    > After these steps you can easily pump the data into the database again.
    >
    > So here in th above given steps , i am creating a new SAPR3 user and why it is not exclusive dba ?
    >  i already have that user SAPR3 can i use the same.
    Yes, you do have the SAPR3 user.
    But you don't seem to like to read documentation or learn about how the tools work or anything like that.
    Therefore I gave you s simple way to reach your goal.
    Of course it's possible to reuse the user.
    But then you would have to deal with already existing tables, already existing data etc.
    You don't seem to be able to do that. So, the easy steps might be better suited for your needs.
    regards,
    Lars

Maybe you are looking for

  • HT1414 The iPOD cannot be restored at this time

    After downloading the 840mb of ios 6.0.1 software update and when i click backup i see the below screen and when i click on Restore iPod . i see the below pop-up why should i loose all my data before updating the software ? this is ridiculous .....an

  • How to do web sharing windows xp to site on macbook?

    I've created a website with iweb '08 on a macbook, 10.6.8.  I've clicked on "web sharing" in system preferences.  My ipad can access the website on the macbook, but my windows xp desktop can't open it but doesn't say "server not found"  it opens a pa

  • Problems with pdf and keys

    I have a third generation iPod and I had no problems with 3.0 firmware, but i decided to update to 4.1 and suddenly, it stopped remembering keys in Safari and opening certain pdf's, because it freezes at the first page and I can't see anymore. I thin

  • Extensions don't work with ff4

    A bunch of my extensions that are supposed to work with FireFox4 just don't work at all. I've lost NoScript, ForecastFox, AdBlock Plus, Roomy Bookmarks Toolbar, Toolbar Buttons, and the SiteAdvisor plugin. I have the latest versions of all of these a

  • How can I restore a folder in SM35 - Very Urgent

    Hi All, How can I restore a folder in SM35 ?, because I created a folder in transaction AFAB and this transaction just process in background (under folder - SM35) and then I deleted this folder wrongly.... I tryed to execute again the transaction AFA