Direct loading of Vendor in 11.5.8
Hi Gurus,
We are currently on Oracle Apps ver.11.5.8 and was wondering if we can create/enter new vendors through loading into the Oracle Payables application to support payables activities from a file
Is there any current API that exist to acheive the above as we will be soon upgrading to Rel 12 and what do you recommend.
Thanks in Advance
Hello,
In R12, there are APIs to create new vendor/vendor Sites:
- ap_vendor_pub_pkg.create_vendor
- ap_vendor_pub_pkg.create_vendor_site
Alternatively, you can also insert your new vendors in the following Interface Tables
- AP_SUPPLIERS_INT
- AP_SUPPLIER_SITES_INT
and afterward launch the following concurrent requests to load the vendors from the interface to the Payables:
- Supplier Open Interface Import
- Supplier site Open Interface Import
HTH,
Vikram
Similar Messages
-
Business partner migration using lsmw direct load .
Hi,
We need to do business partner migration using the direct load method.
We are thinking of using direct input method of lsmw for the data transfer.
Could someone suggest a proper object/ subobject for the same.
Alternatively has anyone done the same using LSMWbapi or LSMWidocs.
Inputs will be highly appreciated <<text removed>>
Dirk.
Edited by: Matt on Dec 19, 2008 1:33 PM - Do not offer rewards. Please read the rules of engagement.Hi ,
Does this solution give any way to update the customer fields too ?
We have few customer fields added to BUT tables ... How can we update these fields using LSMW and IDOC ?
Could you please give your opinion on thesame. -
Can I view, edit and save microsoft documents such as Word, Excel and Powerpoint using iWork applications such as Pages, Keynote and Numbers on my iPad Mini (16GB Wi-Fi)? If Yes, then can I directly load these files to my iPad mini using iTunes? Also, once I finish working on these documents, can I download these back to my Computer.
Sorry, forgot to say:
Thank you so much to EVERYBODY/ANYBODY who may reply with helpful information,
Bob :) :) :) -
Hi every body,
Can any one help me in solving a problem to auto load of vendors from one organisation to another,
thank you,
ramHello Sunder,
What do you mean by auto load from one organization to another? Please provide additional information.
If you mean that you already created vendor data for one purchase organization and wanted to extend for another pur organization, then there is std LSMW object is available for creating vendor master data. but this is not automatic, you need to provide the data.
If you mean updates from legacy system to SAP system, in that case also you need to provide the IDoc/flat file which will upload data in SAP system regularly.
Regards
Arif Mansuri -
Regarding Direct Load or Direct Path Load
Why Direct Load and Direct Path Load is faster?
In which situation Direct Load, Direct Path Load are faster?>
Why Direct Load and Direct Path Load is faster?
In which situation Direct Load, Direct Path Load are faster?
>
See About Direct-Path INSERT in the DBA doc
http://docs.oracle.com/cd/E11882_01/server.112/e17120/tables004.htm#i1009100
>
About Direct-Path INSERT
Oracle Database inserts data into a table in one of two ways:
•During conventional INSERT operations, the database reuses free space in the table, interleaving newly inserted data with existing data. During such operations, the database also maintains referential integrity constraints.
•During direct-path INSERT operations, the database appends the inserted data after existing data in the table. Data is written directly into datafiles, bypassing the buffer cache. Free space in the table is not reused, and referential integrity constraints are ignored. Direct-path INSERT can perform significantly better than conventional insert. -
Direct load insert vs direct path insert vs nologging
Hello. I am trying to load data from table A(only 4 columns) to table B. Table B is new. I have 25 million records in table A. I have debating between direct load insert,, direct path insert and nologging. What is the diference between the three methods of data load? What is the best approach??
Hello,
The fastest way to move data from Table A to Table B is by using direct path insert with no-logging option turned on table B. Meaning this will be produce minimum logging and in case of DR you might not be able to recover data in table B. Now Direct path insert is equivalence of loading data from flat using direct load method. Generally using conventional method it's six phases to move your data from source (table, flat file) to target (table). But with direct path/load it will cut down to 3, and if in addition you will use PARALLEL hint on select and insert you might have faster result.
INSERT /*+ APPEND */ INTO TABLE_B SELECT * from TABLE_A;Regards
Correction to select statement
Edited by: OrionNet on Feb 19, 2009 11:28 PM -
I have few questions in direct path data load.
I have below questions When we load direct path in sqlloader or use
APPEND hint in INSERT statement,
1. Does primary key/UNDIQUE INDEX/UNIQUE KEY disable during the direct load?
I know, check constraint and FK constraints are disabled. I am not sure about
other keys.
2. During the direct path load, does table have lock(exclusive lock)? if so,
others can not apply INSERT/UPDATE/DELETE on the table during the Direct load.
Please clarify?
3. During the direct path load, i believe, the indexes are disabled or unusable?.
Is that true?? Please clarify.This is the original file before the load
100,Stévén,Kíng
101,Nééná,Kóchhár
102,Léx,Dé Háán
103,áléxándér,Húnóld
104,Brúcé,érnst
105,Dávíd,áústín
106,Vállí,Pátábállá
107,Díáná,Lóréntz
Character conversion after the load:
EMPLOYEE_ID LAST_NAME FIRST_NAME
100 StΘvΘn Kφng
101 NΘΘnß K≤chhßr
102 LΘx DΘ Hßßn
103 ßlΘxßndΘr H·n≤ld
104 Br·cΘ Θrnst
105 Dßvφd ß·stφn
106 Vßllφ Pßtßbßllß
107 Dφßnß L≤rΘntz
... -
Hi,
I usually posts threads regarding the internals of oracle. today the topic is direct load inserts.
Here is the mains concept
in direct load insert oracle by pass the buffer cache and inserts in directly into the data file. By by passing the buffer cache oracle avoid redo log generation and other over head. oracle builds a block in the memory and inserts it into the data file above the high water mark.
here are the questions
where does oracle builds the data block in memory is it in server process PGA.
does oracle builds more than one data blocks
does increasing the size of PGA memory can have any effect on direct load inserts.
if oracle by pass buffer cache than rollback segament are still generated. What its mechanism under direct load inserts
regards
NickNick Naughty wrote:
from above document. it seems that oracle streams and array size play an important role in performance but the above document is regarding sql loader what about this statementFor SQL Loader...only
insert /*+ append */ into dest_table
select * from source_table;
here we could not set oracle stream or array size a mentioned in above document. kindly elobrate with respect to above statement. Does not apply.
Perhaps you could explain what you are trying to do, or add some context of the problem. This would aid with responses.
Regards,
Greg Rahn
http://structureddata.org -
While using direct load whether trigger get fired
1. While using direct load whether trigger get fired?
2. Is it possible to log the errors with triggers if any event get failed?
Say for example am have a trigger audit with insert event on a table A.
is it possible to have any error table to monitor when the event get failed. Like the audit table or else track the failure operation with the audit trigger.1. While using direct load whether trigger get fired?
Oracle doesn't use direct-path INSERT if the target table has any triggers or referential constraints.
See the SQL Language doc:
http://docs.oracle.com/cd/E11882_01/server.112/e26088/statements_9014.htm
The target table cannot have any triggers or referential integrity constraints defined on it.
So you would not be using direct-path if the table has any triggers.
2. Is it possible to log the errors with triggers if any event get failed?
Say for example am have a trigger audit with insert event on a table A.
is it possible to have any error table to monitor when the event get failed. Like the audit table or else track the failure operation with the audit trigger.
Yes and No - DML Error logging can put rejected rows into an error table so that the entire INSERT will not get rolled back.
But some datatypes are not supported and there are other restrictions also:
See the SQL Language doc
http://docs.oracle.com/cd/E18283_01/server.112/e17118/statements_9014.htm#SQLRF55101
Restrictions on DML Error Logging
The following conditions cause the statement to fail and roll back without invoking the error logging capability:
Violated deferred constraints.
Any direct-path INSERT or MERGE operation that raises a unique constraint or index violation.
Any update operation UPDATE or MERGE that raises a unique constraint or index violation.
You cannot track errors in the error logging table for LONG, LOB, or object type columns. However, the table that is the target of the DML operation can contain these types of columns.
If you create or modify the corresponding error logging table so that it contains a column of an unsupported type, and if the name of that column corresponds to an unsupported column in the target DML table, then the DML statement fails at parse time.
If the error logging table does not contain any unsupported column types, then all DML errors are logged until the reject limit of errors is reached. For rows on which errors occur, column values with corresponding columns in the error logging table are logged along with the control information.
See also the direct-path and dml error logging sections of the DBA guide
http://docs.oracle.com/cd/E18283_01/server.112/e17120/tables004.htm#InsertDMLErrorLogging -
Hi,
Oracle Newbie Here.
I have a table that can contain up to 100 million rows and periodically this table needs to be copied. The user initiates the copy so I need to copy as fast as possible.
I have been using the direct load statement to perform this action. It performed pretty well when the table was small but as it grows so does the copy time.
I have not been using the parallel execution option when I copy the table. I am running on a multi-core processor. Would using this option really speed up my table copy?
How should I be using it? Are there issues with using it?
thanks
johnWhat do you mean by copy?
i think you can use CTAS ( create table as)
http://download-east.oracle.com/docs/html/A76994_01/transfor.htm
or copy command utility of SQL* Plus
http://download.oracle.com/docs/cd/B10501_01/server.920/a90842/apb.htm -
Hi all,
I have a insert .... select statement and i am using append hint to enforce direct load.
Is it same as the statement without hint on nologging table.
Thanks in advance.
JaggyIf you use insert with append hint, there won't be redo logs amount will be minimal. Of course I don't calculate redo logs generated by indexes. Any index on the table affected by the load will produce logs which requires time. If you want to speed up this process then make indexes on the affected table unusable and rebuild them after the load is done.
If you use insert into without HINT on nologging table, then it is totally different because it is not an direct load method. Redo logs will be generated for the table and all indexes. This option will be much slower.
If you have nologging enabled than it redo logs won't be generated only if you use direct path load (using append hint or import utility) -
Can I use two select statement in Direct-Load INSERT;
I have one table test with
NO NUMBER
NAME VARCHAR2(20)
Columns
I would like insert in no column by SEQUENCE & one from another
table.
Can anybody tell me HOW?
Regards, ChaitanyaIf you do not really care what the sequence is, you can do
something like
INSERT INTO table1 (numbercol,othercol)
SELECT rownum,othercol2
FROM table2; -
Hello,
we want to make bulk insert in C++ faster by avoiding generating redo logs. I plan to use Direct Load Insert. The problem is, DLI works only on insert /*+ APPEND */ into .. select ... but not with 'values' clause. To get rid of this, I want to bulk insert the rows into a temporary table, and than, at the end of the transaction, issue an insert /*+ APPEND */ into target_tab select * from temp_tab. After truncate temp_tab, I commit the transaction.
I know, REDO will be generated on the UNDO, but inserts generate only very low UNDO. By the way, can I avoid generating UNDO for temp tables?
What do you think about this? Any oppinion? Will it be really faster? Do you have any other idea?
Thank you,
BalazsThere is no hint call "NOLOGGING".So /*+ APPEND NOLOGGING */ will be ignored in your case.You may be better performance ,if you do it this way..
1) Create temp_tab with required columns matching target table. using CTAS method.
Create table temp_tab
tablespace <TS_name>
storage( ..................)
PCTFREE 0
NOLOGGING
as
select /*+ PARALLE(a,4) */ col1,col2,col3.....
from soruce_table a;
Comments: Above operation will be with NO REDO and NO UNDO. I used PCTFREE 0 to pack as many rows in the block,so that it helps in querying the temp_tab table in next steps.
2) Load the data into target table .Using parallel direct load , instead of serial-Direct load ( /*+ APPEND */ )
sql > alter session enable parallel dml;
sql > INSERT /*+ PARALLEL(a,4) */ INTO target_table a
SELECT /*+ PARALLEL(b,4) */ from temp_tab b;
sql>commit;
To give to some benchmarks , a 14 GB table is loaded with above method within 9 minutes on 12cpus busy unixbox.
Note : Check the other effect of parallel direct load in 8i concept -
what are the steps for performing direct loading (serially and parallely) , I need the steps.
PKSee Using Direct Path Load chapter from Documentation.
Best Regards,
Alex -
Full data load for vendor line items 0FI_GL_4
Hi Experts,
i need to delete the data from the vendor line items cube and load all the historic data from r/3 production, so how do i proceed, i am assuming that there is no setup table concept for this since its FI, so how do i do this and what are the steps that needs to be done on r/3 side as well as BW?
Thanks,
shetty.Its pretty simple in FI. There are no setup tables like LO extraction.
1. Double click ont he cube in BW and delete all the requests. make sure all of them are deleted. (or) you can right click on the cube and say "delete data".
2. As youa re working on production make sure ur requests and data are completely deleted. Check twice
3. Activate your master data.
4. replicate datasources and activate transfer structure in BW
5. Load data from BW
I would think twice to delete data from a cube particularly in production. I don't know your situation but if I were u, I would create another cube copying the cube and then load all the R/3 data into the new one.
Good luck.
Maybe you are looking for
-
How to prevent flash intro to be played each time
Ive bought recently a flash header for my website. Its mailny purpose is to look good and provide me a menu. So basicly the header is flash, the rest of the site is html. Now, my problem is that id like to like the intro of flash to be played only on
-
Encountering STATUS_CODE_ERROR
Hi, This is the first time I'm encountering this weird error. Every time I try to log in to my Apple ID in the Mac App Store or trying to update an app, an error shows up: "STATUS_CODE_ERROR". Here's a screenshot: https://www.dropbox.com/s/arfk1rpp9e
-
Create automatic white gap/bar with paragraph style
Hello, Thanks to your help and the Paragraph Ruler trick, I managed to get automatic title formatting (the grey background) Now, I would like to get automatically white gap as in line two but I can't figure out how to do it (tried with strikethrough
-
Very slow download speeds with ATT U-Verse
Hello, Experiencing very slow download speeds with Macbook Pro 2011, on Lion, and ATT UVerse Max, which should allow for an avg. 11-12 MBPS speed. Instead, I get a fluctuating 60-370 kbps speed. I am not experiencing this problem on my Windows machi
-
How to repair OS X 10.6.8 without SL DVD
My parents have a intel Mac Mini with OS X 10.6.8 (last possible upgrade). The Disk Utility says that there are minor error and the vlume should be repaired using install DVD. The SL install DVD nor Leopard upgrade DVD won't read for some reason? Onl