How to avoid hardcoding in RFC Lookup and JDBC Lookup during transport...

Hi All,
  We have RFC and JDBC Lookups in some of our scenarios(JDBC to RFC synch)....Now problem is during transport of design objects, we have to change the business system and channel name manually in UDF in target environment....as it wont change although we define the target group in SLD.....we want to avoid this hardcoding.....one idea is we can use a business service inspite of business system and assign the comm channel to it.....this business service will be generic for all environment(dev,qua and prod)...and hence we dnt hav to hardcode....pls let me knw how to achieve or its possible or not....
waiting for ur inputs....
Regds,
Pinangshuk.

Hi,
For RFC lookup you might be doing hardcoding for Business system and RFC channel name.
Options are
1) Create value mapping for Business system. So you can maintain it in Directory part.
2) Keep the RFC channel name same for all system (Development, Quality, Production)
    The details of channel can be changed in the Directory part of the respective system

Similar Messages

  • How to avoid using subfolders in libraries and use views to show all data

    Hi all
    I am very new to Sharepoint design so hopefully I can explain what I need properly.
    I have a sharepoint site which I upload a lot of data to. Most of the data I can categorise into subfolders (Location and Year). Instead of having the same subfolders in each library I wanted the user to select the Location and Year they want from a dropdown
    and then refresh a List webpart based on their requirements.
    Is this easy to do?
    Would I have to create new columns called Location and Year in order to do this?
    Thanks for any help
    Raiyan

    Yes, you can do so. 
    You can create a view to show data based on location and year
    http://blog.pentalogic.net/2008/10/advanced-sharepoint-view-and-filters/
    http://howididit-sharepoint.blogspot.in/2012/04/filtering-lookup-columns.html
    http://stackoverflow.com/questions/18362853/how-do-i-filter-by-today-and-time-in-sharepoint-list-view
    In SharePoint Designer I edit the calm query from my view in Advanced Mode. And I add
    IncludeTimeValue="True" in calm tag Value Type="DateTime". I'm filtering my results by time.  
    <Query>
    <OrderBy>
    <FieldRef Name="Modified" Ascending="FALSE"/>
    </OrderBy>
    <Where>
    <Or>
    <Gt>
    <FieldRef Name="Start"/>
    <Value Type="DateTime" IncludeTimeValue="True">
    <Today/>
    </Value>
    </Gt>
    <Gt>
    <FieldRef Name="TimeOver"/>
    <Value Type="DateTime" IncludeTimeValue="True">
    <Today/>
    </Value>
    </Gt>
    </Or>
    </Where>
    </Query>Also check thishttp://office.microsoft.com/en-in/sharepoint-server-help/create-list-relationships-by-using-unique-and-lookup-columns-HA101729901.aspx

  • How to avoid this full table scan (and index FFS) ?

    Hi All,
    Oracle 11.2 on Linux.
    See this query and its plan below
    SQL> DELETE
      2  FROM  TABLEA APE
      3        WHERE   NOT EXISTS
      4                   (SELECT   1
      5                      FROM   TABLEB AP
      6                     WHERE       AP.col1 = APE.col1
      7                             AND AP.col2 = APE.col2
      8                             AND AP.col3 = APE.col3)
      9  AND ROWNUM < 51 ;
    50 rows deleted.
    Elapsed: 00:12:01.07
    Execution Plan
    Plan hash value: 1740911877
    | Id  | Operation               | Name                  | Rows  | Bytes |TempSpc| Cost (%CPU)| Time  |
    |   0 | DELETE STATEMENT        |                       |    50 |  2650 |       |   573K  (1)| 01:54:40 |
    |   1 |  DELETE                 | TABLEA                |       |       |       |            |       |
    |*  2 |   COUNT STOPKEY         |                       |       |       |       |            |       |
    |*  3 |    HASH JOIN RIGHT ANTI |                       |    80M|  4059M|  1775M|   573K  (1)| 01:54:40 |
    |   4 |     INDEX FAST FULL SCAN| TABLEB_UK             |    47M|  1228M|       | 96480   (1)| 00:19:18 |
    |   5 |     TABLE ACCESS FULL   | TABLEA                |    80M|  1991M|       |   243K  (1)| 00:48:42 |
    ---------------------------------------------------------------------------------------------------------In both tables, TABLEA and TABLEB, there is index on columns col1-col2-col3 as leading columns (TABLEB has few more columns in the index, but after these 3 columns).
    Requirement is, I want to delete first 50 records in TABLEA, which does not exist in TABLEB.
    I tried with various hints, but Oracle is always doing a full scan on one of the tables and index FFS on other. In some cases, Oracle did full scan on both tables and then deleted 50 records. Stats is up-to-date. Doing a full scan on tables with 80 million and 47 million rows is a bit too much for deleting 50 rows.
    How I can make Oracle do
    1) Read TABLEA row-by-row
    2) for each row, check if it exists in TABLEB
    3) If not exists, then delete row from TABLEA, else continue
    4) Stop reading TABLEA after we have deleted 50 records
    Thanks in advance

    >
    >
    Oracle 11.2 on Linux.
    SQL> DELETE
    2  FROM  TABLEA APE
    3        WHERE   NOT EXISTS
    4                   (SELECT   1
    5                      FROM   TABLEB AP
    6                     WHERE       AP.col1 = APE.col1
    7                             AND AP.col2 = APE.col2
    8                             AND AP.col3 = APE.col3)
    9  AND ROWNUM < 51 ;
    50 rows deleted.
    Elapsed: 00:12:01.07
    Execution Plan
    Plan hash value: 1740911877
    | Id  | Operation               | Name                  | Rows  | Bytes |TempSpc| Cost (%CPU)| Time  |
    |   0 | DELETE STATEMENT        |                       |    50 |  2650 |       |   573K  (1)| 01:54:40 |
    |   1 |  DELETE                 | TABLEA                |       |       |       |            |       |
    |*  2 |   COUNT STOPKEY         |                       |       |       |       |            |       |
    |*  3 |    HASH JOIN RIGHT ANTI |                       |    80M|  4059M|  1775M|   573K  (1)| 01:54:40 |
    |   4 |     INDEX FAST FULL SCAN| TABLEB_UK             |    47M|  1228M|       | 96480   (1)| 00:19:18 |
    |   5 |     TABLE ACCESS FULL   | TABLEA                |    80M|  1991M|       |   243K  (1)| 00:48:42 |
    ---------------------------------------------------------------------------------------------------------Requirement is, I want to delete first 50 records in TABLEA, which does not exist in TABLEB.
    Such requirements usually make me curious - what's special about a randomly selected 50 rows ?
    Is this trying to delete the data in batches of 50 rows at a time.
    How I can make Oracle do
    1) Read TABLEA row-by-row
    2) for each row, check if it exists in TABLEB
    3) If not exists, then delete row from TABLEA, else continue
    4) Stop reading TABLEA after we have deleted 50 records
    It look's as if a 'no_unnest' hint in the subquery should do what you want. It should make Oracle run the quey with a FILTER subquery. You could then choose to drive the delete through a tablescan of tableA or an index range scan of the index on tableA. Have you considered the effect of (and requirements relating to) nulls in the three columns of either table ?
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    Author: <b><em>Oracle Core</em></b>

  • How to avoid primary key insert error and pipe those would-be error rows to a separate table?

    Hi All,
    Question: How can I ID duplicate values in a particular column before generating a "Violation of PRIMARY KEY constraint" error?
    Background: this SSIS package pulls rows from a remote server table for an insert to a local table.  The local table is truncated in step 1 of this package.  One of the source columns, "ProductName," is a varchar(50) NOT NULL, with no
    constraints at all.  In the destination table, that column has a primary key constraint.  Even so, we don't expect duplicate primary key inserts due to the source data query.  Nevertheless, I've been
    tasked with identifying any duplicate ProductName values that may try to get inserted, piping them all to a "DuplicateInsertAttempt_ProductName" table, and sending an email to the interested parties.  Since I have no way of knowing which row
    should be imported and which should not, I assume the best method is to pipe all rows with a duplicate ProductName out so somebody else can determine which is right and which is wrong, at which point we'll need to improve the query of the source table.
    What's the proper way to do this?  I assume the "DuplicateInsertAttempt_ProductName" table needs identical schema to the import target table, but without any constraints.  I also assume I must ID the duplicate values before attempting
    the import so that no error is generated, but I'm not sure how to do this.
    Any help would be greatly appreciated.
    Thanks,
    Eric

    agree about preventing a dupe or other error on some inconsequential dimension from killing a data mart load that takes a few hrs to run and is an important reporting system.
    I looked into using the error output before, but i think it came up a bit short...
    This is going from memory from a few years ago, but the columnid that comes out of the error data flow is an internal id for the column in the buffer that can't be easily used to get the column name.
    No 'in flight'/in-process way exists to get the column name via something like thisbuffer.column[columnid].name unfortunately
    In theory, the only way to get the column name was to initialise another version of the package (via loading the .dtsx xml) using the SMO .net libraries. I say in theory because I stopped considering it an option at that point
    And the error code is fairly generic as well if i remember correctly. It's not the error that comes out of the db (Violation of UNIQUE KEY constraint 'x'. Cannot insert duplicate key in object 'dbo.y'. The duplicate key value is (y).)  It's a generic
    'insert failed'/'a constraint failed' type msg.
    I usually leave the default ssis logging to handle all errors (and log them in the sysssislog table), and then I explicitly handle specific exceptions like dupes that I don't want to fail package/parent on error
    Jakub @ Adelaide, Australia Blog

  • How to avoid borderline coming between page and page in Htmlcss format

    Hi there
    I run a report output in HTMLcss format. Output is coming correctly.But what my problem is in the output i am getting some borderline (thick black line) between page and page. While i print the html i am getting this line at end of each page. What is this line and how to remove it htmlcss. i want printout without this line.I am using reports 10g version.
    thanx in advance
    prasanth a.s.

    The line is caused by the default HTML code placed in the After Page Value property found in the Report Escapes section of the report property palette. Clear the value to delete the line

  • How to avoid versioning being turned off and how to avoid file loss?

    Hi,
    I have a customer that requires versioning of documents. So I suggested that all files to be versioned are to be put in a folder that is verison enabled. That works fine.
    The problem now is, that let's say 20 versions already exist, and users are still able to turn of versioning. So if someone turns of versioning in that folder, all 20 versions are lost, aren't they?
    My question now is: How do I prevent users from turning off the versioning? What options do I have in standard to keep the files? Can I somehow only allow the preson who turned on versioning to also turn it off (incl. Admin)?
    I am using a CM Respository.
    Thanks for your help,
    Dieter.

    Hi,
    According to the documentation, to disable or enable the versioning all you need is write permissions for the folder.
    What you could do is play with the visibility of the Versioning command. You could remove it from menus, or set the Roles With Visibility property to only selected roles. For example if you set the Roles With Visibility to super_admin, then only the super admin will be able to see the command.
    Either that or you could create a custom versioning command that will take care of all that you want.
    Best regards,
    Avishai Zamir

  • How to avoid duplication of vendor name and vendor account group

    Let me know the menu path for setting in SAP - controlling / avoid duplication vendor name and vendor account group in one purchasing organisation during creation of vendor / uploading mass data.
    with regards
    vv

    Hi,
    See vendor code is unique no. But the name is not at all unique. so system allow to create duplicate record.
    by giving authorization to single person only you can do it.
    system discipline is actually required

  • Difference between Naming.lookup and Registry.lookup in RMI

    Hi Folks,
    I have an RMI client which tries to look up a remote object using both
    Naming.lookup("//host[:port]/name OR rmi://<host_name>[:<name_service_port>]/<service_name>") and LocateRegistry.getRegistry().lookup("rmi://<host_name>[:<name_service_port>]/<service_name>").
    The first one fails whereas the second one succeeds. I am in a fix as to why it is so.Is there any difference between the lookup methods provided by Naming and Registry?
    The remote object was registered on the server using
    LocateRegistry.createRegistry().rebind("rmi://<host_name>[:<name_service_port>]/<service_name>").

    I suggest you take a look at this site to understand how to ask questions.
    And don't crosspost again

  • How to avoid time wastage in starting & Stopping weblogic server during development ?

    It is very frustrating to stop the server and start the server for any small changes made in the application.
    Is developer supposed to restart the server for for any small changes made along with the testing?
    One restart(Stop and Start) activity takes more than 10-15 mins. in debug mode.
    1) I make some changes then I start server in debug mode
    2) I see some mistakes in my bean code
    3) I change my bean.java or Impl.java
    4) I make a futile effort to reload Jars and reload the page
    5) It never takes my changes
    6) Only changes in properties of jspx components will accept the changes at the page refresh
    6) I add new component or new binding to jspx page, This will not even pay heed to re-deployment, I need restart the server
    7) I work only 7 hrs a day (1 hr-lunch)! 7*60=420 mins.... lets say I make code changes 10 times a day 10*20(200) half day
       is gone in just restarting server/redeploying changes.
    8) So frustrating..... !!!
    Que:
    1) Is the developer is supposed to imagine all the errors beforehand and code so that he wont waste time in restarting?
    2) How nice it would be that you just recompile the changes and you get to see your changes just by refresh...
       that is why pl/sql development is so fast.
    3) Is there any way to do it like that? or Are there are any quick tricks or precautions to avoid restarting the server.
    4) Am I not following the correct development/deployment methods?
    It will be really great help for me if you experts guide me on the practise you follow for restarting/redeploying.
    Thanks
    Rahul

    Rahul, please tell us your jdev version!
    1) this always helps!  In the old days you only compiled once a day and if the compiler found an error you had to wait a whole day until you got the next chance.  Do feel privileged that you can compile multiple times. A restart is not nessecary in every case, but depending on the version you hit to restart more often.  You should try to find out why it takes so long and minimize this time!  Using 11.1.1.7.0 i start my server in less then a minute.
    2) answer depends on the jdev version, but you don't get the speed of pl/sql for wenn application development.
    3) one thing to do is to define the jdbc connections used in the server or then use jndi to access the resource. This way you don't lose a minute or connection stopping the server.  Next it's best to keep the server as clean as possible. Read the log when the server starts and stops you get plenty of info what takes how long.  Then start optimizing.
    4) no, your method is ok.
    Timo

  • How to import logo to the clip and display it during the whole clip

    Hello All,
    I am using Adobe Premiere Pro CS4.
    I am doing a project now from various DVDs I've already made from a while ago.
    Now I will use these them in a commercial things so I want to add a logo to the clips to keep my rights ...etc.
    Here is the situation:-
    1) Importing the clip file into Adobe Premiere Pro CS4.
    2) Importing the logo image file I want to use.
    3) Right click the clip file and choose insert.
    Now I want to know how to import the image file to the clip so it is being there during the whole clip?
    I've tried several times to import it but found it displayed seperatey after or before the clip or in the middle while I want to be displayed during the whole movie.
    Also it is big when I displayed while I want to resize it to be in the top right of the clip.
    Thanks a lot
    Sorry for poor English.

    took longer than I thought...and this is only basic examples...can be finessed better....
    with more thought and time ...but this gives you the basic ideas..read bout track mattes and transparency..etc...
    samples
    white background ...with track matte
    becomes transparent..letters, black opaque, gray transluscent
    transparent background
    letters black and white are shown as is with no track matte
    transparent background with track matte...
    now white letters act as mask and allow image through only white letters

  • Show itemcode on production order lookup and BOM lookup

    When looking up a Production Order or Bill of Material, the ItemCode is not available in the lookup window. Even when you "Product No." in the case of the Production Order or "Parent Item" for BOMs you get the Item description rather than the ItemCode.
    Many clients know their part numbers well and would like to be able to look by ItemCode. I know I could build in a Formatted Search/UDV to one of the fields but it seems like this should be a natural.
    Am I missing something?
    Alan

    Hi Alan.......
    Are you trying to say when you Press Tab on Item Code Field in Production order at parent level there you dont find the Item Code which you want to be produced?
    If yes then It happens if you are doing Standard or Disassembly Production Order and Bill if Material of such product is not present in Database then it wont list up......
    Please check the BOM.......
    Regards,
    Rahul

  • JDBC Lookup problem

    we are facing this problem...
    from SAP we send the idoc: WP_PLU03 to PI through transaction WPMA.
    PI do the data format in: XML Document Format for the Mode XMB2DB_XML (http://help.sap.com/saphelp_sem40bw/helpdata/en/4d/8c103e05df2e4b95cbcc68fed61705/frameset.htm)
    but before to do this format, a couple of RFC Lookups and JDBC lookups are executed.
    the important here are the JDBC lookups.
    when SAP sends a lot of Articles, the WP_PLU03 generates many segments, so only one LUW is not enough,then SAP automatically separates in 2 or more LUWs as needed.
    the first LUW is processed correctly, so, the JDBC adapter does good all the LOOKUPS, and the XMB2DB_XML is well formed and sent to the JDBC adapter for his process.
    but, the error is in the second LUW process , when the JDBC adapter begins to do the LOOKUPS, it STOPS, without error, or any alert...
    any idea?

    >
    Christophe PFERTZEL wrote:
    > Hi Guillermo,
    >
    > 1 LUW = 1 IDOC = 1 msg ? Do you use the same JDBC adapter for lookups AND msg sending ? Are those lookups performed using your own UDF/class or are you using PI feature ?
    >
    > Have you also check for error msgs in XI/PI logs and traces ?
    >
    > Rgds
    > Chris
    hi,
    1 LUW = A LOT of iDoc = MANY MSG.
    I've tried both, first, only one JDBC adapter for the lookup and process the message...
    result = the same problem, JDBC stops with the green circle in the RWB.
    second try, 2 jdbc adapters, one for the lookup and the other for the message process,
    result = the lookup jdbc adapter face the same problem.
    I'm using an UDF for the lookup.
    > Have you also check for error msgs in XI/PI logs and traces ?
    that's the problem, PI doesn't shows error.
    THANKS!

  • How to avoid lo-gin prompt while check in the document and saving doc while opening the document?

    Hello
    I might be missing something to configure or verify but many times I have faced below situation :
    When I click on document in document library either its just opens or first save to disc and open.
    When I check in a document , its asking for credentials but sometimes doesn't
    Can anyone please describe the reason behind above behaviors ? how to avoid saving doc while opening and lo-gin prompt while check in the document?
    Thanks and Regards,
    Dipti Chhatrapati

    It the sharepoint site accessed through firewall
    http://support.microsoft.com/kb/2756625
    You can try adding below registry
    HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\WebClient\Parameters\AuthForwardServerListAdd the site to Trusted Sites Zone instead of Intranet Go to Control Panel >> User Accounts >> Manage Your Credentials and remove any credentials storedMake sure your office and IE Is same build (64 or 32)http://www.techrepublic.com/blog/smb-technologist/prevent-unwanted-credentials-prompts-with-sharepoint-document-libraries/ checked that 'basic authentication' is unchecked under Central Administration -> Application Management -> Authentication
    ProviderOr tryhttp://social.technet.microsoft.com/Forums/office/en-US/45feb132-c304-4521-8b79-42236a829aab/login-prompt-appearing-when-opening-office-documents?forum=sharepointadminlegacy
    If this helped you resolve your issue, please mark it Answered

  • How to resolve duplicate reversal in ods and the purpouse of parallelismtab

    how to avoid duplicate reversal in ods and what is the use of parallelism tab in the ods that appear when we click reconstruction tab in ods manage
    basically my requirement is that 
    we are getting transaction data which will pass through the update rules start routine and will be updated into ods
    but in ods we are getting duplicate reversal means the record with latest version should be updated ut we are getting duplicate reversal and our reveue is showing in minus

    user13310594 wrote:
    Hi,
    I'm having millions of records in the table .Some duplicate values/rows are inserted in it.
    I just want to delete duplicate rows but also want to retain the last duplicate row.
    For eg if one record is found three times ,i want to delete first and second record and retain
    the third one i.e the last inserted one.Hi Paramdeep,
    To start with, since you do not wish to keep multiple rows with same values, why allow them to get inserted in the first place?
    Wouldn't it be easier to block duplicate rows from inserting by creating a Unique constraint on Column(s) that define a row as duplicate, then rather deleting duplicate rows periodically?
    For deleting duplicate rows, there are N number of techniques available. Always remember, you need to have a rigid criteria that marks row as duplicate.
    [url http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:15258974323143]this is one recomended by tom for large tables.
    Vivek L

  • How to avoid the Blank elemts in the List Item

    Dear All,
    I am using a List Item
    List Elements List Item Value
    Yes 1
    No 2
    Only two elements in this List Item.But at Run time there is one Blank element.
    I need to display only two elements from the List Item .How to avoid the Blank.
    Thanks and Regards,
    Fazil

    Fazil,
    I am using a List ItemList Elements List Item Value
    Yes 1
    No 2
    Only two elements in this List Item.But at Run time there is one Blank element.
    I need to display only two elements from the List Item .How to avoid the Blank.>
    Maybe you actually have null entries in your list. You can clean it up using keys as discussed here http://paranoid-engineering.blogspot.in/2008/05/wonderland-of-oracle-forms-builder.html (Ignore the personal opinions of the author if you will ;) ).
    Cheers,

Maybe you are looking for