HOWTO set source and target in WTK KToolbar

I am having trouble getting my MIDlet compiled when pressing the Build button in the WTK's KToolbar. I get the following error:
generics are not supported in -source 1.3
This is because I have JDK 1.6 installed and I'm trying to use a Vector in my code that is typed.
How do I configure the javac settings through the KToolbar so it uses the javac -source and -target arguments?

The thing is like
while practicing in my institute I had used "//ECCserver/SAPMNT" as path for both sender and reciever communication channels...
now i'm unable to write the path for both
please help me out

Similar Messages

  • The source and target structure have a different number of fields

    I am new to workflow and I am trying to create an attachment in Workflow (SAP ECC 6.0) and pass it through to a User Decision (User Decision works fine) however the workflow is failing on the attachment step with u2018The source and target structure have a different number of fieldsu2019. The bindings check ok. Please see details below.
    I have used document u2018Creating Attachments to Work Items or to User Decisions in Workflowsu2019 by Ramakanth Reddy for guidance. Thanks in advance.
    1) Workflow containers (SWDD)
    WORKITEMID (import)
    ZSWR_ATT_ID (export)
    SOFM (export)
    2) Task Container (PFTC)
    1 Import parameter defined u2013 WORKITEMID (swr_struct-workitemid)
    2 Export parameters defined
    - SOFM (Ref. obj. type SOFM)
    - ZSWR_ATT_ID  (swr_att_id-doc_id)
    Binding task -> Method
    Binding for 1 parameter (import) defined
    Task <- Method
    Binding for 2 parameters (export) defined
    3) Z  BOR object created with a Method, Method Parameters and Event (SWO1)
    1 import parameter defined
    2 export parameters defined
    Workflow is triggered by FM SAP_WAPI_CREATE_EVENT, Return_code = 0
    Event_id = 00000000000000000001
    Test results
    A) Triggered by ABAP/ FM SAP_WAPI_CREATE_EVENT - SWI2_DIAG results
    Work item  14791: object <z bor object name> method <method name> cannot be executed. The source and target structure have a different number of fields (this message is repeated 3 times). Error handling for work item 14791. No errors occurred -> details in long text (message is repeated 3 times).
    Message no. WL821, OL383, WL050 in long text
    B) Z BOR Test method <execute>
    Enter workitem id.
    Runtime error - Data objects in Unicode programs cannot be converted. The statement "MOVE src TO dst" requires that the operands "dst" and "src" are convertible. Since this statement is in a Unicode program, the special conversion rules for Unicode programs apply.                                        
    In this case, these rules were violated.   
    Program                             SAPLSWCD                
    Include                                LSWCDF00                
    Row                                    475                     
    Module type                        (FORM)                  
    Module Name                      MOVE_CONTAINER_TO_VALUE           
    C) Z BOR Test method <debugging>
    Enter workitem id.
    SAP_WAPI_ATTACHMENT_ADD, return_code = 0, message_lines  = Attachment created            
    both  swc_set_element container work ok
    Runtime error occurs after end_method executed. Data objects in Unicode programs cannot be converted.
    D) Workflow test
    Enter workitem id <execute>
    Task started> Workflow log> Status = Error
    Workflow errors in Attachment step (however Office document can be viewed in details for step).

    Problem has now been resolved. Problem was related to use of swr_att_id structure and swc_set_element statement in BOR program - problem resolved by only setting w/f container to swr_att_id-doc_id.

  • Is it possible to have source and target schema in the same DB instance ?

    Hi All,
    I'm using Oracle 11gR1.
    I have switched source locations from other server to the one with OWB.
    During deploy I get VLD-3064 and I can't deploy mapping because of many 'table or view does not exist' warnings.
    Is it possible to have source and target schemas in the same instance ??
    How to do it ?

    Hi Jörg,
    Thanks for your help and quick answer.
    I'd like to clarify a little your answer :
    1. Are the any special privilages I need to set ? Currently I have only 'RESOURCE' and 'CONNECT' (I think I didn't have to add any other privs when that schema was located on other instance...)
    2. Also VLD-3064 states that : 'Owning and referenced location of this connector are on the same database instance. Unless explicitly referenced in configuration settings no generated code will use the dataabase link resulting from deployment of this connector'. Does it mean I have to put something in configuration ?

  • LMDB - SLD "Source and target system have the same object server name"

    the system is a SM7.1 SP1
    For some reason (Object-Server-Name changed), the job that syncs the sld and lmdb cancels...
    I set the LMDB-Object-Server-Name to its original Name...
    Now i've deleted the sync-config in the solman_setup and try to start a new sync, but i get the following error, when i try to configure the source-system:
    'Source and target system have the same object server name: CISM318'
    But the names are different: 'SM3' (LMDB) and 'CISM318' (SLD)
    Any suggestions ?
    best regards
    Edited by: Christoph Bastian on Aug 26, 2011 9:37 AM
    Edited by: Christoph Bastian on Aug 26, 2011 11:18 AM

    problem solved...
    Apparently the sync needs some hours to finish a object-server-name-change...
    best regards

  • Issue in mapping the same source and target fields

    Hi All,
    I am working on PI 7.0 and currently I am connecting to PI 7.0 via citrix desktop . Problem here is : in the mapping , direct mapping option that is available to map the same source and target fields is disabled . I do know that this is due to some java version issue .
    What I want to know is there any possibilty of solving this issue without installing an updated java version ?
    Vinay P.

    I've faced the same for many times and all of that were due to Java version.
    1st of all, you have to check the authorizations you have. If you have enough (DEVELOPER_J2EE, CONTENT_ORGANIZER_J2EE, and so), and - obvious but very often forgotten - you have to assure you're trying to change it in the message mapping change mode On.
    Then, last shot should be downgrade your Java version.
    The 1.4.2 JRE version allow you to use not only the Mapp the Identical field but also Dependencies, Clear Mapping and some other tools/buttons. As mentioned above, if there's a fix in the new Java releases, I do not know (hope so) and would appreciate if you could let us know.

  • One problem with constraints missing in source and target database....

    DB Gurus,
    In my database I am going to export and import data using datapump but dont specify CONTENT clause.
    so when I start working on my application, I came to know there are some constraints are missing. now how to check what are the constraints are missing from Source database and how to create those missing constraints on Target database?
    Suggest me if you have any idea.

    Create a database link between source and target schema.
    Use all_/dba_/_user_constraints to generate the difference using MINUS operator.
    something like this:
    select owner,constraint_name,constraint_type from [email protected]
    select owner,constraint_name,constraint_type from dba_constraints

  • Master Data ------  Source and Target Mapping

    I want to know the source and target mapping of some master data elements like plant,vendor,customer,workcenter.
    Where can I get it. Please provide the relevant documents or links.

    Check in RSOSFIELDMAP table.

  • OWB 10g -- Can't Create Database Links for Data Source and Target

    We installed OWB 10g server components on a Unix box running Oracle 10g (R2) database. The Designer Repository is in one instance. The Runtime Repository and the Target are in another instance. The OWB client component was installed on Windows XP. We create a data source module and a target module in OWB. The data source is on another Unix box running Oracle 9i (R2) database. We try to create database links for data source module and target module, respective. But when we created and tested the DB links, the DB links were failed.
    For the database link of data source, we got the following error message:
    SQL Exception
    Repository Error:SQL Exception..
    Class Name: CacheMediator.
    Method Name: getDDEntryFromDB.
    Repository Error Message: ORA-12170: TNS:Connect timeout occurred
    For the database link of target , we got the following error message:
    API2215: Cannot create database link. Please contact Oracle Support with the stack trace and the details on how to reproduce it.
    Repository Error:SQL Exception..
    Class Name: oracle.wh.ui.integrator.common.RepositoryUtils.
    Method Name: createDBLink(String, String, String, String).
    Method Name: -1.
    Repository Error Message: java.sql.SQLException: ORA-00933: SQL command not properly ended.
    However, we could connect to the two databases (data source and target) using the OWB’s utility SQL Plus.
    Please help us to solve this problem. Thank you.

    As I said prior the database link creation should work from within the OWB client (also in 10).
    Regarding your issue when deploying, have you registered your target locations in the deployment manager and did you first deployed your target location's connector which points out to your source?
    I myself had some problems with database link creations in the past and I can't remember exactly what they were but it had something to do with
    - the use of abnormal characters in the database link name
    - long domain name used in as names.default_domain in my sqlnet.ora file
    What you can do is check the actual script created when deploying the database link so see if there's something strange and check if executing the created script manually works or not.

  • No mapping for Same structure in source and target

    Hi Experts,
    I want to transfer the source structure to two receivers . Here source and target structures are same so i dont need mapping.
    My question is how many DT,MT and MI required .. mapping required or not.
    Thanks in Advance

    hi  CS,
                 here in  your  scenario as  u  have  told  that  source  and reciver  are of  same  structure . definitely you didn't need  any  mapping . and you have to create two data type(one for sender and  one for receiver) , two message  types(one for sender and one for receiver)) , and  two message  interfaces(one for sender and one for receiver).
    while creating message types it asks for data  type and while creating message interface it asks for message type only . as the structures of sender and receiver are same then inspite of having two receivers  we  have to create  ony one message interface on sender side . later the adapter will take  care of further proceedings
    but  in  case of configuration you have to import three business systems . one for sender and  two for receiver systems . and  you have to configure the communications  channels  accordingly. as per the receivers accepting the structures you have to use the relevant adaptors .

  • Entering Source and Target Values in Value Mapping

    Is there any auomative way of entering source and target values in Value Mapping or FixValues instead of manually entering the values.
    I have around 400 values to be entered on both sides.....just thinking of an easy way to do it....
    Thanks for suggestions.

    Use value mapping replication
    chk this link as to what SAP docs say on value mapping replication
    Value Mapping replication - Value Mapping replication
    ValueMapping using the Graphical Mapping Tool
    The specified item was not found.

  • Can source and target tables be the same

    Hi there,
    I need to restate profit center column in the delivery fact table. Historical Profit Centers need to be populated since today is the first time we brought in Profit Center field from LIPS table. The source for the delivery fact table is SAP LIPS table. so what I did is:
    joined the LIPS and delivery fact table on ( number = number & line item number = line item number) to get the profit center.
    So the source and the target delivery fact table are the same. Is this a good practice becuase Data services warns me that the source and target tables are the same.
    Please let me know a better alternative approach to this OR a better approach to restate fields for historical data. Thanks in advance.

    actually the LIPS resides on another server and in order to fetch data I have to write an R3 ABAP data flow, so I cant use it directly in a lookup.
    Except what Im thinking based on your reply to use a lookup table is: that I will extract 3 colums (delivery number, delivery line item number and profit center) to a table and then use the lookup.

  • Single Code Page Unicode conversion when source and target are the same

    Hi everyone.  We have recently upgraded our non-unicode 4.7 to ECC6 non-unicode.  Now we have begun researching the next step to convert to Unicode.  I know we could have done the Combined Upgrade and Unicode Conversion but I wanted to do thorough testing after the upgrade to eliminate too many variables if problems arose.
    So, my question is...has anyone done the Unicode conversion when source and target servers are the same?  Most documentation I've seen recommends a system copy to another box first.  I did find one forum thread that gave a high level approach to doing the conversion with just one server and listed this:
    Step 1:- Run Tcode SPUMG to scan
    Step 2:- Export Database
    Step 3:- Drop Database
    Step 4:- Create New Database
    Step 5:- Import database
    Step 6:- Chnage unicode Kernel
    What are your thoughts?  Is it supported?  Am I wasting my time and should I just bite the bullet and buy yet another system?  Thanks in advance.

    Theoretically that's possible, yes. However...
    > Step 5:- Import database
    > Step 6:- Chnage unicode Kernel
    the import must be done using the Unicode kernel.
    So basically you can
    - export the old systm
    - uninstall system + database
    - start as if you were installing "from scratch" on a new box

  • Source and Target Table in the Same schema........will that work in ODI

    Hi ,
    Is it possible to have my source and target tables in the Same achema.......will it have any performance issues or some other issues in our ODI.......
    I would like to know whether it works well or not .........
    ( To my understanding it wont have any issues ......But I am not 100 % sure on my judgement)

    Yes, you're correct.
    Just be sure about you load strategy, it means, if you need only to insert use the KM Control Append, but if need update too, try the IKM Oracle Incremental Update... I believe that will work better for millions of lines.
    Anyway be sure that your database has the necessary tablespaces to manage huge data volume.

  • Error saving map. Stored procedure returned non-zero result. Check if source and target schemas are present.

    I am using VS 2012 and BizTalk 2013 and attempting to deploy an application to BizTalk when I get these errors:
    Error 47
    at Microsoft.BizTalk.Deployment.Assembly.BtsMap.Save()
       at Microsoft.BizTalk.Deployment.Assembly.BtsArtifactCollection.Save()
       at Microsoft.BizTalk.Deployment.Assembly.BtsAssembly.Save(String applicationName)
       at Microsoft.BizTalk.Deployment.BizTalkAssembly.PrivateDeploy(String server, String database, String assemblyPathname, String applicationName)
       at Microsoft.BizTalk.Deployment.BizTalkAssembly.Deploy(Boolean redeploy, String server, String database, String assemblyPathname, String group, String applicationName, ApplicationLog log)
    0 0
    Error 49
    Failed to add resource(s). Change requests failed for some resources. BizTalkAssemblyResourceManager failed to complete end type change request. Failed to deploy map "XXX.BTS2013.XXX.Maps.map_XXXX_R01_InsLabProc".
    Error saving map. Stored procedure returned non-zero result. Check if source and target schemas are present. Error saving map. Stored procedure returned non-zero result. Check if source and target schemas are present.
    0 0
    Error 46
    Failed to deploy map "XXX.BTS2013.XXX.Maps.map_XXXX_R01_InsLabProc".
    Error saving map. Stored procedure returned non-zero result. Check if source and target schemas are present.
    0 0
    I also tried to Import a MSI file from our test environment to see if that would the same errors.  After spending hours (not kidding) looking for an answer, all I could find is that a hotfix would work.  So, I got the hotfix from Microsoft
    Support and applied it then rebooted.  Still getting the same errors.  I'm absolutely at a stand still.  Interesting that I got this application to deploy yesterday and then the next time I deployed it I started getting these errors.  I'm
    ready to pull my hair out!
    Is there an answer for this out there somewhere?  Any help would be appreciated.

    Hi Dave,
    Which hotfix have you applied? I don't think a hotfix of this issue is available for BizTalk 2013 yet. You should create a
    support ticket with Microsoft to get a solution.
    If this answers your question please mark as answer. If this post is helpful, please vote as helpful by clicking the upward arrow mark next to my reply.

  • Best practice deployment - Solaris source and target

    What is the recommended deployment guide for an ODI instance under Solaris. I have a sybase source and an Oracle target, both of which are on Solaris. I plan to put my ODI master and work repository on another Oracle DB on the solaris target machine. Now where does my agent sit, since my source and target are solaris ?? I plan to administer ODI from my windows clients but :-
    Where and how do I configure my agent so that I can schedule scenarios. It would make most sense to be able to run the agent on my target solaris machine , is this possible ?? If not then do I have to have a separate windows server that is used to run the agent and schedule the jobs ??
    Thanks for any assistance,

    thanks for the reply. I cant find anything in the installation guide about Solaris specifically but it mentions to follow the instructions for "Installing the Java Agent on iSeries and AS/400" where the download o/s is not supported.
    So it seems I just need make some directories on the solaris host and to manually copy files into s these directories and as long as java sdk/runtime is there I can use the shell scripts (eg. ) to start and stop the agent.
    So my question I guess is since the os supported downloads are only windows and Linux, where do I copy the files from, the Linux ones ?? Is it right to say that since these are java programs I should be able to copy the Linux ones and use them under Solaris ??
    I dont have the Solaris environment at hand to test this just yet.... hence the questions....
    thanks again

Maybe you are looking for

  • ALV Grid output to Excel Spreadsheet

    I have one that is hopefully easy for you guys and when I search on the forum nothing comes up. Maybe I am searching wrong. When you have a program that outputs to an ALV grid, there is a button you can output to a file. Like an excel document. Icon

  • Calling URL from servlet

    Hi! I am trying to find a solution to this scenario: 1) A cron job will call a servlet. 2) Servlet calls a URL on remote server. 3) Remote server returns an XML document. 4) Servlet processes/parses the XML document. 5) Servlet stores XML fields into

  • Back up software for BB 8900 using macbook air

    Recent;y I have switched onto Mac and although I found syncing software  couldnt manage to find back software. Problem is that I havent backed up my phone for the past 2 months. Does this software exists?  

  • Creating Spool from function module

    Hi friends, I want to create a spool from my function module. I dont knw it is possible or not. Currently I am creating spool by submitting another report from my function module and exporting the values calculated in memory ID and importing in repor

  • Data Dictionary - Primery  Key

    What will be the  Maximum numbers of Primary key, which we can assign in z table.