Data locking issue

Hi,
In our business scenerio we are creating some business transaction through portal by which in background Purchase order gets created based on info record (all are Z development) then immediately Goods receipt is created.
we are getting error "The plant data of the material xxxxx is locked by the user System user ECCu201D or sometimes "The plant data of the material xxxxx is locked by the user yyyy(user name)u201D.Message type E message class M3,message number 897
I have tried to recreate the scenerio by
1) keeping same material master in edit mode-I donot encounter any error
2) keeping purchase order in edit mode having same material code but encounter another error "object requested is currently locked by user yyyy-Error type E error ID MC error number 601
3) keeping MIGO for same material code in creation mode but donot encounter any error.
4) By keeping info record which is used in purchase order in EDIT mode but also could not encounter error.
My query is when plant data of the material is actually locked so that it will not be accessible to other transaction.My thinking was that if i keep material master / concern info record /purchase order having same code in edit mode or MIGO in create mode having same code i should have got this error
Will some one explain when data is locked.
Regards
Atharva G

the data is usually locked when someone is in change mode in material master plant views, or when someone creates a movement that changes stock or price of a material.
Pretty often a so-called late lock is activated, to minimize such  situations. I think this late lock is active in your system, otherwise you would have answered some more of your tests with: error message received.
In this case you will not really be able to manually reenact such situation.
If is probably your database that might be a little to slow for your quick process
in background Purchase order gets created based on info record (all are Z development) then immediately Goods receipt is created.
you may need a little pause between PO posting and receipt posting. I guess the PO is not yet fully stored to the database while the GR tries already to lock the material for his purpose.

Similar Messages

  • InfoPack processing (PSA and data targets in parallel) data and lock issues

    Good morning,
    We are on BW 3.5 and contemplating changing our InfoPackages from option 1 (PSA then data targets) to option 2 (PSA and data targets in parallel) for performance reasons (test have shown a gain of 50% in load time compared to using option 1).
    Issues have been raised that using option 2 may lead to data consistency and lock issues especially for SD infosources where we are dealing with multiple daily status changes for same document that need to be read in a specific order.
    From what I have read, such issues were possible in BW 2.x e.g. OSS Note 384023 and 364577. However, I am under the impression that the issues may have been resolved in BW 3.x and that it is "safe" to use option 2 (PSA and data targets in parallel).
    Would really appreciate if you could please share your experience when using processing option 2 (PSA // DT) especially in a BW 3.xx system.
    Thank you in advance,
    Catherine

    Hi Amit,
    The help text (vers. 3.5) for this method of loading says simply -
    "A process is started to write the data from this data packet into the PSA for each data package. If the data is successfully updated in the PSA, a second parallel process is started. In this process, the transfer rules are used for the package data records, data is adopted by the communication structure, and it is finally written to the data targets. Posting of the data occurs in parallel by packet.
    This method is used to update data into the PSA and the data targets with a high level of performance. The BW system receives the data from the source system, writes it to the PSA, and starts the update immediately and in parallel into the corresponding data target"
    Not sure where you got your description from, but is not explained well and confusing.
    Regards,
    Ken

  • Locking Issue in Planning DSO in SAP BW 7.3(Integrated Planning)

    Hi Experts,
    We have built Aggregation Level on Direct Update (planning enabled) DSO and used the same in the input ready queries.( Its complete Manual planning - new row addition feature in WAD).
    We have 5 characteristic and one key figure for planning.
    We are facing lock issue while planning the data i.e. when one user is planning, other user cannot plan.
    The lock relevant characteristic is Column A and the same has been added to the Lock Characteristic tab(RSPLSE) of the infoprovider. The input ready query is also restricted with a variable on Column A (manual input single variable- mandatory).
    While the user opens the Planning Layout and plans the data, we can see lock entry in RSPLSE(locks tab) for the Column A selection in the infoprovider, yet if another user is trying to plan for data for some other selection of Column A, a lock entry in RSPLSE is shown with the selection but it errors in thelayout with lock on the infoprovider and cannot plan the data.
    Any inputs will be of great help.
    Regards,
    Priyanka

    Post is closed.
    Refer http://scn.sap.com/thread/3564212 for the solution.
    Regards,
    Priyanka

  • Locking Issue in Planning DSO- SAP BW 7.3 Integarted Planning

    Hi Experts,
    We have built Aggregation Level on Direct Update (planning enabled) DSO and used the same in the input ready queries.( Its complete Manual planning - new row addition feature in WAD).
    We have 5 characteristic and one key figure for planning.
    We are facing lock issue while planning the data i.e. when one user is planning, other user cannot plan.
    The lock relevant characteristic is Column A and the same has been added to the Lock Characteristic tab(RSPLSE) of the infoprovider. The input ready query is also restricted with a variable on Column A (manual input single variable- mandatory).
    While the user opens the Planning Layout and plans the data, we can see lock entry in RSPLSE(locks tab) for the Column A selection in the infoprovider, yet if another user is trying to plan for data for some other selection of Column A, a lock entry in RSPLSE is shown with the selection but it errors in thelayout with lock on the infoprovider and cannot plan the data.
    Any inputs will be of great help.
    Regards,
    Priyanka

    Found the answer.
    We had used the Planning Function directly in the WAD layout before saving data without filter.(basically to check some condition).
    Created the Planning Sequence with Column A filter on the Aggregation level and used the same sequence before saving the data in WAD
    Issue resolved.
    Consider this post as closed.
    Regards,
    Priyanka

  • FOR UPDATE cursor is causing Blocking/ Dead Locking issues

    Hi,
    I am facing one of the complex issues regarding blocking / dead locking issues. Please find below the details and help / suggest me the best approach to ahead with that.
    Its core Investment Banking Domain, in Our Day to day Business we are using many transaction table for processing trades and placing the order. In specific there are two main transaction table
    1)     Transaction table 1
    2)     Transaction table 2
    These both the tables are having huge amount of data. In one of our application to maintain data integrity (During this process we do not want other users to change these rows), we have placed SELECT …………….. FOR UPDATE CURSOR on these two table and we have locked all the rows during the process. And we have batch jobs (shell scripts ) , calling this procedure , we will be running 9 times per day 1 hrs each start at 7:15AM in the morn finish it up in the eve 5PM . Let’s say. The reason we run the same procedure multiple times is, our business wants to know the voucher before its finalized. Because there is a possibility that order can be placed and will be updated/cancelled several times in a single day. So at the end of the day , we will be sending the finalized update to our client.
    20 07 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 08 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 09 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 10 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 11 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 12 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 13 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 14 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 15 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 16 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 17 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    Current Program will look like:
    App_Prc_1
    BEGIN
    /***** taking the order details (source) and will be populate into the table ****/
    CURSOR Cursor_Upload IS
    SELECT col1, col2 … FROM Transaction table1 t 1, Source table 1 s
    WHERE t1.id_no = t2.id_no
    AND t1.id_flag = ‘N’
    FOR UPDATE OF t1.id_flag;
    /************* used for inserting the another entry , if theres any updates happened on the source table , for the records inserted using 1st cursor. **************/
    CURSOR cursor_update IS
    SELECT col1, col2 … FROM transaction table2 t2 , transaction table t1
    WHERE t1.id_no = t2.id_no
    AND t1.id_flag = ‘Y’
    AND t1.DML_ACTION = ‘U’,’D’ -- will retrieve the records which are updated and deleted recently for the inserted records in transaction table 1 for that particular INSERT..
    FOR UPDATE OF t1.id_no,t1.id_flag;
    BLOCK 1
    BEGIN
    FOR v_upload IN Cursor_Upload;
    LOOP
    INSERT INTO transaction table2 ( id_no , dml_action , …. ) VALUES (v_upload.id_no , ‘I’ , … ) RETURNING v_upload.id_no INTO v_no -- I specify for INSERT
    /********* Updating the Flag in the source table after the population ( N into Y ) N  order is not placed yet , Y  order is processed first time )
    UPDATE transaction table1
    SET id_FLAG = ‘Y’
    WHERE id_no = v_no;
    END LOOP;
    EXCEPTION WHEN OTHER THEN
    DBMS_OUTPUT.PUT_LINE( );
    END ;
    BLOCK 2
    BEGIN -- block 2 starts
    FOR v_update IN Cursor_Update;
    LOOP;
    INSERT INTO transaction table2 ( id_no ,id_prev_no, dml_action , …. ) VALUES (v_id_seq_no, v_upload.id_no ,, … ) RETURNING v_upload.id_no INTO v_no
    UPDATE transaction table1
    SET id_FLAG = ‘Y’
    WHERE id_no = v_no;
    END LOOP;
    EXCEPTION WHEN OTHER THEN
    DBMS_OUTPUT.PUT_LINE( );
    END; -- block2 end
    END app_proc; -- Main block end
    Sample output in Transaction table1 :
    Id_no | Tax_amt | re_emburse_amt | Activ_DT | Id_Flag | DML_ACTION
    01 1,835 4300 12/JUN/2009 N I ( these DML Action will be triggered when ever if theres in any DML operation occurs in this table )
    02 1,675 3300 12/JUN/2009 Y U
    03 4475 6500 12/JUN/2009 N D
    Sample output in Transaction table2 :
    Id_no | Prev_id_no Tax_amt | re_emburse_amt | Activ_DT
    001 01 1,835 4300 12/JUN/2009 11:34 AM ( 2nd cursor will populate this value , bcoz there s an update happened for the below records , this is 2nd voucher
    01 0 1,235 6300 12/JUN/2009 09:15 AM ( 1st cursor will populate this record when job run first time )
    02 0 1,675 3300 12/JUN/2009 8:15AM
    003 03 4475 6500 12/JUN/2009 11:30 AM
    03 0 1,235 4300 12/JUN/2009 10:30 AM
    Now the issues is :
    When these Process runs, our other application jobs failing, because it also uses these main 2 tranaction table. So dead lock is detecting in these applications.
    Solutin Needed :
    Can anyone suggest me , like how can rectify this blocking /Locking / Dead lock issues. I wants my other application also will use this tables during these process.
    Regards,
    Maran

    hmmm.... this leads to a warning:
    SQL> ALTER SESSION SET PLSQL_WARNINGS='ENABLE:ALL';
    Session altered.
    CREATE OR REPLACE PROCEDURE MYPROCEDURE
    AS
       MYCOL VARCHAR(10);
    BEGIN
       SELECT col2
       INTO MYCOL
       FROM MYTABLE
       WHERE col1 = 'ORACLE';
    EXCEPTION
       WHEN PIERRE THEN
          NULL;
    END;
    SP2-0804: Procedure created with compilation warnings
    SQL> show errors
    Errors for PROCEDURE MYPROCEDURE:
    LINE/COL                                                                          ERROR
         12/9        PLW-06009: procedure “MYPROCEDURE” PIERRE handler does not end in RAISE or RAISE_APPLICATION_ERROR
         :)

  • User Lock issue when processing IDOCS

    Hi Folks,
    We are pushing the data into SAP using IDOCS.During this process some IDOCS are getting failed due to User Lock on shipment header.Anyone here can share their experience in dealing with User Lock issues.
    Thanks,
    Kiran.

    Kiran,
    Only one can edit the document, it's standard fucntion. I think No note can resolve this. I can suggest you to run a back ground job very frequently ,which picks these status 51 IDOCs and reprocess. Report is :RBDMANi2
    We can use Message class & error message number on selection screen, so that we can make sure that only those idocs which were failed due to Lock problem are taken by the job.
    Reddy

  • Data locks in BPC

    Hi,
    In BPS, when user A is modifying a dataset, user B at the same time can't modify the same dataset because of the data locks.
    I wonder how does it work in BPC. In script logic when you add a particular value, say company 1000, in the dataset using XDIM_MEMBERSET instruction, does it mean that now company 1000 is locked and can't be modifed by another user at the same time?
    Thanks, Rahul

    Hi Rahul,
    There is no lock when users are editing data in BPC. Two users could edit the same data, so the second user could overwrite the data of the first one.
    To avoid this situations, you can.
    - Define security accordingly (only one user can write to a particular entity)
    - Define Business Process Flows (so that each user is working only on his data range)
    In the end, this is more a process issue, if two users are responsible to plan the same figures
    Regards,
    Marcel

  • Adobe Cluster Locking issues

    We have two nodes in the cluster. Adobe is installed on just one node. Locator is running on one node. Experiencing locking issues for temporary diskstore. Below is the error description.
    [12/10/14 13:13:44:925 EST] 00000013 webapp   
    I com.ibm.ws.webcontainer.webapp.WebGroupImpl WebGroup SRVE0169I: Loading Web Module: dsc.war.
    [12/10/14 13:13:44:935 EST] 00000013 WASSessionCor I SessionContextRegistry getSessionContext SESN0176I: Will create a new session context for application key default_host/dsc
    [12/10/14 13:13:44:963 EST] 00000013 ServletWrappe I com.ibm.ws.webcontainer.servlet.ServletWrapper init SRVE0242I: [LiveCycleES4] [/dsc] [UploadServlet]: Initialization successful.
    [12/10/14 13:13:44:970 EST] 00000013 ArchiveFileMa A   ArchiveFileManagerImpl:getLocalCacheRootDir():Adobe tmp directory is:/data/exigen_test/PT/jvm/tmp/adobews__-581654660
    [12/10/14 13:14:17:041 EST] 00000013 LocalExceptio E   CNTR0020E: EJB threw an unexpected (non-declared) exception during invocation of method "getDomain" on bean "BeanId(LiveCycleES4#adobe-pof.jar#adobe_POFDataDictionaryLocalEJB, null)". Exception data: com.gemstone.gemfire.cache.DiskAccessException: For DiskStore: ReplicatedGuaranteed_POF_DD_CACHE: Could not lock "/data/exigen_test/PT/jvm/tmp/adobews__-581654660/Caching/Data/ReplicatedGuaranteed_POF_D D_CACHE/DRLK_IFReplicatedGuaranteed_POF_DD_CACHE.lk". Other JVMs might have created diskstore with same name using the same directory., caused by java.io.IOException: No locks available
    at com.gemstone.gemfire.internal.cache.DiskStoreImpl.createLockFile(DiskStoreImpl.java:1800)
    at com.gemstone.gemfire.internal.cache.DiskStoreImpl.loadFiles(DiskStoreImpl.java:1920)
    at com.gemstone.gemfire.internal.cache.DiskStoreImpl.<init>(DiskStoreImpl.java:355)
    at com.gemstone.gemfire.internal.cache.DiskStoreImpl.<init>(DiskStoreImpl.java:270)
    at com.gemstone.gemfire.internal.cache.DiskStoreImpl.<init>(DiskStoreImpl.java:265)
    at com.gemstone.gemfire.internal.cache.DiskStoreFactoryImpl.create(DiskStoreFactoryImpl.java :127)
    at com.adobe.livecycle.cache.adapter.GemfireCacheAdapter.createSubRegion(GemfireCacheAdapter .java:1043)
    at com.adobe.livecycle.cache.adapter.GemfireCacheAdapter.init(GemfireCacheAdapter.java:234)
    at com.adobe.livecycle.cache.adapter.GemfireCacheAdapter.<init>(GemfireCacheAdapter.java:215 )
    at com.adobe.livecycle.cache.adapter.CacheAdapterFactory.getCache(CacheAdapterFactory.java:1 03)
    at com.adobe.pof.POFUtil.getPOFDataDictionaryCache(POFUtil.java:525)
    at com.adobe.pof.adapter.AbstractAdapterContext.getDataDictionaryCache(AbstractAdapterContex t.java:89)
    at com.adobe.pof.adapter.AbstractAdapterContext.getDomainFromCache(AbstractAdapterContext.ja va:127)
    at com.adobe.pof.adapter.JDBCAdapter.getDomain(JDBCAdapter.java:4436)
    at com.adobe.pof.schema.POFDataDictionaryImpl.getDomain(POFDataDictionaryImpl.java:113)
    at com.adobe.pof.schema.POFDataDictionaryRemoteBean.getDomain(POFDataDictionaryRemoteBean.ja va:148)
    at com.adobe.pof.schema.EJSLocalStatelessadobe_POFDataDictionaryLocalEJB_4137caa5.getDomain( Unknown Source)
    at com.adobe.pof.schema.POFDataDictionaryLocalEJBAdapter.getDomain(POFDataDictionaryLocalEJB Adapter.java:69)
    at com.adobe.pof.odapi.POFSchemaManagerRemoteBean.getDomain(POFSchemaManagerRemoteBean.java: 577)
    at com.adobe.pof.odapi.EJSLocalStatelessadobe_POFSchemaManagerLocalEJB_626825cc.getDomain(Un known Source)
    at com.adobe.pof.odapi.POFSchemaManagerLocalEJBAdapter.getDomain(POFSchemaManagerLocalEJBAda pter.java:169)
    at com.adobe.idp.dsc.startup.DSCStartupServlet$1.doInTransaction(DSCStartupServlet.java:137)
    at com.adobe.idp.dsc.transaction.impl.ejb.adapter.EjbTransactionBMTAdapterBean.doRequiresNew (EjbTransactionBMTAdapterBean.java:218)
    at com.adobe.idp.dsc.transaction.impl.ejb.adapter.EJSLocalStatelessEjbTransactionBMTAdapter_ 3af08fdf.doRequiresNew(Unknown Source)
    at com.adobe.idp.dsc.transaction.impl.ejb.EjbTransactionProvider.execute(EjbTransactionProvi der.java:133)
    at com.adobe.idp.dsc.transaction.impl.DefaultTransactionTemplate.execute(DefaultTransactionT emplate.java:79)
    at com.adobe.idp.dsc.startup.DSCStartupServlet.init(DSCStartupServlet.java:132)
    at javax.servlet.GenericServlet.init(GenericServlet.java:161)
    at com.ibm.ws.webcontainer.servlet.ServletWrapper.init(ServletWrapper.java:342)
    at com.ibm.ws.webcontainer.servlet.ServletWrapperImpl.init(ServletWrapperImpl.java:168)
    at com.ibm.ws.webcontainer.servlet.ServletWrapper.loadOnStartupCheck(ServletWrapper.java:136 6)
    at com.ibm.ws.webcontainer.webapp.WebApp.doLoadOnStartupActions(WebApp.java:618)
    at com.ibm.ws.webcontainer.webapp.WebApp.commonInitializationFinally(WebApp.java:587)
    at com.ibm.ws.webcontainer.webapp.WebAppImpl.initialize(WebAppImpl.java:421)
    at com.ibm.ws.webcontainer.webapp.WebGroupImpl.addWebApplication(WebGroupImpl.java:88)
    at com.ibm.ws.webcontainer.VirtualHostImpl.addWebApplication(VirtualHostImpl.java:169)
    at com.ibm.ws.webcontainer.WSWebContainer.addWebApp(WSWebContainer.java:749)
    at com.ibm.ws.webcontainer.WSWebContainer.addWebApplication(WSWebContainer.java:634)
    at com.ibm.ws.webcontainer.component.WebContainerImpl.install(WebContainerImpl.java:422)
    at com.ibm.ws.webcontainer.component.WebContainerImpl.start(WebContainerImpl.java:714)
    at com.ibm.ws.runtime.component.ApplicationMgrImpl.start(ApplicationMgrImpl.java:1165)
    at com.ibm.ws.runtime.component.DeployedApplicationImpl.fireDeployedObjectStart(DeployedAppl icationImpl.java:1369)
    at com.ibm.ws.runtime.component.DeployedModuleImpl.start(DeployedModuleImpl.java:639)
    at com.ibm.ws.runtime.component.DeployedApplicationImpl.start(DeployedApplicationImpl.java:9 67)
    at com.ibm.ws.runtime.component.ApplicationMgrImpl.startApplication(ApplicationMgrImpl.java: 771)
    at com.ibm.ws.runtime.component.ApplicationMgrImpl$5.run(ApplicationMgrImpl.java:2169)
    at com.ibm.ws.security.auth.ContextManagerImpl.runAs(ContextManagerImpl.java:5468)
    at com.ibm.ws.security.auth.ContextManagerImpl.runAsSystem(ContextManagerImpl.java:5594)
    at com.ibm.ws.security.core.SecurityContext.runAsSystem(SecurityContext.java:255)
    at com.ibm.ws.runtime.component.ApplicationMgrImpl.start(ApplicationMgrImpl.java:2174)
    at com.ibm.ws.runtime.component.CompositionUnitMgrImpl.start(CompositionUnitMgrImpl.java:389 )
    at com.ibm.ws.runtime.component.CompositionUnitImpl.start(CompositionUnitImpl.java:123)
      at com.ibm.wsspi.runtime.component.WsComponentImpl$_AsynchInitializer.run(WsComponentImpl.ja va:496)
            at com.ibm.ws.util.ThreadPool$Worker.run(ThreadPool.java:1702)
    Caused by: java.io.IOException: No locks available
            at sun.nio.ch.FileChannelImpl.tryLock(FileChannelImpl.java:924)
            at java.nio.channels.FileChannel.tryLock(FileChannel.java:973)
            at com.gemstone.gemfire.internal.cache.DiskStoreImpl.createLockFile(DiskStoreImpl.java:1781)
            ... 56 more
    Any immediate help would be appreciated to resolve the issue. Thank you.

    Similar issue : https://forums.adobe.com/message/6156015#6156015
    Check the configuration of temp as well as GDS directories for the cluster : http://www.avoka.com/blog/2008/06/livecycle-directories-global-storagetemp-clustering-and- more/
    -Wasil

  • Why BC4J PS_TXN table , sometime may occur data  lock situation ?

    My appliation be develop in Jdev 10.1.2 with BC4J/JSP solution, but in production environment sometime when the application user request is heavy will occur the PS_TXN data lock situation and caused the application be hang and no any other user can access the application, so what situation wrong or event happen may caused the PS_TXN lock issue ? who have meet the situation and have the good idea to avoid the issue ?

    Hi Nies......
    If you want to Import Libraries to your target client then you need BASIS Consultant..........
    For any report writer report we need a library.
    Go to GR23 & specify the library name 1VK & press 'Enter'.
    If it doesn't display......
    Please check with your FI consultant as they are the users of the report writer reports and they only create the same.
    there may be chance that this library has been deleted.
    Is your system newly upgraded?
    Solution..........
    Library 1VK is SAP standard and exists in table T801K in client 000.
    1. Create transport of table with library and transport to target client.
    2. Or in target client, use transaction OKD3 to import the library 1VK from 000.
    3. You can also use transaction GR29 to import individual libraries.
    Thanks
    Dhruv Malhotra

  • Locking issue with BAPI_GOODSMVT_CREATE in userexit_save_document

    Hello All,
    I am performing a separate goods movement for one of the materials in the delivery userexit_save_document during PGI. I have an issue. We might have multiple users trying to PGI deliveries and thereby performing the goods movement for the same material through the BAPI in the userexit_save_document.
    How do I prevent the locking issue when say two users go through the BAPI for goods movement for the same material data at the same time.
    Any answer in this regards is highly appreciated.
    Thanks,
    Mathangi

    Thanks Rich... this make work for me, but one question -
    Do you know if this function releases all lock objects for the user system wide or only for the current session? My concern is that if it is system wide, then I could be releasing lock objects for other sessions unintentionally. For example, if there are multiple RFC calls in parallel for the same function and same user then I could be destroying lock entries that I shouldn't be! I am hoping that it removes lock entries only for the current session...
    any ideas?
    Regards,
    Lee
    > I've used this BAPI in many of my programs,  and
    > always after the call, I code.....
    >
    >
    >
    >     commit work and wait.
    >     call function 'DEQUEUE_ALL'.
    >
    >
    >
    > The DEQUEUE_ALL show release any/all locks.
    >
    >
    > Regards,
    > Rich Heilman

  • JPasswordField locking issue in linux

    Hi,
    password field is working good if its alone in the panel/frame. if added textfield for username before password field, password field is locking and its never allowing to enter the data. this issue is in linux.* the same code works well in windows platform. is this a bug? i am using latest jre though....
    public class test3 extends JFrame implements ActionListener
        JTextField txtUserName=new JTextField(10);
        JPasswordField pwd = new JPasswordField(10);
        public test3()
             setDefaultCloseOperation(JFrame.EXIT_ON_CLOSE);
             getContentPane().setLayout(new GridLayout(2,1));
             getContentPane().add(txtUserName);
         getContentPane().add(pwd);
         txtUserName.addActionListener(this);
         pwd.addActionListener(this);
         pack();
         setVisible(true);
        public void actionPerformed(ActionEvent e)
             System.out.println("UserName=" + txtUserName.getText());
             System.out.println("Password=" + pwd.getText());
        public static void main(String arg[])
             new test3();
    }My environment...
    [root@uma ~]# java -version
    java version "1.6.0_13"
    Java(TM) SE Runtime Environment (build 1.6.0_13-b03)
    Java HotSpot(TM) Server VM (build 11.3-b02, mixed mode)
    [root@uma ~]#
    [root@uma ~]# uname -a
    Linux uma 2.6.18-92.1.13.el5PAE #1 SMP Wed Sep 24 20:07:49 EDT 2008 i686 i686 i386 GNU/Linux
    [root@uma ~]#
    [root@uma ~]# cat /etc/redhat-release
    CentOS release 5.2 (Final)

    Note: This thread was originally posted in the [Java Programming|http://forums.sun.com/forum.jspa?forumID=31] forum, but moved to this forum for closer topic alignment.

  • User1 locking other user (STS Locking Issue SAP IP)

    (STS Locking Issue SAP IP)
    User1 Is locking up other functional org layouts that he does not even have access to it.User of these layouts user2 gets message every time tries to allocate data that the Info provider is locked by user1 .
    even though user1 does not have access to orgs that user 2 has.
    How do i need to resolve this issue?

    Hello,
    if user1 locks all records for a second attribute (besides org) then this will also result in a lock of all records even though the org is distictive. Try to have a look in the lock overview of rsplan and possibly repost the lock arguments.
    Hope that helps!
    Cheers,
    Yann

  • Report Designer Locking issue

    Hi,
       We have found locking issue while accessing same report(Created in Report Designer) by multi users at a time .
       Has anyone face this issue and How to fix it. It's major concern because same report can be accessed by multiusers at  same time
    Thanks
    Siva

    Siva,
    only one person at a time has permission to open the report in the Report designer - just like with Queries in the query designer.  This should not have any effect on you actually executing the report in the portal - multiple people can do it then.
    this is a security issue that other software vendors use as well - just like you cannot open any MS Office files on the network for editing that someone has opened already.
    Thanks

  • Windows 8.1 Data reordering issue with Intel Adaptors

    According to Intel, there is a data reordering issue with their adaptors and probably this dumb WIDI software. This is from Intel site. they say some are fixed, "A future Windows 8 fix will address this issue for other Intel wireless adapters." I
    have one Nope, still broke. I get drops all the time. Brand new Toshiba laptop I7 16 gigs of ram and a SSD and a 2 gig Vid card. Would be nice to be able to play games but I get dropped all the time. Now would Microsoft quit hiding
    about this, and fix the darn thing. Also i'm a system admin for 13 years. I have build over 1000 PCs and servers. I know bad software. Please fix this. PLEASE. Its not going to just go away and its not just Toshiba, I have seen other companies with the
    same problem. If there is a Fix PLEASE POST IT. Or even a workaround I have tried everything.
    http://www.intel.com/support/wireless/wlan/sb/CS-034535.htm
     

    Hi,
    Have your first tried the software fix under this link for your network adapter?
    http://www.intel.com/support/wireless/wtech/proset-ws/sb/CS-034041.htm
    Please Note: The third-party product discussed here is manufactured by a company that is independent of Microsoft. We make no warranty, implied or otherwise, regarding this product's performance or reliability.
    Also, you can try to check if there is any driver update under Device manager from manufacture's website.
    Kate Li
    TechNet Community Support
    Yep didn't work. Still get drops all the time, had to run a Cat 5E cable to my laptop from my modem, because I have Atheros Gigabyte Lan adaptor. Works Great. The Wireless still drops all the time. Has Microsoft let out the patch to fix this or is it coming in
    April in the 8.1 patch that's coming. Funny thing is all for Widi, I don't even use widi, I got the software to do that from Samsung works better on my TV. Intel and Microsoft need to get this fixed. because their driving off gamers and that's the
    people that make sure they buy Microsoft so they can play games. With the wireless link dead and a great laptop worthless what's the point. Ive been in IT for 13 years building PCs  and Servers how I knew how to run a 60 FT Cat 5e line thru
    a 2 story house and terminate it. Most people don't. Fix the problem.  

  • Data Load Issue "Request is in obsolete version of DataSource"

    Hello,
    I am getting a very strange data load issue in production, I am able to load the data upto PSA, but when I am running the DTP to load the data into 0EMPLOYEE ( Master data Object) getting bellow msg
    Request REQU_1IGEUD6M8EZH8V65JTENZGQHD not extracted; request is in obsolete version of DataSource
    The request REQU_1IGEUD6M8EZH8V65JTENZGQHD was loaded into the PSA table when the DataSource had a different structure to the current one. Incompatible changes have been made to the DataSource since then and the request cannot be extracted with the DTP anymore.
    I have taken the follwoing action
    1. Replicated the data source
    2. Deleted all request from PSA
    2. Activated the data source using (RSDS_DATASOURCE_ACTIVATE_ALL)
    3. Re transported the datasource , transformation, DTP
    Still getting the same issue
    If you have any idea please reply asap.
    Samit

    Hi
    Generate your datasource in R/3 then replicate and activate the transfer rules.
    Regards,
    Chandu.

Maybe you are looking for