Issue while reflecting the data to ADF form from ADF Table

Hi All,
I have one scenario as follows:
I have to open a form in entry mode. So I have used a Method Call activity in task flow to call "Create" in method call activity.
Then form opens in a entry mode. I have a adf table in the page showing the entered data.
When form opens in entry mode, instead of entering the values in the field User selects a record in ADF table, but due to the create mode of the form it will through me a error to fill mandatory fields.
Please tell me the way so that if user selects the record(in adf table) instead of entering new record, the value would reflect in the Adf form.
If I use a roll back on selecting the row of ADF table, then it is showing me this error "Row currency has changed since the user interface was rendered. The expected row key was oracle.jbo.Key[null ]" because there will be no row key in the cache when form opens.
Please suggest me the way to complete the task.
I am using jdeveloper 11.1.2.1.0
Thanks,
Gobi
Edited by: gobinkl on Oct 3, 2012 2:06 AM

no no your are complicating yourself. :) better your should provide your usecase.
first thing: go through the books & also adf blogs learn and understand the framework. as above person said.
i hope that you had method call create as activity as default while page rendering it make your af: form empty(that is ready fir insertion format).
so if your are doing like that.am sure it will throw some mandatory errors.
error will not resolve unless without entering any data's.
coming af:table why are using af:table - immediate = true / false(switching).
immediate = true - skip over validation or by pass some validation in some phase - based on the component which you are using. (say as if you are using editable value holder,af:coomand button).
grab the information here in an more legible manner.
http://adfpractice-fedor.blogspot.in/2012/02/understanding-immediate-attribute.html
let me know your usecase.
this will assist you
https://blogs.oracle.com/shay/entry/executing_an_action_on_jsf_pag
http://tanveeroracle.blogspot.in/2009/09/adf-11g-createinsert-to-display-blank.html
Edited by: user707 on Oct 5, 2012 9:02 AM

Similar Messages

  • Idoc issue while loading the data into BI

    Hello Gurus,
    Initially i am having the Source System connection problem. After it was fixed by the basis i had followed the below process.
    I am loading the data using the Generic extractor specifying the selection and its a full load. When i load the data from R3 PP application  i found the below result in the monitor screen of BW.
    1. Job was completed in the R3 system and 1 million records are fetched by extractor.
    2. the records are not posted to the BW side Because of TRFC issue.
    3. i got the idoc number and process it in the BD87 tCode. But it was not process sucessfully it gives the following error " Error when saving BOM. Please see the application log."
    4. when i check the Application Log using the Tcode SLG1 with the help of time and date of that particular process it was in the yelow status.
    Kindly let me know i can i resolve this issue. i have already tried by repeating the infopackage job still i am facing the same issue. I have also check the connection its is ok.
    Regards

    hello veerendra,
    Thanks for your quick response. yes i am able to process it manually. after processing  it was ended with the status 51 application not posted.
    could you pls help me out with the same
    regard
    Edited by: KK on Nov 4, 2009 2:19 AM
    Edited by: KK on Nov 4, 2009 2:28 AM

  • Issue while fetching the file through *.extension by FTP sender file adapte

    Hello Experts,
    I am facing a issue while fetching the data through sender File adapter with  ' *.file extension' .I am illustarting the scenario as below .
    It is a simple scenarion of File to File inboumd scenarion.Here file is getting picked up from a third party system
    through FTP sender chanel and stored in a temp folder of PI through NFS rceiver file adapter .
        The problem is however I am getting, while picking the file with file name as "*.exo"(where exo is the file extension).
    But while fetching the file with particular name like"abcd_10032011*.exo"(file naming has been done by combination of abcd(always same)_currentdate(change according to current date)),file picked successfully .
    So here ,in the prior case file not getting picked up,but in later case it dose .
    Can anyone please let me know what might be the issue?

    Hi Sunit,
    Connect from your PI System to 3rd Party System (where are placed the Source Files)
    FTP <PartySystemHostName>
    eg. FTP 10.2.3.456 (then insert Username & Password to Login)
    Go to source directory
    cd \<SourceDirectory>
    eg. cd \donaldduck\directory\
    Execute a File List command
    ls -la *.<extension>
    eg. ls -la *.exo
    In this way you should be able to view all files with this extension (*.exo), the same action that Sap XI perform to pickup the file.
    Then, try to copy that file to your Local PI System (to check if there're some permissions issue):
    mget <filename>.exo
    eg. mpget File1_01012011.exo

  • Issues with processin the data in Adpater engine

    Dear experts,
    we are facing the below mention issues while processing the data Please let me know what we can do to resolve the issue.
    "Error while reading ID of own business system from the SLD for system DX1 and client 500 Internal error in Landscape Directory"
    regards
    shankar

    Hi,
    As suggested by Aashish, you can go through.
    It seems to like problem with SLD.  It might be wrongly configured. Please check with the T.code SLDAPICUST
    Regards,
    P.Rajesh

  • Error While Activating the Data Source in RSA5 T-Code

    Hi,
    When i tried to inactive the data source in 2lis_11_vahdr it shows "Data Source does not exists errror".
    So i went to RSA5 T-Code and tried to activate the data source but there it is throwing the error like "Event with errors Extract in Data Source 2LIS_11_VAHDR".
    I have checked the above thing with all the data sources every one are throwing the same errors. Can any one please suggest me how to solve the issue.
    Or Is this issue related to Basis People?.
    Regards,
    Kumar.

    Hi Ravi,
    1.When i tried to inactive the data source  in  LBWE T-Code the error is "Data Source 2lis_11_vahdr does not exists ".
    2. In RSA5 T-Code  while activating the data source the error is  "Event with errors Extract in Data Source 2LIS_11_VAHDR".
    Regards,
    Kumar.

  • Error while loading the data from PSA to Data Target

    Hi to all,
         I'm spacing some error while loading the data to data target.
    Error :  Record 1 :Value 'Kuldeep Puri Milan Joshi ' (hex. '004B0075006C0064006500650070002000500075007200690
    Details:
    Requests (messages): Everything OK
    Extraction (messages): Everything OK
    Transfer (IDocs and TRFC): Errors occurred
          Request IDoc : Application document posted
          Info IDoc 2 : Application document posted
          Info IDoc 1 : Application document posted
          Info IDoc 4 : Application document posted
          Info IDoc 3 : Application document posted
          Data Package 1 : arrived in BW ; Processing : Data records for package 1 selected in PSA - 1 er
    Processing (data packet): Errors occurred
          Update PSA ( 2462  Records posted ) : No errors
          Transfer Rules ( 2462  -> 2462  Records ) : No errors
          Update rules ( 2462  -> 2462  Records ) : No errors
          Update ( 0 new / 0 changed ) : Errors occurred
          Processing end : Errors occurred
    I'm totally new to this issue. please help to solve this error.
    Regards,
    Saran

    Hi,
    I think you are facing an invalid character issue.
    This issue can be resolved by correcting the error records in PSA and updating it into the target. For that the first step should be to identify if all the records are there in PSA. You can find out this from checking the Details tab in RSMO, Job log , PSA > sorting records based on status,etc. Once its confirmed force the request to red and delete the particular request from the target cube. Then go to PSA and edit the incorrect records (correcting or blanking out the invalid entries for particular field InfoObject for the incorrect record) and save it. Once all the incorrect records are edited go to RSA1>PSA find the particular request and update to target manually (right click on PSA request > Start update immediately).
    I will add the step by step procedure to edit PSA data and update into target (request based).
    In your case the error message says Error : Record 1 :Value 'Kuldeep Puri Milan Joshi '. You just need to conver this to Capital letter in PSA and reload.
    Edit the field to KULDEEP PURI MILAN JOSHI in PSA and push it to target.
    Identifying incorrect records.
    System wont show all the incorrect records at the first time itself. You need to search the PSA table manually to find all the incorrect records.
    1. First see RSMO > Details > Expand upate rules / processing tabs and you will find some of the error records.
    2. Then you can go to PSA and filter using the status of records. Filter all the red requests. This may also wont show the entire incorrect records.
    3. Then you can go to PSA and filter using the incorrect records based on the particular field.
    4. If this also doesnt work out go to PSA and sort (not filter) the records based on the particular field with incorrect values and it will show all the records. Note down the record numbers and then edit them one by one.
    If you want to confirm find the PSA table and search manually."
    Also Run the report RS_ERRORLOG_EXAMPLE,By this report you can display all incorrected records of the data & you can also find whether the error occured in PSA or in TRANSFER RULES.
    Steps to resolve this
    1. Force the request to red in RSMO > Status tab.
    2. Delete the request from target.
    3. Come to RSMO > top right you can see PSA maintenace button > click and go to PSA .
    4.Edit the record
    5. Save PSA data.
    6. Got to RSA15 > Search by request name > Right click > update the request from PSA to target.
    Refer how to Modify PSA Data
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/40890eda-1b99-2a10-2d8b-a18b9108fc38
    This should solve your problem for now.
    As a long term you can apply some user exit in source system side or change your update rules to ensure that this field is getting blanked out before getting loaded in cube or add that particular char to permitted character list in BW.
    RSKC --> type ALL_CAPITAL --> F8 (Execute)
    OR
    Go to SE38 and execute the program RSKC_ALLOWED_CHAR_MAINTAIN and give ALL_CAPITAL or the char you want to add.
    Check the table RSALLOWEDCHAR. It should contain ALL_CAPITAL or the char you have entered.
    Refer
    /people/sap.user72/blog/2006/07/23/invalid-characters-in-sap-bw-3x-myths-and-reality-part-2
    /people/sap.user72/blog/2006/07/08/invalid-characters-in-sap-bw-3x-myths-and-reality-part-1
    /people/aaron.wang3/blog/2007/09/03/steps-of-including-one-special-characters-into-permitted-ones-in-bi
    http://help.sap.com/saphelp_nw04/helpdata/en/64/e90da7a60f11d2a97100a0c9449261/frameset.htm
    For adding Other characters
    OSS note #173241 – “Allowed characters in the BW System”
    Thanks,
    JituK
    Edited by: Jitu Krishna on Mar 22, 2008 1:52 PM

  • Dump While uploading the data in RSA1 from R3 system

    Hi All,
    We are getting a dump in st22 while uploading the data in RSA1,pls find the details below-:
    Error analysis
        Short text of error message:
        Password logon no longer possible - too many faile d attempts
        Long text of error message:
        Technical information about the message:
        Message class....... "RSAR"
        Number.............. 051
        Variable 1.......... "Password logon no longer possible - too many faile"
        Variable 2.......... "d attempts"
        Variable 3.......... " "
        Variable 4.......... " "
    How to correct the error
        Probably the only way to eliminate the error is to correct the program.
        If the error occures in a non-modified SAP program, you may be able to
        find an interim solution in an SAP Note.
        If you have access to SAP Notes, carry out a search with the following
        keywords:
        "MESSAGE_TYPE_X" " "
        "SAPLRSSM" or "LRSSMU17"
        "RSSM_CALL_RSAR_DATA_REQUEST_1"
    I have checked both R3 and BI system there is no user which is currently locked.
    also I have checked the SAP Notes 1088927,1066180 ,these notes recommends to upgrade the SAP_BW component to 16 level but we are already running at 18.
    Pls suggest
    Regards,
    Prashant

    Hi Prashant
    Your system user is unable to authenticate access on the remote system. This may be an issue for RFC user locking. Check the RFC from SM59 for the commection between the two systems. Check for the User ID and password that is saved on the RFC logon.
    Also check the system user ALEREMOTE.
    I hope this helps
    Regards
    Chen

  • Error "Encountered an error while exporting the data. View the lof file (DPR-10126)

    Hello,
    I would like to have sugestions and help to solve a problem in IS 4.2.
    I've created a view from two tables, joined together with the key from each one.
    Then i've binded the rule with the view (mentioned above) created and calculated score for it.
    The calculation is successful, however when i try to export all failed data i have the following error:
    "Encountered an error while exporting the data. View the lof file (DPR-10126)"
    When exporting failed data from calculations made directly to the datasources i do not encounter any problem, but when i try to export all data to CSV file (not only the 500 results limit) i have this kind of error.
    However, when i visualize the data from the view created, it is also possible to export to csv file.
    Resuming, the error occurs only when exporting all failed data from the view.
    Can anyone help with this issue?
    Thx
    Best Regards,

    Hello,
    Your information is correct.
    I've just confirmed it in the 4.2 SP3 Release Notes and they specifically refer to this issue.
    A user who has 'View objects' rights for the project and a failed rule connection cannot export the failed data. This issue has been fixed in this release.
    Thank you very much for your input.
    Best Regards,

  • Error while retracting the data from BW system

    Dear all,
    while retracting the data from BW we are getting the below dump
    "Inconsistent entry in table TKA07 (Plan Version Maintenance)",Request you to please help to solve the issue
    Thanks,
    Jyoti

    Jyoti,
    Could you please explain in detail, were you using any standard retractor or custom?
    is it erroring out in ECC system or BI?

  • Issues while generating Schema DAT files

    We are facing two type of issues when generating Schema ".dat" files from Informix Database on Solaris OS using the
    "IDS9_DSML_SCRIPT.sh " file.
    We are executing the command on SOLARIS pormpt as follows..
    "IDS9_DSML_SCRIPT.sh <DBName> <DB Server Name> ".
    The first issue is ,after the command is excuted ,while generating the ".dat" files the following error is occuring .This error is occuring for many tables
    19834: Error in unload due to invalid data : row number 1.
    Error in line 1
    Near character position 54
    Database closed.
    This happens randomly for some schemas .So we again shift the script to a different folder in Unix and execute it.
    Can we get the solution for avoiding this error.
    2. The second issue is as follows..
    When the ".dat" files are generated without any errors using the script ,these .dat files are provided to the OMWB tool to load the Source Model.
    The issue here is sometimes OMWB is not able to complete the process of creating the Source Model from the .dat files and gets stuck.
    Sometimes the tables are loaded ,but with wrong names.
    For example the Dat files is having the table name as s/ysmenus for the sysmenus table name.
    and when loaded to oracle the table is created with the name s_ysmenus.
    Based on the analysis and understanding this error is occuring due to the "Delimiter".
    For example this is the snippet from a .dat file generated from the IDS9_DSML_SCRIPT.sh script.The table name sysprocauthy is generated as s\ysprocauthy.
    In Oracle this table is created with the name s_ysprocauthy.
    s\ysprocauthy║yinformixy║y4194387y║y19y║y69y║y4y║y2y║y0y║y2005-03-31y║y65537y║yT
    y║yRy║yy║y16y║y16y║y0y║yy║yy║y╤y
    Thanks & Regards
    Ramanathan KrishnaMurthy

    Hello Rajesh,
    Thanks for your prompt reply. Please find my findings below:
    *) Have there been any changes in the extractor logic causing it to fail before the write out to file, since the last time you executed it successfully? - I am executing only the standard extractors out of the extractor kit so assumbly this shouldnt be a issue.
    *) Can this be an issue with changed authorizations? - I will check this today, bt again this does not seem to be possible as the same object for a different test project i created executed fine and a file was created.
    *) Has the export folder been locked or write protected at the OS level? Have the network settings (if its a virtual directory) changed? - Does not seem so because of the above reason.
    I will do some analysis today and revert back for your help.
    Regards
    Gundeep

  • Issue while playing the content of Azure Media Services after deleting the channel.

    Hi,
    We are having the issue while playing the content of Azure Media Services after deleting the channel.
    Please find detail scenarios below:
    We have done live streaming of an event using Azure Media Services, event was around 5-6 hrs. After completion of event channel was deleted, when we checked content data
    it was only 1hr. Content has only last 1 hr. recording, we are unable to get full recoding.
    We divided event into two parts to avoid above issue. After completion of 1st part which was of 2 hrs., we stopped the streaming and deleted the channel. New channel was
    created and streaming was started for 2<sup>nd</sup> part, after completion of event channel was deleted. Now we are unable to play the content data for 1st part but able to play the 2nd part of event through content publish URL.
    We have done streaming of 3 hrs. event on 18th April with same process and full recoding was available. 
    Can you let me know what might be the issue?

    Hi,
    Please refer to Justin’s blog and see if it helps:
    http://www.justinkobel.com/post/2012/04/30/SharePoint-2010-Claims%E2%80%93Operation-is-not-valid-due-to-the-current-state-of-the-object.aspx
    As suggested by Alex, please provide more information about your process of changing password.
    Here is an article for updating passwords on SharePoint 2010:
    http://blogs.technet.com/b/seanearp/archive/2011/01/25/updating-passwords-on-sharepoint-2010.aspx
    Regards,
    Rebecca Tu
    TechNet Community Support

  • Issue related with the Date Format --- error when activating the DSO

    Hi Friends..
    plz help me to fix the issue ...im getting when activating the DSO...
    Actually im not getting error when loading the data thru DTP..im facing this issue while activating the DSO..
    is there any routine to solve this or i have to go for data selection in info package ......
    I have searched in the forum also..but im not able to fix..plz help
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0CRMPOSTDAT*
      Value '31122009' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject ZKITRDT*
      Value '01/20/20' of characteristic 0DATE is not a number with 000008 spaces
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject ZMNS_DATE*
      Value '20101502' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0CRM_EXCDAT*
      Value '31122009' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0CRM_PRIDAT*
      Value '31122009' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject ZCOMPDAT*
      Value '31122009' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject ZDUEDAT*
      Value '20101402' of characteristic 0DATE is not plausible
    Thanks
    Baskar
    Edited by: tjgnanabaskar on Aug 6, 2010 5:35 PM

    Hi Friends..
    plz help me to fix the issue ...im getting when activating the DSO...Actually im not getting error when loading the data thru DTP..im facing this issue while activating the DSO..is there any routine to solve this or i have to go for data selection in info package ......I have searched in the forum also..but im not able to fix..plz help
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0CRMPOSTDAT
    Value '31122009' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject ZKITRDT
    Value '01/20/20' of characteristic 0DATE is not a number with 000008 spaces
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject ZMNS_DATE
    Value '20101502' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0CRM_EXCDAT
    Value '31122009' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0CRM_PRIDAT
    Value '31122009' of characteristic 0DATE is not plausible
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject ZCOMPDAT
    Value '31122009' of characteristic 0DATE is not plausible
    Thanks
    Baskar

  • Fire Fox sent not all the data in text form. Reached only piece of information. Where can I find the cache of sent text?

    Fire Fox sent not all the data in text form. Reached only piece of information. Where can I find the cache sent the text?
    For example, I posted an article on a news site, but found that only a small part of my article was posted.

    I'm not sure whether Firefox saved that information. There is a chance that it is in the session history file, especially if you haven't closed that tab yet. To check that:
    Open your currently active Firefox settings folder (AKA your Firefox profile folder) using
    Help > Troubleshooting Information > "Show Folder" button
    Find all files starting with the name sessionstore and copy them to a safe working location (e.g., your documents folder). In order to read a .js file, it is useful to rename it to a .txt file. Or you could drag it to Wordpad or your favorite word processing software.
    The script will look somewhat like gibberish, but if you search for a word in your article that doesn't normally appear in web pages, you may be able to find your data.
    Any luck?
    For the future, you might want to consider this add-on, although I don't know how long it saves data after you submit a form: [https://addons.mozilla.org/en-US/firefox/addon/lazarus-form-recovery/].

  • Error while extracting the data in R/3 production system,

    Hi Team,
    We got the following error while extracting the data in R/3 production system,
    Error 7 When Sending an IDoc R3 3
    No Storage space available for extending the inter 44 R3 299
    No storage space available for extending the inter R3 299
    Error in Source System RSM 340
    Please guide us to fix the issue

    It´s very difficult to help you without knowing
    - what is going to be transferred
    - where you get this error
    - system configuration
    - actual memory usage
    - operating system
    - database and configuration etc. etc.etc. etc.
    I suggest you open an OSS call and let the support have a look on your system. It´s much easier if one has system access to find out the cause for that problem.
    Markus

  • Error while activating the Data in DSO

    Hello,
    We moved huge amount of FI data from datasource to DSO. After moving all the data, we tried to activate the data. But the mistake we did is we activated all the requests at one go. (my assumption) It turned all the request to Red. The job log says SQL error.
    Is there any way of solving this issue without deleting the data. Data load will take too much time and we would like to avoid it if possible.
    Waiting for inputs!!!
    Regards,
    Abhishek

    yes a Dump was triggered.
    Database error text........: "ORA-12801: error signaled in parallel query
    server P007#ORA-01652: unable to extend temp segment by 2560 in tablespace
    PSAPTEMP"
    Internal call code.........: "[RSQL/FTCH//BIC/AYEPSPFI140 ]"
    Please check the entries in the system log (Transaction SM21).
    If the error occures in a non-modified SAP program, you may be able to
    find an interim solution in an SAP Note.
    If you have access to SAP Notes, carry out a search with the following
    keywords:
    "DBIF_RSQL_SQL_ERROR" "CX_SY_OPEN_SQL_DB"
    "CL_RSODSO_ACTIVATE_REQUESTS===CP" or "CL_RSODSO_ACTIVATE_REQUESTS===CM00D"
    "ACTIVATE_DATA_FROM_QUEUE"
    Edited by: Abhishek Nishis on Oct 12, 2009 10:54 PM

Maybe you are looking for