Data inconsistency in table port_inventory

We want to use the 'Open database schema support in CiscoWorks' to read data from the view port_inventory.
But we run in some trouble:
The data should present the data from the cisco works inventory jobs.
The job is running on our CiscoWorks once a day, but the data in the view port_inventory differ sometimes in time intervals of 10 minutes.
Also the data are in conflict with data retrived by Cisco Works GUI:
  e.g.
    in the view port_inventory a certain port on a certain switch has the VLAN-ID 'null'
    but in CiscoWorks this port appears as proper configured with a valid VLAN-ID
In addition: the data in the view changes sometimes approximately in intervals of 10 minutes.
Our runtime-environment is:
- Windows Server 2008 R2
- Cisco Works 4.0.1
Could anyone help us in understanding this problem?
Thanks for answers, Elmar

Hi AnshulR,
I think I know the cause, the inconsistency is due to the aggregation has compress, so even I deleted the requests from cube and reloaded, the compressed aggregation did not adjusted aotomatically. right?
And then, do you have some suggestion to recovery this issue? I would greatly appreciate if you could help me to resolve that.
note that the system is 3.5 and I did not use selective deletion just deleted the requests and reloaded them.
Thanks,
Daniel
Edited by: Delve on Mar 6, 2012 6:25 PM

Similar Messages

  • Data inconsistency problem TABLES: VBUK/VBUP vs VBAK/VBAP

    Hi,
    I have the fun task of trying to solve a data inconsistency problem caused by a faulty network connection between R/3 and our SQL server.
    To add to the fun my knowledge of the outbound systems is pretty close to zero.
    Apparently our warehouse was booking stock and after booking they process the STO NO in VL10B.
    While posting the outbound in VL02n some of the lines did not generate.
    So they deleted the outbound and tried again.
    After they finished playing around and our connections acted funky, I now have data for the outbound in tables VBUP/VBUK (Outbound status tables), but seemingly nowhere else.
    There is no data in ekpo, lips, vbak or vbap.
    So the question is:
    How on earth do I get rid of these entries safely, or otherwise try to generate the missing data?
    Just changing the statuses to completed did not work as they end up having problems when trying to pick more stock on articles that had entries on the missing outbound.
    Any other ideas?
    What other tables do I need to check for entries?
    Thanks
    Charl
    Message was edited(typos) by:
            Charl Holtzhausen

    Hi,
    With the help of SAP we solved the problem.
    Apart from VBUP/VBUK there was data in table VBBE (stock movement).
    Deleting the entries in these cleared the problem.
    It was important to check that further STOs were not adversely affected, which it seems it was not.
    Regards

  • Newbie: GUI front end to update data in a table

    Hi: I have some tables in a oracle database to which I want to provide a web based GUI that end users can update, delete and add new records. Is there an application (preferably freeware) that can do this? Basically the user interface should show the current list of records and allow user to delete existing records or add new records.
    Thanks
    Ray

    This is not a Web interface, but SQL Developer can be used without any Oracle client installation, just a runtime.
    And you have iSQL*Plus which allow you to run any query, almost like the overknown SQL*Plus.
    By the way, why did you want to allow all your users to modify data without any application ? You may have some problem of data inconsistency for the business logic, if the data are updated without any application code.
    Nicolas.

  • Copying large amount of data from one table to another getting slower

    I have a process that copies data from one table (big_tbl) into a very big archive table (vb_archive_tbl - 30 mil recs - partitioned table). If there are less than 1 million records in the big_tbl the copy to the vb_archive_table is fast (-10 min), but more importantly - it's consistant. However, if the number of records is greater than 1 million records in the big_tbl copying the data into the vb_archive_tbl is very slow (+30 min - 4 hours), and very inconsistant. Every few days the time it takes to copy the same amount of data grows signicantly.
    Here's an example of the code I'm using, which uses BULK COLLECT and FORALL INSERST to copy the data.
    I occasionally change 'LIMIT 5000' to see performance differences.
    DECLARE
    TYPE t_rec_type IS RECORD (fact_id NUMBER(12,0),
    store_id VARCHAR2(10),
    product_id VARCHAR2(20));
    TYPE CFF_TYPE IS TABLE OF t_rec_type
    INDEX BY BINARY_INTEGER;
    T_CFF CFF_TYPE;
    CURSOR c_cff IS SELECT *
    FROM big_tbl;
    BEGIN
    OPEN c_cff;
    LOOP
    FETCH c_cff BULK COLLECT INTO T_CFF LIMIT 5000;
    FORALL i IN T_CFF.first..T_CFF.last
    INSERT INTO vb_archive_tbl
    VALUES T_CFF(i);
    COMMIT;
    EXIT WHEN c_cff%NOTFOUND;
    END LOOP;
    CLOSE c_cff;
    END;
    Thanks you very much for any advice
    Edited by: reid on Sep 11, 2008 5:23 PM

    Assuming that there is nothing else in the code that forces you to use PL/SQL for processing, I'll second Tubby's comment that this would be better done in SQL. Depending on the logic and partitioning approach for the archive table, you may be better off doing a direct-path load into a staging table and then doing a partition exchange to load the staging table into the partitioned table. Ideally, you could just move big_tbl into the vb_archive_tbl with a single partition exchange operation.
    That said, if there is a need for PL/SQL, have you traced the session to see what is causing the slowness? Is the query plan different? If the number of rows in the table is really a trigger, I would tend to suspect that the number of rows is causing the optimizer to choose a different plan (with your sample code, the plan is obvious, but perhaps you omitted some where clauses to simplify things down) which may be rather poor.
    Justin

  • Data inconsistency: posting with restricted transaction type (FAA_POST047)

    Hi colleagues!
    I have the following error:
    Data inconsistency: posting with restricted transaction type
    Message no. FAA_POST047
    Diagnosis
    You are trying to post in company code 1000 with transaction type 130 in ledger group 0N. According to table TABWA, this transaction type is limited. Nonetheless, according to your entries, you want to post to area 11.
    System Response
    Using limited transaction types is obsolete and would lead to an inconsistent posting in this situation. The system rejects the posting.
    Procedure
    Use a transaction type that is not limited. And, if possible, choose the ledger group and/or depreciation area on the initial screen.
    According to specific accounting we shouldn't post to one of the depr. area, so most of the tr.types are restricted to post to all areas except this one.
    Because of this, posting from MM with acc.assignment 'A' is impossible - I have this error.
    Any ideas how to solve?
    Thanks in advance

    Hi Anton
    This looks like your transaction type -130 is limited to post only in ledger group-0N as per the transaction type configuration. Due to which system is unable to post all the ledgers and hence this lead to inconsistency.
    Please verify the transaction type 130- in configuration T code - OAYA and see if it is restricted by depreciation area (eventually allow to post only ledger group- 0N.
    Tcode - OAYA (Limit Transaction Types to Depreciation Areas)
    please verify and let us know the update. hope this helps
    Best Regards
    Jomon

  • Migration CC and FF data from Virsa 4.0 to GRC 10.1 Data Inconsistency

    Hello Guru's
    We have done Migration from VIRSA 4.0 Compliance Calibrator (CC) and Firefighter (FF) data from  to GRCAC 10.1 system.
    After Migration, data inconsistency is coming while validation.
    Followed standard sap migration guide virsa 4.0 to 10.0
    Patch level  - GRC AC 10.1 system
    SAP_BASIS
    740
    0008
    SAPKB74008
    0000
    @6H\QAchieved with Support Package@
    SAP Basis Component
    SAP_ABA
    740
    0008
    SAPKA74008
    0000
    @6H\QAchieved with Support Package@
    Cross-Application Component
    SAP_GWFND
    740
    0008
    SAPK-74008INSAPGWFND
    0000
    @6H\QAchieved with Support Package@
    SAP Gateway Foundation 7.40
    SAP_UI
    740
    0010
    SAPK-74010INSAPUI
    0000
    @6H\QAchieved with Support Package@
    User Interface Technology 7.40
    PI_BASIS
    740
    0008
    SAPK-74008INPIBASIS
    0000
    @6H\QAchieved with Support Package@
    Basis Plug-In
    ST-PI
    740
    0000
    0000
    SAP Solution Tools Plug-In
    SAP_BW
    740
    0008
    SAPKW74008
    0000
    @6H\QAchieved with Support Package@
    SAP Business Warehouse
    GRCFND_A
    V1100
    0006
    SAPK-V1106INGRCFNDA
    0000
    @6H\QAchieved with Support Package@
    GRC Foundation ABAP
    GRCPINW
    V1100_731
    0006
    SAPK-11506INGRCPINW
    0000
    @6H\QAchieved with Support Package@
    SAP GRC NetWeaver Plug-In
    IW_FNDGC
    100
    0002
    SAPK-10002INIWFNDGC
    0000
    @6H\QAchieved with Support Package@
    SAP IW FNDGC 100
    WEBCUIF
    747
    0005
    SAPK-74705INWEBCUIF
    0000
    @6H\QAchieved with Support Package@
    SAP Web UI Framework
    IW_CBS
    200
    0009
    SAPK-20009INIWCBS
    0000
    @6H\QAchieved with Support Package@
    SAP IW CBS 200
    IW_CNT
    200
    0009
    SAPK-20009INIWCNT
    0000
    @6H\QAchieved with Support Package@
    SAP IW CNT 200
    IW_SCS
    200
    0009
    SAPK-20009INIWSCS
    0000
    @6H\QAchieved with Support Package@
    Screen Scraping
    POASBC
    100_731
    0001
    SAPK-10201INPOASBC
    0000
    @6H\QAchieved with Support Package@
    POA Shared Business Components
    Patch level VIRSA 4.0 system
    GRCPINW
    V1100_731
    0008
    SAPK-11508INGRCPINW
    SAP GRC NetWeaver Plug-In
    Kindly let us know after migration, any sap notes need to install.
    Regards,
    karunakar

    Hello Prasant,
    We have migrated VIRSA 4 0 to GRC 10.1 system, after migration we have below issues.
    1. In table GRACSODRISK (not able to see the medium risk, but in NWBC screen (Access Risks) we are seeing the medium risks.
    2. In Function Permission.txt file ( only AND is coming for all, but some OR is avaiable in VIRSA 4 0 system.
    And for gap (it automatically taken AND
    Example:
    AM02
    FB02
    F_BKPF_BED
    ACTVT
    2
    AND
    0
    AM02
    FB02
    F_BKPF_BED
    ACTVT
    3
    AND
    1
    AM02
    FB02
    F_BKPF_BED
    BRGRU
    AND
    1
    AM02
    FB02
    F_BKPF_BEK
    ACTVT
    2
    AND
    0
    AM02
    FB02
    F_BKPF_BEK
    ACTVT
    3
    AND
    1
    AM02
    FB02
    F_BKPF_BEK
    BRGRU
    AND
    1
    kindly suggest.
    Regards,
    Karunakar

  • Data service for table in Oracle 8.0.6

    Hi,
    Using WebLogic 8.1.4 and LiquidData 8.5 I am trying to create physical data services for tables in a DB in Oracle 8.0.6. I am aware that that Oracle version is not supported by Oracle anymore, but I need to work with that version anyway (you know how it is sometimes).
    I managed to create a connection pool for this through the WebLogic Server Console by providing the JDBC driver for 8.0.6., but when I want to create a data source using the new connection pool and WebLogic tries to get the metadata, I get pop up windows with messages like:
    "Bigger type length than maximum"
    and
    "OALL8 in an inconsistent state"
    and
    "Protocol violation"
    One more thing to mention: I also added the Oracle 8.0.6 JDBC driver to the WebLogic Server classpath (Tools -> WebLogic Server -> Server Properties ... -> WebLogic Server: added classes12.zip to Server classpath additions) and restarted WebLogic Workshop and Server. Still I get those error messages.
    Is there a special procedure how to provide/configure a specific driver for a DBMS that is not natively supported by WebLogic?
    Any help is appreciated.
    Thanks,
    Wilko

    Hi Mike,
    Thanks for the quick reply. Below the contents of the console window from starting the Workshop and Server. I'll try your next hint next and let you know about the outcome. As far as I see there were no errors issued by the Server while I tried to connect to Oracle 8.0.6 to upload metadata. (I am not sure whether anything was printed out while I started the server). My address is w.eschebach at vsnlinternational dot com.
    Thanks,
    Wilko
    This is how my workshop.cfg looks like:
    C:\bea\weblogic81\workshop
    C:\bea\jdk142_05\jre\bin\java.exe
    -XX:-UseThreadPriorities -Xmx256m -Xms64m -Xss256k -client -Dsun.io.useCanonCaches=false -Dsun.java2d.noddraw=true -Dsun.java2d.d3d=false -Djava.system.class.loader="workshop.core.AppClassLoader" -cp "C:\bea\weblogic81\workshop\wlw-ide.jar" workshop.core.Workshop
    Console output:
    DEBUG: extensions=C:\bea\weblogic81\workshop\\extensions
    INFO: Registering extension com.bea.portal.ide.CommonServices
    INFO: Service com.bea.portal.ide.findrefs.FindRefsSvc registered
    INFO: Handler for urn:com-bea-portal-ide:ref-finders registered
    INFO: Registering extension workshop.control.ControlServices
    INFO: Service com.bea.ide.control.ControlSvc registered
    INFO: Registering extension com.crystaldecisions.integration.weblogic.workshop.r
    eport.Bootstrap
    INFO: Registering extension workshop.debugger.DebuggerServices
    INFO: Exit Handler found
    INFO: Service com.bea.ide.debug.DebugSvc registered
    INFO: Handler for urn:com-bea-ide:debugExpressionViews registered
    INFO: Registering extension workshop.jspdesigner.JspDesignerServices
    INFO: Service com.bea.ide.ui.browser.BrowserSvc registered
    INFO: Service com.bea.ide.jspdesigner.PaletteActionSvc registered
    INFO: Handler for urn:com-bea-ide-jspdesigner:tags registered
    INFO: Registering extension workshop.liquiddata.LiquidDataExtension
    INFO: Registering extension workshop.pageflow.services.PageFlowServices
    INFO: Exit Handler found
    INFO: Service workshop.pageflow.services.PageFlowSvc registered
    INFO: Service com.bea.ide.ui.palette.DataPaletteSvc registered
    INFO: Handler for urn:workshop-pageflow-wizard:extension registered
    INFO: Registering extension com.bea.portal.ide.portalbuilder.PortalBuilderServic
    es
    INFO: Service com.bea.portal.ide.portalbuilder.laf.LookAndFeelSvc registere
    d
    INFO: Service com.bea.portal.ide.portalbuilder.laf.css.CssSvc registered
    INFO: Service com.bea.portal.codegen.CodeGenSvc registered
    INFO: Registering extension com.bea.portal.ide.PortalServices
    INFO: Service com.bea.portal.ide.cache.CacheInfoSvc registered
    INFO: Registering extension workshop.process.ProcessExtension
    INFO: Service workshop.process.ProcessSvc registered
    INFO: Service workshop.process.broker.channel.ChannelManagerSvc registered
    INFO: Handler for urn:com-bea-ide-process:process registered
    INFO: Registering extension workshop.shell.ShellServices
    INFO: Exit Handler found
    INFO: Service com.bea.ide.ui.frame.FrameSvc registered
    INFO: Service com.bea.ide.core.datatransfer.DataTransferSvc registered
    INFO: Service com.bea.ide.actions.ActionSvc registered
    INFO: Service com.bea.ide.document.DocumentSvc registered
    INFO: Service com.bea.ide.core.HttpSvc registered
    INFO: Service com.bea.ide.ui.help.HelpSvc registered
    INFO: Service com.bea.ide.ui.output.OutputSvc registered
    INFO: Service com.bea.ide.core.navigation.NavigationSvc registered
    INFO: Service com.bea.ide.filesystem.FileSvc registered
    INFO: Service com.bea.ide.filesystem.FileSystemSvc registered
    INFO: Service com.bea.ide.refactor.RefactorSvc registered
    INFO: Service com.bea.ide.security.SecuritySvc registered
    INFO: Handler for urn:com-bea-ide:actions registered
    INFO: Handler for urn:com-bea-ide:document registered
    INFO: Handler for urn:com-bea-ide:frame registered
    INFO: Handler for urn:com-bea-ide:encoding registered
    INFO: Handler for urn:com-bea-ide:help registered
    INFO: Registering extension workshop.sourcecontrol.SCMServices
    INFO: Service com.bea.ide.sourcecontrol.SourceControlSvc registered
    INFO: Handler for urn:com-bea-ide:sourcecontrol registered
    INFO: Registering extension workshop.sourceeditor.EditorServices
    INFO: Service com.bea.ide.sourceeditor.EditorSvc registered
    INFO: Service com.bea.ide.sourceeditor.compiler.CompilerSvc registered
    INFO: Handler for urn:com-bea-ide:sourceeditor:sourceinfo registered
    INFO: Registering extension com.bea.wls.J2EEServices
    INFO: Service com.bea.wls.ejb.EJBSvc registered
    INFO: Service com.bea.wls.DBSvc registered
    INFO: Registering extension workshop.workspace.WorkspaceServices
    INFO: Exit Handler found
    INFO: Service com.bea.ide.workspace.WorkspaceSvc registered
    INFO: Service com.bea.ide.workspace.ServerSvc registered
    INFO: Service com.bea.ide.workspace.SettingsSvc registered
    INFO: Service com.bea.ide.build.AntSvc registered
    INFO: Service com.bea.ide.workspace.RunSvc registered
    INFO: Handler for urn:com-bea-ide:settings registered
    INFO: Handler for urn:com-bea-ide:project registered
    INFO: Registering extension workshop.xml.XMLServices
    INFO: Service com.bea.ide.xml.types.TypeManagerSvc registered
    INFO: Service com.bea.ide.xml.types.TypeResolverSvc registered
    INFO: Service com.bea.ide.xmlmap.XMLMapSvc registered
    DEBUG: Workshop temp dir: C:\DOCUME~1\TR003137\LOCALS~1\Temp\wlw-temp-18920
    DEBUG: ExtensionsLoaded: 8329ms
    DEBUG: UI Displayed: 11563ms
    DEBUG: Time to load XQuery Functions (in seconds) - 0
    DEBUG: Time to load repository (in seconds) - 0
    DEBUG: LdBuildDriver loaded
    DEBUG: project ProvisioningDataServices activated
    DEBUG: Setting active project to: ProvisioningDataServices
    DEBUG: Workspace Activated: 17126ms
    DEBUG: Document Panel initialized: 17501ms
    DEBUG: *** CompilerProject constructor 1
    DEBUG: WorkspaceLoaded: 17594ms
    DEBUG: getClasspathMapping initiated with 29 item list.
    DEBUG: getClasspathMapping returning 29 item map.
    INFO: Startup Complete
    DEBUG: Time to load repository (in seconds) - 1
    DEBUG: Loading template file wsrp-producer-project.zip
    DEBUG: Loading template file wli-tutorial.zip
    DEBUG: Loading template file wli-schemas.zip
    DEBUG: Loading template file wli-newprocess.zip
    DEBUG: Loading template file wli-helloworld.zip
    DEBUG: Loading template file webflow-project.zip
    DEBUG: Loading template file tutorial-webservice.zip
    DEBUG: Loading template file tutorial-pageflow.zip
    DEBUG: Loading template file tutorial-jbc.zip
    DEBUG: Loading template file tutorial-ejb.zip
    DEBUG: Loading template file portal-project.zip
    DEBUG: Loading template file portal-application.zip
    DEBUG: Loading template file pipeline-application.zip
    DEBUG: Loading template file oag-schemas.zip
    DEBUG: Loading template file netui-webapp.zip
    DEBUG: Loading template file liquiddata-project.zip
    DEBUG: Loading template file liquiddata-application.zip
    DEBUG: Loading template file ejb-template.zip
    DEBUG: Loading template file default-workshop.zip
    DEBUG: Loading template file datasync-template.zip
    DEBUG: Loading template file crystalreports.zip
    DEBUG: Loading template file commerce-project.zip
    DEBUG: Loading template file commerce-application.zip
    DEBUG: URI is null. Delete Version will not show up in the menu
    DEBUG: URI is null. Delete Version will not show up in the menu
    DEBUG: GCThread: performing gc while idle

  • Do I have to use lock when I am reading data from a table

    Hi,
    When i am reading data from a table , do I have to set a lock on that table .
    Is it necessary for us to set the lock on a table when I am reading data from the table.
    When I am updating the table , do I have to set a lock on the table ?
    If yes, then what sort of lock-read lock,write lock or shared lock?
    Regards,
    Sushanth H.S.

    check it out,
    Lock objects are use in SAP to avoid the inconsistancy at the time of data is being insert/change into database.
    SAP Provide three type of Lock objects.
    - Read Lock(Shared Locked)
    protects read access to an object. The read lock allows other transactions read access but not write access to
    the locked area of the table
    - Write Lock(exclusive lock)
    protects write access to an object. The write lock allows other transactions neither read nor write access to
    the locked area of the table.
    - Enhanced write lock (exclusive lock without cumulating)
    works like a write lock except that the enhanced write lock also protects from further accesses from the
    same transaction.
    You can create a lock on a object of SAP thorugh transaction SE11 and enter any meaningful name start with EZ Example EZTEST_LOCK.
    Use: you can see in almost all transaction when you are open an object in Change mode SAP could not allow to any other user to open the same object in change mode.
    Example: in HR when we are enter a personal number in master data maintainance screen SAP can't allow to any other user to use same personal number for changes.
    Technicaly:
    When you create a lock object System automatically creat two function module.
    1. ENQUEUE_<Lockobject name>. to insert the object in a queue.
    2. DEQUEUE_<Lockobject name>. To remove the object is being queued through above FM.
    You have to use these function module in your program.
    check this link for example.
    http://help.sap.com/saphelp_nw04s/helpdata/en/cf/21eea5446011d189700000e8322d00/content.htm
    tables:vbak.
    call function 'ENQUEUE_EZLOCK3'
    exporting
    mode_vbak = 'E'
    mandt = sy-mandt
    vbeln = vbak-vbeln
    X_VBELN = ' '
    _SCOPE = '2'
    _WAIT = ' '
    _COLLECT = ' '
    EXCEPTIONS
    FOREIGN_LOCK = 1
    SYSTEM_FAILURE = 2
    OTHERS = 3
    if sy-subrc 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    endif.
    Normally ABAPers will create the Lock objects, because we know when to lock and how to lock and where to lock the Object then after completing our updations we unlock the Objects in the Tables
    http://help.sap.com/saphelp_nw04s/helpdata/en/cf/21eea5446011d189700000e8322d00/content.htm
    purpose: If multiple user try to access a database object, inconsistency may occer. To avoid that inconsistency and to let multiple user give the accessibility of the database objects the locking mechanism is used.
    Steps: first we create a loc object in se11 . Suppose for a table mara. It will create two functional module.:
    1. enque_lockobject
    1. deque_lockobject
    before updating any table first we lock the table by calling enque_lockobject fm and then after updating we release the lock by deque_lockobject.
    Select the radio button "Lock object"..
    Give the name starts with EZ or EY..
    Example: EYTEST
    Press Create button..
    Give the short description..
    Example: Lock object for table ZTABLE..
    In the tables tab..Give the table name..
    Example: ZTABLE
    Save and generate..
    Your lock object is now created..You can see the LOCK MODULES..
    In the menu ..GOTO -> LOCK MODULES..There you can see the ENQUEUE and DEQUEUE function
    Reward if helpful

  • Data Inconsistency: Parallel Dep area is not posted Message no. AAPO104

    Team
    Could you please help on this.
    We tried settling a PS to Asset for AUC T code CJ88 but recd error message Data Inconsistency: Parallel Dep area is not posted Message no. AAPO104.
    Dep area are maintained correctly, parallel currencies & Dep area setting for that are maintained correctly, in Customising conssistency check no error.
    but it is not settling, could you please let me know if I have missed something.
    Transaction type is also not limited to any Dep area.
    Regards
    Ajay

    what should be checked in TOBJ
    This table basically relates to User-defined (additional to SAP-defined) objects are documented.  So you need to check with your basis.  Otherwise, for the error message you indicated, only three notes are available which has already been indicated above.
    thanks
    G. Lakshmipathi

  • Data inconsistency, Partition /BIC/B00004240000 not deleted

    Hi,
    We've got some problems when we are trying to delete a PSA request in our QA system.
    The error message is: Data inconsistency, Partition /BIC/B00004240000 not deleted
    Does anyone know what this means? We've never had this problem before.
    Thanks!
    BR,
    Linda

    Hello Linda,
    Please try this ABAP program RSAR_PSA_PARTITION_CHECK and while executing set the following
    Object = 'RSAR'
    Subobject = 'METADATA'
    Ext. Identif. = 'RSAR_PSA_PARTITION_CHECK'
    Please refer to the OSS
    Note 856097 - Check and repair program for Note 849857
    Note 733371 - CleanUp for inconsistent PSA directory entries
    Note 1042182 - Secondary index is inconsistent for PSA tables
    Note 1102625 - Repair tool for Note 1102623 (PARTNO = 9999)
    Thanks
    Chandran
    Edited by: Chandran Ganesan on Feb 29, 2008 4:45 PM

  • Data inconsistency in release procedure

    Dear All,
    I have one query regrading data inconsistency it is as follows.
    I deleted one release strategy in development system and transported the same to Quality system.After transporting to Quality system I found some inconsistency in data in Quality system.
    So now if I create deleted release strategy again in Development system and if I transport the same to quality will this get updated in all the tables as it was earlier(before transporting the deleted release strategy to quality).
    Please help me.
    Thanks in advance,
    With Rgds,
    Babu

    will this get updated in all the tables as it was earlier
    Pretty much Yes, it will have a different version though
    Regards
    Juan

  • Not able to refresh the data in a table

    Hi In my application i fill data in a table on clikc of a button ..
    Following are the line of code i have user
    RichColumn richcol = (RichColumn)userTableData.getChildren().get(0);                                           (fetch the first column of my table)
    richcol.getChildren().add(reportedByLabel);                                                                   (reportedByLabel is a object of RichInputText)
    AdfFacesContext adfContext1 = AdfFacesContext.getCurrentInstance();
    adfContext1.addPartialTarget(richcol);
    adfContext1.addPartialTarget(userTableData);
    But on submit of that button table data is not refreshed after adding partial trigger on that table as well as that column also .... any idea??
    Edited by: Shubhangi m on Jan 27, 2011 3:50 AM

    Hi,
    The Code that you have shown adds an additional inputText component to the first column of a table.
    Is that your intention?
    If yes, please use the following code snippet to achieve your functionality:
    JSPX Code:
    <af:form id="f1">
    <af:commandButton text="Add Column" id="cb1"
    actionListener="#{EmployeesTableBean.onAddColumn}"/>
    <af:table value="#{bindings.Employees.collectionModel}" var="row"
    rows="#{bindings.Employees.rangeSize}"
    emptyText="#{bindings.Employees.viewable ? 'No data to display.' : 'Access Denied.'}"
    fetchSize="#{bindings.Employees.rangeSize}"
    rowBandingInterval="0"
    selectedRowKeys="#{bindings.Employees.collectionModel.selectedRow}"
    selectionListener="#{bindings.Employees.collectionModel.makeCurrent}"
    rowSelection="single" id="t1"
    binding="#{EmployeesTableBean.table}">
    <af:column sortProperty="EmployeeId" sortable="false"
    headerText="#{bindings.Employees.hints.EmployeeId.label}"
    id="c1">
    <af:outputText value="#{row.EmployeeId}" id="ot2">
    <af:convertNumber groupingUsed="false"
    pattern="#{bindings.Employees.hints.EmployeeId.format}"/>
    </af:outputText>
    </af:column>
    <af:column sortProperty="FirstName" sortable="false"
    headerText="#{bindings.Employees.hints.FirstName.label}"
    id="c2">
    <af:outputText value="#{row.FirstName}" id="ot1"/>
    </af:column>
    <af:column sortProperty="LastName" sortable="false"
    headerText="#{bindings.Employees.hints.LastName.label}"
    id="c3">
    <af:outputText value="#{row.LastName}" id="ot3"/>
    </af:column>
    </af:table>
    </af:form>
    Bean:
    public class EmployeesTableBean {
    private RichTable table;
    public EmployeesTableBean() {
    public void setTable(RichTable table) {
    this.table = table;
    public RichTable getTable() {
    return table;
    public void onAddColumn(ActionEvent actionEvent) {
    RichInputText newRichInputText = new RichInputText();
    newRichInputText.setId("new");
    newRichInputText.setValue("Name:");
    RichColumn richcol = (RichColumn)table.getChildren().get(0);
    richcol.getChildren().add(newRichInputText);
    AdfFacesContext adfContext1 = AdfFacesContext.getCurrentInstance();
    adfContext1.addPartialTarget(table);
    Thanks,
    Navaneeth

  • How can I insert data in different table dynamically?

    I have one requirement ,In these  I have transaction table in which payLoad contain data like employee record,PO record,SO record  in XML format.Now I have transfer the data from transaction table to particular table(Emp,SO,PO) dynamical ? how can I do it.. Give me your suggestion guys?? Thanks in Advance

    Hello,
    Still no luck.I am surely doing silly mistakes.Anyway,Here are my workings-
    1> student_mst (id(pk),studentname) and student_guard_mst(id(fk),guardianname)
    2> created EO from both of the tables,made id in both EO as DBSequence and an association was also generated.
    3> i made that association composite by clicking the checkbox
    4> i created 2 VO from 2 EO.
    5> put those VO in Application Module.
    6> dragged and dropped 2 VO on my jspx page and dropped them as ADF Form.
    Now what to do please?

  • How to delete the data in a table using function

    hi all,
    i need to delete the data in a table using four parameters in a function,
    the parameters are passed through shell script.
    How to write the function
    Thanks

    >
    But the only thing is that such function cannot be used in SQL.
    >
    Perhaps you weren't including the use of autonomous transactions?
    CREATE OR REPLACE FUNCTION remove_emp (employee_id NUMBER) RETURN NUMBER AS
    PRAGMA AUTONOMOUS_TRANSACTION;
    tot_emps NUMBER;
    BEGIN
    SELECT COUNT(*) INTO TOT_EMPS FROM EMP3;
    DELETE FROM emp3
    WHERE empno = employee_id;
    COMMIT;
    tot_emps := tot_emps - 1;
    RETURN TOT_EMPS;
    END;
    SQL> SELECT REMOVE_EMP(7499) FROM DUAL;
    REMOVE_EMP(7499)
                  12
    SQL> SELECT REMOVE_EMP(7521) FROM DUAL;
    REMOVE_EMP(7521)
                  11
    SQL> SELECT REMOVE_EMP(7566) FROM DUAL;
    REMOVE_EMP(7566)
                  10
    SQL>

  • Adding Data From One Table to Another

    Now, this doesn't strike me as a particularly complex problem, but I've either strayed outside the domain of Numbers or I'm just not looking at the problem from the right angle. In any case, I'm sure you guys can offer some insight.
    What I'm trying to do is, essentially, move data from one table to another. One table is a calendar, a simple two column 'date/task to be completed' affair, the other is a schedule of jogging workouts, i.e, times, distances. Basically, I'm trying to create a formula that copies data from the second table onto the first but only for odd days of the week, excepting Sundays (and assuming Monday as the start of the week). Now, this isn't the hard part, I can do that. The problem comes when I replicate the formula down the calendar. Even on the days when the 'if' statement identifies it as an 'even day', the cell reference to the appropriate workout on the second table is incremented, so when it comes to the next 'odd day', it has skipped a workout.
    I can't seem to see any way of getting it to specifically copy the NEXT line in the second table, and not the corresponding line.
    This began as a distraction to try and organise my running so I could see at a glance what I had to do that day and track my progress, but now it's turned into an obsession. SURELY there's a solution?
    Cheers.

    Hi Sealatron,
    Welcome to Apple Discussions and the Numbers '09 forum.
    Several possible ways to move the data occur to me, but the devil's in the details of how the data is currently arranged.
    Is it
    • a list of three workouts, one for each of Monday, Wednesday and Friday, then the same three repeated the following week?
    • an open-ended list that does not repeat?
    • something else?
    Regards,
    Barry

Maybe you are looking for

  • What is the best practice for developing web service?

    Hi All, I'm a newbee to web services... I was wondering what would be the best approach in developing a web service, using tools or programmatic approach? If I use WebLogic Workshop, am I tied to a vendor? Is it possible for me to develop web service

  • Adobe Air on Windows XP Embedded?

    Hi All, I am working on a project where we may need to use Windows XP Embedded. We are using Adobe Air for the the UI. Will Adobe Air install on Windows XP Embedded? Thanks in advance for any advice.

  • Table

    Hi How to find table for PO release statergy status.Blocked or released ? Suppose 3 Release Code and PO released by only one person. vIVEK

  • Copied ITunes from Old to New Comp. but Files NOT showing up in new Library

    SO, I used my IPod to copy my ITunes Library from my work computer to my home computer. Now in my Music folder both of the libraries are there- but when I open ITunes it is still all of my old music... none of the new music is showing up. I can manua

  • SYSLOGD gone wild - where are the files?

    Hey all, Yesterday I had 15GB of free HD yesterday, it told me today that I was down to 100MB. Confused - because I haven't added any files, I check the activity monitor and found that SYSLOGD was running, and using an absurd amount of processor and