Data object at user level missing

Hai all,
10.2.0.4 on solaris 10.
recently I imported a 600 GB table through data pump from one user to
a different user in the same database. now, logged in as the newly imported user, I can able to see the contents of the table that is imported, but when I queried user_objects or user_tab_partitions , I cannot be able to see the table ?
What could be the problem
Kai

Hi,
Try to check the log of import and check for any Issues. then we can figure out what might the thing happened.
what is the version of db..??
- Pavan Kumar N
- ORACLE OCP - 9i/10g
https://www.oracleinternals.blogspot.com

Similar Messages

  • How to replicate session data at the user level?

    Hi all,
    my client has users who use laptop/tablet computers in the field with wireless connections. The application requires the users to complete long multi-page forms where the data is only submitted to the EIS layer at the end of the process. The connections regularly drop out which understandably is a cause of some grief. In addition, the users want to be able to switch to another machine in order to complete their session.
    The 'switching machine' requirement kills the idea of using persistent cookies and session replication unfortunately. So I am faced with the idea of storing the session data (keyed by user id), in serialized form, either on a stand-alone disk which is accessible by each webserver in the cluster or is the database. The problem with the database idea is the performance hit of traversing the EJB and JDBC layers.
    Can anyone suggest the best practice in this case? Or point to an article or tool that covers the problem?
    BTW, the app serves a small (<50) and stable base of authenticated users so scalability is not really an issue whereas failover is.
    Regards,
    Dave.

    > Is it possible to close a PO at the header level
    Po can be closed at Item level and not Header level. You can set Deletion Indicator and further can be archived thru SARA MM_EKKO
    or a way to deactivate the PO so that no new items can be added?
    You can activate Release strategy and Choose a Proper Release indicatory so that no new changes to be carried out.
    Else go for User Exits

  • How to Convert Date Object at Universe Level with out Timestamp

    Hi,
    I have a Object  called "PCREGISTERDATE" at universe where the data type is date .but the dates are coming in the following format:9/26/2007 9:48:40 PM but i want to show the date as with out time stamp.
    how can i create a object by at universelevel which shoul show only date with out Time Stamp.
    please help me on this ASAP.
    Thanks & Regards,
    Kumar

    Please try below date fucntion:
    select convert(varchar(10),getdate(),101)  - this is for sybase
    syntex will change based on your source database.
    Thanks
    Ponnarasu .K

  • Delete a row from a data object using an AQ-driven EMS

    Hello, technetwork.
    I need to show the contents of a table that changes frequently (inserts, updates and deletes) at ORDBMS 10g in my BAM dashboard.
    What i have done is:
    - Create two different queue tables, queues.
    - Create two triggers: one for "AFTER insert or UPDATE" and the other for "BEFORE DELETE".
    - Create two EMS, one configured to Upsert operations and the other to Delete.
    - The Upsert EMS works, but the Delete EMS does not.
    - Both EMS Metrics say that they work fine and commit the messages i send to the ADC.
    - Testing showed records are populated and updated in the ADC but never deleted.
    There is a detailed user case for "Creating an EMS Against Oracle Streams AQ JMS Provider" in the Fusion Midleware Developer's Guide for SOA Suite, but it deals only with Upserts. I am using the last versions of SOA suite and Weblogic. The official support has no information either.
    I hope writing a web service client isn't the only way to delete a row from a data object. Am i missing something? Any help will be much appreciated.
    My EMS config:
    Initial Context Factory: weblogic.jndi.WLInitialContextFactory.
    JNDI Service Provider URL: t3://<spam>:80.
    Topic/Queue Connection Factory Name: jms/BAMAQTopicCF.
    Topic/Queue Name: jms/ProdGlobalBAMD_flx.
    JNDI Username: .
    JNDI Password: .
    JMS Message Type: TextMessage.
    Durable Subscriber Name (Optional): BAM_ProdGlobalBAMD.
    Message Selector (Optional): .
    Data Object Name: /bam/ProdGlobalBAM_flx.
    Operation: Delete.
    Batching: No.
    Transaction: No.
    Start when BAM Server starts: No.
    JMS Username (Optional): .
    JMS Password (Optional): .
    XML Formatting
    Pre-Processing
    Message Specification
    Message Element Name: row
    Column Value
    Element Tag
    Attribute
    Source to Data Object Field Mapping
    Key Tag name Data Object Field
    . BARCODE. BarCode.
    Added my EMS config

    Ram_J2EE_JSF wrote:
    How to accomplish this using JavaScript?Using Javascript? Well, you know, Javascript runs at the client side and intercepts on the HTML DOM tree only. The JSF code is completely irrelevant. Open your JSF page in your favourite webbrowser and view the generated HTML source. Finally just base your Javascript function on it.

  • Reg Differrences between SynBo and Data Object

    Hi,
            I want to know the architectural diffrences between a SyncBo of MI 2.5 and Data Object of NWM 7.1.  As it is given in help.sap.com documentation I think A data object support multiple levels of parent- child hirearchy where as a SyncBo is limited to one level of Header and Items.Similarly do any other differences exist?
    2) In what way A data Object concept is better than that of a Synco.?
    3) For manipulating the data of a data Object on Front end, Can we Use the existing SmartSync API of MI 2.5, or a seperate API for client side is going to be released?
    Thanks in Advance,
    Rajesh

    Hello Rajesh,
    1. The differences are small. You pretty much highlighted the biggest one (multi child node). Other differences exist in the fact that a SyncBO is linked to a backend BAPI as the DO is not (you would usually create the DO with the fields you need and then have backend adapters to fill it). Lifecycle of the DO is also different. Even release, non destructive change can happen and old application as well as new application will work at the same time. I can list a couple more, but I think you get that a SyncBO is different than a DO (even if in our development we still get mix both wording:))
    2. Yes DO is a better concept than SyncBO.
    3. There are 3 ways of accessing API on the client side. The first is the old SmartSync API which is only used in backward compatibility application (MAM for example). It should not be used for new development, but only to leverage existing one. After that you have two new frameworks for laptop and for PDA that you can find in NWDS 7.1. They reviewed the naming so I am not sure of the current ones, but I think it is something like Mobile Application for Handheld and Mobile Application for Laptop. Those two new framework uses WebDynPro flavored UI and provide access to the sync/data layer.
    Thank you,
    Julien.

  • DOBJ.UPDATE_FAILED error while attaching adapters to Users Data Object

    PROBLEM DESCRITPION:-
    The following error occured while trying to attach entity adapters to Users Data Object (com.thortech.xl.dataobj.tcUSR) in Post-Update section.
    Error Message: The security level for this data item indicates that it cannot be updated.
    Update failed.
    Update failed.
    Error Keyword: DOBJ.UPDATE_FAILED
    Description: Update failed.
    Action: E
    Severity: C
    Kindly suggest to fix this issue.

    Resolved the DOBJ.UPDATE_FAILED problem - I modified the dvt_data_level to 0 for dvt_post_update_sequence selective entries.
    Able to attach the custom event handlers to Users Data Object - No more DOBJ.INVALID_UPDATE error!

  • Some DB objects like User-defined datatypes being missed out?

    Dear all.
    I create an Oracle model for a captured SQL Server model and then right click -> click on 'Generate' to create the DB creation script.
    (1) But few objects like user-defined datatypes are missed out in the script.
    (2) Also, a simple stored proc as defined below missed out:
    CREATE PROCEDURE [SalesLT].[SayHi]
    AS
    SELECT 'Hi I am a job scheduler'
    GO
    Rgds, Abhi

    Hello,
    An OracleDI Dataserver connection to an RDBMS is, most of the time, defined with a single set of parameters: a user account name, a password and a URL.
    OracleDI may need to access numerous tables on the RDBMS which may not all be stored in the same database schema, catalog, library, ...and for which access privileges may not be identical.
    For this reason, it is necessary to ensure that the RDBMS user account has sufficient access privileges for these schema, catalog, librairies ...
    The user account should have at least read/right access permission for any OracleDI Physical Schema referenced for the concerned Dataserver (depending on the project within which it is used, a Physical Schema may store both source and/or target tables).
    Moreover, for project requirements, it may be necessary to use/create/drop certain RDBMS
    components such as stored procedures, views, temporary tables, system tables ...
    For this reason, make sure the user account referenced in the dataserver connection parameters is attributed sufficient privileges to any object it may manipulate with OracleDI.
    2nd point: you can use any logical name for the data server Name.

  • Ideas for Providing User Level Data Backup and Restore

    I'm looking for ideas for implementing a user level application data backup and restore in an Apex app.
    What would be great is to have a user be provided an export file and a way to import this file. A bit overkill but hopefully never needed.
    Another option that is perfectly doable is a report that simply provides a means to create an export of the data. Since I already have an interface I can use an export to interface an export.
    Any thoughts?
    Hopefully I'm missing something already there for an end user to use.

    jlincoln wrote:
    "Do you mean "export" and "import" colloquially, or in the specific sense of the exp/imp/datapump utilities?"': I mean as in imp/exp Oracle utilities. Generally speaking, it would be neat to be able to export and import via an Apex an application. In this hosted environment I don't have that access but would this be a bad idea if you don't care about the existing data in the schema in which the data resides?I can envisage a mechanism using <tt>exp/imp</tt>, but since it requires <tt>dbms_scheduler</tt> external jobs and access to the file system it's highly unlikely to be possible in a hosted environment. (Unless you're doing the hosting?)
    Backup: Necessary for piece of mind and flexibility. I am working on a VB/Access user who does this today to get to the point when they can be comfortable with the backups occurring regularly and by the hosting site's DBA group.
    Restore: Like I said. I am working on a VB/Access user who does this today to get to the point when they can be comfortable with the backups occurring regularly and by the hosting site's DBA group. This is a very small data set. A restore would simply remove existing data and replace it with the new data.My opinion is that time would be better spent working on the user rather than a redundant backup and restore feature. Involve them in a disaster recovery exercise with whoever is hosting the environment to prove that their data is safe. Normally the inclusion of data in regular, effective database backups is sold as a major feature of APEX solutions.
    "What about security/privacy when this data ends up in uncontrolled environments?": I don't understand the point of this question. The data should not end up in uncontrolled environments. Just like the data in the database or its backups.Again, having data in a central, shared location protected by multiple levels of application, database, and OS security is usually seen as a plus for APEX over VB/Access. Exporting the data in toto to a PC/laptop that can be stolen or lost, and where it can be copied to USB drives/phones/email loses this protection.
    User Level: Because the end user must have access to the backup and restore mechanisms of the application.
    Application Data: The application data. Less than 10MB. Very small. It can be exported in a flat file downloaded by the end user. This file can then be used to upload and import via an existing application interface. For example.
    "I'm struggling to parse this for meaning.": When I say I have an existing interface I am referring to a program residing in the Apex application that will take data from a flat table structure (i.e. interface table), validate the data, derive data, and load into the target table structure.Other than the report export capability linked to above, there's nothing built-in to APEX that comes close to your requirement. If the data is simple enough that it can be handled in such a report, and you have a process that can read and recreate this export, then you have your backup/restore capability. If the data can't be handled in a simple report, then you'll need a more complex PL/SQL process to generate the file.

  • Changing The Validity Dates In Org Model in CRM- User Level

    Hi
    How can we restrict the Validity Dates in Organisation Structure at User Level.
    I need that the user should be given some specific dates<b> ( Valid till or assigned till )</b> while assigning in the org. And it should be visible always in the Org Model.
    We tried using Delimited Date Function, but the User gets ellapsed if the Validity is expired.With the business point of view, <b>the requirement is that all user should be listed in Org model with respect to any dates.this would mean that if a user is terminated we would want to change the Valid till date in the Org</b>

    Hi Amrita,
    Business partner number and Object ID is automatically determines by system when creating a org unit.
    I dont find any configuration to create org unit with your own BP number and object ID. You have option to make repairs to the existing org unit.
    Please find the below path to repair the org unit
    IMG -> CRM -> Master Data -> Business partner -> Integration Business partner -Org management -> Create Business partner initially
    Select the org unit and execute to get all the org units.
    You can select the org units with errors /warnings and click on start repair to get automatic assignment of BP.
    Hope it helps
    Reward points if it helps
    Regards,
    Madhu

  • Is there any object in labview that contains a list of data for the user to select (selection one at a time) or add a new data?

    Is there any object in labview that contains a list of data for the user to select (selection one at a time) or add a new data?

    List and table controls -> listbox..is that what you are thinking of?
    The listbox presents the user with a list of options, and you can set it to only accept one selection at a time...Adding new data to the list can not be done directly by the user but if you make e.g. a text control and a button you can programatically insert new objects described in the text box when the button is pressed...(see example).
    If you need more than one column you have the multicolumn listbox. If you want the users to write new entries directly yu can use a table and read selected cells using it's selection start property to read what cell has been selected.
    MTO
    Attachments:
    Listbox_example.vi ‏34 KB

  • Comman to take user level backup(tables and data in table for a particular)

    Hi ,
    I need to create a sql script to take user level back up i.e i need to take back up ,(export the tables and data in those tables in particular user) to my pc..
    i am able to do the same from unix command ,but not able to do the samq in sql script
    Please advice.

    Hi ,
    I need to create a sql script to take user level back up i.e i need to export the tables and data in those tables in particular user to my pc..
    i am able to do the same from unix command ,but not able to do the samq in sql script
    Please advice

  • BSP Exception: Missing reference when converting data object ZZxx

    Hi there,
    I have created these new Z fields and tried to include them in a field group. These fields did appear, but they are highlighted with a red box, with the above text in the tool tip.
    Does anyone know what it means??
    Thanks

    I believe they were generated by EEWB - or could've been added manually to the CRMD_CUSTOMER_H. This error only occurs on the Currency fields...something to do with reference field but I just can't find what is wrong with this. I searched up SAP notes and found the following the closest match: -
    <b>Symptom</b>
    when trying to scroll down in the result list in the F4 for the Ibase the error error "BSP exception : Missing reference when converting data object amount.  Correct the entry" occurs.
    <b>Other terms</b>
    CRMT_BSP_IBASE_TREE_NF, data object amount
    <b>Reason and Prerequisites</b>
    This happens due to inconsistancy in structure
    <b>Solution</b>
    Please follow the below steps manually.
    1. Go to tx:SE11
    2. Select data type "CRMT_BSP_IBASE_TREE_NF" in change mode
    3. Select component AMOUNT in 'Components' tab
    4. Go to "Currency/quantity fields" tab
    5. Replace 'Reference table' entry from 'CRMT_BSP_IBASE_DETAIL'
       to 'CRMT_BSP_IBASE_TREE_NF'
    6. Save and activate

  • Need help about Financials Operating Unit Level Setup data object transforming

    Hi All,
    We met some issues while using iSetup to create new OU from existing one via 'Financials Operating Unit Level Setup’ template and need your help.
    We can perform Extracting successfully, but when we’re tring to proceed Transform, it seems iSetup can’t meet our expectation.
    Firstly, after succesful creation of the Transform for 'Financials Operating Unit Level Setup’.
    We clicked the transform icon and it led us to the Details page.
    However, we can’t search out any result on that page, is it a bug?
    If we didn’t performed ‘Transform All’, then there’s no row shows on Transform Details page so we can't make any transform, as below:
    But after performing a ‘Transform All’ task, revisite the Details page and unfreeze, Operating Units and Inventory Organizations became transformable, as shown in below screenshots.
    Secondly, we found the transformable data objects are so limited, even the whole Payable Options and Finacial Options doesn’t support, as well as the Operating Unit short code.
    Is there any documents indicating which data objects are transform supported? Or if there’s any sql query for it?
    Will be very appreciated if anyone can help.
    Regards,
    Genghua Guo

    Hi,
    According to your description, my understanding is that machines failed to ping the server name, and successfully to ping its IP address.
    First check to see if DNS related port 53 is opened on the DC, or disable any anti-various software temporally, then ping the DC’s
    FQDN and confirm that if the name would be resolved.
    >The host 7acf7675-12f6-4280-84ff-ca225bc99373._msdcs.REGENCYVILLAGE.LOCAL could not be resolved to an IP address.
    You domain name is REGENCYVILLAGE.LOCAL. It seems that the DNS server failed to resolve your DC’s name. check the zone named REGENCYVILLAGE.LOCAL to see if an A record is corresponding to your DC.
    >75.75.75.75 (<name unavailable>) [Invalid], 75.75.76.76 (<name unavailable>) [Invalid]
    It seems that invalid IP address of DNS server on the TCP/IP properties of DC. 
    I want to confirm with you that if you have installed the DNS with installing AD DS. Just the DC’s name can’t be resolved? Or any other device also has the same question?
    =======
    You may use AD DS Best Practices Analyzer (BPA), which is a server management tool that can help you implement best practices in the configuration of your Active Directory environment. Detailed information reference:
    https://technet.microsoft.com/en-us/library/dd391875(v=ws.10).aspx
    besides, suggestion about Active Directory Best practices, just for you reference:
    https://technet.microsoft.com/en-us/library/cc778219%28v=ws.10%29.aspx?f=255&MSPPError=-2147217396
    Best Regards,
    Eve Wang
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Debug privilege - is it at object level or user level or both?

    Just want to know at what level Debug privilege can be granted? At object level or user-level or both?
    I'm under the impression that if an user say 'A' is given this privilege, then he can debug all the procedures/packages/objects in his schema . I guess it's like granting the privilege at user level.
    On the other hand when we say 'grant debug on proc_name to xyz' , it's at object level.
    Please check and correct my assumptions.

    in order to debug (i.e, activate dbms_debug) in a session you need:
    debug connect session
    in order to debug objects that you are not the owner you need:
    debug any procedure
    or
    grant debug on procedure from the user that own the pl/unit
    hope it helps,
    Amiel Davis

  • Unable to Multi Level Drill Down on Single Data Object in BAM Report

    Hi,
    We need to have multi level drill down in a BAM report.
    We are able to create single level drill down using 'Collapsed List' view in a Tiled Report, but unable to drill further.
    We are using BAM (11.1.1.2).
    Please let us know the steps to do this, your input is highly appreciated.
    Thanks,
    Arijit
    Edited by: user2917352 on May 4, 2011 11:21 PM

    Have you used Dimension? By defining Dimensions, you can create multi level drill down on the same data object, in the same report.
    1. Define the Dimension in the data object
    2. In the Studio, choose Drilling tab under the report
    3. Drill Hierarchies will be available. Choose the Dimension
    The drilling level will depend on the depth of the Dimension you have specified!
    -AR

Maybe you are looking for

  • Compressor on Intel vs. PPC

    On my old dual 1GHz G4, using the version of Compressor that came with DVD Studio Pro 2, it takes about five hours to encode one hour of video at the "high quality" setting. I'm contemplating the purchase of a Mac Pro tower and the current version of

  • Account number range in EBS

    Hi ,    When do EBS configuration we need to assign account symbol to GL accounts. My question is when ++++++++++ is populated in GL account field for Bank Deposit symbol what is the account number range for '+++++++++'. Namely,  which accounts will

  • LDAP Authorization for OBIEE 10.1.3

    Hello, We have setup LDAP authentication (ADSI LDAP) using OBIEE standalone. I'm trying to figure out the best way to manage Authorization - user to group assignment in OBIEE. Options: 1. Using external table Challange: The client doesn't have other

  • BEA WLS 9.0  java.lang.NoClassDefFoundError: utils/ImportPrivateKey

    Hi I'm newbie on this forum and on BEA WebLogic 9.0. I have to import my certificate in the keystore in order to setup SSL connection. I'm trying to use the java utils provided by BEA WL, but I have this error: bash-3.00# java utils.ImportPrivateKey

  • SyncRead moving files before it is completely available

    Hi, Have a scenario where the BPEL File Adapter with Sync Read operation will move a file from source directory to target directory. For small files testing seems to be OK. In case of large files the moving operation is started by Sync Read before th