EXPDP and statistics

Hi,
in 10g R2, it seems there is no option for statistics. Is there any ? How is the syntaxe ?
Thank you.

You can use EXCLUDE parameter.
Also have in mind that statistics are by default always saved for indexes and tables during the data pump export.
So , to not export them, use EXCLUDE=statistics.

Similar Messages

  • Full expdp and impdp: one db to another

    Hi! Good Day!
    I Would like to ask for any help regarding my problem.
    I would like to create a full export of one database to and import it on another database. This 2 database are on a separate machines.
    I am trying to use the expdp and impdp tool to do this task. However, i have been experiencing some problems during the import task.
    here is the detail of my problems:
    When i try to impdp the dump file, it seems that i was not able to iimport the user's data and metadata.
    Here is the exact command that i've used during the import and export task:
    export (server #1)
    expdp user01/******** directory=ora3_dir full=y dumpfile=db_full%U.dmp filesize=2G parallel=4 logfile=db_full.log
    import (server #2)
    impdp user01/******** directory=ora3_dir dumpfile=db_full%U.dmp full=y log=db_full.log sqlfile=db_full.sql estimate=blocks parallel=4
    here is the log that was generated during the impdp running:
    Import: Release 10.2.0.1.0 - 64bit Production on Friday, 27 November, 2009 17:41:07
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP and Data Mining options
    Master table "SSMP"."SYS_SQL_FILE_FULL_01" successfully loaded/unloaded
    Starting "SSMP"."SYS_SQL_FILE_FULL_01": ssmp/******** directory=ora3_dir dumpfile=ssmpdb_full%U.dmp full=y logfile=ssmpdb_full.log sqlfile=ssmpdb_full.sql
    Processing object type DATABASE_EXPORT/TABLESPACE
    Processing object type DATABASE_EXPORT/PROFILE
    Processing object type DATABASE_EXPORT/SYS_USER/USER
    Processing object type DATABASE_EXPORT/SCHEMA/USER
    Processing object type DATABASE_EXPORT/ROLE
    Processing object type DATABASE_EXPORT/GRANT/SYSTEM_GRANT/PROC_SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/GRANT/SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/ROLE_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/DEFAULT_ROLE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLESPACE_QUOTA
    Processing object type DATABASE_EXPORT/RESOURCE_COST
    Processing object type DATABASE_EXPORT/SCHEMA/DB_LINK
    Processing object type DATABASE_EXPORT/TRUSTED_DB_LINK
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/SEQUENCE
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/DIRECTORY/DIRECTORY
    Processing object type DATABASE_EXPORT/DIRECTORY/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/DIRECTORY/GRANT/CROSS_SCHEMA/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/CONTEXT
    Processing object type DATABASE_EXPORT/SCHEMA/PUBLIC_SYNONYM/SYNONYM
    Processing object type DATABASE_EXPORT/SCHEMA/SYNONYM
    Processing object type DATABASE_EXPORT/SCHEMA/TYPE/TYPE_SPEC
    Processing object type DATABASE_EXPORT/SYSTEM_PROCOBJACT/PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    Processing object type DATABASE_EXPORT/SYSTEM_PROCOBJACT/POST_SYSTEM_ACTIONS/PROCACT_SYSTEM
    Processing object type DATABASE_EXPORT/SCHEMA/PROCACT_SCHEMA
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/PRE_TABLE_ACTION
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/GRANT/CROSS_SCHEMA/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/INDEX
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/COMMENT
    Processing object type DATABASE_EXPORT/SCHEMA/PACKAGE/PACKAGE_SPEC
    Processing object type DATABASE_EXPORT/SCHEMA/PACKAGE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/FUNCTION/FUNCTION
    Processing object type DATABASE_EXPORT/SCHEMA/FUNCTION/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/PROCEDURE/PROCEDURE
    Processing object type DATABASE_EXPORT/SCHEMA/PROCEDURE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/FUNCTION/ALTER_FUNCTION
    Processing object type DATABASE_EXPORT/SCHEMA/PROCEDURE/ALTER_PROCEDURE
    Processing object type DATABASE_EXPORT/SCHEMA/VIEW/VIEW
    Processing object type DATABASE_EXPORT/SCHEMA/VIEW/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/VIEW/GRANT/CROSS_SCHEMA/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/VIEW/COMMENT
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/CONSTRAINT/REF_CONSTRAINT
    Processing object type DATABASE_EXPORT/SCHEMA/PACKAGE_BODIES/PACKAGE/PACKAGE_BODY
    Processing object type DATABASE_EXPORT/SCHEMA/TYPE/TYPE_BODY
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/FUNCTIONAL_AND_BITMAP/INDEX
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDEX_STATISTICS
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/STATISTICS/TABLE_STATISTICS
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/POST_TABLE_ACTION
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TRIGGER
    Processing object type DATABASE_EXPORT/SCHEMA/VIEW/TRIGGER
    Processing object type DATABASE_EXPORT/SCHEMA/JOB
    Processing object type DATABASE_EXPORT/SCHEMA/DIMENSION
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/POST_INSTANCE/PROCACT_INSTANCE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/POST_INSTANCE/PROCDEPOBJ
    Processing object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCOBJ
    Processing object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    Job "SSMP"."SYS_SQL_FILE_FULL_01" successfully completed at 17:43:09
    Thank you in advance.

    I believe there are lines like that in the export log.
    here's an extract from the log file:
    Processing object type DATABASE_EXPORT/SCHEMA/VIEW/TRIGGER
    Processing object type DATABASE_EXPORT/SCHEMA/JOB
    Processing object type DATABASE_EXPORT/SCHEMA/DIMENSION
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/POST_INSTANCE/PROCACT_INSTANCE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/POST_INSTANCE/PROCDEPOBJ
    Processing object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCOBJ
    Processing object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    . . exported "SSMP"."SMALLBOX" 450.2 MB 2489790 rows
    . . exported "SSMP"."LTCONFSB" 397.9 MB 3122899 rows
    . . exported "SSMP"."INFRJCT" 317.3 MB 3615810 rows
    . . exported "SSMP"."ICTEST1" 202.5 MB 789618 rows
    . . exported "SSMP"."EQUIPERF" 315.3 MB 3008636 rows
    . . exported "SSMP"."INFOPE2" 190.7 MB 569039 rows
    . . exported "SSMP"."SMALLBSB" 189.8 MB 2659468 rows
    . . exported "PRS1"."STKDTLOLD" 193.4 MB 983322 rows
    . . exported "SSMP"."ICTEST11" 185.2 MB 553916 rows
    . . exported "PRS1"."PURCHASE" 166.5 MB 122787 rows
    . . exported "SSMP"."JIGREC" 183.5 MB 2664488 rows
    . . exported "SSMP"."SALES" 112.4 MB 294134 rows
    . . exported "SSMP"."P_INFOPE" 160.4 MB 485773 rows
    . . exported "SSMP"."CIOP" 128.1 MB 385160 rows
    . . exported "SSMP"."SMALLBOX2" 126.0 MB 589964 rows
    . . exported "SSMP"."SHIPOUT" 123.4 MB 281778 rows
    . . exported "SSMP"."INFHISTORY" 129.9 MB 786844 rows
    . . exported "SYSMAN"."MGMT_TARGETS" 14.90 KB 11 rows
    . . exported "SSMP"."LOGCNT" 4.937 KB 1 rows
    . . exported "SSMP"."WEEKINF" 45.07 KB 700 rows
    . . exported "SSMP"."DEPTNM" 12.33 KB 81 rows
    . . exported "SSMP"."WFRID" 5.953 KB 2 rows
    . . exported "SSMP"."MATDTLS" 85.53 KB 199 rows
    . . exported "SSMP"."ICRATTACH" 9.257 KB 11 rows
    . . exported "SSMP"."WIP2" 20.34 KB 67 rows
    . . exported "SSMP"."MATINFO" 17.83 KB 43 rows
    . . exported "SSMP"."ICRFAPP" 11.48 KB 11 rows
    . . exported "SSMP"."WMXINF" 43.60 KB 500 rows
    . . exported "SSMP"."MLDMACHINE" 32.30 KB 228 rows
    . . exported "SSMP"."ICRNAME" 10.48 KB 23 rows
    . . exported "SSMP"."XLODCNT" 4.937 KB 1 rows
    . . exported "SSMP"."IJPLODCNT" 5.234 KB 1 rows
    . . exported "SSMP"."MONTHINF" 14.67 KB 152 rows
    . . exported "SSMP"."YYSTMP" 15.57 KB 59 rows
    . . exported "SSMP"."MPRICE" 39.95 KB 365 rows
    . . exported "SYSMAN"."ESM_COLLECTION" 8.585 KB 91 rows
    . . exported "SSMP"."INFCHKSHT" 6.406 KB 4 rows
    . . exported "SSMP"."MRKLNO" 4.929 KB 1 rows
    . . exported "SSMP"."JPNMARK" 5.890 KB 2 rows
    . . exported "SSMP"."LODCNT" 4.921 KB 1 rows
    . . exported "SSMP"."MRPNAME" 29.73 KB 310 rows
    . . exported "SYSMAN"."EXPORT000381" 49.68 KB 244 row

  • Performance Problems - Index and Statistics

    Dear Gurus,
    I am having problems lossing indexes and statistics on cubes ,it seems my indexes are too old which in fact are not too old just created a month back and we check indexes daily and it returns us RED on the manage TAB.
    please help

    Dear Mr Syed ,
    Solution steps I mentioned in my previous reply itself explains so called RE-ORG of tables;however to clarify more on that issue.
    Occasionally,ORACLE <b>Cost-Based Optimizer</b> may calculate the estimated costs for a Full Table Scan lower than those for an Index Scan, although the actual runtime of an access via an index would be considerably lower than the runtime of the Full Table Scan,Some Imperative points to be considered in order to perk up the performance and improve on quandary areas such as extensive running times for Change runs & Aggregate activate & fill ups.
    Performance problems based on a wrong optimizer decision would show that there is something serious missing at Database level and we need to RE_ORG  the degenerated indexes in order to perk up the overall performance and avoid daily manual (RSRV+RSNAORA)activities on almost similar indexes.
    For <b>Re-organizing</b> degenerated indexes 3 options are available-
    <b>1) DROP INDEX ..., and CREATE INDEX …</b>
    <b>2)ALTER INDEX <index name> REBUILD (ONLINE PARALLEL x NOLOGGING)</b>
    <b>3) ALTER INDEX <index name> COALESCE [as of Oracle 8i (8.1) only]</b>
    Each option has its Pros & Cons ,option <b>2</b> seems to be having lot of advantages to
    <b>Advantages- option 2</b>
    1)Fast storage in a different table space possible
    2)Creates a new index tree
    3)Gives the option to change storage parameters without deleting the index
    4)As of Oracle 8i (8.1), you can avoid a lock on the table by specifying the ONLINE option. In this case, Oracle waits until the resource has been released, and then starts the rebuild. The "resource busy" error no longer occurs.
    I would still leave the Database tech team be the best to judge and take a call on these.
    These modus operandi could be institutionalized  for all fretful cubes & its indexes as well.
    However,I leave the thoughts with you.
    Hope it Helps
    Chetan
    @CP..

  • Which Event Classes i should use for finding good indexs and statistics for queries in SP.

    Dear all,
    I am trying to use pro filer to create a trace,so that it can be used as workload in
    "Database Engine Tuning Advisor" for optimization of one stored procedure.
    Please tel me about the Event classes which i  should use in trace.
    The stored proc contains three insert queries which insert data into a table variable,
    Finally a select query is used on same table variable with one union of the same table variable, to generate a sequence for records based on certain condition of few columns.
    There are three cases where i am using the above structure of the SP, so there are three SPS out of three , i will chose one based on their performance.
    1) There is only one table with three inserts which gets  into a table variable with a final sequence creation block.
    2) There are 15 tables with 45 inserts , which gets into a tabel variable with a final
    sequence creation block.
    3)
    There are 3 tables with 9 inserts , which gets into a table variable with a final
    sequence creation block.
    In all the above case number of record will be around 5 lacks.
    Purpose is optimization of queries in SP
    like which Event Classes i should use for finding good indexs and statistics for queries in SP.
    yours sincerely

    "Database Engine Tuning Advisor" for optimization of one stored procedure.
    Please tel me about the Event classes which i  should use in trace.
    You can use the "Tuning" template to capture the workload to a trace file that can be used by the DETA.  See
    http://technet.microsoft.com/en-us/library/ms190957(v=sql.105).aspx
    If you are capturing the workload of a production server, I suggest you not do that directly from Profiler as that can impact server performance.  Instead, start/stop the Profiler Tuning template against a test server and then script the trace
    definition (File-->Export-->Script Trace Definition).  You can then customize the script (e.g. file name) and run the script against the prod server to capture the workload to the specified file.  Stop and remove the trace after the workload
    is captured with sp_trace_setstatus:
    DECLARE @TraceID int = <trace id returned by the trace create script>
    EXEC sp_trace_setstatus @TraceID, 0; --stop trace
    EXEC sp_trace_setstatus @TraceID, 2; --remove trace definition
    Dan Guzman, SQL Server MVP, http://www.dbdelta.com

  • How does Index fragmentation and statistics affect the sql query performance

    Hi,
    How does Index fragmentation and statistics affect the sql query performance
    Thanks
    Shashikala
    Shashikala

    How does Index fragmentation and statistics affect the sql query performance
    Very simple answer, outdated statistics will lead optimizer to create bad plans which in turn will require more resources and this will impact performance. If index is fragmented ( mainly clustered index,holds true for Non clustred as well) time spent in finding
    the value will be more as query would have to search fragmented index to look for data, additional spaces will increase search time.
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers
    My TechNet Wiki Articles

  • CRM workload and statistics tables (ST03N)

    Hi All,
          can anyone please give the name of the table used for storing the CRM workload and statistics data (Tx: ST03N).
    Thanks in advance

    here you go...
    SAPBWPRFLS                     BW Workload: Relation of Workload Profile and
                                                Class Names
    SAPWLDMTHD                     SAP Workload: Dynamically-Called Method for
                                                Events
    SAPWLROLED                     SAP Workload: Storage of User-Specific ST03N
                                                Role
    SAPWLROLES                     SAP Workload: Table for User-Specific Parameters
    SAPWLTREEG                     SAP Workload: Nodes for ST03G Tree Controls
    SAPWLTREET                     SAP Workload: Text Field for ST03G Trees
    SGLWLCUST1                     User Parameters for Workload Transactions
    SGLWLPRFLS                     SAP Workload: Profile Codes and Class Names
    SGLWLSLIST                      SAP Workload: User-Specific System List for Global
                                              Analysis
    Julius

  • Logs and Statistics Page in Airport

    What does "rotated ccmp group key" mean on Airport routers? What do other saying mean on the logs and statistics page? Thanks helps alot

    Your signal and noise readings are excellent; these are numbers in decibel (logarithmic) scale, with 0 (zero) being the maximum; the signal levels reported indicate how much below the theoretical maximum signal strength you are receiving. As a point of comparison, as I type this right now I'm getting a signal of -45 db and a noise of -9 db. Rate, I believe, is in megabits per second (I'm showing 117 at the moment, though it often goes up to 130 or so).

  • I am trying the expdp and impdp utility getting error

    I am trying the expdp and impdp utility, as per the documentation I logged in as sys and created a directory datapumpt in the location 'C:\oracle\odp' , i did grant the read and write privilege to my user_id.
    when i tried the expdp command in the prompt its throwing errors:
    Ora:39002 invalid operation
    ora: 39070 unable to open the log file
    ora: 29283 invalid file operation
    ora:06512 : at sys.utl_file line 475
    ora: 29283 invalid file operation.
    Operating system : windows / oracle version : 10.1.0.3.0
    Operating system : Linux AS4
    please help me on this.

    I'm afraid there is a big problem...
    C:\oracle\odp is a windows path
    If the Oracle server is running under Linux, your directory is invalid.
    A directory is an "alias" for a folder on the server, not on the client !!!
    The expdp/utl_file/... won't access folder on the client !

  • Move all data from one database to other using oracle expdp and impdp tool

    I need to move all of the data from one database to another. For that I am using oracle expdp and impdp tool.
    I have the full database export dump file. I moved that dump file to my other database.
    Now when I try to use impdp tools it gives lots of error regarding path mismatch,user not exists, role not exists.
    As my datafile path of source database does not match with the target database path , also the users are not there in target database. I dont want to create all the tablespaces that in my source database as there are 82 tablespaces.Is there any way by which I can move my all data without using remap option for tablespaces or users and grants.

    The FULL parameter indicates that a complete database export is required. The following is an example of the full database export and import syntax.
    The user must have the privielge of EXP_FULL_DATABASE
    expdp system/password@db10g full=Y directory=TEST_DIR dumpfile=DB10G.dmp logfile=expdpDB10G.log
    impdp system/password@db10g full=Y directory=TEST_DIR dumpfile=DB10G.dmp logfile=impdpDB10G.logGO through below linkfor more details...
    http://www.oracle-base.com/articles/10g/OracleDataPump10g.php
    Regards
    Umi
    Edited by: Umi on Feb 4, 2011 2:27 AM

  • Expdp and user_sdo_geom_metadata

    Dear all,
    I am trying to export data with expdp, and import it into another database. I can however see problems with my geometry indexes, which fail on import. user_sdo_geom_metadata also seems to be empty after import. Is there an extra option I should use in expdp to support geometries?
    Regards

    enhancement request 2419788 has been added to Oracle's database.

  • How to impl result match percentage and statistics on # of visit of docs

    Can someone please suggest on how to impl result match percentage and statistics on # of visit of docs.
    Can this be achieved without using ultra search?
    For example, by entering 'merge', I would like to see
    result such as
    "Party Merge High Level Design", 100%, # of hits:125.

    Can someone please suggest on how to impl result match percentage and statistics on # of visit of docs.
    Can this be achieved without using ultra search?
    For example, by entering 'merge', I would like to see
    result such as
    "Party Merge High Level Design", 100%, # of hits:125.

  • Using expdp and impdp instead of exp and imp

    Hi All,
    I am trying to use expdp and impdp instead of exp and imp.
    I am facing few issues while using expdb. I have a Job which exports data from one DB server and then its imported so another DB server. Both DB servers are run on separate machines. Job runs on various clients machine and not on any of DB server.
    For using expdp we have to create DIRECTORY and as I understand it has to be created on DB server. Problem here is Job can not access DB Server or files on DB server. Also dump file created is moved by Job to other machines based on requirement( Usually it goes to multiple DB server).
    I need way to create dump files on server where job runs.
    If I am not using expdp correctly please guide. I am new to expdp/impdp and imp/exp.
    Regards,

    Thanks for quick reply ..
    Job executing expdb/impdp runs on Red Hat Enterprise Linux Server release 5.6 (Tikanga)
    ORACLE server Release 11.2.0.2.0
    Job can not access the ORACLE server as it does not have privileges (In fact there is no user / password to access ORACLE server machines). Creating dump on oracle server and moving is not an option for this JOB. It has to keep dump with itself.
    Regards,

  • Siebel and parameters and statistics

    Hi,
    I work on SIEBEL 8 with 10gr2 database on i want to know the preconisations SIEBEL about specific parameter and statistics.
    Thanks for your help

    Read Technical Note 582 for general information.
    Additionally, this technical note also summarizes/compares 9i and 10g server initialization parameters. Be aware that some of the settings for 9i and 10g are completely different.
    Install/Upgrade/Manage
    http://download.oracle.com/docs/cd/B40099_02/books/UPG/UPG_Prep_Oracle3.html
    HTH
    -Anantha

  • LIFO Valuation for two financial books Tax and Statistics

    Dear Sir,
    Please let me know if my following requirement is possible in SAP or not.
    We would like to generate LIFO valuation for two financial books Tax and Statistics
    Tax book begins from April to March and
    Stat book begins from August to July.
    We are able to execute LIFO valuation only for Statistic books, as our financial books are from Aug to July.
    Also in LIFO configuration there is no setting for two fiscal year periods.
    Please let me know if our requirement is feasible for not.
    Regards,
    Sandeep

    Guys,
    Can anyone through some light on my query?
    Regards,
    Sandeep parab

  • Gnome-Power-Manager don't save history and statistics

    I notice that g-p-m doesn't save history and statistics during a session:
    when I reboot the pc I loose every information and the graphs start by 0.
    I checked every configuration on gconf-editor, but nothing helped me.
    There is the section /apps/gnome-power-manager/info/ but there isn't a schema for any configuration, I don't know what test I could make.
    And I don't find the folder ~/.gnome2/gnome-power-manager on my Arch installations.
    I found this bug with Ubuntu:
    https://bugs.launchpad.net/ubuntu/+sour … bug/302570
    So I run lshal | grep battery to have information about my asus eeepc battery.
    This is the result:
    [root@e3pc luca]# lshal | grep battery
    udi = '/org/freedesktop/Hal/devices/computer_power_supply_battery_BAT0'
      battery.charge_level.current = 20871  (0x5187)  (int)
      battery.charge_level.design = 51034  (0xc75a)  (int)
      battery.charge_level.last_full = 50638  (0xc5ce)  (int)
      battery.charge_level.percentage = 41  (0x29)  (int)
      battery.charge_level.rate = 16993  (0x4261)  (int)
      battery.is_rechargeable = true  (bool)
      battery.model = '901'  (string)
      battery.present = true  (bool)
      battery.rechargeable.is_charging = true  (bool)
      battery.rechargeable.is_discharging = false  (bool)
      battery.remaining_time = 6306  (0x18a2)  (int)
      battery.reporting.current = 2691  (0xa83)  (int)
      battery.reporting.design = 6580  (0x19b4)  (int)
      battery.reporting.last_full = 6529  (0x1981)  (int)
      battery.reporting.rate = 2191  (0x88f)  (int)
      battery.reporting.technology = 'Li-ion'  (string)
      battery.reporting.unit = 'mAh'  (string)
      battery.serial = ''  (string)
      battery.technology = 'lithium-ion'  (string)
      battery.type = 'primary'  (string)
      battery.vendor = 'ASUS'  (string)
      battery.voltage.current = 7756  (0x1e4c)  (int)
      battery.voltage.design = 8400  (0x20d0)  (int)
      battery.voltage.unit = 'mV'  (string)
      info.capabilities = {'battery'} (string list)
      info.category = 'battery'  (string)
      info.udi = '/org/freedesktop/Hal/devices/computer_power_supply_battery_BAT0'  (string)
    Eeepc hasn't the serial. Ok...
    But on my hp 6735s there is the serial of the battery while I run "lshal | grep battery".
    And it has the same problem.
    The version of g-p-m is 2.26.2-1 on both of them.
    I had many difficulties to find documentation about this gnome daemon!!
    Somebody found a solution?
    Thank you all!

    up

Maybe you are looking for

  • Sap integration with third party software

    hey all, we are actually trying to integrate sap with EAI( IBM WEBSPERE third party software) and here date needs to flow from EAI - SAP. and we have done that through BAPI and in the EAI we got a message saying that the data base has been created bu

  • Front Row compatabilty

    I currently own a Pc with Vista, can i install Front Row on it, or is there a alternative for a Pc, as windoes Media Center is rubbish Thanks replies are very appreciated. Nathan

  • Extract date from time stamp

    hi , how to extract DATE from timestamp column thanks

  • Adding attachment through ABAP screen

    Hi experts, I have a requirement where i need to add attachment from a ABAP screen or through SM30 screen. That is, we provide a transaction code for the users to maintain table entries. In additon to maintain values, the user wants to add attachment

  • Installing Windows 8 Pro on Macbook pro 5,4

    Tried to install Win Pro 8 on my Macbook pro 5,4. Had Win 7 pro and bought the pro version for $15.  Bootcamp partition used.  Installation was going fine and it completely crashed.  I do not want to format that drive since it has valuable materials