Reading data volume in Set-up tables - LO Purchasing

Hi Gurus,
I performed OLI3BW and it took almost a day to complete for HDR, ITM & SCL structures. I would like to know is there a way we can find out volume of records tranferred to set-up tables. I used TCode: NPRT, this is showing only records from EKKO table. Please advise.
Thank You,

Hi,
I think you can only goto the application tables ekko, ekpo... using transaction se16 to get an estimation of records posted to the setup tables.
regards
Siggi

Similar Messages

  • Reading data from a set of tables

    Hi,
    I am writing code to extract the data from several tables in a database. For each table, I am dynamically getting the number of fields or columns in the table and using Resultset obtaining the data in every field for each row. I do this using the getObject() method. Once I get the data for every field, I then want to insert this data into another table and for this I am using setObject() method. I am using these methods because the type of each field will be known only at runtime. After I do this, I do executeUpdate() in order to insert the data into the new table.
    The problem is here -
    At times it just inserts a few rows / records in the new table or it just gets and sets the Object for the first Row/Record in the table and then it freezes.
    Is there any problem with the getObject() and setObject() methods?
    while(rs.next())
    for (int i =1; i <= rsmd.getColumnCount(); i++)
    obj = rs.getObject(i);
    System.out.println("Object received " + obj);
    pstmt.setObject(i,obj);
    System.out.println("Object set : "+ obj);
    pstmt.executeUpdate();
    } // end of while
    Thanks in advance,
    Nik

    I don't know which DBMS/JDBC driver you are using. But there are drivers out there which do not permit two open Statements on the same connection. Maybe that is your problem. Try to close the ResultSet and the SELECT statement before the executeUpdate() of the second statement
    Another thing to check: are you sure the column order for the SELECT and the INSERT is identical (in other words, are you building the INSERT based on the information received from the ResultSet?)
    When it freezes, what do you mean with that? Where exactly does it "freeze"?
    This might be a DB lock which your program is waiting for - some DBMS do a page/table lock when updating, so a different statement could block your statement.
    Thomas

  • How to extract data into the set-up table for 2LIS_06_INV LIS structure

    We are using ECC 6.0 and SAP BI NW 2004S. I activated the 2LIS_06_INV  (Invoice Verification) structure. Interestingly, I don't see any Events under this structure  (MC06M_0ITM) - my understanding is the events usually determine what type of data is generated for a given structure.
    I see Invoice Verification when I use the Inventory Management -Perform Setup  option when doing the set-up tables. However, when I use this option, I get a message saying " No extraction structure active or no BW connected".
    Can someone list the pre-requisites and the steps to load the set-up table for the 2LIS_06_INV structure.
    Thanks,
    Sanjay

    1: RSA5 Activate Data Source
    2: LBWE Activate datasource again
    3: SBIW Fill setup table 'Settings for Application-Specific DataSources (PI)'-> 'Initialization'->'Filling in the Setup Table'->'Application-Specific Setup of Statistical Data'->'Invoice Verification - Execute Reconstruction'

  • Oracle DB: How to read data from a corrupted database table.

    Hi All,
    Wanted to know if there is a way I could read data from a corrupted Oracle database table?
    Are there any tools that I can use? Or does Oracle provides any mechanism to do that?
    Any pointers in this regard would be helpful.
    Thanks in Advance.

    user10600611 wrote:
    Table shows inconsistent data.You are going to have to be more specific. Inconsistent in what sense? A correct data model should not allow for inconsistencies to crop up.
    However, you may be able to use one of the many FLASHBACK features of Oracle to look at the table at a prior point in time before the corruption.
    HTH!

  • Reading data from XML to Internal table

    Hi Experts,
    I got a requirement to read the data fom xml to internal table.is it possible to store deep internel table data into flat internal tables?my internel table contains 4 internel table and these 4 internel table contains 2 internel tables each...can any one help me o this...
    points will be rewarded...
    Regards,
    Rakhi

    Hi Rakhi,
    * Structure to Get the Client Details as in XML format
      DATA: BEGIN OF client,
              BEGIN OF Personal_Details,
                kunnr like kna1-kunnr,
                name1 like kna1-name1,
                adrnr like kna1-adrnr,
              END OF Personal_Details,
              BEGIN OF Address,
                street like adrc-street,
                city1  like adrc-city1,
              END OF Address,
              BEGIN OF Communication,
                fax_number like adrc-fax_number,
                tel_number like adrc-tel_number,
              END OF Communication,
            END OF client.
      DATA: it_client TYPE TABLE OF client WITH HEADER LINES,
            result LIKE client.
      DATA: xml_string TYPE string.
    *  Get the Client Details into the structure to be made as an XML string
      TRY.
    *  Convert the Structure to XML string
          CALL TRANSFORMATION  ('ID')
            SOURCE para = it_client
            RESULT XML xml_string.
    *  Convert the XML string to structure - result
          CALL TRANSFORMATION  ('ID')
            SOURCE XML xml_string
            RESULT para = result.
      CATCH cx_st_error.
      ENDTRY.
    best regards,
    Thangesh

  • Read statement to read data and display in output table

    hi
    consider a scenerio as below
    int table in which data i already tehre gt_sagmeld (here primary key is guid_lclic)
    and matching field for next select is guid_mobj
    not corpar table is joined to butoo by  parno.
    and primary key of corpar is guid_corpar which has no link to gt_sagmeld
    and primary key of but00 is partner and it isjoined to corpar by but00-partner = corpar-parno
    adn i need to read but00-bpext
    pls see the below seelct statment for it
    gt_Saglemd has data...
    data: gt_sagmeld_temp like gt_sagmeld.
    gt_sagmeld_temp[] = gt_sagmeld[].
    delete adjacent duplicates from gt_sagmeld_temp[] comparing guid_mobj
    select       corpar~PAFCT 
                  corpar~parno     
                into table gt_corpar
        from    corpar
        for all entries in gt_sagmeld_temp
        where  /sapsll/corpar~guid_mobj = gt_sagmeld_temp-guid_mobj
        and    /sapsll/corpar~PAFCT = 'SH'.
    sort gt_corpar by parno.
    delete adjacent duplicates from gt_corpar comparing parno.
    select but000~partner
           but000~bpext
       from but000 into table gt_but001
    for all entries in gt_corpar
    where  but000~partner = gt_corpar-parno.
    now the table gt_but001 contans the required bpext
    and i want to read it to outtab
    reading...
    loop at gt_satmeld into wa_gt_sagmeld
    some field selection....
    now how to read the bpext in this loop from but001
    as there is table corpar also and then but001
    read table gt_but000 into wa_but000
      with key  partner = corpar-parno
    ...but this will not work as firs i need to read corpar also so how to use the loop or what is procedure to read ...
    pls suggest

    Hi,
    It is very much possible.
    1. You can create a new table in SE11/SE12 and u can create this table as a maintainable table( able to maintain entries). An ABAPer can help you to enable this table as maintainable one.
    2. If you know the name of the maintainable table, you can maintain entries directly in SM30. If you need a tcode for this, you can also do this by creating a tcode in SE91 for the table.
    3. If you know the tcode, you can directly view and download the entries using the tcode itself.

  • How to read data and load into a table

    Hi,
    I have a client requirement as below.
    Client will send the xmlpayload, based on this xml I want to read the data and store it into a database table. Could you please let me know how to achieve this functionality.
    sample xmlpayload:
    <gesws:localeAndSend xmlns:gesws="https://services.gmail.com" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="https://services.gmail.com docs/xsd/localeandsend.xsd" schemaVersion="1.0">
    <gesws:localer>
    <gesws:divisionstate>localeD</gesws:divisionstate>
    <gesws:division>Retail</gesws:division>
    <gesws:attributes>
    <gesws:attribute>
    <gesws:name>email</gesws:name>
    <gesws:value>[email protected]</gesws:value>
    </gesws:attribute>
    <gesws:attribute>
    <gesws:name>firstName</gesws:name>
    <gesws:value>Robin</gesws:value>
    </gesws:attribute>
    <gesws:attribute>
    <gesws:name>lastName</gesws:name>
    <gesws:value>Dan</gesws:value>
    </gesws:attribute>
    <gesws:attribute>
    <gesws:name>postalCode</gesws:name>
    <gesws:value>56302</gesws:value>
    </gesws:attribute>
    <gesws:attribute>
    <gesws:name>Mobileapp</gesws:name>
    <gesws:value>-6</gesws:value>
    </gesws:attribute>
    <gesws:attribute>
    <gesws:name>CodedString</gesws:name>
    <gesws:value>1550</gesws:value>
    </gesws:attribute>
    <gesws:attribute>
    <gesws:name>CodedBoolean</gesws:name>
    <gesws:value>true</gesws:value>
    </gesws:attribute>
    </gesws:attributes>
    </gesws:localer>
    <gesws:localerMessage>
    <gesws:mId>120098</gesws:mId>
    </gesws:localerMessage>
    </gesws:localeAndSend>
    Thanks in advance

    You can use XMLTABLE to extract data from your XMLTYPE datatype...
    SQL> ed
    Wrote file afiedt.buf
      1  with t as (select xmltype('<gesws:localeAndSend xmlns:gesws="https://services.gmail.com" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
      2  xsi:schemaLocation="https://services.gmail.com docs/xsd/localeandsend.xsd" schemaVersion="1.0">
      3    <gesws:localer>
      4      <gesws:divisionstate>localeD</gesws:divisionstate>
      5      <gesws:division>Retail</gesws:division>
      6      <gesws:attributes>
      7        <gesws:attribute>
      8          <gesws:name>email</gesws:name>
      9          <gesws:value>[email protected]</gesws:value>
    10        </gesws:attribute>
    11        <gesws:attribute>
    12          <gesws:name>firstName</gesws:name>
    13          <gesws:value>Robin</gesws:value>
    14        </gesws:attribute>
    15        <gesws:attribute>
    16          <gesws:name>lastName</gesws:name>
    17          <gesws:value>Dan</gesws:value>
    18        </gesws:attribute>
    19        <gesws:attribute>
    20          <gesws:name>postalCode</gesws:name>
    21          <gesws:value>56302</gesws:value>
    22        </gesws:attribute>
    23        <gesws:attribute>
    24          <gesws:name>Mobileapp</gesws:name>
    25          <gesws:value>-6</gesws:value>
    26        </gesws:attribute>
    27        <gesws:attribute>
    28          <gesws:name>CodedString</gesws:name>
    29          <gesws:value>1550</gesws:value>
    30        </gesws:attribute>
    31        <gesws:attribute>
    32          <gesws:name>CodedBoolean</gesws:name>
    33          <gesws:value>true</gesws:value>
    34        </gesws:attribute>
    35      </gesws:attributes>
    36    </gesws:localer>
    37    <gesws:localerMessage>
    38      <gesws:mId>120098</gesws:mId>
    39    </gesws:localerMessage>
    40  </gesws:localeAndSend>') as xml from dual)
    41  --
    42  -- end of test data
    43  --
    44  select x.mID, x.divisionstate, x.division
    45        ,y.att_name, y.att_val
    46  from   t
    47        ,xmltable(xmlnamespaces('https://services.gmail.com' as "g")
    48                 ,'g:localeAndSend'
    49                 passing t.xml
    50                 columns mID           number       path './g:localerMessage/g:mId'
    51                        ,divisionstate varchar2(10) path './g:localer/g:divisionstate'
    52                        ,division      varchar2(10) path './g:localer/g:division'
    53                        ,atts          xmltype      path './g:localer/g:attributes'
    54                 ) x
    55        ,xmltable(xmlnamespaces('https://services.gmail.com' as "g")
    56                 ,'g:attributes/g:attribute'
    57                 passing x.atts
    58                 columns att_name      varchar2(20) path './g:name'
    59                        ,att_val       varchar2(20) path './g:value'
    60*                ) y
    SQL> /
           MID DIVISIONST DIVISION   ATT_NAME             ATT_VAL
        120098 localeD    Retail     email                [email protected]
        120098 localeD    Retail     firstName            Robin
        120098 localeD    Retail     lastName             Dan
        120098 localeD    Retail     postalCode           56302
        120098 localeD    Retail     Mobileapp            -6
        120098 localeD    Retail     CodedString          1550
        120098 localeD    Retail     CodedBoolean         true
    7 rows selected.
    SQL>Once you're getting the data out, you can do what you want with it, e.g. pivot it to a single row...
    SQL> ed
    Wrote file afiedt.buf
      1  with t as (select xmltype('<gesws:localeAndSend xmlns:gesws="https://services.gmail.com" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
      2  xsi:schemaLocation="https://services.gmail.com docs/xsd/localeandsend.xsd" schemaVersion="1.0">
      3    <gesws:localer>
      4      <gesws:divisionstate>localeD</gesws:divisionstate>
      5      <gesws:division>Retail</gesws:division>
      6      <gesws:attributes>
      7        <gesws:attribute>
      8          <gesws:name>email</gesws:name>
      9          <gesws:value>[email protected]</gesws:value>
    10        </gesws:attribute>
    11        <gesws:attribute>
    12          <gesws:name>firstName</gesws:name>
    13          <gesws:value>Robin</gesws:value>
    14        </gesws:attribute>
    15        <gesws:attribute>
    16          <gesws:name>lastName</gesws:name>
    17          <gesws:value>Dan</gesws:value>
    18        </gesws:attribute>
    19        <gesws:attribute>
    20          <gesws:name>postalCode</gesws:name>
    21          <gesws:value>56302</gesws:value>
    22        </gesws:attribute>
    23        <gesws:attribute>
    24          <gesws:name>Mobileapp</gesws:name>
    25          <gesws:value>-6</gesws:value>
    26        </gesws:attribute>
    27        <gesws:attribute>
    28          <gesws:name>CodedString</gesws:name>
    29          <gesws:value>1550</gesws:value>
    30        </gesws:attribute>
    31        <gesws:attribute>
    32          <gesws:name>CodedBoolean</gesws:name>
    33          <gesws:value>true</gesws:value>
    34        </gesws:attribute>
    35      </gesws:attributes>
    36    </gesws:localer>
    37    <gesws:localerMessage>
    38      <gesws:mId>120098</gesws:mId>
    39    </gesws:localerMessage>
    40  </gesws:localeAndSend>') as xml from dual)
    41  --
    42  -- end of test data
    43  --
    44  select mID, divisionstate, division
    45        ,max(decode(att_name, 'email', att_val)) as email
    46        ,max(decode(att_name, 'firstName', att_val)) as firstname
    47        ,max(decode(att_name, 'lastName', att_val)) as lastname
    48        ,max(decode(att_name, 'postalCode', att_val)) as postalcode
    49        ,max(decode(att_name, 'CodedString', att_val)) as codedstring
    50        ,max(decode(att_name, 'CodedBoolean', att_val)) as codedboolean
    51  from (
    52        select x.mID, x.divisionstate, x.division
    53              ,y.att_name, y.att_val
    54        from   t
    55              ,xmltable(xmlnamespaces('https://services.gmail.com' as "g")
    56                       ,'g:localeAndSend'
    57                       passing t.xml
    58                       columns mID           number       path './g:localerMessage/g:mId'
    59                              ,divisionstate varchar2(10) path './g:localer/g:divisionstate'
    60                              ,division      varchar2(10) path './g:localer/g:division'
    61                              ,atts          xmltype      path './g:localer/g:attributes'
    62                       ) x
    63              ,xmltable(xmlnamespaces('https://services.gmail.com' as "g")
    64                       ,'g:attributes/g:attribute'
    65                       passing x.atts
    66                       columns att_name      varchar2(20) path './g:name'
    67                              ,att_val       varchar2(20) path './g:value'
    68                       ) y
    69       )
    70  group by mID, divisionstate, division
    71* order by 1
    SQL> /
           MID DIVISIONST DIVISION   EMAIL                FIRSTNAME            LASTNAME             POSTALCODE           CODEDSTRING          CODEDBOOLEAN
        120098 localeD    Retail     [email protected]    Robin                Dan                  56302                1550                 true

  • How to read data from idoc to internal table?

    Hi
    How to get data from idoc segments to internal table?

    Hi
    Check this ex
    The following coding sample, shows how you may read a MATMAS IDoc and extract the data for the MARA and MARC segments to some internal variables and tables.
    DATA: xmara LIKE mara.
    DATA: tmarc LIKE marc
                OCCURS 0
                WITH HEADER LINE.
    LOOP AT edidd.
       CASE edidd-segnam.
         WHEN 'E1MARAM'.
              MOVE edidd-sdata TO xmara.
         WHEN 'E1MARCM'.
           MOVE edidd-sdata TO tmarc.
              APPEND tmarc.
       ENDCASE.
    ENDLOOP.
    now do something with xmara and tmarc.
    hope this helps you...
    Reward points if useful..
    Regards
    Sreenivas

  • How to read data from field symbol containing Table

    I defined a field symbols for reading output of a BRF plus rule however since fielk symbol is of type any
    I cannot read its contents
    Any inputs are highly appreciated
    Please see attachment explaining problem

    Thanks a lot for your help
    However It did not work for me as it started giving error incompatible types
    Here is my code
    constants:lv_function_id type if_fdt_types=>id value '3440B5B078B21EE3BC9EB53C42F84A45'.
    data:lv_timestamp type timestamp,
          lt_name_value type abap_parmbind_tab,
          ls_name_value type abap_parmbind,
          lr_data type ref to data,
          lv_counter type i value 0,
          lr_target type ref to data,
          lx_fdt type ref to cx_fdt,
          lo_trace type ref to if_fdt_trace,
          lo_lean_trace type ref to if_fdt_lean_trace,
          la_z_string type if_fdt_types=>element_text,
          lv_role_name  type grac_role_name,
          lv_result type string.
    field-symbols <la_any> type any.
    field-symbols <role_name> type grac_s_od_role_detail.
    field-symbols <system> type string.
       types:
        begin of ys_access ,
                  role_name type  string,
                  system_name   type string,
                end of ys_access .
      types:
        yt_access type table of ys_access .
    field-symbols: <access_item> type table,
                    <ls_req_access> type ys_access.
    data con_name type if_fdt_types=>element_text.
    data ls_any type ys_access.
    * All method calls within one processing cycle calling the same function must use the same timestamp.
    * For subsequent calls of the same function, we recommend to use the same timestamp for all calls.
    * This is to improve the system performance.
    * If you are using structures or tables without DDIC binding, you have to declare the respective types
    * by yourself. Insert the according data type at the respective source code line.
    * GET TIME STAMP FIELD lv_timestamp.
    * Process a function and record trace data, passing context data objects via a name/value table.
    * Prepare function processing:
    ls_name_value-name = 'Z_STRING'.
    la_z_string = 'P1'.
    get reference of la_z_string into lr_data.
    ls_name_value-value = lr_data.
    insert ls_name_value into table lt_name_value.
    * Create the data to store the result value after processing the function
    * You can skip the following call, if you already have
    * a variable for the result. Please replace also the parameter
    * EA_RESULT in the method call CL_FDT_FUNCTION_PROCESS=>PROCESS
    * with the desired variable.
      cl_fdt_function_process=>get_data_object_reference( exporting iv_function_id      = lv_function_id
                                                                    iv_data_object      = '_V_RESULT'
                                                                    iv_timestamp        = lv_timestamp
                                                                    iv_trace_generation = abap_true
                                                          importing er_data             = lr_data ).
      assign lr_data->* to <la_any>.
      try.
          cl_fdt_function_process=>process( exporting iv_function_id = lv_function_id
                                                      iv_timestamp   = lv_timestamp
                                                      iv_trace_mode  = if_fdt_constants=>gc_trace_mode_lean
                                                     "iv_trace_mode  = if_fdt_constants=>gc_trace_mode_lean_required
                                            importing ea_result      = <la_any>
                                                      eo_trace       = lo_trace
                                            changing  ct_name_value  = lt_name_value ).
          lo_lean_trace ?= lo_trace.
          lo_lean_trace->save( ).
          catch cx_fdt into lx_fdt.
    * You can check CX_FDT->MT_MESSAGE for error handling.
      endtry.
    * Get context values after processing:
      cl_fdt_function_process=>get_context_value( exporting  iv_function_id = lv_function_id
                                                             iv_trace_generation = abap_true
                                                             iv_timestamp   = lv_timestamp:
                                                             iv_data_object = '3440B5B078B21EE3BC9EFF599C110ADD' "Z_STRING
                                                   importing ev_data             = lv_result ). " Suggested variable: la_z_string
      create data lr_target type grac_s_od_role_detail.
    *  assign <la_any> to <role_name>.
      assign lr_target->* to <la_any>.
       loop at lt_name_value into ls_name_value.
        if ls_name_value-name cs 'RESULT'.
          assign ls_name_value-value  to <la_any>.
            ls_any = ls_name_value-value  .
        endif.
        endloop.
        access_item = <la_any>.
    * assign ( )<la_any> to <access_item>.
    * assign component 1 of structure <la_any> to <rolename>.
    *   assign <la_any> to <ls_req_access>.
    * lr_target = <la_any>.
    * assign lr_target->* to <role_name>.
    * lv_role_name = <role_name>.
      lv_result = lv_result.

  • How to copy a set of tables from a database to another periodically?

    We have a 4 node RAC primary database(10.2.0.2) with a physical standby(10.2.0.2) on our production site. Offlate we noticed that one of the applications(APP2) is causing heavy loads due large data downloads on the primary database servers. Our primary database has 2 schemas,
    1) one being the main schema with all objects, (USER1)
    2) and the other has views that query some set of tables from the main schema. (USER2)
    The application APP2 uses USER2 views to query and download huge data periodically. We need to be able to give accurate data results to APP2, but in the same time take off the load from the database, as APP2 is not our main application.
    We would like to know if there are any cost effective options in oracle to do this, and if so, what is the best option? Anyone has any experience setting up something like this before?
    We have thought of creating another 10.2.0.2 database on a different server and giving it regular updates(like data feeds) from the current database. The current database data changes quiet often, so the data feeds would have to be done often to keep the data current on the new database. So, we are not exactly sure how to go about it. Would a COPY command help?
    Please advice.

    user623066 wrote:
    Our 4 node RAC is already busy with our main application, which has its connections spread across all 4 nodes.
    Our main applications services are the same on all nodes and use all 4 nodes in the same way.
    There are some other utilities that we run from one of the app servers that connect to only 1 of the nodes.
    APP2 uses all 4 servers, which is again controlled by connection pooling and distributes the load.Wouldn't separate services be more beneficial here? If APP2 is locked down to one node during normal operation, that ensures that other connections aren't going to be competing for hardware with APP2 on 3 of the 4 nodes. If APP2 is generating less than 25% of the total load, you can let the other applications use whatever hardware resources are left idle on the node APP2 is locked down to.
    By Large data downloads, I meant both increase in network traffic and the CPU load on the database nodes.
    We are already using resouce manager to limit the resources allocated to USER2 that APP2 uses.
    And we have also limited the large downloads to take place in the early hours of the day when the traffic from our main application is less.
    But this has still not been optimal for the usage requirements for APP2. APP2 is also doing queries all through the day, but has a limit for the number of rows downloaded during peak hours.Can you explain a bit more about why using Resource Manager hasn't been sufficient? That's normally a pretty good way to prevent one hungry user from drastically affecting everyone else. Perhaps you just need to tweak the configuration here.
    Logical Standby seems a good option. But we need to keep our physical standby in place. Is it possible to have a logical standby and a physical standby? (ofcourse on separate servers)Sure. You can have as many standby servers of whatever type you'd like.
    Could we use a COPY command to copy data for the set of tables to a new database? Or is that also a complex option?You could, yes. COPY is a SQL*Plus command that has been depricated for copying data between Oracle databases for quite a while. It only works from SQL*Plus and would only be designed for one-time operations (i.e. there is no incremental COPY command). I can just about guarantee that's not what you want here.
    How do materialized views work? Wouldn't they still reside on the main database? Or is it possible to have remote materialized views?You probably don't want materialized views, but if you decide to go down that path
    - You'd create materialized view logs on the base tables to track changes
    - You'd create materialized views on the destination database that select data over a database link back to the source database
    - You'd put those materialized views into one or more refresh groups that are scheduled to refresh periodically
    - During a refresh, assuming incremental refreshes, the materialized view logs would be read and applied to the materialized views on the destination system to update the materialized views.
    Justin

  • Mdworker deny file-read-data, what does it really mean?

    I get numerous log posts like this:
    3/17/14 7:22:57.561 AM sandboxd[5688]: ([4412]) mdworker(4412) deny file-read-data /Volumes/Time Machine Backups/Backups.backupdb/Jack Shelton’s iMac/2014-03-17-042247/JackHD/System/Library/Fonts/TimesLTMM (import fstype:hfs fsflag:4809018 flags:48200000056 diag:0 uti:com.apple.font-suitcase plugin:/Library/Spotlight/Font.mdimporter - find suspect file using: sudo mdutil -t 235288)
    My questiions are:
    What is being sandboxed, the mdworker or the file Times LTMM or the mdimporter?
    Which one has the sudo command in it?
    What does the -t flag/option mean?  Looking it up via man mdutil in the terminal, it does't exist.
    I saw this type of post quite frequently in Mavericks so much so that I decided to start over from scratch.  I wiped the HD including the HD Recovery partition and installed OS X 10.6.8.  I am now in Mountain Lion and have reinstalled MS Office 2011 and loaded the main identity data from my recovered files.  I did not see any adverse log posts until I ran the first time machine backup.
    I hope someone can clear the air for me with perhaps a way to fix it.

    Thank you Baltwo.  It's a major relief to know that the sudo command is being used by an Apple app/process and not by a file that's not supposed to have it..  However a few more questions if you don't mind.
    Is the font file the suspect file or is it this particular mdworker?
    If the font file is the suspect file will a process or app that calls for it be allowed to use it?
    Which process/app/file invoked the sandbox and deny file-read-data?
    What is the role of the mdimporter?
    These questions are part of my attempt to understand how a large part of the files on my computer and Time Machine become corrupted when I first installed Mavericks.  Mainly .plist (and other plain text files), .png, jpeg (and other picture types) and pkg info bundles would have thier icons and file Kind changed to Unix Executable files, spreadssheets and doc/docx would become genrandom (or something similar).  I became aware of this when my machine began freezing up.  I made several clean reinstalls and the problems would reoccur on plugging in the Time Machine.  I eventually recovered the files I wanted from the Time Machine but I had to use another iMac with an older OS X.  I repartitioned my machine and rebuilt it to Mountain Lion and i'm in the process of reloading with what I hope are non corrupted files. Although I still see log posts like "Locum[6230]: Connection with distnoted server was invalidated" every time I paste a folder or file.  I don't know what that means but it doesn't seem to cause any further reaction or posts.
    My apologizes for bothering you with alll this and thanks for your insight.

  • Read Data From DSO - In ABAP Program

    HI Friends,
    Can any one tell how to read Data from a DSO in a Abap Program on BI System itself......
    Which is the better way to read data from DSO in Abap Program... Is it
    1. Reading Data Directly From the Active Table Of the DSO ... or
    2. Reading Data From DSO....
    In case of reading data from DSO.... Which function module is used to do this work....When i searched the forum i came across 2 function Modules.... BAPI_ODSO_READ_DATA_UC and RSDRI_INFOPROV_READ..... Please let me know which one should be used....
    Regards,
    Shyam.

    Hi Shyam,
    The active content of any ods will be stored in the database tables( SE11) with the naming convention /BIC/A<odsname>00 for Custom-defined DSO and /BI0/A<odsname>00 for Business content delivered DSO.
    For example if the technical name of the ods is TEST , then the active content table name in SE11 will be /BIC/ATEST00.
    You can write your own program to read the contents of this database table.
    Regards,
    Krishna.

  • Read data randomly

    How can I read data records from an internal table in a random order?
    My aim is to select records randomly from an internal table without speciafying a key.

    I think the function RANDOM_AMOUNT gives decimal numbers
    Use this code to get only integers
    *& Report  ZKRIS_RANDOM
    report  zkris_random.
    selection-screen begin of block b1 with frame title title1.
      selection-screen begin of block b2 with frame title title2.
        parameters: low type i.
        parameters: high type i.
      selection-screen end of block b2.
      parameters: num_vals type i.
    selection-screen end of block b1.
    initialization.
    title1 = 'Parameters'.
    title2 = 'Limits'.
    start-of-selection.
      types: begin of random_numbers_type,
        random_nr type i,
        end of random_numbers_type.
      data: random_numbers type random_numbers_type  occurs 0 with header line.
      constants: a type i value 21,
                 c type i value 101,
                 m type i value 32768,
                 start type i value 500.
      data: new_num type i,
            interval type i.
      data: last_num type i.
      if high < low or num_vals < 1.
        write: 'parameter error'.
        exit.
      endif.
      if last_num is initial.
        last_num = start.
      endif.
      clear random_numbers[].
      interval = high - low + 1.
      do num_vals times.
        new_num = ( a * last_num + c ) mod m.
        last_num = new_num.
        random_numbers-random_nr = ( new_num mod interval ) + low.
        append random_numbers.
      enddo.
      loop at random_numbers.
        write:/ random_numbers-random_nr.
      endloop.

  • Extracting similar data from a dynamic set of tables

    I want to select common fields from an arbitrary set of tables, adding an extra field that is the table name, and return all of these rows in a single table. I know the steps but I am getting bogged down in the details (being new to Oracle):
    --Create a temporary table.
    --Generate a dynamic SQL statement to grab the data for each table of interest.
    --Execute that query and store the results in the temporary table.
    --Output the entire temporary table and delete it.
    Here is my latest attempt:
    ================
    Create global temporary table myTempTable
         Title varchar2(30),
         IdType number(8),
         Count number(8)
    ) on commit delete rows;
    declare
    sql_stmt VARCHAR2(200);
    tname VARCHAR2(30);
    begin
    for tname in (
    select distinct table_name from sys.ALL_TAB_COLS
              where owner='me' and table_name like '%ASSIGNS'
    loop
    sql_stmt :=
    'insert into myTempTable(Title, IdType, Count) ' ||
         'select '':1'', IdType, count(*) from me.:1 group by IdType';
    EXECUTE IMMEDIATE sql_stmt USING tname;
    end loop;
    select * from myTempTable;
    commit; -- to delete the temp table
    end;
    ================
    The error messages are:
    --expressions have to be of SQL types
    --an INTO clause is expected in this SELECT statement
    What am I missing here?

    (1) First of all I am executing the whole script in SQL*Plus, thats why I use SQL*Plus command like the SLASH Operator.
    <br>
    (2)
    SQL> VAR cur refcursor
    SQL> CREATE TABLE mytemptable AS SELECT ename table_name, deptno idtype, sal cnt FROM emp WHERE 1=0
    Table created.
    SQL> DECLARE
       sql_stmt   VARCHAR2 (200);
       tname      VARCHAR2 (30);
    BEGIN
       FOR tname IN (SELECT DISTINCT table_name
                                FROM SYS.all_tab_cols
                               WHERE owner = USER AND table_name LIKE '%EMP')
       LOOP
          sql_stmt :=
                'insert into myTempTable select '''
             || tname.table_name
             || ''' table_name, deptno IdType, count(*) cnt from '
             || tname.table_name
             || ' group by deptno';
          EXECUTE IMMEDIATE sql_stmt;
       END LOOP;
       OPEN :cur FOR
          SELECT *
            FROM mytemptable;
       EXECUTE IMMEDIATE 'drop table myTempTable';
    END;
    PL/SQL procedure successfully completed.
    SQL> PRINT :cur
    TABLE_NAME     IDTYPE        CNT
    EMP                10          3
    EMP                20          5
    EMP                30          6
    3 rows selected.Since we DROP the table at the end we don't need an explicit commit, since all DDL statement will implicitly commit anyway.

  • Reading Data from a Table With an Expert

    Hi all - I'm trying to write an Expert in OWB that would read data from a table and create a new table based on that information. I have the table creation part down, but how can I read data from a table with an Expert? From what I've read, Experts only deal with metadata, and there is no mechanism in Experts to actually read data in the tables. Does OraTcl work in Experts? Has anyone ever came across this problem, and if so, how did you solve it? Thanks in advance for your time! - Don

    Hi Don
    Can also use Java and JDBC from within Tcl, see the routines in this file below, take modify or whatever;
    http://blogs.oracle.com/warehousebuilder/ombora.tcl
    source <dir>\ombora.tcl
    set g_user scott
    set g_upwd zzzzzz
    set g_hostname localhost
    set g_port 1521
    set g_srvname ora111
    oraconnect $g_user $g_upwd $g_hostname:$g_port:$g_srvname
    oraselect "select * from emp" ""
    As well as dumping the results to stdout (you probably want to comment that part out and change any other part you don't like), g_res is a ResultSet object you can use to do whatever.
    Cheers
    David

Maybe you are looking for

  • SY-DBCNT QUESTION ?

    CAN ANYONE TELL ME HOW TO USE SY-DBCNT AFTER THE SELECT STATEMENT ? I NEED THIS TO KNOW HOW MY RECORDS MY SELECT QUERY IS RETURNING ? THKS!       select * from ekbe into ekbe where                  ebeln = it_out-ebeln              and ebelp = it_out

  • My apple tv is not working on the computer tab as it says to turn on homesharing and it is on

    my apple tv says to turn on homesharing which it already is i am unable to access the photos in the computer tab any suggestions thanks i have restarted and returned on both computer and apple tv btw

  • How to fix this error in webutil configuration

    hi all, I'm new in forms now i'm trying to configure webutil in forms 10g, actually i search in this forum and i found this link http://baigsorcl.blogspot.com/2010/01/open-file-dailog-box-example-using.html for BaiG member but i stucked in this step

  • Compatibility with Office for Mac

    I'll ask this in the Pages section, though it also applies to Numbers. I currently use MS Office for Mac 2011, but would like to uninstall it and use iWork 09.  I do have to interactively work on basic text documents and basic spreadsheets with other

  • Missing Thumbnails and Missing Photos

    **** iLife!! I updated iphoto to 6.02, and now I'm getting blank spaces for a bunch of photos in the thumbnail section, and I'm getting the empty grey square with the exclamation point in regular sized photos. Everything was just fine before the upgr