SLT Logging table issue

Hi,
We have configured SLT and was running fine. However, we found that some of the existing logging tables are created as a table in the configuration and that is in error (obviously).
Here is the screenshot:
The logging table for Table EKBZ is /1CADMC*637. Now the same logging tables has been created as Transp table and hence it is in error.
Did anybody come across such situation? Please help to resolve the same.
Regards,
SS

Hi,
have you been able to identify the cause of the issue and resolve it?
If yes, Can you share more details

Similar Messages

  • Deleting obsolete logging tables in SLT system

    Hi Experts,
    Need help to sort out below issue,
    There are many obsolete logging tables in SLT system ,might have been left out after dropping SLT schemas, previous replications.
    These logging tables exist in ABAP dictionary but not in database. Due to this, it shows inconsistencies in DB02 > Diagnostics > Missing Tables and Indexes. There are in total 252 missing tables in DB. Currently DB migration is stuck due to this issue.
    Is there any way to remove the tables from dictionary without deleteing them manually through SE14 ?
    How to avoid having such tables ? Am I missing any steps while deleting the previous schemas?
    Regards
    Kartik

    Hi Kartik,
    these obsolete objects/logging could be also there becasue someone refreshed/copied the systems.
    There is also one transaction IUUC_REMOTE in the source system to delete the logging tables. But I am not sure if it distinguish between db or ddic level. Have a try and post the result.
    Best,
    Tobias

  • Huge number of unprocessed logging table records found

    Hello Experts,
    I am facing one issue where huge number of unprocessed logging table records were found in SLT system for one table. I have check all setting and error logs but not found any evidence that causing the unprocessed records. In HANA system also it shows in replicated status. Could you please suggest me something other than to replicate same table again, as that option is not possible at this moment.

    Hi Nilesh,
    What are the performance impacts on the SAP ECC system when multiple large SAP tables like BSEG are replicated at the same time? Is there a guideline for a specific volume or kind of tables?
    There is no explicit guideline since aspects as server performance as well as change rate of the tables are also relevant. As a rule of thumb, one dedicated replication job per large table is recommended.
    from SLT
    How to enable parallel replication before DMIS 2011 SP6    do not ignore its for SP06 == go through
    How to improve the initial load
    Regards,
    V Srinivasan

  • Need suggestion to get data from change log table of ODS.

    Hello,
    There is a case where i am loading opportunity header data from header ODS and opportunity item data from item ODS in the opportunity cube.
    Status (1= OPEN, 2= WON ETC) of the opportunity are available only in header ODS and not in item ODS.
    While loading data from header ODS to cube, I am loading it directly but while loading data from item ODS to cube i am using active data table of header ODS as a lookup in the update rule from item ODS to cube. I am selecting status from the active data table of header ODS while loading data from item ODS to cube.
    Since active data table will have only after image records, there is some data mismatch in the report as i am selecting data from active data table of header ODS while loading data from item ODS to cube.
    I need to select data from Change log in order to get before image also instead of active data table in order to overcome this issue. Is there any way by which i can do selection from Change log instead of active data table as change logs are generated at run time.
    Please let me know if you have any suggestions.
    Regards,
    Sanjay Chaurasia.

    Hi,
    You can use the changelog table of the DSO.
    Right click manage the Header DSO, go to the contents tab and click Change Log table. There you can see the technical name of the Change Log table.
    In the update rule Routine, give the tech name of Change log table instead of Active table name.
    Hope it helps.
    Krishna

  • DS 4.2 get ECC CDHDR deltas in ABAP data flow using last run log table

    I have a DS 4.2 batch job where I'm trying to get ECC CDHDR deltas inside an ABAP data flow.  My SQL Server log table has an ECC CDHDR last_run_date_time (e.g. '6/6/2014 10:10:00') where I select it at the start of the DS 4.2 batch job run and then update it to the last run date/time at the end of the DS 4.2 batch job run.
    The problem is that CDHDR has the date (UDATE) and time (UTIME) in separate fields and inside an ABAP data flow there are limited DS functions.  For example, outside of the ABAP data flow I could use the DS function concat_date_time for UDATE and UTIME so that I could have a where clause of 'concat
    _date_time(UDATE, UTIME) > last_run_date_time and concat_date_time(UDATE, UTIME) <= current_run_date_time'.  However, inside the ABAP data flow the DS function concat_date_time is not available.  Is there some way to concatenate UDATE + UTIME inside an ABAP data flow?
    Any help is appreciated.
    Thanks,
    Brad

    Michael,
    I'm trying to concatenate date and time and here's my ABAP data flow where clause:
    CDHDR.OBJECTCLAS in ('DEBI', 'KRED', 'MATERIAL')
    and ((CDHDR.UDATE || ' ' || CDHDR.UTIME) > $CDHDR_Last_Run_Date_Time)
    and ((CDHDR.UDATE || ' ' || CDHDR.UTIME) <= $Run_Date_Time)
    Here are DS print statements showing my global variable values:
    $Run_Date_Time is 2014.06.09 14:14:35
    $CDHDR_Last_Run_Date_Time is 1900.01.01 00:00:01
    The issue is I just created a CDHDR record with a UDATE of '06/09/2014' and UTIME of '10:48:27' and it's not being pulled in the ABAP data flow.  Here's selected contents of the generated ABAP file (*.aba):
    PARAMETER $PARAM1 TYPE D.
    PARAMETER $PARAM2 TYPE D.
    concatenate CDHDR-UDATE ' ' into ALTMP1.
    concatenate ALTMP1 CDHDR-UTIME into ALTMP2.
    concatenate CDHDR-UDATE ' ' into ALTMP3.
    concatenate ALTMP3 CDHDR-UTIME into ALTMP4.
    IF ( ( ALTMP4 <= $PARAM2 )
    AND ( ALTMP2 > $PARAM1 ) ).
    So $PARAM1 corresponds to $CDHDR_Last_Run_Date_Time ('1900.01.01 00:00:01') and $PARAM2 corresponds to $Run_Date_Time ('2014.06.09 14:14:35').  But from my understanding ABAP data type D is for date only (YYYYMMDD) and doesn't include time, so is my time somehow being defaulted to '00:00:00' when it gets to DS?  I ask this as a CDHDR record I created on 6/6 wasn't pulled during my 6/6 testing but this 6/6 CDHDR record was pulled today.
    I can get  last_run_date_time and current_run_date_time into separate date and time fields but I'm not sure how to build the where clause using separate date and time fields.  Do you have any recommendations or is there a better way for me to pull CDHDR deltas in an ABAP data flow using something different than a last run log table?
    Thanks,
    Brad

  • Oracle 11g - External Table Issue SQL - PL/SQL?

    Oracle 11g - External Table Issue?
    =====================
    I hope this is the right forum for this issue, if not let me, where to go.
    We are using Oracle 11g (11.2.0.1.0) on (Platform : solaris[tm] oe (64-bit)), Sql Developer 3.0.04
    We are trying to use oracle external table to load text files in .csv format. Here is our data look like.
    ======================
    Date1,date2,Political party,Name, ROLE
    20-Jan-66,22-Nov-69,Democratic,"John ", MMM
    22-Nov-70,20-Jan-71,Democratic,"John Jr.",MMM
    20-Jan-68,9-Aug-70,Republican,"Rick Ford Sr.", MMM
    9-Aug-72,20-Jan-75,Republican,Henry,MMM
    ALL NULL -- record
    20-Jan-80,20-Jan-89,Democratic,"Donald Smith",MMM
    ======================
    Our Expernal table structures is as follows
    CREATE TABLE P_LOAD
    DATE1 VARCHAR2(10),
    DATE2 VARCHAR2(10),
    POL_PRTY VARCHAR2(30),
    P_NAME VARCHAR2(30),
    P_ROLE VARCHAR2(5)
    ORGANIZATION EXTERNAL
    (TYPE ORACLE_LOADER
    DEFAULT DIRECTORY P_EXT_TAB_D
    ACCESS PARAMETERS (
    RECORDS DELIMITED by NEWLINE
    SKIP 1
    FIELDS TERMINATED BY "," OPTIONALLY ENCLOSED BY '"' LDRTRIM
    REJECT ROWS WITH ALL NULL FIELDS
    MISSING FIELD VALUES ARE NULL
    DATE1 CHAR (10) Terminated by "," ,
    DATE2 CHAR (10) Terminated by "," ,
    POL_PRTY CHAR (30) Terminated by "," ,
    P_NAME CHAR (30) Terminated by "," OPTIONALLY ENCLOSED BY '"' ,
    P_ROLE CHAR (5) Terminated by ","
    LOCATION ('Input.dat')
    REJECT LIMIT UNLIMITED;
    It created successfully using SQL Developer
    Here is the issue.
    It is not loading the records, where fields are enclosed in '"' (Rec # 2,3,4,7)
    It is loading all NULL value record (Rec # 6)
    *** If we remove the '"' from input data, it loads all records including all NULL records
    Log file has
    KUP-04021: field formatting error for field P_NAME
    KUP-04036: second enclosing delimiter not found
    KUP-04101: record 2 rejected in file ....
    Our questions
    Why did "REJECT ROWS WITH ALL NULL FIELDS" not working?
    Why did Terminated by "," OPTIONALLY ENCLOSED BY '"' not working?
    Any idea?
    Thanks in helping.
    Edited by: qwe16235 on Jun 11, 2011 11:31 AM

    I'm not sure, but maybe you should get rid of the redundancy that you have in your CREATE TABLE statement.
    This line covers all fields:
    FIELDS TERMINATED BY "," OPTIONALLY ENCLOSED BY '"'
    {code}
    So I would change the field list to:
    {code}
    DATE1 CHAR (10),
    DATE2 CHAR (10),
    POL_PRTY CHAR (30),
    P_NAME CHAR (30),
    P_ROLE CHAR (5)
    {code}
    It worked on my installation.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Oracle 11g - External Table Issue?

    Oracle 11g - External Table Issue?
    =====================
    I hope this is the right forum for this issue, if not let me, where to go.
    We are using Oracle 11g (11.2.0.1.0) on (Platform : solaris[tm] oe (64-bit)), Sql Developer 3.0.04
    We are trying to use oracle external table to load text files in .csv format. Here is our data look like.
    ======================
    Date1,date2,Political party,Name, ROLE
    20-Jan-66,22-Nov-69,Democratic,"John ", MMM
    22-Nov-70,20-Jan-71,Democratic,"John Jr.",MMM
    20-Jan-68,9-Aug-70,Republican,"Rick Ford Sr.", MMM
    9-Aug-72,20-Jan-75,Republican,Henry,MMM
    ------ ALL NULL -- record
    20-Jan-80,20-Jan-89,Democratic,"Donald Smith",MMM
    ======================
    Our Expernal table structures is as follows
    CREATE TABLE P_LOAD
    DATE1 VARCHAR2(10),
    DATE2 VARCHAR2(10),
    POL_PRTY VARCHAR2(30),
    P_NAME VARCHAR2(30),
    P_ROLE VARCHAR2(5)
    ORGANIZATION EXTERNAL
    (TYPE ORACLE_LOADER
    DEFAULT DIRECTORY P_EXT_TAB_D
    ACCESS PARAMETERS (
    RECORDS DELIMITED by NEWLINE
    SKIP 1
    FIELDS TERMINATED BY "," OPTIONALLY ENCLOSED BY '"' LDRTRIM
    REJECT ROWS WITH ALL NULL FIELDS
    MISSING FIELD VALUES ARE NULL
    DATE1 CHAR (10) Terminated by "," ,
    DATE2 CHAR (10) Terminated by "," ,
    POL_PRTY CHAR (30) Terminated by "," ,
    P_NAME CHAR (30) Terminated by "," OPTIONALLY ENCLOSED BY '"' ,
    P_ROLE CHAR (5) Terminated by ","
    LOCATION ('Input.dat')
    REJECT LIMIT UNLIMITED;
         It created successfully using SQL Developer
    Here is the issue.
    It is not loading the records, where fields are enclosed in '"' (Rec # 2,3,4,7)
    It is loading all NULL value record (Rec # 6)     
    *** If we remove the '"' from input data, it loads all records including all NULL records
    Log file has
    KUP-04021: field formatting error for field P_NAME
    KUP-04036: second enclosing delimiter not found
    KUP-04101: record 2 rejected in file ....
    Our questions
    Why did "REJECT ROWS WITH ALL NULL FIELDS" not working?
    Why did Terminated by "," OPTIONALLY ENCLOSED BY '"' not working?
    Any idea?
    Thanks in helping.
    Edited by: qwe16235 on Jun 10, 2011 2:16 PM

    The following worked for me:
    drop table p_load;
    CREATE TABLE P_LOAD
    DATE1 VARCHAR2(10),
    DATE2 VARCHAR2(10),
    POL_PRTY VARCHAR2(30),
    P_NAME VARCHAR2(30),
    P_ROLE VARCHAR2(5)
    ORGANIZATION EXTERNAL
    (TYPE ORACLE_LOADER
    DEFAULT DIRECTORY scott_def_dir1
    ACCESS PARAMETERS (
    RECORDS DELIMITED by NEWLINE
    badfile scott_def_dir2:'p_load_%a_%p.bad'
    logfile scott_def_dir2:'p_load_%a_%p.log'
    SKIP 1
    FIELDS TERMINATED BY "," OPTIONALLY ENCLOSED BY '"' LDRTRIM
    MISSING FIELD VALUES ARE NULL
    REJECT ROWS WITH ALL NULL FIELDS
    DATE1 CHAR (10) Terminated by "," ,
    DATE2 CHAR (10) Terminated by "," ,
    POL_PRTY CHAR (30) Terminated by "," ,
    P_NAME CHAR (30) Terminated by "," OPTIONALLY ENCLOSED BY '"' ,
    P_ROLE CHAR (5) Terminated by ","
    LOCATION ('Input.dat')
    REJECT LIMIT UNLIMITED;
    Note that I had to interchange the two lines:
    MISSING FIELD VALUES ARE NULL
    REJECT ROWS WITH ALL NULL FIELDS
    Just to get the access parameters to parse correctly.
    I added two empty lines, one in the middle and one at the end - both were rejected.
    In the log file, you will see the rejectiions:
    $ cat p_load_000_9219.log
    LOG file opened at 07/08/11 19:47:23
    Field Definitions for table P_LOAD
    Record format DELIMITED BY NEWLINE
    Data in file has same endianness as the platform
    Reject rows with all null fields
    Fields in Data Source:
    DATE1 CHAR (10)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    DATE2 CHAR (10)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    POL_PRTY CHAR (30)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    P_NAME CHAR (30)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    P_ROLE CHAR (5)
    Terminated by ","
    Enclosed by """ and """
    Trim whitespace same as SQL Loader
    KUP-04073: record ignored because all referenced fields are null for a record
    KUP-04073: record ignored because all referenced fields are null for a record
    Input Data:
    Date1,date2,Political party,Name, ROLE
    20-Jan-66,22-Nov-69,Democratic,"John ", MMM
    22-Nov-70,20-Jan-71,Democratic,"John Jr.",MMM
    20-Jan-68,9-Aug-70,Republican,"Rick Ford Sr.", MMM
    9-Aug-72,20-Jan-75,Republican,Henry,MMM
    4-Aug-70,20-Jan-75,Independent
    Result:
    SQL> select * from p_load;
    DATE1 DATE2 POL_PRTY P_NAME P_ROL
    20-Jan-66 22-Nov-69 Democratic John MMM
    22-Nov-70 20-Jan-71 Democratic John Jr. MMM
    20-Jan-68 9-Aug-70 Republican Rick Ford Sr. MMM
    9-Aug-72 20-Jan-75 Republican Henry MMM
    Regards,
    - Allen

  • DMIS 2010 SP10 replication is not clearing the logging table

    Hello,
    We have installed the DMIS2010 SP10 on our test environment and for some table the replication is getting extreme slow. Initially I had a case when I started to load a new table and all the logs collected during the initial load was not processed after the load turned into replication. Every new entries to the table were replicated to HANA but the entries which were created during the initial load were still in the logging table.
    After a while some other table got this issue, entries got left in the logging table and no replication done to HANA for those lines.
    In the ltro I see really high latency for these tables where the replication is not ok.
    Did anybody had similar issue?
    OSS is opened, but no reply yet from SAP, so any help would be appreciated.
    Thanks for any advice
    Tamas

    Hi Tobias,
    we need to upgrade the ECC kernel to be able to use DMIS2011 and that is not possible at the moment due to other dependencies. I requested our basis team to review the notes, maybe they missed to install any.
    BR
    Tamas

  • The Connection String for the Query Log table is automatically encrypted.

    When I try to use the Usage Based Optimization to apply Aggregation Design to my measure group, it shows me the following
    error message.
    The connection string cannot be found. Open Microsoft SQL Server Management Studio and, in the Analysis Server Properties
    page, check the value of the Log\QueryLog\QueryLogConnectionString
    property.
    I encountered this error like two weeks ago.  At that time I just reset the connection string and every things seem
    to be fine.  A week ago, I successfully applied the Usage Based Optimization for one of my cubes.  However when I tried to apply UBO for my other cubes today, I encountered the same issue again!  I believe no one has changed the property of
    the connection string.
    Also if I query the Query Log table, I can see those latest queries made by the users.  I'm sure the queries are still
    logging into this table.
    This is really strange.  Anyone else has encountered the same issue?  Thanks.

    Hello Thomas,
    I encounterd this issue. And I am struggling trying to solve this problem. If you have resolved this issue and I guess you must've, because this post is two years old, could you kindly post how you resolved this issue?
    Thanks in advance
    Best Regards,
    Neeraja

  • Materialized View Log table with sequence

    I have a materialized view log in Database A..a very simple one...similar to "create materialized view log on t WITH PRIMARY KEY"...The primary key is a composite key of 2 columns...Description of the mlog table looks like this...
    desc mlog$_t
    emp_ctr NUMBER
    emp_date DATE
    SNAPTIME$$ DATE
    DMLTYPE$$ VARCHAR2(1)
    OLD_NEW$$ VARCHAR2(1)
    CHANGE_VECTOR$$ RAW(255)
    The Materialized view is in database B...again a simple one...no aggregation...This gets refreshed on demand manually evey week. The records that get accumulated in the MLOG every week is ~ 300K..The manual refresh used to finish in about 20mins but since last week it is running for ever. Upon reviewing the contents of MLOG I noticed that it contained a mixture of inserts and updates. I'm not sure if the MLOG contanied only inserts in the past when the MV used to refresh very quickly...Based on the documentation, Oracle recommends creating MLOGS with "sequence" whenever inserts/updates/deleted are expected. I'm planning to drop and recreate the MLOG with sequence with the hope that this would fix the slow refresh issue...Appreciate if you could let me know your thoughs...
    I'm also planning to trace the session to get some wait event statistics to find out the reason for the slow refresh...

    There are a few reasons why refreshes slow down over time.
    1. Is the log table being cleared out after a refresh? If not, you may have another mview (maybe even one that doesn't exist any more) registered against this log. When this happens, the log table never stops growing.
    2. How big is the segment occupied by the table? If it has grown very large at some time in the past (e.g., because refresh was delayed), then when the refresh does a FTS on the log it will scan the whole segment - even if only a few rows are present.
    3. It may help to add an index to the PK columns, and another to SNAPTIME$$. Without them, the refresh will FTS the log. For a big segment, that can take a long time.
    -- Phil

  • Capture PO errors to log table

    Please let me know if there is a way to capture SAP standard errors raised during PO creation into a Z log table.
    EX: SAP raises Message 06, message number 882, is there a way to capture this error into a log table.
    Thanks for your help.

    In PO processing using ME21N/ME22N a lot messages are not issued individually, but are presented in a list in a pop-up window.  Not sure if this is true of the messages you want to capture.
    If you look at the following code implemented in the CHECK method of "ME_PROCESS_PO_CUST" BADI, you will see it adds a customer message to this list using the macro MMPUR_MESSAGE from the MM_MESSAGES_MAC include.
    Perhaps there is also a macro for reading the messages. 
    Even if not, the macro code should show how you could read them and then you could place code in this user exit to get the messages and log the relevent ones to a Z table
    [code]
    METHOD if_ex_me_process_po_cust~check.
    NB: Add Type Group MMMFD to implementing class properties tab
    DATA: lt_items TYPE purchase_order_items,
    l_item_wa TYPE purchase_order_item,
    l_item TYPE mepoitem,
    l_total TYPE brtwr,
    l_matkl TYPE matkl,
    l_flag TYPE c.
    INCLUDE mm_messages_mac. "useful macros for message handling
    CLEAR: l_total, l_flag.
    CALL METHOD im_header->get_items
    RECEIVING
    re_items = lt_items.
    Loop through items and calculate total
    LOOP AT lt_items INTO l_item_wa.
    CALL METHOD l_item_wa-item->get_data
    RECEIVING
    re_data = l_item.
    l_matkl = l_item-matkl.
    l_total = l_total + l_item-brtwr.
    Check if any lines missing contract number
    IF l_item-konnr IS INITIAL.
    l_flag = 'X'.
    ENDIF.
    ENDLOOP.
    IF l_flag = 'X'.
    IF l_total > 5000.
    mmpur_metafield mmmfd_agreement.
    mmpur_message 'E' '06' '999'
    'Purchase order over $5000.00 requires a contract' '' '' ''.
    invalidate the object
    CALL METHOD im_header->invalidate( ).
    LOOP AT lt_items INTO l_item_wa.
    CALL METHOD l_item_wa-item->invalidate( ).
    ENDLOOP.
    ENDIF.
    ENDIF.
    ENDMETHOD.
    [/code]
    Hope this is of some use.
    Andrew

  • IMac 27" Late 2013 Wi Fi Log in Issue

    Hello everyone,
    I apologize if my questions have been asked in the past. I'm completely new to the Apple Mac OS. I bought a new iMac 27" in the middle of 2014 and use it about once or twice a week and have had no issues to-date. I'm still in the process of learning though so my use of it is limited to basic emailing, web browsing and Skype.
    In the last 2 weeks I have had problems logging into my Wifi network at home. It would not log on even after I had reset the modem. After trying and rebooting my iMac 3-4 times it would finally do so without changing anything in the it. I'm not sure what is causing it, but thought I would ask here.
    What would be causing the log in issue to my Wi Fi network?
    Do I need to be concern with the findings in the Etre check?
    Startup after logging in my password would take longer than usual...what would be causing it?
    Here is the result from the Etre scan. I guess what concerns me is in red...and also if it shows the Wi Fi log in issue.
    Thank you in advance for your help and advice.
    EtreCheck version: 2.1.8 (121)
    Report generated 9 February 2015 11:23:18 am AEST
    Download EtreCheck from http://etresoft.com/etrecheck
    Click the [Click for support] links for help with non-Apple products.
    Click the [Click for details] links for more information about that line.
    Hardware Information: ℹ️
        iMac (27-inch, Late 2013) (Technical Specifications)
        iMac - model: iMac14,2
        1 3.4 GHz Intel Core i5 CPU: 4-core
        32 GB RAM Upgradeable
            BANK 0/DIMM0
                8 GB DDR3 1600 MHz ok
            BANK 1/DIMM0
                8 GB DDR3 1600 MHz ok
            BANK 0/DIMM1
                8 GB DDR3 1600 MHz ok
            BANK 1/DIMM1
                8 GB DDR3 1600 MHz ok
        Bluetooth: Good - Handoff/Airdrop2 supported
        Wireless:  en1: 802.11 a/b/g/n/ac
    Video Information: ℹ️
        NVIDIA GeForce GTX 780M - VRAM: 4096 MB
            iMac 2560 x 1440
    System Software: ℹ️
        OS X 10.9.5 (13F34) - Time since boot: 0:24:23
    Disk Information: ℹ️
        APPLE SSD SD0128F disk0 : (121.33 GB)
            EFI (disk0s1) <not mounted> : 210 MB
            Boot OS X (disk0s3) <not mounted> : 134 MB
            Macintosh HD (disk2) / : 1.11 TB (1.04 TB free)
                Encrypted AES-XTS Unlocked
                Core Storage: disk0s2 120.99 GB Online
                Core Storage: disk1s2 999.35 GB Online
        APPLE HDD ST1000DM003 disk1 : (1 TB)
            EFI (disk1s1) <not mounted> : 210 MB
            Recovery HD (disk1s3) <not mounted>  [Recovery]: 650 MB
            Macintosh HD (disk2) / : 1.11 TB (1.04 TB free)
                Encrypted AES-XTS Unlocked
                Core Storage: disk0s2 120.99 GB Online
                Core Storage: disk1s2 999.35 GB Online
    USB Information: ℹ️
        VIA Labs, Inc.          USB3.0 Hub             
            VIA Labs, Inc.          USB3.0 Hub             
                Seagate  Backup+  SL 2 TB
            EFI (disk3s1) <not mounted> : 210 MB
            Seagate Backup Plus Drive (disk3s2) /Volumes/Seagate Backup Plus Drive : 2.00 TB (1.82 TB free)
            Recovery HD (disk3s3) <not mounted>  [Recovery]: 650 MB
                Initio   INIC-3609       
        Apple Inc. FaceTime HD Camera (Built-in)
        Apple Inc. BRCM20702 Hub
            Apple Inc. Bluetooth USB Host Controller
        Logitech Inc. Logitech USB Headset H340
        Apple, Inc. Keyboard Hub
            Apple Inc. Apple Keyboard
        VIA Labs, Inc.          USB2.0 Hub             
            VIA Labs, Inc.          USB2.0 Hub             
    Thunderbolt Information: ℹ️
        Apple Inc. thunderbolt_bus
    Gatekeeper: ℹ️
        Mac App Store and identified developers
    Startup Items: ℹ️
        ProTec6b: Path: /Library/StartupItems/ProTec6b
        Startup items are obsolete in OS X Yosemite
    Launch Agents: ℹ️
        [loaded]    com.google.keystone.agent.plist [Click for support]
    Launch Daemons: ℹ️
        [loaded]    com.adobe.fpsaud.plist [Click for support]
        [loaded]    com.bombich.ccc.plist [Click for support]
        [loaded]    com.google.keystone.daemon.plist [Click for support]
        [loaded]    com.microsoft.office.licensing.helper.plist [Click for support]
    User Launch Agents: ℹ️
        [failed]    com.spotify.webhelper.plist [Click for support]
        [not loaded]    jp.co.canon.Inkjet_Extended_Survey_Agent.plist [Click for support]
    User Login Items: ℹ️
        iTunesHelper    Application  (/Applications/iTunes.app/Contents/MacOS/iTunesHelper.app)
        Alfred 2    Application  (/Applications/Alfred 2.app)
        Spotify 08-26-32-444    Application  (/Users/[redacted]/.Trash/Spotify 08-26-32-444.app)
        Dropbox    Application  (/Applications/Dropbox.app)
        Canon IJ Network Scanner Selector EX    Application Hidden (/Applications/Canon Utilities/IJ Network Scanner Selector EX/Canon IJ Network Scanner Selector EX.app)
        Speech Startup    Application  (/System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/SpeechRecogn ition.framework/Versions/A/Speech Startup.app)
    Internet Plug-ins: ℹ️
        FlashPlayer-10.6: Version: 16.0.0.296 - SDK 10.6 [Click for support]
        QuickTime Plugin: Version: 7.7.3
        Flash Player: Version: 16.0.0.296 - SDK 10.6 Outdated! Update
        EPPEX Plugin: Version: 3.0.5.0 [Click for support]
        Default Browser: Version: 537 - SDK 10.9
        SharePointBrowserPlugin: Version: 14.4.3 - SDK 10.6 [Click for support]
        Silverlight: Version: 5.1.20125.0 - SDK 10.6 [Click for support]
    3rd Party Preference Panes: ℹ️
        Flash Player  [Click for support]
    Time Machine: ℹ️
        Auto backup: YES
        Volumes being backed up:
            Macintosh HD: Disk size: 1.11 TB Disk used: 75.54 GB
        Destinations:
            Seagate Backup Plus Drive [Local]
            Total size: 2.00 TB
            Total number of backups: 33
            Oldest backup: 2014-04-16 03:45:27 +0000
            Last backup: 2015-02-09 01:07:51 +0000
            Size of backup disk: Adequate
                Backup size 2.00 TB > (Disk used 75.54 GB X 3)
    Top Processes by CPU: ℹ️
             3%    WindowServer
             2%    Safari
             2%    Google Chrome
             1%    fontd
             1%    coreaudiod
    Top Processes by Memory: ℹ️
        481 MB    WindowServer
        309 MB    Airmail
        206 MB    mds_stores
        206 MB    Safari
        172 MB    Google Chrome
    Virtual Memory Information: ℹ️
        27.13 GB    Free RAM
        4.39 GB    Active RAM
        1.08 GB    Inactive RAM
        1.75 GB    Wired RAM
        912 MB    Page-ins
        0 B    Page-outs
    Diagnostics Information: ℹ️
        Feb 9, 2015, 10:58:33 AM    Self test - pass

    Hi, Jiyujin.  
    Thank you for visiting Apple Support Communities.
    I would need some clarification of the reference to log on and a little more information about the behavior to provide a better answer.  However, here is a troubleshooting article that may help.  
    Wi-Fi: How to troubleshoot Wi-Fi connectivity
    http://support.apple.com/HT202222
    Cheers,
    Jason H.

  • Creating a new logging table in source

    Hi ,
    I have two relication setup , one is ECC to Ent hana and the other Ecc to BW on hana.
    For ECC to BW on hana , i want to create a seperate logging table in source  to diffrenciate from the two replication.
    Could you please let me know how to do this.
    Thanks,
    Rajiv

    Hi,
    this is unfortunately the wrong community, pls. have a look here: SAP LT Replication Server
    There you can post your question.
    Best,
    Heike

  • Deleting data from a very large log table (custom table in our namespace)

    Hello,
    I have been tasked with clearing a log table in our landscape to only include the most recent entries.  Is it possible to do this given that the table has already got 230 000 000 entries and will need to keep around 600 000 recent entries?
    Should I do this via ABAP and if so, how?  Thanks,
    Samir

    Hi,
    so you are going to keep 0,3 % of your data?
    If you should do it in ABAP or on the database is your decission.
    In my opinion doing things on the database directly should be done
    exceptional cases only e.g. for one time actions or actions that have to
    be done very rarely and with different parameters / options. Regular
    and similar tasks should be done in ABAP i think.
    In any case i would not delete the majority of the records but copy
    the records to keep in an empty table with the same structure, delete the
    table as a whole (check clients!) and "copy" the new table back in ABAP
    or rename the new table to the old table after droping the old table on the database.
    If you have only one client you can copy the data you need in a new
    table and truncate the old table (fast deletion for all clients). If you have
    data to keep  for other clients as well check how much data it is per client
    in comparison to the total number of lines (if only a small fraction, pefer copying
    them too).
    On the database you can use CTAS (create table as select) and drop table and
    rename table. Those commands shoudl be very efficient but work client independently.
    If you have to consider clients SELECT; INSERT; DELETE or TRUNCATE (depends on if you have
    copied all data considering clients) are
    your friends.
    Kind regards,
    Hermann

  • Error log table and output type

    Hi All,
    While creating an invoice(while saving billing document) is it poosible to capture the error in a log in table?
    And how is this error log table related to the output type?
    Thanks

    You seem to be confusing saving of the billing document with creation of the output. These are two independent updates (LUWs). A document may be saved without any output.
    If there are any errors in the billing document creation, the document just won't be created. Such errors should be captured in the log of the Blling Due list, if it's been run properly.
    To update the processing log for the output, use FM NAST_PROTOCOL_UPDATE.

Maybe you are looking for

  • CP5 Checking to see if a slide has been viewed between projects

    Honestly, I'm not the one thinking of some of my questions...... Here is the scenario.  We have a course made up of lessons that have been aggregated together.  This course has an Intro, X number of lessons and a Completion.  In the Intro there is a

  • Seeing iTunes library from MacPro with ethernet only

    I upgraded from an Apple TV 1st Generation so I could use Netflix.  I've managed to successfully set up home sharing with my MacBook Pro, but it is not seeing the content on my MacPro. This does not have a wi-fi card, but is connected to a wi-fi rout

  • Distorted sound on Apple TV 3

    Hi, I just bought an Apple TV (gen. 3). But I seem to have some problems with the sound. It is connected to my TV through HDMI and I have a speaker set connected directly to my TV. The problem is that the sound is distorted and the bass is all wrong.

  • QuickTime problem and question

    Hi I have been a mac convert for a few weeks (mac mini) and I have been experiencing a very bizarre situation with QuickTime 7.03 and 7.04. My Windows box is used as a PVR. I record some TV shows and movies on it and whenever I try to watch them thro

  • Table maintenance in ECC6

    HI All, I have created a table in ECC 6, now after creating table maintenance generator when i try to maintain data this error comes when 'View/table ZHRPMT_REQPRO can only be displayed and maintained with restrictions'  I don't know why this error i