Report RTCCTOOL, Schedule RSCOLL00 as an hourly

hello  I have a problem When I run RTCCTOOL in satelite system I got  this  red light
1. Note 12103 Collectors and TCOLL
Planning table TCOLL for SAP workload collectors for 700 (note 966309 )
Press the lights icon to get the detailed check log. Add missing entries to table TCOLL using ST03N > Expert Mode > Collectorv and Performance DB > Performance Monitor Collector > Execution Times. Schedule RSCOLL00 as an hourly batch job if not scheduled.
So I cheked the  red light ->  Only 2 jobs found during last 200 periods.
i did not find the solution in note 966309
what i have to do? Please
Regards,
Danny

Hi,
Click the traffic singal on the left most side. This will give u clear picture of the issue. Today I've faced the same problem and the issue actually is to correct the entry in the table TCOLL.
If possible post the list in the resulting screen(ater double clicking the traffic signal)
NB: Mark the thread as anwered if ur issue is resolved.
Regards,
Varadharajan M

Similar Messages

  • Dataset can not be open error while running report rtcctool

    Hello Experts,
    While running report rtcctool we are gettin following dumps the says
    ABAP runtime errors    DATASET_CANT_OPEN
           Occurred on     05/27/2011 at 10:00:52
    Unable to open file "/usr/sap/B**/DVEBMGS00/sec/SAP_AGS_OLCNT_VERIFY.pse".
    I am aware that the eror is happening because of this file but not sure from where to get this file. Please let me know any of you aware about this.
    Regards
    Yogesh

    Hi all.
    I have the same problem and explore all possible solutions found in this discussion and others without satisfactory result.
    Everything points that may be an inconsistency error external certificate which is in the DB. Anyone know how to fix this.
    Greetings.

  • Reports-Job-Scheduler: create job with startoption later: wrong date-format

    Hi,
    I test reports 11.1.1.2.0 (weblogic11 g) and try to start a report job with Reports-Job-Scheduler.
    Startoption: later and date = 09:48 Jan 12, 2010
    When I submit the job I get following Error:
    REP-50006: Das Argument SCHEDULE ist in der Befehlszeile ungültig: 09:48 Jän 12, 2010
    in english:
    Rep-50006: argument Schedule is wrong in command line: 09:48 Jän 12, 2010
    Why is Jän instead of Jan.
    When I start the report in February then it works.
    I defined also environment variables in basic configuration for the report server:
    NLS_DATE_FORMAT DD.MM.YYYY
    NLS_DATE_LANGUAGE GERMAN
    NLS_LANG GERMAN_GERMANY.WE8ISO8859P1
    I would be very pleased to get support.
    Many thanks
    Alois

    I created a SR and oracle support found a bug in EM.

  • Can you create report and schedule in EVDRE

    Hi,
    I was asked by someone that if it is possible to create report and schedule in one single worksheet?  As you set report/schedule in workbook options for the file my assumption was its not possible. Is there any work around?
    Also i was asked about mandate in evsnd. I havent come across that either. what is it?

    Hi,
    Yes! You can create report and schedule in single work sheet. If you insert 2 different evdre's , one for report and other for schedule then use [Get only range|http://help.sap.com/saphelp_bpc75_nw/helpdata/en/47/b9a7ccae664edb9034c3579ebec486/content.htm] for report area.
    Or using a single evdre also you can acheive this by selecting part of range not to send.
    hope it works...
    regards,
    Raju

  • Report is executing more than one hour

    Hi All,
    I have an issue with my report,its taking more than one hour for execution.
    This report is called through an package,in package there are cursors which retrive data and sends to mail.The problem is with this select statement in cursor,its taking a lot of time to retrive the data.
    In my cursor select statement i am using the following tables
    HZ_PARTY_SITES HPS,
    HZ_CUST_ACCOUNTS CUST,
    HZ_CUST_ACCT_SITES_ALL HCASA,
    HZ_CUST_SITE_USES_ALL HCSUA,
    HZ_LOCATIONS HL,
    OE_ORDER_HEADERS_ALL OOH,
    OE_TRANSACTION_TYPES_TL OT,
    OE_AUDIT_ATTR_DESC_V AADV
    when i use OE_AUDIT_ATTR_DESC_V ,its taking a lot of time to retrive.So any one can help me out in this issue,how to add this view to other tables
    Regards,
    RR

    RR,
    Before you add any indexes to the listed tables, check with Oracle support to ensure you will be violating your Oracle Enterprise Business Suite agreement. Oracle has specific rules on modifying EMS tables. Additionally, any EBS patch or update that gets applied will likely drop any custom changes to these tables. It might be better for you to open a Service Request with Oracle Support and have them tell you how to improve the performance of your query since all of your tables are EBS tables.
    Craig...
    Message was edited by:
    CraigB
    RR, you may also want to check with your DBA(s) to see when the Statistics were last gathered on these tables.

  • Max Instance of One DESKI Report for Schedule

    Hi,
    Can someone pls let me know how many Max Instance we can create for One DESKI Report for Scheduling.
    Thanks in advance.
    Prabhat

    Jason,
    Thanks for the info. The query builder o/p and the application o/p are the same. When this situation happened last time I just simply deleted those older instances from the history and left only the current one.
    This has forced the query to return the only available instance, which is the current one.
    My suspect here is that, the SI_ID of a new instance is smaller than the older instance...
    for example,
    Instance 1 - 10023
    Instance 2 - 12344
    Instance 3 - 9892
    Is this possible?? Since I have deleted the older instances I have no way to confirm this case unless it happens again.
    The BO documentation about SI_ID reads like this...
    "This property is used to uniquely identify an object. The number is assigned at creation time, and is never changed until the object is destroyed. You should never hard-code this number, however, since if the object is destroyed the number may be assigned to another object. No two objects have the same ID, even if they are different types of objects; that is, you will never find a report that has the same ID as a folder."
    As it claims that an ID can be re-used if the object is destoryed I assume that though the new ID will be unique there is no guarantee that it will be bigger than the previous instance of the same object.
    But the problem is that I dont have a documentation proof to claim this hence asking help from the guru's here...
    Regards,
    Elan.

  • Assymetric report(input schedule) in SAP BPC

    What I mean!?
    I want  place in row expansion two dimension. Connection between members of these dimensions i described on example 1.
    I would like to build report(input schedule) like in example in analytical systems it call assymetic report. By means standart functional I can achieve only example2.
    Any suggestions will be appreciable.
    Regards
    Example 1                            
    Division1 
       ---         Cost1 100
      ---           Cost2 200
    Division2  
    ---           Cost3 200
    ---            Cost4 300
    ---            Cost5 100
    Division3  Cost6 500
    Example 2
    Division1  
    ---Cost1  100
        ---        Cost2   200
        ---        Cost3
        ---        Cost4
        ---        Cost5
        ---        Cost6
    Division2   Cost1
         ---            Cost2
         ---            Cost3  200
         ---            Cost4  300
         ---            Cost5 100
         ---            Cost6
    Division3  Cost1
         ---            Cost2
         ---            Cost3
         ---            Cost4
         ---            Cost5
         ---            Cost6  500

    Vijay,
       you  can  achieve  this with a  single evdre saved  as  an  Input  schedule...
    couple of pointers
    1)if you  do not  want   to  have the  col  names  changing  with  a  change  in current  view......put a  NOEXPAND   for  the  TIme  dimension  in the  expansion  range(where  you  have  SELF,DEP etc)
    2)for  the report  part....use  getonlyrange    to  have it  pulled  from the  DB  instead  of  user  entering data...
    go to helpsite   for the  exact  usage  of NOEXPAND  and GetOnlyRange..play  around enough...i'm sure  you'll figure  it  out!!!!
    hope  it  helps....

  • SSRS report email Scheduling omit zero record

    Is that a way to omit zero record for SSRS report email scheduling?
    That means if zero record happened it shouldn't trigger emailing.

    Hi SAMSUNG TECH,
    Per my understanding that you don't want to send the scheduled email subscription if the report have no data, right?
    I have tested on my local enviroment that you can modify the commend in the job step include some condition to send the email or not.
    Detials information in below for you reference:
    Open SQL Server Management Studio (SSMS) to connect to Database Engine ->expand SQL Server Agent->expand Jobs, double-click the Job.
    you can use query below to find the Subscription you have just created and the “ScheduleID” in the table is equal to the Jobs Name.
    use ReportServer;
    SELECT USR.UserName AS SubscriptionOwner
    ,SCH.ScheduleID
    ,SCH.NextRunTime
    ,SCH.LastRunTime
    ,SCH.EndDate
    ,SCH.EventData
    ,SUB.EventType
    ,SUB.SubscriptionID
    ,SUB.ModifiedDate
    ,SUB.[Description]
    ,SUB.DeliveryExtension
    ,SCH.Name AS ScheduleName
    ,CAT.[Path] AS ReportPath
    ,CAT.[Description] AS ReportDescription
    FROM dbo.Subscriptions AS SUB
    INNER JOIN dbo.Users AS USR
    ON SUB.OwnerID = USR.UserID
    INNER JOIN dbo.[Catalog] AS CAT
    ON SUB.Report_OID = CAT.ItemID
    INNER JOIN dbo.ReportSchedule AS RS
    ON SUB.Report_OID = RS.ReportID
    AND SUB.SubscriptionID = RS.SubscriptionID
    INNER JOIN dbo.Schedule AS SCH
    ON RS.ScheduleID = SCH.ScheduleID
    ORDER BY USR.UserName
    ,CAT.[Path]
    2.  Select "Steps" on the left pane on the Job Properties window.
    3.  Click Edit button in the bottom of the window and then change the command
    Existing command is:
    EXEC [ReportServer].dbo.AddEvent @EventType='TimedSubscription', @EventData='9be28f07-3784-4070-802a-b7ca0aec4c7c'
    Change the command as below: (Remember to check the Database Name and table name):
    If exists
    (Select top 1 * from [DBName].TableName )
    EXEC   [ReportServer].dbo.AddEvent @EventType='TimedSubscription',   @EventData='9be28f07-3784-4070-802a-b7ca0aec4c7c'
    BEGIN
    END
    Note: Change the @EventType and @EventData to the values in your job,the value of this two fields can be found by execute the query in step1.
    More details information in below articles:
    http://www.jasonyousef.com/2012/02/dont-send-that-empty-ssrs-report.html
    http://blogs.msdn.com/b/bimusings/archive/2005/07/29/445080.aspx
    If you still have any question, please feel free to ask.
    Regards
    Vicky Liu

  • Schedule setups outside working hours

    Dear Experts,
    We are implementing PPDS in a process industry, setup matrixs are going to be used and in order to create a  setup constraint based plan for the next 3 weeks, we use the optimizer. During the production, many of the setup changes are really long, about 5 hours and customer would like to run those setups outside the working hours, which are Monday to Friday from 06:00 to 17:00.
    Manual adjustment of the operations is not a nice solution because almost all the setups are that long so we should manually re-schedule many operations. We cannot have an strategy with the option consider non working time because then we risk to have production, not only setups, outside the working hours.
    Please advice.
    Thanks
    CTF

    Hi,
    I am planning to reduce the setup times that they are actually performing outside working time to 10 min so we dont need to schedule them outside working hours.
    The problem is that they never produce those materials (which require long setups) during a working day if they have produced others product before, so they skip the huge setup time. How can we set up the system in a way that it actually plans to produce those articles at the  beginning of a day and produce them the entire day.
    Thanks
    Regards
    CTF

  • Reports Queue Scheduler

    Hi ,
    I am trying to use the reports queue scheduler to schedule reports , the scheduling parts works fine, but when I try to write the output to a network drive, it is giving me error saying " Error with I/O operation, unable to open the file"
    Any help is appreciated.
    ~hari.

    Are you running on IAS or locally.
    If IAS I don't think Oracle can write to a network drive - what we do is to write to a Unix directory that Oracle can see and then samba share that out so that Windows explorer can access it

  • If the reporting is scheduled, can the schedule be managed via Autosys

    If the reporting is scheduled, can the schedule be managed via Autosys?  Please describe.

    Yes,
    For ex: You have Information Broadcasting scheduled via process chain. Now you can have process chain trigger set by AutoSys. Using Autosys you can schedule SAP jobs across applications and platforms.

  • Back ground job - report, is scheduled 49 hrs ago, still ACTIVE??

    Hi Experts,
    Pls. clarify one of my simple doubt that, Weather a report(which has only one SELECT statement, with key in WHERE clause) wuld take 49 hours to execute? bcoz, I scheduled a back ground job for my_report, 49 hours ago, its still ACTIVE, OR does it went to infinitive loop?
    thanq

    Hi
    no one select query won't take that mucj of time
    may be that program went in infinite loop
    please check with that
    reward if useful
    <b>below is the best way to write a select query</b>
    Ways of Performance Tuning
    1.     Selection Criteria
    2.     Select Statements
    •     Select Queries
    •     SQL Interface
    •     Aggregate Functions
    •     For all Entries
    Select Over more than one Internal table
    Selection Criteria
    1.     Restrict the data to the selection criteria itself, rather than filtering it out using the ABAP code using CHECK statement. 
    2.     Select with selection list.
    Points # 1/2
    SELECT * FROM SBOOK INTO SBOOK_WA.
      CHECK: SBOOK_WA-CARRID = 'LH' AND
             SBOOK_WA-CONNID = '0400'.
    ENDSELECT.
    The above code can be much more optimized by the code written below which avoids CHECK, selects with selection list
    SELECT  CARRID CONNID FLDATE BOOKID FROM SBOOK INTO TABLE T_SBOOK
      WHERE SBOOK_WA-CARRID = 'LH' AND
                  SBOOK_WA-CONNID = '0400'.
    Select Statements   Select Queries
    1.     Avoid nested selects
    2.     Select all the records in a single shot using into table clause of select statement rather than to use Append statements.
    3.     When a base table has multiple indices, the where clause should be in the order of the index, either a primary or a secondary index.
    4.     For testing existence , use Select.. Up to 1 rows statement instead of a Select-Endselect-loop with an Exit. 
    5.     Use Select Single if all primary key fields are supplied in the Where condition .
    Point # 1
    SELECT * FROM EKKO INTO EKKO_WA.
      SELECT * FROM EKAN INTO EKAN_WA
          WHERE EBELN = EKKO_WA-EBELN.
      ENDSELECT.
    ENDSELECT.
    The above code can be much more optimized by the code written below.
    SELECT PF1 PF2 FF3 FF4 INTO TABLE ITAB
        FROM EKKO AS P INNER JOIN EKAN AS F
          ON PEBELN = FEBELN.
    Note: A simple SELECT loop is a single database access whose result is passed to the ABAP program line by line. Nested SELECT loops mean that the number of accesses in the inner loop is multiplied by the number of accesses in the outer loop. One should therefore use nested SELECT loops  only if the selection in the outer loop contains very few lines or the outer loop is a SELECT SINGLE statement.
    Point # 2
    SELECT * FROM SBOOK INTO SBOOK_WA.
      CHECK: SBOOK_WA-CARRID = 'LH' AND
             SBOOK_WA-CONNID = '0400'.
    ENDSELECT.
    The above code can be much more optimized by the code written below which avoids CHECK, selects with selection list and puts the data in one shot using into table
    SELECT  CARRID CONNID FLDATE BOOKID FROM SBOOK INTO TABLE T_SBOOK
      WHERE SBOOK_WA-CARRID = 'LH' AND
                  SBOOK_WA-CONNID = '0400'.
    Point # 3
    To choose an index, the optimizer checks the field names specified in the where clause and then uses an index that has the same order of the fields . In certain scenarios, it is advisable to check whether a new index can speed up the performance of a program. This will come handy in programs that access data from the finance tables.
    Point # 4
    SELECT * FROM SBOOK INTO SBOOK_WA
      UP TO 1 ROWS
      WHERE CARRID = 'LH'.
    ENDSELECT.
    The above code is more optimized as compared to the code mentioned below for testing existence of a record.
    SELECT * FROM SBOOK INTO SBOOK_WA
        WHERE CARRID = 'LH'.
      EXIT.
    ENDSELECT.
    Point # 5
    If all primary key fields are supplied in the Where condition you can even use Select Single.
    Select Single requires one communication with the database system, whereas Select-Endselect needs two.
    Select Statements           contd..  SQL Interface
    1.     Use column updates instead of single-row updates
    to update your database tables.
    2.     For all frequently used Select statements, try to use an index.
    3.     Using buffered tables improves the performance considerably.
    Point # 1
    SELECT * FROM SFLIGHT INTO SFLIGHT_WA.
      SFLIGHT_WA-SEATSOCC =
        SFLIGHT_WA-SEATSOCC - 1.
      UPDATE SFLIGHT FROM SFLIGHT_WA.
    ENDSELECT.
    The above mentioned code can be more optimized by using the following code
    UPDATE SFLIGHT
           SET SEATSOCC = SEATSOCC - 1.
    Point # 2
    SELECT * FROM SBOOK CLIENT SPECIFIED INTO SBOOK_WA
      WHERE CARRID = 'LH'
        AND CONNID = '0400'.
    ENDSELECT.
    The above mentioned code can be more optimized by using the following code
    SELECT * FROM SBOOK CLIENT SPECIFIED INTO SBOOK_WA
      WHERE MANDT IN ( SELECT MANDT FROM T000 )
        AND CARRID = 'LH'
        AND CONNID = '0400'.
    ENDSELECT.
    Point # 3
    Bypassing the buffer increases the network considerably
    SELECT SINGLE * FROM T100 INTO T100_WA
      BYPASSING BUFFER
      WHERE     SPRSL = 'D'
            AND ARBGB = '00'
            AND MSGNR = '999'.
    The above mentioned code can be more optimized by using the following code
    SELECT SINGLE * FROM T100  INTO T100_WA
      WHERE     SPRSL = 'D'
            AND ARBGB = '00'
            AND MSGNR = '999'.
    Select Statements       contd…           Aggregate Functions
    •     If you want to find the maximum, minimum, sum and average value or the count of a database column, use a select list with aggregate functions instead of computing the aggregates yourself.
    Some of the Aggregate functions allowed in SAP are  MAX, MIN, AVG, SUM, COUNT, COUNT( * )
    Consider the following extract.
                Maxno = 0.
                Select * from zflight where airln = ‘LF’ and cntry = ‘IN’.
                 Check zflight-fligh > maxno.
                 Maxno = zflight-fligh.
                Endselect.
    The  above mentioned code can be much more optimized by using the following code.
    Select max( fligh ) from zflight into maxno where airln = ‘LF’ and cntry = ‘IN’.
    Select Statements    contd…For All Entries
    •     The for all entries creates a where clause, where all the entries in the driver table are combined with OR. If the number of entries in the driver table is larger than rsdb/max_blocking_factor, several similar SQL statements are executed to limit the length of the WHERE clause.
         The plus
    •     Large amount of data
    •     Mixing processing and reading of data
    •     Fast internal reprocessing of data
    •     Fast
         The Minus
    •     Difficult to program/understand
    •     Memory could be critical (use FREE or PACKAGE size)
    Points to be must considered FOR ALL ENTRIES
    •     Check that data is present in the driver table
    •     Sorting the driver table
    •     Removing duplicates from the driver table
    Consider the following piece of extract
    Loop at int_cntry.
           Select single * from zfligh into int_fligh
    where cntry = int_cntry-cntry.
    Append int_fligh.
    Endloop.
    The above mentioned can be more optimized by using the following code.
    Sort int_cntry by cntry.
    Delete adjacent duplicates from int_cntry.
    If NOT int_cntry[] is INITIAL.
                Select * from zfligh appending table int_fligh
                For all entries in int_cntry
                Where cntry = int_cntry-cntry.
    Endif.
    Select Statements    contd…  Select Over more than one Internal table
    1.     Its better to use a views instead of nested Select statements.
    2.     To read data from several logically connected tables use a join instead of nested Select statements. Joins are preferred only if all the primary key are available in WHERE clause for the tables that are joined. If the primary keys are not provided in join the Joining of tables itself takes time.
    3.     Instead of using nested Select loops it is often better to use subqueries.
    Point # 1
    SELECT * FROM DD01L INTO DD01L_WA
      WHERE DOMNAME LIKE 'CHAR%'
            AND AS4LOCAL = 'A'.
      SELECT SINGLE * FROM DD01T INTO DD01T_WA
        WHERE   DOMNAME    = DD01L_WA-DOMNAME
            AND AS4LOCAL   = 'A'
            AND AS4VERS    = DD01L_WA-AS4VERS
            AND DDLANGUAGE = SY-LANGU.
    ENDSELECT.
    The above code can be more optimized by extracting all the data from view DD01V_WA
    SELECT * FROM DD01V INTO  DD01V_WA
      WHERE DOMNAME LIKE 'CHAR%'
            AND DDLANGUAGE = SY-LANGU.
    ENDSELECT
    Point # 2
    SELECT * FROM EKKO INTO EKKO_WA.
      SELECT * FROM EKAN INTO EKAN_WA
          WHERE EBELN = EKKO_WA-EBELN.
      ENDSELECT.
    ENDSELECT.
    The above code can be much more optimized by the code written below.
    SELECT PF1 PF2 FF3 FF4 INTO TABLE ITAB
        FROM EKKO AS P INNER JOIN EKAN AS F
          ON PEBELN = FEBELN.
    Point # 3
    SELECT * FROM SPFLI
      INTO TABLE T_SPFLI
      WHERE CITYFROM = 'FRANKFURT'
        AND CITYTO = 'NEW YORK'.
    SELECT * FROM SFLIGHT AS F
        INTO SFLIGHT_WA
        FOR ALL ENTRIES IN T_SPFLI
        WHERE SEATSOCC < F~SEATSMAX
          AND CARRID = T_SPFLI-CARRID
          AND CONNID = T_SPFLI-CONNID
          AND FLDATE BETWEEN '19990101' AND '19990331'.
    ENDSELECT.
    The above mentioned code can be even more optimized by using subqueries instead of for all entries.
    SELECT * FROM SFLIGHT AS F INTO SFLIGHT_WA
        WHERE SEATSOCC < F~SEATSMAX
          AND EXISTS ( SELECT * FROM SPFLI
                         WHERE CARRID = F~CARRID
                           AND CONNID = F~CONNID
                           AND CITYFROM = 'FRANKFURT'
                           AND CITYTO = 'NEW YORK' )
          AND FLDATE BETWEEN '19990101' AND '19990331'.
    ENDSELECT.
    1.     Table operations should be done using explicit work areas rather than via header lines.
    2.     Always try to use binary search instead of linear search. But don’t forget to sort your internal table before that.
    3.     A dynamic key access is slower than a static one, since the key specification must be evaluated at runtime.
    4.     A binary search using secondary index takes considerably less time.
    5.     LOOP ... WHERE is faster than LOOP/CHECK because LOOP ... WHERE evaluates the specified condition internally.
    6.     Modifying selected components using “ MODIFY itab …TRANSPORTING f1 f2.. “ accelerates the task of updating  a line of an internal table.
    Point # 2
    READ TABLE ITAB INTO WA WITH KEY K = 'X‘ BINARY SEARCH.
    IS MUCH FASTER THAN USING
    READ TABLE ITAB INTO WA WITH KEY K = 'X'.
    If TAB has n entries, linear search runs in O( n ) time, whereas binary search takes only O( log2( n ) ).
    Point # 3
    READ TABLE ITAB INTO WA WITH KEY K = 'X'. IS FASTER THAN USING
    READ TABLE ITAB INTO WA WITH KEY (NAME) = 'X'.
    Point # 5
    LOOP AT ITAB INTO WA WHERE K = 'X'.
    ENDLOOP.
    The above code is much faster than using
    LOOP AT ITAB INTO WA.
      CHECK WA-K = 'X'.
    ENDLOOP.
    Point # 6
    WA-DATE = SY-DATUM.
    MODIFY ITAB FROM WA INDEX 1 TRANSPORTING DATE.
    The above code is more optimized as compared to
    WA-DATE = SY-DATUM.
    MODIFY ITAB FROM WA INDEX 1.
    7.     Accessing the table entries directly in a "LOOP ... ASSIGNING ..." accelerates the task of updating a set of lines of an internal table considerably
    8.    If collect semantics is required, it is always better to use to COLLECT rather than READ BINARY and then ADD.
    9.    "APPEND LINES OF itab1 TO itab2" accelerates the task of appending a table to another table considerably as compared to “ LOOP-APPEND-ENDLOOP.”
    10.   “DELETE ADJACENT DUPLICATES“ accelerates the task of deleting duplicate entries considerably as compared to “ READ-LOOP-DELETE-ENDLOOP”.
    11.   "DELETE itab FROM ... TO ..." accelerates the task of deleting a sequence of lines considerably as compared to “  DO -DELETE-ENDDO”.
    Point # 7
    Modifying selected components only makes the program faster as compared to Modifying all lines completely.
    e.g,
    LOOP AT ITAB ASSIGNING <WA>.
      I = SY-TABIX MOD 2.
      IF I = 0.
        <WA>-FLAG = 'X'.
      ENDIF.
    ENDLOOP.
    The above code works faster as compared to
    LOOP AT ITAB INTO WA.
      I = SY-TABIX MOD 2.
      IF I = 0.
        WA-FLAG = 'X'.
        MODIFY ITAB FROM WA.
      ENDIF.
    ENDLOOP.
    Point # 8
    LOOP AT ITAB1 INTO WA1.
      READ TABLE ITAB2 INTO WA2 WITH KEY K = WA1-K BINARY SEARCH.
      IF SY-SUBRC = 0.
        ADD: WA1-VAL1 TO WA2-VAL1,
             WA1-VAL2 TO WA2-VAL2.
        MODIFY ITAB2 FROM WA2 INDEX SY-TABIX TRANSPORTING VAL1 VAL2.
      ELSE.
        INSERT WA1 INTO ITAB2 INDEX SY-TABIX.
      ENDIF.
    ENDLOOP.
    The above code uses BINARY SEARCH for collect semantics. READ BINARY runs in O( log2(n) ) time. The above piece of code can be more optimized by
    LOOP AT ITAB1 INTO WA.
      COLLECT WA INTO ITAB2.
    ENDLOOP.
    SORT ITAB2 BY K.
    COLLECT, however, uses a hash algorithm and is therefore independent
    of the number of entries (i.e. O(1)) .
    Point # 9
    APPEND LINES OF ITAB1 TO ITAB2.
    This is more optimized as compared to
    LOOP AT ITAB1 INTO WA.
      APPEND WA TO ITAB2.
    ENDLOOP.
    Point # 10
    DELETE ADJACENT DUPLICATES FROM ITAB COMPARING K.
    This is much more optimized as compared to
    READ TABLE ITAB INDEX 1 INTO PREV_LINE.
    LOOP AT ITAB FROM 2 INTO WA.
      IF WA = PREV_LINE.
        DELETE ITAB.
      ELSE.
        PREV_LINE = WA.
      ENDIF.
    ENDLOOP.
    Point # 11
    DELETE ITAB FROM 450 TO 550.
    This is much more optimized as compared to
    DO 101 TIMES.
      DELETE ITAB INDEX 450.
    ENDDO.
    12.   Copying internal tables by using “ITAB2[ ] = ITAB1[ ]” as compared to “LOOP-APPEND-ENDLOOP”.
    13.   Specify the sort key as restrictively as possible to run the program faster.
    Point # 12
    ITAB2[] = ITAB1[].
    This is much more optimized as compared to
    REFRESH ITAB2.
    LOOP AT ITAB1 INTO WA.
      APPEND WA TO ITAB2.
    ENDLOOP.
    Point # 13
    “SORT ITAB BY K.” makes the program runs faster as compared to “SORT ITAB.”
    Internal Tables         contd…
    Hashed and Sorted tables
    1.     For single read access hashed tables are more optimized as compared to sorted tables.
    2.      For partial sequential access sorted tables are more optimized as compared to hashed tables
    Hashed And Sorted Tables
    Point # 1
    Consider the following example where HTAB is a hashed table and STAB is a sorted table
    DO 250 TIMES.
      N = 4 * SY-INDEX.
      READ TABLE HTAB INTO WA WITH TABLE KEY K = N.
      IF SY-SUBRC = 0.
      ENDIF.
    ENDDO.
    This runs faster for single read access as compared to the following same code for sorted table
    DO 250 TIMES.
      N = 4 * SY-INDEX.
      READ TABLE STAB INTO WA WITH TABLE KEY K = N.
      IF SY-SUBRC = 0.
      ENDIF.
    ENDDO.
    Point # 2
    Similarly for Partial Sequential access the STAB runs faster as compared to HTAB
    LOOP AT STAB INTO WA WHERE K = SUBKEY.
    ENDLOOP.
    This runs faster as compared to
    LOOP AT HTAB INTO WA WHERE K = SUBKEY.
    ENDLOOP.

  • Setting Up Earned Value Reporting in Project Server 2010, Using Hours Not Dollars

    Hello All,
    Are there instructions for setting up Project Server 2010 for Earned Value data capture and reporting?  I'm not certain that we are capturing the correct data to report EVM (i.e., Physical % Complete), and I was unable to find any videos on YouTube.
    Also, I noticed that Project Server 2010 defaults to reporting EV in Dollars.  Can we set it up to report in Hours instead?
    Julie  

    Hi Julie,
    If you allow to jump in to complement Dale's excellent reply and comment some of your concern...
    5. The status date is a fundamental parameter while talking about the EVM (it is ANYWAY a key setting for tracking projects with MS Project). The EV field will take into account actuals up to this date. Meaning that if your project plans have been updated
    by the PM on tuesday and you are running the analysis on friday, you set the status date to tuesday and the uncomplete work between tuesday and friday will not be considered as late work since you are running the analysis as of tuesday. The status date can
    be set from the "project" tab either in the project information dialog box or directly in the project tab, status group.
    6. MS Project proposes 3 EVM tables by default. A table is a predefined set of column which composes a view, together with a Gantt chart or a timephased grid. If you go to the view tab, click on "tables" then "more tables",
    you will find EV, EV cost indicators and EV Schedule indicators tables which will display the out of the box EVM fields calculated by MS Project based on the costs, actuals, Baseline, status date.
    Then about the cumulative data, you might be refering to S-curves which are a representation allowing you showing the cumulative values of the EV. To do that, you'll have severals ways:
    Use the cumulative calculated field in the visuals reports
    Create
    Dashboards from the report tab using the cumulative data (actual, Baseline, cost)
    Create
    MDX mesures in the OLAP cubes for cumulative data and create excel reports
    Develop custom SSRS reports
    Buy a 3rd-party addon
    Finally please let me give you an excellent reference for the EVM in MS Project with EPM Central blog articles:
    http://www.epmcentral.com/msproject/evintro.php
    http://www.epmcentral.com/msproject/evprereq.php
    http://www.epmcentral.com/msproject/evterms.php
    http://www.epmcentral.com/msproject/evexplain.php
    Hope this helps,
    Guillaume Rouyre, MBA, MVP, P-Seller |

  • LMS 3.2 - RME 4.3.1 - Report Generator scheduled Syslog-Reports

    Hi all,
    within Report Generator in RME 4.3.1 I discovered an issue with newly added devices in scheduled Syslog Severtiy Level Summary Reports.
    Settings for the report
    - Device Selector: All Devices
    - Data Range: 24 hours
    - Schedule: daily, at xx:xxh
    Our purpose for this job is to get a daily summary of syslogs of all managed devices.
    But there seems to be a problem with the dynamic update of newly added devices to DCR and RME-Device-DB.
    The report only creates output for all managed device at timestamp of report creation, not for newly added devices.
    Is this a know bug, or is this the planned behavior?
    Any suggested workarounds/fixs are appreciated.
    Thanks a lot for your help.
    Lothar

    This is documented in the online help for the report generator, sync archive, etc. pages.
    # Group Selector — To schedule a job for dynamic group of devices.
    The job is scheduled only for the devices that are present in the selected group at the time when the job is run. The customizable group selector for jobs evaluates static groups also as dynamic during run time.

  • Refresh portlet report more frequently than once per hour

    When you publish a report from discoverer in the portal, you can tell it to automatically refresh, but the most frequent refresh it allows is once per hour. Is there some way I can refresh a report more frequently than that? Can I use the scheduled workbook functionality to achieve this, since in there you can schedule it to run as frequently as once per minute.
    I'm using Discoverer 10.1.2.
    Thanks
    -Nissim

    These are two different schedulers...
    We are considering changing the UI to let the users specify a schedule frequency more frequent than once an hour. But that is in a future, as yet undecided release.
    For the time being, you can resort to hacks to refresh Discoverer portlers more frequently than an hour. Some information on the same is available on Metalink. I shall see if I can publish in the next few days a blog post on how to do that. Of course you have to realize that something like this is not encouraged or supported.
    Thanks
    Abhinav

Maybe you are looking for

  • Downloading pictures from an ipod?

    I have several photos on my ipod touch that I want to download onto y macbook pro. I plug the ipod into the computer but I don't know what to do after that. Any help would be much appreciated.

  • Reversing subcontracted finished goods that came from customer returns.

    Dear Gurus, I have a scenario where the finished goods that came from subcontracting needs to be reversed back to its components without reversing the subcontracting charges. How to process this scenario in SAP. The subcontracting job is packing. The

  • SWF requested multiple times from server

    Hi, We have implemented an SWF in our homepage, using normal object tags. The SWF source is called from our database. Loading works fine, and the SWF plays OK on the page. However, using firebug or httpwatch, we can see that the same SWF is requested

  • MPLS TE Network Design - Need Advice

    Please refer to attached diagram. TE tunnels is configured between Site A and Site B, fully meshed. But I have Site C, due to geographical near to Site B , the link is connect to site B PE routers, not to P router. Q: To reach site A shall I just bui

  • Big Network

    Hi I want to calculate the paths from certain starting points in my network. The network (logical) has 3500000 edges and 52000 nodes. I am using the Java APIs but it doesn't even read the network and it throws memory exception. My RAM is 1G. Is it fe