Automatic rule and package size

Hi Guys,
We've set an automatic rule for windows patching... to download on patch Tuesdays and deploy to our pilot machines. However, the deployment package is getting bigger and bigger. It's up to 2gb now. I only want the package to have last months patches
and purge the rest. Is there anyway to accomplish this?
Thanks in advance.

For the package size, you simply create a new one manually every 3, 6, 9, ... months (take your pick) and then reassign the package in the ADR. Pre-R2 you will have to do this with PowerShell. Post R2 you can do it in the UI.
As for the compliance groups for older updates, what happens when a system comes online that needs an update? Older updates typically already have exploits in the wild for the vulnerability that they fix so that system is at risk (as is the rest of the organization)
because the update isn't available.
Jason | http://blog.configmgrftw.com

Similar Messages

  • Report for all RAP advertisements with Package size.

    Hi,
    I have received a requirement to get the information about all the advertisements published in RAP along with package details like (Package Name/version/ and Package size).
    I have two different working reports, merging these two reports would suffice my requirements.
    Please help to merge below two reports.
    1.  
    SELECT DISTINCT p.PackageID, p.Name,p.Version, p.SourceVersion, 'Source Size (MB)' = n.SourceSize/1024
     FROM v_Package p
     LEFT JOIN
     v_PackageStatusRootSummarizer n
     ON
     p.PackageID = n.PackageID
    order by Name
    2.
    SELECT v_Advertisement.AdvertisementName,
    v_Advertisement.Comment, 
    v_Package.Name AS C062,
    ProgramName,
    v_Advertisement.SourceSite,
    CASE WHEN AssignedScheduleEnabled != 0 or
                                          (AdvertFlags & 0x720) != 0
    THEN '*'
    ELSE ''
    END AS C063,
    v_Collection.Name as C064, 
    CASE IncludeSubCollection
    WHEN 1 THEN '*'
    WHEN 0 THEN ''
    END AS C065,
    AdvertisementID
    FROM v_Advertisement 
    INNER JOIN v_Package ON v_Advertisement.PackageID = v_Package.PackageID
    INNER JOIN v_Collection ON v_Advertisement.CollectionID = v_Collection.CollectionID
    ORDER BY v_Advertisement.AdvertisementName
    The Lightning Khan

    SELECT
    p
    .PackageID,
    p.Name,p.Version,
    p.SourceVersion,
    'Source Size (MB)'
    = n.SourceSize/1024,
    v_Advertisement
    .AdvertisementName,
    v_Advertisement
    .Comment,
    p
    .Name
    AS C062,
    ProgramName
    v_Advertisement
    .SourceSite,
    CASE
    WHEN AssignedScheduleEnabled
    != 0
    or
    (AdvertFlags
    & 0x720)
    != 0
    THEN
    ELSE
    END
    AS C063, 
    v_Collection
    .Name
    as C064,
    CASE
    IncludeSubCollection
    WHEN
    1 THEN
    WHEN
    0 THEN
    END
    AS C065,
    AdvertisementID
    FROM v_Advertisement
    INNER
    JOIN v_Package p
    ON v_Advertisement.PackageID
    = p.PackageID
    INNER
    JOIN v_Collection
    ON v_Advertisement.CollectionID
    = v_Collection.CollectionID
    LEFT
    JOIN  v_PackageStatusRootSummarizer n 
    ON  p.PackageID
    = n.PackageID
    ORDER
    BY v_Advertisement.AdvertisementName
    John Marcum | Microsoft MVP - Enterprise Client Management
    My blog: System Center Admin | Twitter:
    @SCCM_Marcum | Linkedin:
    John Marcum

  • Problem with balance carry forward business rule and automatic adjustments

    Hi All,
    I faced with issue when run BCF business rule and consolidation process. According sap courses BPC440 members with A(automatic adjustments) type should carry forward automatically, and I(imputable), M (adjustments) members should carry forwarded by bcf business rule.
    When I set up my BCF business rule I use property "Category_for_ope" in Category dimension. But this rule not work properly, then I use notes:
    1784365 - Category_for_ope property should not effect Consolidation.
    1902677 - The property category_for_ope shall work for consolidation
    After applying those notes, system work as I aspect. After i run consolidation first time system carry forward my automatic adjustment from Category "Expected" to "PLAN_Y0" and this is correct!!!. But when i run consolidation process second and third time system duplicated automatic adjustments in Category "PLAN_Y0".
    Someone can help me with this Issue?
    And sorry for my english

    Hi,
    Can you try as follows :
    *RUN_PROGRAM COPYOPENING
    CATEGORY = %C_CATEGORY_SET%
    CURRENCY = %RPTCURRENCY_SET%
    TID_RA = %TIME_SET%
    *ENDRUN_PROGRAM
    hope it helps...
    regards,
    Raju

  • Maximum package size for data packages was exceeded and Process terminated

    Hello Guru,
    When i am execute the process chain i got this message Maximum package size for data packages was exceeded and Process terminated,any body help to me in this case how can i proceed.
    Thanks & Regards,
    Suresh.

    Hi,
    When the load is not getiing processed due to huge volume of data, or more number of records per data packet, Please try the below option.
    1) Reduce the IDOC size to 8000 and number of data packets per IDOC as 10. This can be done in info package settings.
    2) Run the load only to PSA.
    3) Once the load is succesfull , then push the data to targets.
    In this way you can overcome this issue.
    You can also try RSCUSTV* where * is an integer to change data load settings.
    Change Datapackage size for extraction, use Transaction RSCUSTV6.
    Change Datapackage size when upload from an R/3 system, set this value in R/3 Customizing (SBIW -> General settings -> Control parameters for data transfer).
    IN R/3, T-Code SBIW --> Genaral settings --> Maintain Control Parameters for Data Transfer (source system specific)
    Hope this helps.
    Thanks,
    JituK

  • Support package SAPKB46C57 - package info shows blank fields and zero size

    Hello everybody,
    We have problem with SAP 4.6C and we found that note 1100728 can help us. We found that all corrections are in Support package SAPKB46C57, but we are not able to download it. We tried to check package info and we received blank fields and zero size of support package SAPKB46C57.
    Does anybody have same experiences with it?
    SAP support is very slow in solving this issue.
    Does anybody know how we can implement corrections from note 1100728?
    Thank you very much for answer.
    Best Regards
    Ing. Lukas Jarmar

    When i compiled the package, it compiled with following error.
    Warning: Package Body created with compilation errors.
    SQL> show error
    Errors for PACKAGE BODY STATSPACK:
    LINE/COL ERROR
    2045/3 PLS-00201: identifier 'SYS.DBMS_SHARED_POOL' must be declared
    2045/3 PL/SQL: Statement ignored
    SQL>

  • Distribution rules and projects - automatic checks

    Hello,
    I haveve three issues/questions regarding distribution rules and projects:
    - Are we able to implement an automatic control which does not allow bookings without distribution rules? The
    control already exists for booking which we'll be entered manually. But for automatically generated postings
    is not always in place. For example by closing a production order when the appropriate distribution rule is
    missing in the BOM.
    - Additional for projects: Either the distribution rule which is defined in the project master data should be
    proposed automatically and/or only the defined distribution rule is allowed for any bookings. This is mainly
    necessary for invoices which are registered only in SAP and not in PM (for example sales commission invoices,
    transport costs etc.). There people often use distribution rules which are not suitable.
    Our understanding is that once a project is set-up with a distribution rule and an entry is posted against
    that item, the distribution rule can not be changed. Is this correct?
    With kind regards,
    Amine Z.

    Closed so I can open a new thread

  • Data package size-LAN and WAN

    HI Experts,
    Could anybody give explanation/Document for the  below query?
    When transfering R/3 to BW.How the data pacakage size is determined.?Would the size of data packets be different over LAN and WAN? Why?
    Thanks
    Pradeep

    In transaction SBIW -> General Settings -> Maintain Control Parameters for Data Transfer (in the OLTP System), you can see and edit the default values for the source system.
    If you display the infopackage in the BW system and click on the menu option Scheduler -> DataS. Default Data Transfer, you will be able to edit the settings for the infopackage and also see what's the default configuration for the source system.
    You may also consult the following SAP Notes for more information (including exceptions):
    [417307 - Extractor package size: Collective note for applications.|https://websmp107.sap-ag.de/sap/support/notes/417307]
    [409641 - Examples of packet size dependency on ROIDOCPRMS.|https://websmp107.sap-ag.de/sap/support/notes/409641]
    I don't see what difference it could have over LAN or WAN, though...

  • Unicode export:Table-splitting and package splitting

    Hi SAP experts,
    I know there are lot of forums related to this topic, but I have some new questions and hence posting a new thread.
    We are in the process of doing unicode conversion in our landscape(CRM 7.0 system based on NW 7.01) and we are running on AIX 6.1 and DB2 9.5. The database size is around 1.5 TB and hence, we want to go in for optimization for export and import in order to reduce the downtime.As a part of the process, we have tried to do table-splitting and parallel export-import to reduce the downtime.
    However, we are having some doubts whether this table-splitting has actually worked in our scenario,as the export has exeucted for nearly 28 hours.
    The steps followed by us :
    1.) Doing the export preparation using SAPINST
    2.) Doing table splitting preparation, by creating a table input file having entries in the format <tablename>%<No.of splits>.Also, we have used the latest R3ta file and the dbdb6slib.o(belonging to version 7.20 even though our system is on 7.01) using SAPINST.
    3.) Starting with the export using SAPINST.
    some observations and questions:
    1.) After completion of tablesplitting preparation, there were .WHR files that were generated for each of the tables in DATA directory of export location. However, how many .WHR files should be created and on what basis are they created?
    2.) I will take an example of a table PRCD_CLUST(cluster table) in our environment, which we had split. We had 29 *.WHR files that were created for this particular table. The number of splits given for this table was 36 and the table size is around 72 GB.Also, we noticed that the first 28 .WHR files for this table, had lots of records but the last 29th .WHR file, had only 1 record. But we also noticed that, the packages/splits for the 1st 28 splits were created quite fast but the last one,29th one took a long time(serveral hours) to get completed.Also,lots of packages were generated(around 56) of size 1 GB each for this 29th split. Also, there was only one R3load which was running for this 29th split, and was generating packages one by one.
    3.) Also,Our question here is that is there any thumb rule for deciding on the number of splits for a table.Also, during the export, are there any things that need to be specified, while giving the inputs when we use table splitting,in the screen?
    4.) Also, what exactly is the difference between table-splitting and package-splitting? Are they both effective together?
    If you have any questions and or need any clarifications and further inputs, please let me know.
    It would be great, if we could get any insights on this whole procedure, as we know a lot of things are taken care by SAPINST itself in the background, but we just want to be certain that we have done the right thing and this is the way it should work.
    Regards,
    Santosh Bhat

    Hi,
    First of all please ignore my very first response ... i have accidentally posted a response to some other thread...sorry for that 
    Now coming you your questions...
    > 1.) Can package splitting and table-splitting be used together? If yes or no, what exactly is the procedure to be followed. As I observed that, the packages also have entries of the tables that we decided to split. So, does package splitting or table-splitting override the other, and only one of the two can be effective at a time?
    Package splitting and table splitting works together, because both serve a different purpose
    My way of doing is ...
    When i do package split i choose packageLimit 1000 and also split out the tables (which i selected for table split)  into seperate package (one package per table). I do it because that helps me to track those table.
    Once the above is done i follow it up with the R3ta and wheresplitter for those tables.
    Followed by manual migration monitor to do export/import , as mentioned in the previous reply above you need to ensure you sequenced you package properly ... large tables are exported first , use sections in the package list file , etc
    > 2.) If you are well versed with table splitting procedure, could you describe maybe in brief the exact procedure?
    Well i would say run R3ta (it will create multiple select queries) followed by wheresplitter (which will just split each of the select into multiple WHR files)  ...  
    Best would go thought some document on table spliting and let me know if you have specific query. Dont miss the role of hints file.
    > 3.) Also, I have mentioned about the version of R3ta and library file in my original post. Is this likely to be an issue?Also, is there a thumb rule to decide on the no.of splits for a table.
    Rule is use executable of the kernel version supported by your system version. I am not well versed with 7.01 and 7.2 support ... to give you an example i should not use 700 R3ta on 640 system , although it works.
    >1.) After completion of tablesplitting preparation, there were .WHR files that were generated for each of the tables in DATA directory of export location. However, how many .WHR files should be created and on what basis are they created?
    If you ask for 10 split .... you will get 10 splits or in some case 11 also, the reason might be the field it is using to split the table (the where clause). But not 100% sure about it.
    > 2) I will take an example of a table PRCD_CLUST(cluster table) in our environment, which we had split. We had 29 *.WHR files that were created for this particular table. The number of splits given for this table was 36 and the table size is around 72 GB.Also, we noticed that the first 28 .WHR files for this table, had lots of records but the last 29th .WHR file, had only 1 record. But we also noticed that, the packages/splits for the 1st 28 splits were created quite fast but the last one,29th one took a long time(serveral hours) to get completed.Also,lots of packages were generated(around 56) of size 1 GB each for this 29th plit. Also, there was only one R3load which was running for this 29th split, and was generating packages one by one.
    Not sure why you got 29 split when you asked for 36, one reason might be the field (key) used for split didn't have more than 28 unique records. I dont know how is PRCD_CLUST  split , you need to check the hints file for "key". One example can be suppose my table is split using company code, i have 10 company codes so even if i ask for 20 splits i will get only 10 splits (WHR's).
    Yes the 29th file will always have less records, if you open the 29th WHR you will see that it has the "greater than clause". The 1st and the last WHR file has the "less than" and "greater than" clause , kind of a safety which allows you to prepare for the split even before you have downtime has started. This 2 WHR's ensures  that no record gets missed, though you might have prepared your WHR files week before the actual migration.
    > 3) Also,Our question here is that is there any thumb rule for deciding on the number of splits for a table.Also, during the export, are there any things that need to be specified, while giving the inputs when we use table splitting,in the screen?
    Not aware any thumb rule. First iteration you might choose something like 10 for 50 GB , 20 for 100 GB. If any of the tables overshoots the window. They you can give a try by  increase or decrease the number of splits for the table. For me couple of times the total export/import  time have improved by reducing the splits of some tables (i suppose i was oversplitting those tables).
    Regards,
    Neel
    Edited by: Neelabha Banerjee on Nov 30, 2011 11:12 PM

  • Will it improve performance if PACKAGE SIZE added to a select query?

    Hi Gurus
    While doing the performance optimization I could see that in the code it is splitting the BKPF entries selected to a package of 500 and selecting from BSEG cluster table. In effect the remaining tables are also getting hit based on the no. of entries(packets) in the selected BKPF internal table.
    Cant we use the package size 500  in the select statement instead of this spliting and hiting the tables in a do loop.
    eg:-
          SELECT x  v z
            FROM bseg
            APPENDING TABLE lt_bseg
            PACKAGE SIZE 500
            FOR ALL ENTRIES IN lt_coep
            WHERE bukrs = lt_coep-bukrs
            AND   belnr = lt_coep-belnr_fi
            AND   gjahr = lt_coep-gjahr.
    When I read the keyword documentation I saw  a note in that as " If the addition PACKAGE SIZE is specified together with FOR ALL ENTRIES, it is not passed to the database system, but is applied to the result set on the application server, after all selected rows have been read"  I did not understand the meaning of this.
    So I am confused whether I can go ahead with this approach or will go along with the already used BKPF split logic and selct from BSEG and other tables in DO loop.
    Can you please help me << removed >>.
    Thanks & Regards,
    Shinimol.
    Edited by: Rob Burbank on Sep 23, 2010 9:09 AM

    Hi,
    regarding your second doubt:
    > Second doubt is on the note I saw in key word documentation. " If the addition PACKAGE SIZE is specified together with FOR ALL ENTRIES, it is not passed to the database system, but is applied to the result set on the application server, after all selected rows have been read.  "  I couldn't understnd this. When I debugged the select with F5 it was selecting 500 each  till it reaches th end.
    > So what does it mean, we can use this option or we should not use?
    the FAE does your select in pieces and collects the result in the database interface in an inermediate buffer and
    eliminates the duplicates there. AFTER this it gives you the number of rows back to your ABAP program that are specified with package size. It simply means that the package size addition has no effect on the memory consumptionfor the intermediate
    buffer in the database interface.
    Kind regards,
    Hermann

  • Issue when number of records in a DSO exceeds DTP package size

    Hi all,
    I'm having a strange problem with the package size of my DTPs.
    I move data from a DSO to another one while performing some transformations in an Export routine. The DTP has a package size of 6 million records.
    When the number of records in the first DSO exceeds the package size, some of the records seem not to be processed properly by the transformation and this forces me to choose a very large number for the package size so that the ABAP code in the transformation is processed for all the records of the source DSO.
    I can't understand why this is happening because package size is only supposed to determine the number of records to be processed in a single step and nothing else.
    Am I right?
    Thanks

    My tip would also be the summary item

  • What is the best Practice to improve MDIS performance in setting up file aggregation and chunk size

    Hello Experts,
    in our project we have planned to do some parameter change to improve the MDIS performance and want to know the best practice in setting up file aggregation and chunk size when we importing large numbers of small files(one file contains one record and each file size would be 2 to 3KB) through automatic import process,
    below is the current setting in production:-
    Chunk Size=2000
    No. Of Chunks Processed In Parallel=40
    file aggregation-5
    Records Per Minute processed-37
    and we made the below setting in Development system:-
    Chunk Size=70000
    No. Of Chunks Processed In Parallel=40
    file aggregation-25
    Records Per Minute processed-111
    after making the above changes import process improved but we want to get expert opinion making these changes in production because there is huge number different between what is there in prod and what change we made in Dev.
    thanks in advance,
    Regards
    Ajay

    Hi Ajay,
    The SAP default values are as below
    Chunk Size=50000
    No of Chunks processed in parallel = 5
    File aggregation: Depends  largely on the data , if you have one or 2 records being sent at a time then it is better to cluster them together and send it at one shot , instead of sending the one record at a time.
    Records per minute Processed - Same as above
    Regards,
    Vag Vignesh Shenoy

  • Package Size parameter in Partner Profile

    Hello guys,
    I'd need your advice. We have 2 systems (SAP and Non-SAP) and we need to connect them by mean of HR-PDC interface (IDoc communication is supported). We can create new Partner Profile in SAP and we can add a few IDoc types as Inbound Parameters of the partner profile. Then it is possible to define maximal package size for several IDoc types. I suppose that the Non-SAP system reads the Partner Profile and the "Package Size" parameter for the particular IDoc type before it sends some IDoc to SAP system. IDoc's size is limited by the parameter.
    And now our problem. We need to insert XI between these 2 systems. XI should only forward IDocs and nothing else. How could I set up limitation of IDoc's size? Do you thing that following scenarion could be working?
    I would define Partner Profile in XI same way as in the previous case in SAP. XI is configured to process IDocs in Integration Server by means of IDoc adapter. My expectation is that the Non-SAP system would read the partner profile in XI (including "Package Size" parameter) and then it would send IDoc to XI. Then XI would process the received IDocs in IDoc adapter.
    How can be "Package Size" parameter useful for inbound IDocs?
    Best Regards,
    Zbynek

    Hello ,
                Check the blog below
    /people/michal.krawczyk2/blog/2007/12/02/xipi-sender-idoc-adapter-packaging
    Rajesh

  • Business Rules and BPM

    Anybody here knows the difference between Business Rules and BPM..
    We can do pretty much everything with a Business Rules Engine as we would do with BPM , then in which scenario should we use BPM and in which scenario should we use Business Rules Engine?
    Ofcourse Business Rules Engine has more complex set of rules than any BPM product in the market for .eg. ILOG Business Rules...

    BPM and a rules engine are different. Both are important, but they have different uses.
    Oracle BPM Studio is used by Business Analysts and Developers to model processes, integrate underlying components (including business rules) and develop user interfaces.
    There's an Engine that executes the processes that have been modeled. It sends the work item instances to the right person or persons with the right data when the work item reaches an interactive activity in the process. Similarly the Engine has invokes the underlying components specified in the process when the work item instance reaches an automatic activity.
    There's an out of the box Workspace runtime that lets end users manage and run their work item instances as they flow throughout the processes.
    We can business rules inside the processes, but one does not replace the other.
    Dan

  • Need you help !!! -- How to set PACKAGE SIZE in CURSOR?

    Dear experts:
    Does anybody know how to set the package size if you use cursor in your abap program?
    My scenario is to select all the data from MARA table in which the data volumn is very big, and then process for some specific logic. When I try to select them into one internal table or use CURSOR without package, my abap program will abort after running for around 1 hour. So I was thinking if I can use cursor and set the PACKAGE SIZE like 10,000 and process the data by batch.
    Here is my sample code:
    OPEN CURSOR: CUR_MARA_TBL FOR
    SELECT * FROM MARA.
    DO.
    FETCH NEXT CURSOR CUR_MARA_TBL INTO CORRESPONDING FIELDS OF WA_MARA_TBL.
    IF sy-subrc = 0.
    ......Detail logic......
    ENDIF.
    ENDDO.
    I would appreciate if anybody can let me know how to set the package size based on the sample code above, or any webpage/article about that.
    Any reply would be appreciated and thank you all in advance!
    Best Regards.
    Tim
    Moderator message: please study F1 help for FETCH statement.
    locked by: Thomas Zloch on Sep 8, 2010 4:10 PM

    The Spark List's layout property has the requestedRowCount and requestedMinRowCount properties.
    You can either set a new layout property on your List like this:
    <s:List>
            <s:layout>
                <s:VerticalLayout requestedMinRowCount="1" />
            </s:layout>
        </s:List>
    or reskin the List and modify the VerticalLayout in your copy of ListSkin.
    Jason

  • HT4623 I have updated my iphone 5 with IOS7. Now there is an update 7.0.2 on my cell and the size of update file is showing around 21.0 MB. I tried to update it ITunes but it seems it is downloading the entire IOS. Is there any separate update for the sam

    I have updated my iphone 5 with IOS7. Now there is an update 7.0.2 on my cell and the size of update file is showing around 21.0 MB. I tried to update it ITunes but it seems it is downloading the entire IOS. Is there any separate update for the same or I have to download the entire IOS again?

    Well when you download the update with iTunes it will always download the whole new version (in your case 7.0.2) even if its just an bug fix. If you already have 7.0 download the update over-the-air.
    You already wrote that the update is only 21 MB, so its not the whole firmware package of 7.0.2 you get with iTunes.

Maybe you are looking for

  • IPhone will not come out of Recovery Mode. At. All.

    Hi, my 1 year old iPhone has gone into Recovery Mode after a failed restore and im at my wits end. I get error message, after error message. I have windows xp so its a lot harder to use all the downloads that are available. I havent jailbroken my iPh

  • TS3989 new photos are not uploading to photostream

    i have a iphone 4s, ipad 2 and a MacPro using iPhoto.  Yesterday my devices stopped uploading NEW pictures to PhotoStream.  I've tested it by taking pictures with each device and waiting to see if it shows up in photostream.  Nothing works.  Is still

  • Disabling individual wireless laptop

    Help! Urgent! - my teenage daughter is out of control with her wireless laptop, using it excessively. I need to take hers off the nework, but still be able to run mine on wireless. How do I do this? Do I need to get into hers (she's pinned it) or can

  • ESB Webservice call from Java client

    Hi, i need to call webservice from Java client which inturn will call ESB I tried creating proxy from the WSDL in jDeveloper and tried setting the following end points My concrete WSDL path :http://10.237.25.63:8889/esb/wsil/SubroCaseESBSystem/InputR

  • My computer (s) don't see IPOD

    We've had this ipod for two years. Two weeks ago it would no longer update with itunes, in fact, itunes, ipod updater, system tools, nothing would see it. Tried 4 different usb ports, two different cords, nuttin! The cords would still charge it and i