Tool for mass modification of KM / collaboration permissions

Hi,
do you know if there is a tool in portal for mass modification of permissions for documents, folders and collaboration rooms? It would be quite handy.
For example I have a folder which contains 100 documents and I want to change permissions for only some of the documents, but I don't want to change them one by one (go through the details dialog box for each document and change permissions).
The same question is for collaboration rooms.
Regards,
Ladislav

Hi Ladislav,
There is a KM report named Propagation of Properties, under Content Administration -> KM Content -> Toolbox -> Reports -> Mass Operations -> Propagation of Properties, which allows you to propagate the existing properties of a resource, to other folders and their contents or to individual documents.
Refer to [http://help.sap.com/saphelp_nw04/helpdata/EN/44/8d6f90d85c6572e10000000a11466f/content.htm] for more information on how to use this KM report.
Hope this Helps.
Regards,
Adren

Similar Messages

  • Performance Tool for Mass Tests

    Hello everybody,
    does anybody know, if there are any tools (best would be freeware), besides  Mercury Loadrunner,  which you can make mass tests over the ITS ?
    We just need 1 license or one installation to measure the performance when alot of users work.
    We make a performance project and we need to simulate mass tests in the ITS.(WEB)
    Thanks already for help!
    Jasmin

    please search for eCATT
    http://help.sap.com/saphelp_nw04s/helpdata/en/1b/e81c3b84e65e7be10000000a11402f/frameset.htm
    Nesimi

  • Tool for Mass deletion Idoc

    Hello everybody,
    We have a little problem with our ALE....We have generated 2 000 000 idoc by error in our Sender System..all the idoc have the statut 30 (generated)...
    We're looking for a programm or a transaction to edit or delete these idocs...(we didn't find it)
    Is it possible ?
    We are in 4.6C version
    Thansk for your help

    If you do not want the IDocs, which I am guessing is the case based on your question, you could use transaction code SARA to archive the IDocs.  If you are not familiar with SARA, there are some steps you will need to complete to setup variants for the Archive run.
    1. Go to WE47 and check whether you have status code 30 setup for archiving possible or not.  When you double click the status code it brings up the details for the status.  There is a section for "Archiving"; it should be set to "Poss.".
    2. Go to transaction code SARA, enter Object Name = IDOC and enter.
    3. You will need to go to the Customizing section of Archive Management and setup the Test run variant and Production run variant for the Delete Program (RSEXARCD) .
    4. When you click Archive, you will be able to create a variant for the Archive program run (RSEXARCA).  At this point you can specify "Do not create archive file" if you do not want the IDocs archived to a file.
    Hope this helps.  Best of luck.
    Regards,
    Stan

  • Tools for Variant and Batch Schedule

    Dear all,
    do you know wehther there is tools with following functionality
    - create variant with a input file (information for report names, variant names, variant parameter names and values)
    - schedule SAP jobs with a input file (information about job name, job class, server which the batch should excuted, Steps with batch user name, program name and variant name)
    - tools for mass download and upload variant.
    Thanks and regards
    Tao

    Hi kiran,
    Pl check,
    http://www.sap-img.com/sap-sd/steps-for-sd-variant-configuration.htm
    http://www.sap-img.com/sap-sd/some-light-on-batch-determination.htm
    https://websmp207.sap-ag.de/~sapidp/011000358700003276182004
    Regards,
    SDS

  • Using AE(htmldb 2.0)  as for reporting tool for other oracle databases.

    Maybe I was mistaken, but when I first saw the HTMLDB demo's I saw this as a slick way to build a browser based reporting tool for the masses for quasi-ad-hoc reporting on some of our larger databases that currently have limited reporting capability, and also as a possible replacement for the many MSAccess applets we have.
    I just got 2.0 installed and working on a 9i database and went through the basic tutorials, but I can't find much info on how to connect to other oracle databases.
    Are my options as follows?:
    create dblinks from the local database that houses htmldb to all the other databases?
    or
    install htmldb schemas on each database I want to access from htmldb?
    or
    I'll keep looking, but if any of you have any pointers or know of a tutorial that explains the best way to do this it would be greatly appreciated!
    tia

    As for question 1, yep, that's what I mean. Example:
    In a schema in the App Express instance, create a user called "APP1_USER", and then
    create view app1_user.people as select * from people@link_user@remotedb
    As for the next question, using the built-in insert/update/delete processes will give you
    ORA-01461: can bind a LONG value only for insert into a LONG column
    among other errors. Apparently version 2.1 will fix this. See Cannot Update VARCHAR column in Remote DB
    We address this problem by creating pl/sql APIs for DML, and then writing our own insert/update processes. Slows us down a little, but even with that, App Express remains a productive choice for us.
    Another weird problem: auto-row fetch (built-in to HTML DB) doesn't work over db links when you have two columns for a primary key. Many ways around this, e.g. create sequence-populated primary key, create a view with single column that concatenates primary key column values (key1|key2)
    Hope this helps!
    -John

  • For mass upload of documents

    Hi Everybody
    Is there any tools available for mass upload of documents into server? please provide me the information.
    Thanks
    Prasad

    Hi Prasad,
    There are several options that you can explore for mass document upload and see which best fits your requirements and constraints:
    -You can upload documents by using LSMW(Legacy System Migration Workbench). Using this, you can safely upload 1000-1100 documents at one go.
    - Ask your ABAPer to develop a BDC for mass document upload.
    - Use 'DMS Loader' solution from SEALS systems. http://www.sealsystems.de/fileadmin/mediaFiles/Factsheets/DMS_Loader_en.pdf
    Regards,
    Pradeepkumar Haragoldavar
    Edited by: Pradeepkumar  Haragoldavar on Jun 17, 2010 11:09 AM

  • Can anyone help with a good system maintenance tool for ML

    Does any one have any suggestions for a good dependable all-in-one system maintenance tool for ML and is it straight forward and easy to use?
    Thanks in advance .....

    You really have no need for one. That said:
    Kappy's Personal Suggestions for OS X Maintenance
    For disk repairs use Disk Utility.  For situations DU cannot handle the best third-party utilities are: Disk Warrior;  DW only fixes problems with the disk directory, but most disk problems are caused by directory corruption; Disk Warrior 4.x is now Intel Mac compatible. Drive Genius provides additional tools not found in Disk Warrior.  Versions 1.5.1 and later are Intel Mac compatible.
    OS X performs certain maintenance functions that are scheduled to occur on a daily, weekly, or monthly period. The maintenance scripts run in the early AM only if the computer is turned on 24/7 (no sleep.)
    If you are using a pre-Leopard version of OS X, then an excellent solution is to download and install a shareware utility such as Macaroni, JAW PseudoAnacron, or Anacron that will automate the maintenance activity regardless of whether the computer is turned off or asleep.  Dependence upon third-party utilities to run the periodic maintenance scripts was significantly reduced since Tiger.  (These utilities have limited or no functionality with Snow Leopard, Lion, or Mountain Lion and should not be installed.)
    OS X automatically defragments files less than 20 MBs in size, so unless you have a disk full of very large files there's little need for defragmenting the hard drive. As for virus protection there are few if any such animals affecting OS X. You can protect the computer easily using the freeware Open Source virus protection software ClamXAV. Personally I would avoid most commercial anti-virus software because of their potential for causing problems. For more about malware see Macintosh Virus Guide.
    I would also recommend downloading a utility such as TinkerTool System, OnyX 2.4.3, or Cocktail 5.1.1 that you can use for periodic maintenance such as removing old log files and archives, clearing caches, etc.
    For emergency repairs install the freeware utility Applejack.  If you cannot start up in OS X, you may be able to start in single-user mode from which you can run Applejack to do a whole set of repair and maintenance routines from the command line.  Note that AppleJack 1.5 is required for Leopard. AppleJack 1.6 is compatible with Snow Leopard. (There is no confirmation that this version also works with Lion or later.)
    When you install any new system software or updates be sure to repair the hard drive and permissions beforehand. I also recommend booting into safe mode before doing system software updates.
    Get an external Firewire drive at least equal in size to the internal hard drive and make (and maintain) a bootable clone/backup. You can make a bootable clone using the Restore option of Disk Utility. You can also make and maintain clones with good backup software. My personal recommendations are (order is not significant):
      1. Carbon Copy Cloner.
      2. Data Backup
      3. Deja Vu
      4. SuperDuper!
      5. SyncTwoFolders
      6. Synk Pro
      7. Synk Standard
      8. Tri-Backup
    Visit The XLab FAQs and read the FAQs on maintenance, optimization, virus protection, and backup and restore.
    Always have a current backup before performing any system updates or upgrades.
    Additional suggestions will be found in
       1. Mac Maintenance Quick Assist,
       2. Mac OS X speed FAQ,
      3. Speeding up Macs,
      4. Macintosh OS X Routine Maintenance,
      5. Essential Mac Maintenance: Get set up,
      6. Essential Mac Maintenance: Rev up your routines,
      7. Maintaining OS X, 
      8. Five Mac maintenance myths,
      9. How to Speed up Macs, and
    10. Myths of required versus not required maintenance for Mac OS X.
    Referenced software can be found at CNet Downloads or MacUpdate.
    Be sure you have an adequate amount of RAM installed for the number of applications you run concurrently. Be sure you leave a minimum of 10% of the hard drive's capacity or 20 GBs, whichever is greater, as free space.

  • What is a good cleanup tool for Mavericks?

    What is a good cleanup tool for Mavericks?

    Kappy's Personal Suggestions About Mac Maintenance
    For disk repairs use Disk Utility.  For situations DU cannot handle the best third-party utility is: Disk Warrior;  DW only fixes problems with the disk directory, but most disk problems are caused by directory corruption. Drive Genius provides additional tools not found in Disk Warrior for defragmentation of older drives, disk repair, disk scans, formatting, partitioning, disk copy, and benchmarking. 
    Four outstanding sources of information on Mac maintenance are:
    1. OS X Maintenance - MacAttorney.
    2. Mac maintenance Quick Assist
    3. Maintaining Mac OS X
    4. Mac Maintenance Guide
    Periodic Maintenance
    OS X performs certain maintenance functions that are scheduled to occur on a daily, weekly, or monthly period. The maintenance scripts run in the early AM only if the computer is turned on 24/7 (no sleep.) See Mac OS X- About background maintenance tasks. If you are running Leopard or later these tasks are run automatically, so there is no need to use any third-party software to force running these tasks.
    If you are using a pre-Leopard version of OS X, then an excellent solution is to download and install a shareware utility such as Macaroni, JAW PseudoAnacron, or Anacron that will automate the maintenance activity regardless of whether the computer is turned off or asleep.  Dependence upon third-party utilities to run the periodic maintenance scripts was significantly reduced after Tiger.  (These utilities have limited or no functionality with Snow Leopard, Lion, or Mountain Lion and should not be installed.)
    Defragmentation
    OS X automatically defragments files less than 20 MBs in size, so unless you have a disk full of very large files there's little need for defragmenting the hard drive except when trying to install Boot Camp on a fragmented drive. Malware Protection
    As for malware protection there are few if any such animals affecting OS X. Starting with Lion, Apple has included built-in malware protection that is automatically updated as necessary. To assure proper protection, update your system software when Apple releases new OS X updates for your computer.
    Helpful Links Regarding Malware Protection:
    1. Mac Malware Guide.
    2. Detecting and avoiding malware and spyware
    3. Macintosh Virus Guide
    For general anti-virus protection I recommend only using ClamXav, but it is not necessary if you are keeping your computer's operating system software up to date. You should avoid any other third-party software advertised as providing anti-malware/virus protection. They are not required and could cause the performance of your computer to drop.
    Cache Clearing
    I recommend downloading a utility such as TinkerTool System, OnyX 2.4.3, Mountain Lion Cache Cleaner 7.0.9, Maintenance 1.6.8, or Cocktail 5.1.1 that you can use for periodic maintenance such as removing old log files and archives, clearing caches, etc. Corrupted cache files can cause slowness, kernel panics, and other issues. Although this is not a frequent nor a recurring problem, when it does happen there are tools such as those above to fix the problem.
    If you are using Snow Leopard or earlier, then for emergency cleaning install the freeware utility Applejack.  If you cannot start up in OS X, you may be able to start in single-user mode from which you can run Applejack to do a whole set of repair and maintenance routines from the command line.  Note that AppleJack 1.5 is required for Leopard. AppleJack 1.6 is compatible with Snow Leopard. (AppleJack works with Snow Leopard or earlier.)
    Installing System Updates or Upgrades
    Repair the hard drive and permissions beforehand.
    Update your backups in case an update goes bad.
    Backup and Restore
    Having a backup and restore strategy is one of the most important things you can do to maintain your computer. Get an external Firewire drive at least equal in size to the internal hard drive and make (and maintain) a bootable clone/backup. You can make a bootable clone using the Restore option of Disk Utility. You can also make and maintain clones with good backup software. You can never have too many backups. Don't rely on just one. Make several using different backup utilities. My personal recommendations are (order is not significant):
         1. Carbon Copy Cloner
         2. Get Backup
         3. Deja Vu
         4. SuperDuper!
         5. Synk Pro
         6. Tri-Backup
    Visit The XLab FAQs and read the FAQs on maintenance and backup and restore.
    Always have a current backup before performing any system updates or upgrades.
    Final Suggestions
    Be sure you have an adequate amount of RAM installed for the number of applications you run concurrently. Be sure you leave a minimum of 10% of the hard drive's capacity or 20 GBs, whichever is greater, as free space. Avoid installing utilities that rely on Haxies, SIMBL, or that alter the OS appearance, add features you will rarely if ever need, etc. The more extras you install the greater the probability of having problems. If you install software be sure you know how to uninstall it. Avoid installing multiple new software at the same time. Install one at a time and use it for a while to be sure it's compatible.
    Additional reading may be found in:    
    1. Mac OS X speed FAQ
    2. Speeding up Macs
    3. Macintosh OS X Routine Maintenance
    4. Essential Mac Maintenance: Get set up
    5. Essential Mac Maintenance: Rev up your routines
    6. Five Mac maintenance myths
    7. How to Speed up Macs
    8. Myths of required versus not required maintenance for Mac OS X
    Referenced software can be found at CNet Downloads or MacUpdate.
    Most if not all maintenance is for troubleshooting problems. If your computer is running OK, then there isn't really a thing you need to do except repair the hard drive and permissions before installing any new system updates.

  • I want to change Material Type for mass upload.(MMAM)

    hi all,
    I want to change Material Type for mass upload.(MMAM)
    Suggestion pls? or Any BAPI is availablre?
    Edited by: abap_friends on Jan 6, 2011 7:12 AM
    Moderator message: please do your own research before asking.
    Edited by: Thomas Zloch on Jan 6, 2011 11:27 AM

    This may be little tough for 10000 sales orders.
    1. You have to remove the output record for the irrelevant output record using tcode VV12 for the relevant document type. So now for the document type, the old output wont be triggered.
    2. Now, create output records for hte correct output type in VV11 tcode(for the output type BA03) for the document type. Now this output will be triggered for all the document types you are creating from now onwards.
    3. To trigger the output type BA03, in all the old orders, you just have to open these sales orders in change mode (VA02), and just save it. When u open it the output type BA03 will be automatically triggered as the master data is maintained in VV11. But because just opening all the 10000 sales orders in VA02 one by one be a mechanical task, you have to think of some tool for this.
    May be a CATT procedure can do the trick. Else, you can just create a small program, where you can upload these sales orders and the program will just open these orders in change mode and just saves it (without making any changes). Then the output will be retriggered.
    Here again, if you are printing the output directly, it is advisable to trigger the output in batch mode(Date/Time field is 1 in VV11), and then you can again trigger the printing again using the program RSNAST00 after office hours for these 10k orders.
    I hope I answered ur query.
    Pls lemme know if you need some more info. (You can reward me too...)

  • Tool for Support

    Hello,
    I am searching for a tool for SCCM.
    I have some IT responsible persons (who are not realy it technicans) in diffrent locations.
    I want to give them a possibility to deploy Software to their clients with GUI without using ConfigMgr Console.
    Does anyone knows such tools?

    Is there a reason you don't want them to have the console? You should be able to setup permissions that will only allow them to admin their area if that is the worry.
    Also are you wanting them to push the install or would they be able to go to the machine, and do the install. If they would be able to go to machine to do the install then the solution below should work for you, I have it in place in my environment.
    AD group that has the supporting team members in it
    then created a User Collection in SCCM, and based its membership on the AD group I created
    then created deployments for each of the applications that I want the support team members to have access to.
    The support team member can go to the Application Catalog on the web, and install the application from there. This allows for you to control who all has access to installs by their AD group.
    Also to make the online application catalog easier to get to I made a redirect site in IIS on the application catalog server.
    Without the redirect you have to tell your support team to go to http://ServerName/CMApplicationCatalog"
    With the redirect you can name it whatever you want like http://ServerName/apps"
    Had to add the quote to make it not link.

  • Tool to create XML definition, a tool for creating XSD

    I need a tool to create XML definition, a tool for creating XSD.
    I was trying to do this with Oracle packages. But again there are some problems. Could you tell me which one, I should use?
    Can I do it with the Oracle SQL Developer?

    Hello,
    i get the error message and I don't know, what I can do
    The user have the permissions to use DBMS_LOB package.
    Is it the sofware, do I need a license version tool for XML generation
    SQL> select DBMS_XMLSCHEMA.GENERATESCHEMA( 'HR', 'EMP_DATA' )
    2 from dual;
    select DBMS_XMLSCHEMA.GENERATESCHEMA( 'HR', 'EMP_DATA' )
    ERROR at line 1:
    ORA-00904: "DBMS_XMLSCHEMA"."GENERATESCHEMA": invalid identifier

  • Monitoring and backup/restore tools for TT databases

    Hi there,
    I am new to TimesTen. I need to set up automatic monitoring for TT databases. Are there any tools available? Also would like to know if there are any tools for tracking backups for these datastores.
    Any standards & best practices for managing TimesTen environment available?
    Appreciate any help/advise.
    Regards

    Hi Vidya,
    You did not mention what type of Unix/Linux you are using or what version of TimesTen. I will assume it is a mainstream Unix (Solaris, HP/UX or similar) or Linux (RedHat, SuSE etc.) and a recent (6.0 or 7.0) version of TimesTen. I will also assume that you want to put everything back 'as is' after the rebuild (i.e. you are not upgrading TimesTen versions or changing anything else about the configuration). IF YOU ARE PLANNING TO UPGRADE TIMESTEN AS PART OF THE REBUILD THEN THE STEPS LISTED BELOW WILL NOT WORK 'AS IS' AND WILL NEED MODIFICATION.
    Assuming that the machine is going to be 'wiped' during the rebuild then the main things you need to preserve are:
    1. Kernel parameter and other system level configuratiopn for TimesTen and indeed any other software running on the box. You would do this by taking copies of the relevant system configuration files and re-applying the customisations after the rebuild. As well as kernel parameters this might include things like users and groups.
    2. The port numbers used by each installed TimesTen instance. You can find these out from the ttVersion and ttStatus commands in each instance. Make a note of these as you will need them afterwards for re-installation.
    3. TimesTen configuration files for each installed TimesTen instance. You may have multiple TimesTen instances of the same or different versions installed. You can find out by looking at
    the file /etc/TimesTen/instance_info which contains details of every TimesTen installation (instance) on the machine. The instance configuration files are found in <install_dir>/info and are called; ttendaemon.options, sys.odbc.ini, sys.ttconnect.ini, snmp.ini and ttpasswd. You should make safe copies of all these files for each instance.
    4. Any user configuration files that e.g. configure TimesTen environment variables and any user ODBC configuration files such as .odbc.ini in user's home directory.
    5. Any datastores (databases) that you need to preserve across the rebuild. For each of these, take a backup of the datastore using the TimesTen ttBackup command and preserve the output files/directories. You need to do this separately for each datastore in each TimesTen instance.
    After you have secured the above then you should be okay to do the rebuild. After the rebuild you should:
    a) Re-apply all system level config changes (kernel parameters, users, groups etc.) preserved from (see step 1 above).
    b) Re-install all required TimesTen instances into the same installation directories, and using the same port numbers, as they had previous to the rebuild (see step 2 above). The installations should be performed as the same user as the original installatuions (the ttVersion output from step 2 above would tell you that). Prior to performing the installation you would need a sysadmin to create the /etc/TimesTen directory and assign suitable permissiosn to it (see TimesTen installation Guide for more details).
    c) Re-instate the saved TimesTen configuration files for each instance (see step 3 above).
    d) Re-instate any user level configuration files (see step 4 above).
    e) Re-create any datastore directories as required ready to receive the restored datastores (check datastore and log file paths in ODBC configuration files).
    f) Restore the datastore backup directories (created in step 5 above) to somewhere on disk.
    g) Use ttRestore in each instance to restore each datastore from the backup files restored in step (f) above. This re-creates all your datastores with all their data etc.
    If you follow this procedure and are not doing anything like changing TimesTen major versions (e.g. 6.0 -> 7.0) as part of the rebuild then you should not need to relink any application binaries. If for some reason you want to relink any application binaries you should consult the application developers who should have suitable makefiles to do this.
    I think this is pretty complete but of course it is possible there are additional things specific to your installation or operational regime of which I am unaware and so of course we can not accept any responsibility for any errors or omissions...
    I will send you the document on monitoring TimesTen via your e-mail address.
    Chris

  • Backpac: A package state snapshot and restore tool for Arch Linux

    backpac:
    A package state snapshot and restore tool for Arch Linux with config file save/restore support.
    https://aur.archlinux.org/packages.php?ID=52957
    https://github.com/altercation/backpac (see readme on the github repository for more information)
    Summary & Features
    It's a common method of setting up a single system: take some notes about what packages you've installed, what files you've modified.
    Backpac creates those notes for you and helps back up important configuration files. Specifically, backpac does the following:
    maintains a list of installed groups (based on 80% of group packages being installed)
    maintains a list of packages (including official and aur packages, listed separately)
    maintains a list of files (manually created)
    backs up key config files as detailed in the files list you create
    The package, group and files lists along with the snapshot config files allows system state to be easily committed to version control such as git.
    Backpac can also use these lists to install packages and files. Essentially, then, backpac takes a snapshot of your system and can recreate that state from the files and lists it archives.
    Use Cases
    Ongoing system state backup to github
    Quick install of new system from existing backpac config
    Conform current system to given state in backpac config
    Backpac is a very, very lightweight way of saving and restoring system state.
    It's not intended for rolling out and maintaining multiple similar systems, it's designed to assist individual users in the maintainance of their own Arch Linux box.
    Status
    Alpha, release for testing among those interested. Passing all tests right now but will continue to rework and refine. Bug reports needed.
    Why?
    There are a lot of 'big-iron' solutions to maintaining, backing up and restoring system state. Setting these up for a single system or a handful of personal systems has always seemed like overkill.
    There are also some existing pacman list making utilities around, but most of them seem to list either all packages or don't separate the official and aur packages the way I wanted. Some detect group install state, some don't. I wanted all these features in backpac.
    Finally, whatever tool I use, I'd like it to be simple (c.f. the Arch Way). Lists that are produced should be human readable, human maintainable and not different from what I'm using in non-automated form. Backpac fulfills these requirements.
    Regarding files, I wanted to be able to backup arbitrary system files to a git repository. Tools like etckeeper are interesting but non /etc files in that case aren't backed up (without some link trickery) and there isn't any automatic integration with pacman, so there is no current advantage to using a tool like that. I also like making an explicit list of files to snapshot.
    Sample Output
    This is the command line report. Additionally, backpac saves this information to the backpac groups, packages and files lists and the files snapshot directory.
    $ backpac -Qf
    backpac
    (-b) Backups ON; Files will be saved in place with backup suffix.
    -f Force mode ON; No prompts presented (CAUTION).
    (-F) Full Force mode OFF; Prompt displayed before script runs.
    (-g) Suppress group check OFF; Groups will be checked for currency.
    (-h) Display option and usage summary.
    (-p) Default backpac: /home/es/.config/backpac/tau.
    -Q Simple Query ON; Report shown; no changes made to system.
    (-R) Auto-Remove OFF; Remove/Uninstall action default to NO.
    (-S) System update OFF; No system files will be updated.
    (-U) backpac config update OFF; backpac files will not be updated.
    Sourcing from backpac config directory: /home/es/.config/backpac/tau
    Initializing.................Done
    GROUPS
    ============================================================================
    /home/es/.config/backpac/tau/groups
    GROUPS UP TO DATE: group listed in backpac and >80% local install:
    base base-devel xfce4 xorg xorg-apps xorg-drivers xorg-fonts
    GROUP PACKAGES; MISSING?: group member packages not installed:
    (base: nano)
    (xfce4: thunar xfdesktop)
    PACKAGES
    ============================================================================
    /home/es/.config/backpac/tau/packages
    PACKAGES UP TO DATE: packages listed in backpac also installed on system:
    acpi acpid acpitool aif alsa-utils augeas cowsay cpufrequtils curl dialog
    firefox gamin git ifplugd iw mesa mesa-demos mutt netcfg openssh rfkill
    rsync rxvt-unicode sudo terminus-font vim wpa_actiond wpa_supplicant_gui
    xmobar xorg-server-utils xorg-twm xorg-utils xorg-xclock xorg-xinit xterm
    yacpi yajl youtube-dl zsh
    AUR UP TO DATE: aur packages listed in backpac also installed on system:
    flashplugin-beta freetype2-git-infinality git-annex haskell-json
    package-query-git packer wpa_auto xmonad-contrib-darcs xmonad-darcs
    AUR NOT IN backpac: installed aur packages not listed in backpac config:
    yaourt-git
    FILES
    ============================================================================
    /home/es/.config/backpac/tau/files
    MATCHES ON SYSTEM/CONFIG:
    /boot/grub/menu.lst
    /etc/acpi/handler.sh
    /etc/rc.conf
    /etc/rc.local

    firecat53 wrote:I think your plan for handling an AUR_HELPER is good. If AUR_HELPER is defined by the user, then either you might need a list of major AUR helpers and their command line switches so you can pick the correct switch for what needs to be done (most use some variation of -S for installing, but not all), or have the user define the correct switch(es) somehow for their chosen AUR helper.
    That's a good idea. I'll add that to my AUR refactoring todo.
    I also found directory tracking to be a weakness in other dotfile managers that I tried. I think you would definitely have to recursively list out the contents of a tracked directory and deal with each file individually. Wildcard support would be nice...I just haven't personally found a use case for it yet.
    I've been thinking that I could just add the directory and scan through it for any non-default attribute files. If those are found then they get automatically added to the files list. That's pretty close to what etckeeper does.
    Edit: I just compiled the dev version and removed my comments for already fixed things...sorry!
    The master branch should have those fixes as well, but I didn't update the version number in the package build. I'll have to do that.
    1. Still apparently didn't handle the escaped space for this item: (the file does exist on my system)
    Ok, good to know. This wildcard directory business will require some new code and refactoring so I'll also rework my filenames handling.
    2. Suggestion: you should make that awesome README into a man page!
    I was working on one (the pkgbuild has a commented out line for the man page) but I had to leave it for later. Definitely want a man page. Once this stabilizes and I'm sure there aren't any big structural changes, I'll convert it to man format.
    3. Suggestion: add the word 'dotfile' into your description somewhere on this page, the github page, and in the package description so people looking for dotfile managers will find it. You could also consider modularizing the script into a dotfile manager and the package manager, so people on other distros could take advantage of your dotfile management scheme.
    I actually have a different script for dotfile management that doesn't touch packages, but there is definitely overlap with this one. That script isn't released yet, though, and if people find this useful for dotfile management that's great. I'll add that in.
    4. Suggestion: since -Q is a read-only operation, why not just make it run with -f automatically to avoid the prompt?
    Originally, running backpac without any command line options produced the Query output. I was concerned that since it is a utility that can potentially overwrite system files, it is important to give users a clear statement prior to execution about what will be done. Since the Query output is essentially the same as the Update and System reports in format and content, I wanted to be explicit about the Query being a passive no-change operation. The current command line options aren't set in stone though. If you feel strongly about it being different, let me know.
    Long answer to a short question
    5. Another suggestion: any thought to providing some sort of 'scrub' function to remove private information from the stored files if desired? This would be cool for publishing public dotfiles to github. Perhaps a credentials file (I did this with python for my own configs). Probably detecting email addresses and passwords without a scrub file would be rather difficult because dotfiles come in so many flavors.
    Yes, absolutely. In fact, if you look at the lib/local file (pretty sure it's in both master and dev branches in this state) you'll see some references to a sanitize function. The idea there is that the user will list out bash associative arrays like this:
    SANITIZE_WPA_=(
    [FILE]='/etc/wpa_supplicant.conf'
    [CMD]='sed s/expungepattern/sanitizedoutput/g'
    Question: am I missing an obvious option to remove a file from the files.d directory if I delete it from the files list? Or do I have to delete it manually? It might be helpful to add a section to the README on how to update and delete dotfiles from being tracked, and also a more detailed description of what the -b option does (and what is actually created when it's not used).
    You are only missing the function I didn't finish. There should be either dummy code or a TODO in the backpac main script referencing garbage collection, which isn't difficult but I just haven't finished it. The idea being another loop of "hey I found these old files in your files.d, mind if I delete them?" It's on my list and I'll try to get it in asap.
    And finally, just out of curiosity, why did you choose to actually copy the files instead of symlink like so many other dotfile managers do?
    git not following symlinks, hardlinks also out for permissions issues (git wouldn't be able to read the files, change them, etc.)
    I definitely would prefer to not make an entire copy of the file, but I haven't come up with a better option. Shout with ideas, though. Also, if there is a way around the link issues I noted above, let me know. I don't see one but that doesn't mean it's not there.
    edit: I think a Seattle area Arch meetup would be cool! Perhaps coffee someplace? Bellevue? U-district? Anyone else? BYOPOL (bring your own pimped out laptop)
    A general meetup sounds good. I was also thinking it would be fun to do a mini archcon with some demos.

  • Are there any recovery tools for optical media?

    I've had a couple of DVD's  turned in to coasters recently and I was hoping that Linux could come to my rescue with some awesome 'raw-read' / recovery tools for optical media?
    I burnt the discs in Windows 7 using InfraRecorder on a brand new Plextor drive (a 'PX-891SA') and these are my first attempts to burn discs with it so I'm a bit worried it could be a hardware fault!
    I started off trying to back up some files to clear some space on my hard-drive, burning them to a dual-layer dvd, but the software kept failing to even start the burn process -- that was until I realised that I needed to run the program "as Administrator" (in Windows). It did then manage to burn the disc (I was using "Session-at-Once" mode) however when it came to the end of the burn process / the start of the 'verify data' phase it threw an error saying the write process had failed! Windows couldn't access the disc / data and (I think) still thought the disk was empty. So I restarted the burning application and used a function to just 'close' the whole disc. This seemed to work because Windows was then able to recognise the disc and read the data, although accessing the disc was slow and sounded 'stutter' ish.
    I then tried to write a single layer dvd to finish off the last couple of gig I had to back up, but again the software completed the burn phase only to report a 'fail' error right at the end (I'm assuming) before finalising the disc. Although this time I'd tried to create a multi-session disc by using the '2XXA' (or something) 'Multi-Session' iso9660 format. The result was that Windows couldn't read the disc and still thought it was empty, even after trying a forced 'close disc' operation! -- Linux however could read the disc and see that the data had in fact been burnt to the disc (although I've not yet tried copying it in Linux!).
    Finally I needed to burn a load of archives to a couple of discs (totaling ~9GiB) -- starting with a dual-layer dvd first everything seemed to be working fine, the burn process started okay and the write buffer was full etc., until at what I assume was the layer change-over it all came to a grinding halt and threw another 'write operation failed' error, only this time I can't even mount the disk in Linux....see 'dmesg | tail' output below:
    [ 5758.353005] ISO 9660 Extensions: Microsoft Joliet Level 3
    [ 5758.353015] ISOFS: Interleaved files not (yet) supported.
    [ 5758.353016] ISOFS: File unit size != 0 for ISO file (1472).
    [ 5758.353019] ISOFS: changing to secondary root
    [ 5758.353035] ISOFS: Interleaved files not (yet) supported.
    [ 5758.353036] ISOFS: File unit size != 0 for ISO file (1536).
    [ 5758.353037] isofs_fill_super: root inode is not a directory. Corrupted media?
    Does anyone have any thought or suggestions that might help?
    - Is it likely to be a hardware fault? (I bought a Plextor because I thought they were one of the best!)
    Maybe even if I could just close the session all would not be lost?
    -- Thanks in advance --
    >>>> UPDATE >>>>
    I tried to use ddrescue....it was fine until it got to the end of the disk and then it sounded like the drive was chewing on a bag of pork scratchings
    $ sudo ddrescue -v /dev/sr0 /mnt/odd/ddr.file ~/ddr.log
    GNU ddrescue 1.16
    About to copy 4029 MBytes from /dev/sr0 to /mnt/odd/ddr.file
    Starting positions: infile = 0 B, outfile = 0 B
    Copy block size: 128 sectors Initial skip size: 128 sectors
    Sector size: 512 Bytes
    Press Ctrl-C to interrupt
    rescued: 4028 MB, errsize: 983 kB, current rate: 0 B/s
    ipos: 4028 MB, errors: 1, average rate: 3310 kB/s
    opos: 4028 MB, time since last successful read: 10.8 m
    Splitting failed blocks...
    Interrupted by user
    ...it was taking forever and getting nowhere so I just pressed 'ctrl+c'.....here's the log file:
    # Rescue Logfile. Created by GNU ddrescue version 1.16
    # Command line: ddrescue -v /dev/sr0 /mnt/odd/ddr.file /home/shodan/ddr.log
    # current_pos current_status
    0xF0213000 /
    # pos size status
    0x00000000 0xF0208000 +
    0xF0208000 0x0000B000 -
    0xF0213000 0x0000D000 /
    0xF0220000 0x00000400 -
    0xF0220400 0x0001FE00 /
    0xF0240200 0x00000400 -
    0xF0240600 0x0003FE00 /
    0xF0280400 0x00000400 -
    0xF0280800 0x00077600 /
    0xF02F7E00 0x00000200 -
    When I tried to mount the output file using $ sudo mount -ro loop /mnt/odd/ddr.file /mnt/vodd I got the same error as with the actual disc (natch)...
    mount: wrong fs type, bad option, bad superblock on /dev/loop0,
    missing codepage or helper program, or other error
    In some cases useful info is found in syslog - try
    dmesg | tail or so
    Last edited by wwgfd (2012-12-14 13:18:41)

    @chris_l -- yeah I was worried that might be the case, still I'm not sure if that explains why it keeps failing the burn process part way through?
    I've tried this and that in Linux but without much luck. I'm still quite a noob with a lot of Linux things and this is the first time I'm trying anything to do with optical disks/drives etc. so I'm learning as I go...
    The disk that failed in Windows but was 'readable' in Linux did just pass a run through ddrescue -- i.e. it created an output file okay and did not report any errors in the process.
    However when I mounted the resulting image and tried to copy the files across the first folder reported an "input/ouput error" (just using gui copy+paste in Thunar file manager)
    ...the other folders did copy okay though!
    It seems as though it's the lead-in / lead-out and disc finalizing process that are the main fail points with the burn operation. I don't know if InfraRecorder is any good or likely to be the problem? Anyway the drive cam with Nero (inc. Linux version) so might give either or both of these a go too, if they don't work either then at least I wont be short of something to rest a mug o' tea on
    @ chris_l (again) -- the Plextor drives do have a "Vari-Rec" feature which allows a user to 'tweak' the laser power / output level so I might try this too.
    @alphaniner -- thanks I've not seen this package before, I'm guessing it's a bit like cdparanoia but for dvds?!
    Thanks for the help!
    (pre-post) UPDATE >>> I've just mounted my Windows partition  in Linux so I'll have a crack at burning the files using cdrtools before I resort to throwing the dvd drive out of the window and chasing it around the yard with a lump hammer

  • Is the iCloud a backup tool for your computer?  How do I get my data back -- off the iCloud?

    Is the iCloud a backup tool for your computer?  ---- Apple care suggested using iCloud for a backup to my documents.  I do not have an iphone, ipad or laptop.  Just an iMac.  Nothing to share the data with.
    Now I am unable to access some of my documents on my computer and saving documents is cumbersome because it always wants to save it in iCloud, not my computer.  The documents I am unable to open are my old MS-Word.  One is a macro enabled word doc.   I am using os x mountain lion.
    After over an hour chat with apple care, the issue is not resolved.  I did not have more time to resolve the issue today and it will be continued next week.  Per apple care, I do not have any documents in the iCloud however I am almost at the free capacity with  contacts and documents.
    My question is -- If I purchase backup from a clould provider - other than apple, back up my files in documents in their cloud (make sure I can open them) and then delete the link to iCloud.  Would I preserve my data? Has anyone tried this?  I am open to suggestions.
    Thanks,
    Phyllis

    Is the iCloud a backup tool for your computer?  ---- Apple care suggested using iCloud for a backup to my documents.
    Not really. You can store "Documents in the Cloud" but this applies only to documents created by Pages, Numbers, Preview, Keynote, and Microsoft Office documents created on your computer stored in iCloud's "iWork". Within that limited purpose, iCloud works very well.
    My question is -- If I purchase backup from a clould provider - other than apple, back up my files in documents in their cloud (make sure I can open them) and then delete the link to iCloud.  Would I preserve my data?
    I'm not quite sure I understand what you mean by "purchase backup from a cloud provider" but if you're considering a service such as Carbonite, I don't advocate any of them. To back up your Mac, use your own backup device - Time Capsule or an external hard disk. It's inexpensive and reliable: two elements essential for a backup strategy, that no subscription service can offer.

Maybe you are looking for

  • Can't load my PS to a new computer from my backup disk.....?

    My MacBook died. I had the entire adobe suite on it. I have it backed up tho on an external drive. I am trying to put it on the used Macbook I purchased but it won't load from the external drive to the used Macbook I purchased. Is there a file or fil

  • How to make an apk from an indesign document

    hello, I made a newspaper in indesign. And it works fully on the Ipad. now is my question if you can make an apk file from an indesign file so you can see my site on a android tablet. My android tablet doesn't have an android market so I can't downlo

  • [ADF BC | ADF Faces] Updating a table with values updated in transaction

    Summary: A table is based on a view with a bind variable. Based on some action, I update rows in the view so that they no longer satisfy the query. How do I update the table to reflect this? I have an ADF Faces page with a multi-select table componen

  • Quick Question: package names com.whatever.blah.ClassName

    I want to put a bit more time aside this week to explain things I see in Java to my old brain (I don't quite 'get it' like you younger people :-)) I often see things like this in example code: import com.macfaq.io.SafeBufferedReader;and it's crossed

  • Can't Sign in to my CC account

    Hi, I can't sign in to my CC account. This fix http://helpx.adobe.com/creative-cloud/kb/unable-login-creative-cloud-248.html still does not work? Thanks