Systemd migration, some questions esp. rc.local

Hello
Although I'm not a fan of dzillions of config files I decided to try systemd.
Most of the things seem to work fine but some things are unclear to me:
1.) Things are working fine with rc-local.service, however I tried to make unit files for my entries in rc.local:
hdparm -B 254 /dev/sda
My unit file:
[Unit]
Description=hdparm control script
After=local-fs.target
[Service]
ExecStart=/usr/local/sbin/hdparm.sh
[Install]
WantedBy=multi-user.target
Here I'm not sure about the After=local-fs.target. AFAIU I've to make sure that /dev/sda is ready before using hdparm. I read in another thread to use After=dev-sda.device. But systemctl does not show me dev-sda.device, instead it's: sys-devices-pci0000:00-0000:00:1f.2-ata1-host0-target0:0:0-0:0:0:0-block-sda.device Should I enter this in the unit file ?
Could I enter local-fs.target instead?
2.) Then I've got a big problem with vgaswitcheroo. There seems no appropriate solution to this ? I tried to put this in a unit too, resulting in  some.. well not complete crashes but very grave errors on boot...
Again, could I add After=systemd-user-sessions.service
but Before=xinit.service ?
3.) Then I use acpi_call (Yes, I used both vgaswitcheroo and acpi_call, first vgaswitcheroo then acpi_call in rc.local. I know it's doubling things but worked fine for me, after a kernel update acpi_call is not immediately availiable, AFAIK I have to recompile it first, that's when vgaswitcheroo came to it's use.) Same problem here, what should I use for After= Before etc... ?
4.) Then I've a very small script that get's started at boot and should run forever, this failed:
[Unit]
Description=brightness control script
[Service]
Type=forking
ExecStart=/usr/local/sbin/brightness.sh
[Install]
WantedBy=multi-user.target
Ok this one's not very important...
I copied rc-local.service to /etc/systemd/system. As said above everything works with this well at the moment.
Will this still work when I switch to systemd completely ?
5.) Then nfs-common.service failed:
nfs-common.service - Legacy unit for nfs-common
Loaded: loaded (/etc/rc.conf)
Active: failed (Result: exit-code) since Sun, 16 Sep 2012 22:32:04 +0200; 8s ago
Docs: man:arch-daemons(8)
Process: 1490 ExecStart=/etc/rc.d/nfs-common start (code=exited, status=1/FAILURE)
CGroup: name=systemd:/system/nfs-common.service
Sep 16 22:32:04 hplaptop nfs-common[1490]: [53B blob data]
However I can mount and unmount nfs partitions perfectly fine without it. This is a client only ! So, I don't really care about this one. In the wiki nfs it says nfs-common needs to be started, is this obsolete now ?
6.) Ah yes, same for cpufreq. This one's really not needed I think. What was it needed before ? (Ok, I saw that the wiki is in rework.)
7.) And one little thing I noticed: I use xinit.service as explained in the wiki/systemd (thanks) I copied graphical.target from lib to etc and changed Wants= from display-manager.service to xinit.service , I think that's correct ? (Did not reboot yet though..) Maybe this could be mentioned in the wiki if so.
Regards
Last edited by rebootl (2012-09-16 20:56:52)

skanky wrote:
Roken wrote:
Mine works as a one shot service - the unit loads a script which loads my server in a gnu screen session. However, the script is written to monitor various aspects of the server, and loops continuously every 30 seconds, so the script never exits.
I guess since you only need the script to run through once at startup you should be good to go, too.
Okay, that's great thanks.
Edit: meant to say, I'll give it a try.
Worked a treat. I should probably have just tried it, but I'd just spent half an hour or so getting another service file working correctly, thought it worth checking first.
Many thanks Roken.

Similar Messages

  • Migrating from Aperture to LR - some questions

    Hey all,
    I'm a current user of Aperture and I'm considering going to LightRoom but I have some questions/concerns.  I downloaded the demo and I'm playing with that but given my lack of knowledge, I wanted to ask some questions to get a feel for the product.
    I'm not sure if my current aperture library is considered large or not but I have about 75gig of images to which I have them organized by project.  Since the birth of my girls I've been focusing a lot on them - go figure   Anyways the setup I have for them is a good example in how I typically setup up my images within Aperture. I'm not sure how easy (or not so easy this can be done with LR)
    As you can see I have a folder called twins, then the year and finally the month.  The purple icons are smart albums, basically images that pass muster and I upload to smugmug.
    I basically would like to have something along those lines in LR - have my images organized by event, date (if need be) and then a smart album of images that I'll be publishing. 
    I'm still getting used to the interface, there are things I definitely like and some things I need to work on a little more.  Any advice on moving from aperture to LR would be helpful though I do want to qualify that I'm still checking it out and have not made a decision to jump ship quite yet.
    Finally are there any deals or such to lower the price of LR?

    "while LR shows you your folders, offers the temptation of trying to categorise by them (resist it!), and allows you to organise and categorise by collections. You you aren't forced to use virtual folders, but can and should."
    So you're saying don't organize the images in the physical folders - that's going to be a hard habit to break.  In playing with LR last night, I created a separate catalog called twins09 and I exported my masters (with the sidecar) out of aperture by folder, i.e., January09 goes into a separate folder, as does June09 etc.  I then created my smart collections of keepers by using the 5 stars and date constraints that LR provides, i.e, greater the 20090430 and less then 20090601 for May09 images.
    What you're recommending is to create a twins 2009 folder and store all images there and then use collections/smart collections to organize the images.  Interesting. Given the habits I've picked up using aperture that will be a difficult adjustment.

  • Some questions on versioning and synchronizing metadata

    Hy all!
    I am quite new to warehousing and Oracle Warehouse Builder, and so i would have some questions regarding on some common issues. I would appriciate if you guys would who have experience in this domain to share some good practice knowledge :)
    I am using OWB 10.2
    So first of all i would like to know if you have some proposal of the way of versioning control and synchronizing projects between team memebers when working on a bigger project, team memebers that don't work on the same repository (cause i saw that OWB has an integrated multiuser support for handeling object locks and user sessions).
    I saw that a way of migrating data from one place to a nother is using the import/export options integrated in OWB. This creates mdl files wich are some kind of "dumps" of the metadata informations, but the thing with these mdl files wich i don't think is a good way to synchronize is that first of all the .mdx and .xml files contained in the .mdl (wich is kind of a zip) contains many informations in it (like creation date, some timestamps, etc) wich are always updated when exporting, and if synchronizing these files maybee using CVS, we always will get differences between the files alltough they would contain the same thing, only timestamps changed.
    Then a nother issue with this, is that we could have 2 alternatives: dump the whole project, wich is odd to have to synchronize a single file between users, especialy on a big project, then the orher way would be doing for each object from the project (each mapping, each table, etc) an separate .mdl filem then to synchronize each file of each object, wich will be unefficient on reimporting each file in part.
    So please if you can share the way you work on a big project with many implementers with OWB, i would really appriciate.
    A nother thing i would like to know is: is there a way to generate from an existing project (like one created with OWB) the OMB commands dump (maybee in a tcl script)? Cause i saw that the way the exeprienced users implement warehousing is using TCL with OMB language. I downloaded the example from oracle for warehouse project, and i saw that is entirely made from tcl scripts (so no mdl file involved). And this i think would be nice, to have the OMB commands generated from an existing projects.
    I see this OWB projects like a database wich can be built up from only OMB commands and OWB a graphical tool to do this (same as constructing a database only from DDL commands or using SQL developer to do this), this is why i am asking about a way of dumping the OMB commands for creating an OWB project.
    Please give me some advices, and correct me if i sad some dumb things :D but i really am new to warehousing and i would really appriciate if you guys with experience could share some informations.
    Thank you verry much!
    Alex21

    Depends. Having everyone working on the same project certainly simplifies things a lot regarding merging and is generally my preference. But I also recognize that some projects are complex enough that people wind up stepping on each other's toes if this is the case. In those cases, though, I try to minimize the issue of merging changes by having common structural objects (code libraries, tables, views, etc) retained in a single, strictly controlled, central project schema and having the developer's personal work areas reference them by synonym, thus being unable to alter them to the detriment of others.
    If they want to change a common object then need to drop their synonym and make a local copy which they can alter, and then there is a managed process by which these get merged back into the main project schema.
    This way any changes MUST go through a central schema, we can put processes in place to notify all of the team of any impending changes, and can also script updates across the team.
    Every hour a script runs automatically that checks for dropped synonyms and notifies the project leader. It especially checks for two developers who have built local copies of the same object and notifies each that they need to coordinate with each other as they are risking a conflict. When a structural change is submitted back to the central shared schema, it is added to a batch that is installed at end of business and a list of those impending changes is circulated to the team along with impact analysis for dependencies. The install script updates the main schema, then also drops the local copy of the object in the developer's schema who made the change and re-establishes the synonym there to get back to status quo for the change monitoring. Finally, it then updates itself in all of the developer areas via OMBPlus. So, each morning the developers return to an updated and synched environment as far as the underlying structure.
    This takes care of merging structural issues, and the management of the team should minimize other metadata merging by managing the worklist of who is to be working on a given mapping or process flow at a given time. Anyone found to be doing extraneous changes to a mapping or process flow when it is not in their job queue without getting pre-approval will be spoken to VERY firmly as this is counter to policy. And yes, OWB objects such as mappings are then also coordinated to the central project via import/export. OMBplus scripts also propogate these changes daily across the team as well.
    Yep, there is a whole lot of scripting involved to get set up.... but it saves a ton of time merging things and solvinv conflicts down the road.
    Cheers,
    Mike

  • Some questions on SAP DMS

    Hello
    some questions on SAP DMS:-
    - can versions of same document be shown graphically. ?
    - can links be created for documents. can we make sure that links are not broken when document is moved ?(links to document on 1st location shall not be broken)
    - is there a notification mechanism(user can recieve e-mail when document gets changed)
    - is there a review mechanism, before doc. is archived.(sort life cycle)
    Thanks

    Hi Shovon,
    1. can versions of same document be shown graphically?
    You can display all existing versions of a document info record by using the menu 'Extras' >> 'Versions'. Then you will see all existing document version for this specific document number. Further you will see which is the current version at the moment.
    2. can links be created for documents. can we make sure that links are not broken when document is moved ?(links to document on 1st location shall not be broken)
    I'm not quite sure what you mean by links to documents. If you mean linking document info records together, I can confirm that this works also in DMS. You can maintain the object DRAW for object links in the customizing and then you will be able to link one document info record to another. The data for all document info records is stored in table DRAW.
    If you are talking about originals which are linked to document info records, please note that if you check in an original the system creates a unique LOIO- and PHIO-ID which is stored in the Content Server and in the database tables DMS_DOC2LOIO and DMS_PHIO2FILE. But there are also different reports to migrate original data from one content Server to another (e.g. DMS_RELOCATE_CONTENT).
    3. Is there a notification mechanism(user can recieve e-mail when document gets changed)
    Yes you can activate an event type linkage for object
    DRAW     CHANGED     DOKST
    DRAW     CHANGED     DWNAM
    in transaction SWE2. Here you just have to set the flag for 'Type linkage' (linkage activated). Then everytime the document is changed a notification is sent to the entered USER of this document info records.
    4. Is there a review mechanism, before doc. is archived.(sort life cycle)
    Normally there is no special lifecycle. You can control this by maintaining a released status and if there are more versions always version with the release status is the current one. But there is no specific date when the document info record expires or something like that.
    Best regards,
    Christoph
    P.S.: Please reward points for useful information.

  • Some questions on JFormattedTextField

    I more or less manage to work with JFormattedTextField, but have not mastered this control fully. I hope that someone can help me with some questions I still have.
    JFormattedTextField field = new JFormattedTextField(); // no-args constructor
    double d = 1.23456; // primitive type will be autoboxed to Double
    field.setValue(d); // AbstractFormatterFactory based on DoubleNow, when the field has the focus, a text value of +1.2+ is displayed. As soon as it looses focus, the text value is changed into +1.235+. I find that weird behaviour and would rather see it the other way around or see the same text in both situations. Can anybody explain me what's happening here? I do not observe this behavior when using one of the other constructors.
    Another problem has to do with internationalization. The Dutch locale, for example, uses a comma as a decimal separator. Users having a scientific background, however, often override the default settings and use a period as a decimal separator (on Windows by changing the regional options). Is there a way of using whatever separator a user has chosen for his system when formatting numbers in Java? I know you can use the DecimalFormatSymbols class to change symbols for a specific locale, but that does not take into account user-specific settings.

    Hi.
    Try this:
    DecimalFormatSymbols fs = new DecimalFormatSymbols();
        fs.setDecimalSeparator('.');
        fs.setGroupingSeparator(',');
    DecimalFormat format = new DecimalFormat("##.#", fs);
    AbstractFormatter formatter    = new NumberFormatter( format );
    AbstractFormatterFactory ff = new DefaultFormatterFactory( formatter, formatter, formatter);
    JFormattedTextField field = new JFormattedTextField();
    field.setFormatterFactory(ff);

  • Question about /usr/local/bin in Mavericks ...

    I just did a clean install of Mavericks on a 2010 MacPro. I understand Mavericks does not replicate this path (/usr/local/bin) when doing a clean install. Prior to this I was running 10.7.xx and I had a few compiled binaries installed in the bin folder. How do I address this? Can I simply recreate the path manually without risk?
    Thanks.
    -paul.

    etresoft The 905
    Re: Question about /usr/local/bin in Mavericks ... 
    Jan 15, 2014 9:30 AM (in response to Paul Figgiani)
    So why would you expect /usr/local/bin to be there? If you install custom, low-level software like this on one system, it will not get migrated to the new system.
    As I stated in my original post I did not expect it to be there. All I'm asking is how do I adreess the issue? Do I recreate it manually without any risk of screwing something up?

  • Thinking about purchasing a Macbook, but I need some questions answered

    Hey everybody,
    I got a response to my Craigslist ad, so it's looking like selling my iMac G5 and buying a new Macbook might actually happen. My only problem is that I have some questions about what to do with my money. I'm not a power user, or someone who really plays games (other than my Nintendo emulators) so I already know that a Macbook is the right computer for me.
    (1) I don't have external storage, so how should I go about transferring my stuff? Buying the Macbook before selling my iMac is not an option because I just don't have that much money. I've read that some people buy the Macbook and then upgrade the HD, using the 60 Gig that comes with it as their external storage by buying a $20 case for it. I have about 50 GB taken up on my iMac, so I'm going to probably need around 100 GB at least. I know upgrading the memory doesn't, but does upgrading the HD void the warranty? Any brand recommendations for HD's? Heat and noise are very important to me. (7200 RPM is out)
    (2) I never really used the Superdrive on my iMac, but it was nice to know it was there. Is the performance difference from 1.83 to 2Ghz noticeable? I hear it's not, but I also read that under Rosetta it matters. Since Office isn't universal yet, I'd like to know. Can anyone verify this?
    (3) I read on MacOSRumors that Leopard is probably going to do some nice little graphic effects when it's released, and that the next revision of the Macbook Core Duo will probably still have integrated graphics. Since I don't plan on doing this again for at least another year, do you think the Intel graphics will be enough to keep up for a little while?
    (4) Anybody have some experience with selling a computer on Craigslist? I'm smart enough to request cash, local buyers only, and have them meet me at my house to try the computer out. Any advice on how to make sure this runs smoothly? I might end up leaving some things on the computer for the buyer. Anything that I should make sure to delete (i.e. preferences, passwords, etc.?) A checklist of what I should remember to do would really help.
    I think that about covers all of it...Thanks in advance.

    I sold my computer on Craigslist. No problems. I think it's only really dodgy if you are buying one from there - then you should be very careful!
    I'm not sure exactly the best way to go about deleting your stuff, but you have the right idea: get rid of your preferences and passwords, and your internet history as well.
    I'm not sure how to look at 'Keychain', but you might want to make sure that this is empty.
    In terms of processor speed, I can't help you: I have just got the 2Ghz, but partly because not having a DVD burner isn't an option for me. As file sizes and your storage needs increase, your need for a DVD burner will probably increase as well.
    Leopard is not far away now, and I would guess that Apple would make sure that their new OS will work adequately on their new computers. I would be surprised if there is much of a noticeable difference in anything but the most graphics heavy applications. Could be wrong about this, but I think that the basic GUI will perform well on any MacBook.
    In terms of moving info from your iMac... the cheapest thing apart from buying the MacBook before you sell is borrowing a HD or iPod from a friend. 50G isn't THAT much info, you could always burn it to DVD-Rs in a pinch.

  • Grüße Fräulein, Sturm-Verbot-Führer, Hello, we are with some questions

    Grüße Fräulein, Sturm-Verbot-Führer, Hello, we are with some questions about the thought of GBP 77.336, apparently some of our local chapters of chicken producers in Mexico, should be changed, put a small bed of sand and tents in the city. One of the Mexicans, all connections Poncho sniffing glue every time you see him with a hatchet to cut the data cable with an ax. We called immigration and the nature of the problem and Poncho and his producer was a kind of rebellion, threw the chicken **bleep** on the government and is working like crazy NAP. Bryan and our engineers, it seems that the costs of these persons from Mexico, is excited because he said he misses his family in the Middle East, Verizon, I called our 1-800 number and were glad we had rotary phones , but said that someone would go to see these issues with Peter Nam and poncho, if you write an element upstream of the cut of radiation accelerate our DSLAM. Now, with all kinds of components, is accelerating as bowling, and then started kind of breakdown of carbon molecules locally is not always the Internet is like living in Mount St. Helens.

    I'm afraid I don't understand the question.
    If a forum member gives an answer you like, give them the Kudos they deserve. If a member gives you the answer to your question, mark the answer as Accepted Solution so others can see the solution to the problem.
    "All knowledge is worth having."

  • [Urgent] Some questions about OID/OSSO 10g - 11g upgrade

    Dear all,
    We are under doing upgrading assessment of OID/OSSO 10g to 11g for a customer. After reviewed the 'upgrading guide', we still have some questions as below:
    1.     Whether the ‘10g DIP profiles’ will be still available after the OID 11g upgrade? Currently there are some sync of AD<->OID and DB->OID.
    2.     Whether the ‘WNA’ function will still work after the upgrade?
    3.     Is there a big change of OID API from 10g to 11g? If so, I think a big effort maybe on application modification.
    4.     I found that there is a OAM Basic version for OSSO 10g upgrade. So if this OAM Basic will migrate the OSSO configuration (like external application) automatically, or it must be re-configed after the upgrade?
    5.     Currently customer config OID 10g as BPEL/ESB’s identity store. So does BPEL/ESB 10g is certificated with OID 11g also? I didn’t find the certification so far.
    Thank you in advance and any comment are welcome.

    Dear all,
    We are under doing upgrading assessment of OID/OSSO 10g to 11g for a customer. After reviewed the 'upgrading guide', we still have some questions as below:
    1.     Whether the ‘10g DIP profiles’ will be still available after the OID 11g upgrade? Currently there are some sync of AD<->OID and DB->OID.
    2.     Whether the ‘WNA’ function will still work after the upgrade?
    3.     Is there a big change of OID API from 10g to 11g? If so, I think a big effort maybe on application modification.
    4.     I found that there is a OAM Basic version for OSSO 10g upgrade. So if this OAM Basic will migrate the OSSO configuration (like external application) automatically, or it must be re-configed after the upgrade?
    5.     Currently customer config OID 10g as BPEL/ESB’s identity store. So does BPEL/ESB 10g is certificated with OID 11g also? I didn’t find the certification so far.
    Thank you in advance and any comment are welcome.

  • [Urgent] Some questions about Oracle Portal 10g - 11g upgrade

    Dear friends,
    We are under doing upgrading assessment of Oracle Portal from 10g to 11g. After reviewed the 'upgrading guide', we still have some questions as below:
    1. Whether the 'Instant Portal' feature remain in Oracle Portal 11g? If yes, how to migrate the Instant Portal 10g sites? If no, any workaround?
    2. Whether the old 10g Page Group and Pages will be still available after 11g upgrade? Also can customer import the old 10g transportset into Portal 11g?
    3. How about the customized Portal 10g objects (such as customized template, style, attributes) after 11g upgrade? Still be usable?
    4. Customer use PDK to develop many portlets in 10g. Can they use JDeveloper 11g to migrate their old project to 11g automatically?
    Thank you in advance and any comment are welcome.

    1. Whether the 'Instant Portal' feature remain in Oracle Portal 11g? If yes, how to migrate the Instant Portal 10g sites? If no, any workaround?Instant Portal is not available anymore in Oracle Portal 11g. The Instant Portal page groups will be migrated to standard Portal page groups. Maintenance of the Instant Portal can be done with the standard Portal tools. The Instant Portal tools are not available anymore.
    Whether the old 10g Page Group and Pages will be still available after 11g upgrade? Also can customer import the old 10g transportset into Portal 11g?Old 10g Page Groups are migrated to Portal 11g. They will still be available after the upgrade.
    Export/Import through transports sets has always been limited to instances of the same version. It is not supported between versions as documented in the Administration Guide :
    [11.2.1|http://download.oracle.com/docs/cd/E14571_01/portal.1111/e10239/cg_imex.htm#CCJBCCGD] System Requirements
    Before exporting and importing content, ensure that your system meets the minimum system requirements, as described in this section.
    Notes:
    * Export and import functions only within the same release of Oracle Portal and the same patch release, for example, release 10.1.4 to release 10.1.4 or release 11.1.1 to release 11.1.1. You cannot export and import between two different releases, such as release 10.1.2 to release 10.1.4 or release 10.1.4 to release 11.1.1.
    How about the customized Portal 10g objects (such as customized template, style, attributes) after 11g upgrade? Still be usable?Customized objects get migrated as well. Certain types (e.g. PL/SQL item types) need to be checked after upgrade as the behavior of the PL/SQL code may differ between database versions. This is particularly of concern when the 10.1.4.x Portal uses a 10.1 database. A database upgrade to either 10.2, 11.1 or 11.2 is necessary as Portal 11g is not supported with RDBMS 10.1.
    Customer use PDK to develop many portlets in 10g. Can they use JDeveloper 11g to migrate their old project to 11g automatically?Never done this, but the Portal framework is able to consume JPDK providers which are running in older versions of the toolkit. If the providers are running in standalone OC4J containers, you would be able to upgrade the framework and keep the providers in their OC4J containers. This will allow you to focus on the framework first and worry about your providers later.
    Thanks,
    EJ

  • Hi team , I have some questions/Problem for my product apple (iPad,iPhone) , I want to employee speak and type thai language

    Hi team , I have some questions/Problem for my product apple (iPad,) , I want to employee that can  speak  or response me in thai language
    1. ผม อาคเนย์  พำนักอยู่ประเทศไทย กรุงเทพฯ  มีปัญหาสอบถาม ดังต่อไปนี้
       - กระผมได้ทำการตัดบัตร เครดิต เพื่อซื้อเกมส์ผ่าน itune store ผ่าน apple itune ID : misskor.yaprom@*** เพื่อซื้อเกมส์ Eleves Realm ในวันที่18 ก.ค. 56 เวลา 17.07น. ซึ่งทางบัตรเครดิตได้แจ้งเรียกเก็บเงินมายอดเงิน 39.99$ ซึ่งในระบบจริงๆ ทางกระผมต้องการตัดในยอด 99.99$ แต่พอได้ประสานงานไปยังธนาคาร ได้รับการแจ้งกลับมาว่า ได้ทำการตัดบัตรในยอดเงิน 39.99$ ซึ่งในความเป็นจริงนั้น กระผมไม่ได้สั่งซื้อเกมส์ในยอด 39.99$ ซึ่งในยอด 99.99$ นั้นพยายามตัดในระบบบัตรเครดิตอยู่ แต่ทางกระผมได้ยืนยันกลับไปว่าไม่ให้ระบบตัดนะครับ เพราะว่าเนื่องจากมีปัญหาในการชำระเงินระหว่าง Apple itune store อยู่
       - ทั้งนี้ขอให้ทางเจ้าหน้าที่ประสานงานตรวจสอบ apple itune ID : misskor.yaprom@*** เพื่อซื้อเกมส์ Eleves Realm ตามที่ได้ให้รายละเอียดโดยด่วนว่าเป็นเพราะว่าระบบมีปัญหาหรือว่ามีอะไรเกิดขึ้นในข ั้นตอนการชำระเงินครับ
    รบกวนประสานงานกลับมายังกระผม อาคเนย์ ที่หมายเลขโทรศัพท์มือถือ +**** / reply feedback  email : lekod1@*** โดยด่วน ในวันศุกร์ที่ 19 ก.ค. 2556 ครับ
    ขอบคุณครับ
    อาคเนย์  อุดปิน
    กด
    <Edited By Host>

    Google translation:
    พนักงานของ iTunes Store จะไม่ได้อ่านข้อความในเว็บบอร์ดนี้ ถ้าคุณต้องการความช่วยเหลือสำหรับปัญหาที่มีใน iTunes Store, คุณจะต้องติดต่อกับพวกเขาผ่านทางแบบฟอร์มเว็บนี้:
    http://www.apple.com/emea/support/itunes/contact.html

  • I have some questions regarding setting up a software RAID 0 on a Mac Pro

    I have some questions regarding setting up a software RAID 0 on a Mac pro (early 2009).
    These questions might seem stupid to many of you, but, as my last, in fact my one and only, computer before the Mac Pro was a IICX/4/80 running System 7.5, I am a complete novice regarding this particular matter.
    A few days ago I installed a WD3000HLFS VelociRaptor 300GB in bay 1, and moved the original 640GB HD to bay 2. I now have 2 bootable internal drives, and currently I am using the VR300 as my startup disk. Instead of cloning from the original drive, I have reinstalled the Mac OS, and all my applications & software onto the VR300. Everything is backed up onto a WD SE II 2TB external drive, using Time Machine. The original 640GB has an eDrive partition, which was created some time ago using TechTool Pro 5.
    The system will be used primarily for photo editing, digital imaging, and to produce colour prints up to A2 size. Some of the image files, from scanned imports of film negatives & transparencies, will be 40MB or larger. Next year I hope to buy a high resolution full frame digital SLR, which will also generate large files.
    Currently I am using Apple's bundled iPhoto, Aperture 2, Photoshop Elements 8, Silverfast Ai, ColorMunki Photo, EZcolor and other applications/software. I will also be using Photoshop CS5, when it becomes available, and I will probably change over to Lightroom 3, which is currently in Beta, because I have had problems with Aperture, which, until recent upgrades (HD, RAM & graphics card) to my system, would not even load images for print. All I had was a blank preview page, and a constant, frozen "loading" message - the symbol underneath remained static, instead of revolving!
    It is now possible to print images from within Aperture 2, but I am not happy with the colour fidelity, whereas it is possible to produce excellent, natural colour prints using its "minnow" sibling, iPhoto!
    My intention is to buy another 3 VR300s to form a 4 drive Raid 0 array for optimum performance, and to store the original 640GB drive as an emergency bootable back-up. I would have ordered the additional VR300s already, but for the fact that there appears to have been a run on them, and currently they are out of stock at all, but the more expensive, UK resellers.
    I should be most grateful to receive advice regarding the following questions:
    QUESTION 1:
    I have had a look at the RAID setting up facility in Disk Utility and it states: "To create a RAID set, drag disks or partitions into the list below".
    If I install another 3 VR300s, can I drag all 4 of them into the "list below" box, without any risk of losing everything I have already installed on the existing VR300?
    Or would I have to reinstall the OS, applications and software again?
    I mention this, because one of the applications, Personal accountz, has a label on its CD wallet stating that the Licence Key can only be used once, and I have already used it when I installed it on the existing VR300.
    QUESTION 2:
    I understand that the failure of just one drive will result in all the data in a Raid 0 array being lost.
    Does this mean that I would not be able to boot up from the 4 drive array in that scenario?
    Even so, it would be worth the risk to gain the optimum performance provide by Raid 0 over the other RAID setup options, and, in addition to the SE II, I will probably back up all my image files onto a portable drive as an additional precaution.
    QUESTION 3:
    Is it possible to create an eDrive partition, using TechTool Pro 5, on the VR300 in bay !?
    Or would this not be of any use anyway, in the event of a single drive failure?
    QUESTION 4:
    Would there be a significant increase in performance using a 4 x VR300 drive RAID 0 array, compared to only 2 or 3 drives?
    QUESTION 5:
    If I used a 3 x VR300 RAID 0 array, and installed either a cloned VR300 or the original 640GB HD in bay 4, and I left the Startup Disk in System Preferences unlocked, would the system boot up automatically from the 4th. drive in the event of a single drive failure in the 3 drive RAID 0 array which had been selected for startup?
    Apologies if these seem stupid questions, but I am trying to determine the best option without foregoing optimum performance.

    Well said.
    Steps to set up RAID
    Setting up a RAID array in Mac OS X is part of the installation process. This procedure assumes that you have already installed Mac OS 10.1 and the hard drive subsystem (two hard drives and a PCI controller card, for example) that RAID will be implemented on. Follow these steps:
    1. Open Disk Utility (/Applications/Utilities).
    2. When the disks appear in the pane on the left, select the disks you wish to be in the array and drag them to the disk panel.
    3. Choose Stripe or Mirror from the RAID Scheme pop-up menu.
    4. Name the RAID set.
    5. Choose a volume format. The size of the array will be automatically determined based on what you selected.
    6. Click Create.
    Recovering from a hard drive failure on a mirrored array
    1. Open Disk Utility in (/Applications/Utilities).
    2. Click the RAID tab. If an issue has occurred, a dialog box will appear that describes it.
    3. If an issue with the disk is indicated, click Rebuild.
    4. If Rebuild does not work, shut down the computer and replace the damaged hard disk.
    5. Repeat steps 1 and 2.
    6. Drag the icon of the new disk on top of that of the removed disk.
    7. Click Rebuild.
    http://support.apple.com/kb/HT2559
    Drive A + B = VOLUME ONE
    Drive C + D = VOLUME TWO
    What you put on those volumes is of course up to you and easy to do.
    A system really only needs to be backed up "as needed" like before you add or update or install anything.
    /Users can be backed up hourly, daily, weekly schedule
    Media files as needed.
    Things that hurt performance:
    Page outs
    Spotlight - disable this for boot drive and 'scratch'
    SCRATCH: Temporary space; erased between projects and steps.
    http://en.wikipedia.org/wiki/StandardRAIDlevels
    (normally I'd link to Wikipedia but I can't load right now)
    Disk drives are the slowest component, so tackling that has always made sense. Easy way to make a difference. More RAM only if it will be of value and used. Same with more/faster processors, or graphic card.
    To help understand and configure your 2009 Nehalem Mac Pro:
    http://arstechnica.com/apple/reviews/2009/04/266ghz-8-core-mac-pro-review.ars/1
    http://macperformanceguide.com/
    http://www.macgurus.com/guides/storageaccelguide.php
    http://www.macintouch.com/readerreports/harddrives/index.html
    http://macperformanceguide.com/OptimizingPhotoshop-Configuration.html
    http://kb2.adobe.com/cps/404/kb404440.html

  • Some questions about configuration in MAX.

    Hello,everyone!
    I have some questions about configuration in MAX(I am a jackaroo for motion control development),I hope I can get your help.
    I use PCI-7344+UMI-7764+Servo amplifier+Servo motor,my MAX version is 4.2 and I use NI-Motion7.5
    My question as following:
    1,In Axis Configuration,for motor type,why I must select stepper but not servo?my motor is servo motor!If I select Servo,my motor can't run,I don't know why.
     If I select stepper,though motor can work but I can't test encoder in MAX.
    2,In Stepper settings,for stepper loop mode,why I must select open-loop but not close-loop?If I select close-loop,the servo motor doesn't work too.
    3,If I want my two servo motors run at different velocity,How shoud I do?It seems I just can set the same velocity in MAX for my two servo motors.
     My English is poor,Pls pardon me!I come from China.
    Thank you for your help!
    EnquanLi
    Striving is without limit!

    Hi,Jochen,
    Thank you for your kindly help!
    The manufacturer of the drive and motor that I am using now is Japan SANYO DENKI,drive type is RS1A01AA,motor type is R2AA06020FXP00.
    And I use position control mode,thehe encoder's counts per revolution is 131072.I set the electronic gear ratio to 1:1 for drive.
    Now,I can use Close-Loop to control the motor but still has some problems.When I configure it to run in closed loop mode, the motors behave strangely and never move to the target position.When I configure it to run in closed loop mode, the motors behave strangely and never move to the target position.The detail situation is as following
    1,Motor can't run.
    2, Or motor moves to a position, then moves in the same direction agian and eventually stops.
    Except for the  two points mentioned above,"Following Error" is  occured frequently,I don't know why.
    I am still not clear why I must set the motor type be stepper in MAX .
    And I have another question:what the relationship between the steps and the counts?They have the proportion relations?I notice that there are a section said like this in help document: For proper closed-loop and p-command operation, steps per revolution/counts per revolution must be in the range of 1/32,767 < steps/counts < 32,767. An incorrect counts to steps ratio can result in failure to reach the target position and erroneous closed-loop stepper operation.
    I am verry sorry I have too many questions!
    I am very appreciate for your kingly help!Thanks again!
    EnquanLi
    China
    Striving is without limit!

  • Some Questions on Adobe PDF Forms

    Hi,
    I have some questions on Adobe Forms development (especially Adobe Forms in ABAP)
    1)     In a form interface, in the Code Initialization can we use Object Oriented Syntaxes such as class method calls, etc?
    2)     Can we declare Global Data variables with reference to Class References / Interfaces
    3)     Can we use the latest ECC 6.0 enhancement framework to add code in the standard form interface to enhance / enrich some logic within the form? This is to avoid copying the form interface into a Z in order to enrich the form via coding?
    Please let me know if this is possible at all and how
    Will award points
    Thanking You in advance
    Regards,
    Aditya

    That is possible of course. You should use the PDFObject API for that requirements.
    1. define a xml schema with all data given by SAP system.
    2. create a form template which uses this schema as data connection
    3. implement a java application (i.e. J2EE application) which:
    - gets the data from SAP system and generates some pdf files (using PDFObject API)
    - reads pdf files back to xml and submits the data to a BAPI
    Regards
    Sebastian

  • Some questions on whitespace and &

    Dear all,
    I have some questions on whitespace and & that need you kind help:
    1. Except /n /t /r and space, is there any characters that are whiteSpace characters?
    2. When parsing XML document, when the ignorableWhitespace() method is called? Will it be called from characters() ?
    e.g. Where the white space after <a> is called, and where is the white space between test1 and test 2 indied <b> is called?
    <a>
       <b>test1 test2</b>
    </a>  3. When the & should be escaped in a well-formed (or validated) XML file by replacing &?
    It should be excaped in any element content, such as <b>test1 & test2< /b> - except for <![CDATA[ section?
    It should NOT always be used in attributes? e.g. <a b="test1 & test2"/>  - has to do escaped by entity & - am I right?
    Thanks!
    Thanks
                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

    Thanks DrClap.
    >
    2. When parsing XML document, when the
    ignorableWhitespace() method is called? Will it be
    called from characters() ?It will be called when the document has a DTD and the
    parser is a validating one that is actually
    validating, when the parser sees whitespace that is
    not part of a DTD element. It will not be called from
    characters() because that is a method that you write.I wrote a contentHandler and using Xerces parser to parse an XML document. When I set validating feature to false, I found that for:
    <a>
       <b>test1 test2</b></a>The white space between <a> and <b> (including newline) is actually called from characters(), not in ignorableWhitespace(). So does this mean that if validating - white space is always called from ignorableWhitespace(), and if not validating, if it is always called from characters()?
    >
    3. When the & should be escaped in a well-formed(or
    validated) XML file by replacing &? The XML Recommendation says "The ampersand character
    (&) and the left angle bracket (<) may appear
    in their literal form only when used as markup
    delimiters, or within a comment, a processing
    instruction, or a CDATA section. If they are needed
    elsewhere, they must be escaped using either numeric
    character references or the strings "&" and
    "<" respectively."So in this sense, both & in <b>test1 & test2< /b> and <a b="test1 & test2"/> is illegal unless replacing them by & - right? What does the markup delimiters mena in above XML Recommendation ? Example?
    Thanks!

Maybe you are looking for