XDK -Performance best practices etc

All ,
Am looking for some best practices with specific emphasis on performance
for the Oracle XDK ..
can any one share any such doc or point me to white papers etc ..
Thanks

The following article discusses how to choose the most performant parsing strategy based on your application requirements.
Parsing XML Efficiently
http://www.oracle.com/technology/oramag/oracle/03-sep/o53devxml.html
-Blaise

Similar Messages

  • CE Benchmark/Performance Best Practice Tips

    We are in the early stages of starting a CE project where we expect a high volume of web service calls per day (e.g. customer master service, material master service, pricing service, order creation service etc).
    Are there any best-practice guidelines which could be taken into account to avoid any possible performance problems within the web service u201Cinfrastructureu201D? 
    Should master data normally residing in the backend ECC server be duplicated outside ECC? 
    e.g. if individual reads of the master data in the backend system take 2 seconds per call, would it be more efficient to duplicate     the master data on the SAP AS Java server, or elsewhere u2013 if the master data is expected to be read thousands of times each    day.
    Also, what kind of benchmarking tools (SAP std or 3rd party) are available to assess the performance of the different layers of the infrastructure during integration + volume testing phases?
    I've tried looking for any such documentation on SDN, OSS, help.sap.com, but to no avail.
    Many thanks in advance for any help.
    Ali Crawshaw

    Hi Ali,
    For performance and benchmarking have you had a look at Wiley Introscope?
    The following presentation has some interesting information [Wiley Introscope supports CE 7.1|http://www.google.co.za/url?sa=t&source=web&ct=res&cd=7&ved=0CCEQFjAG&url=http%3A%2F%2Fwww.thenewreality.be%2Fpresentations%2Fpdf%2FDay2Track6%2F265CTAC.pdf&ei=BUGES-yyBNWJ4QaN7KzXAQ&usg=AFQjCNE9qA310z2KKSMk4d42oyjuXJ_TfA&sig2=VD1iQvCUmWZMB5OB-Z4gEQ]
    With regards to best practice guidelines, if you are using PI for service routing try to keep to asynch services as far as possible, asynch with acknowledgments if need be. Make sure your CE Java AS is well tuned according to the SAP best practice.
    Will you be using SAP Global Data Types for your service development? If you are then the one performance tip i have regarding the use of GDT's is to keep your GDT structures as small (number of fields) as possible, as large GDT structures have an impact on memory consumption at runtime.
    Cheers
    Phillip

  • Design Patterns/Best Practices etc...

    fellow WLI gurus,
    I am looking for design patterns/best practices especially in EAI / WLI.
    Books ? Links ?
    With patterns/best practices I mean f.i.
    * When to use asynchronous/synchronous application view calls
    * where to do validation (if your connecting 2 EIS, both EIS, only in WLI,
    * what if an EIS is unavailable? How to handle this in your workflow?
    * performance issues
    Anyone want to share his/her thoughts on this ?
    Kris

              Hi.
              I recently bought WROX Press book Professional J2EE EAI, which discusses Enterprise
              Integration. Maybe not on a Design Pattern-level (if there is one), but it gave
              me a good overview and helped me make some desig decisions. I´m not sure if its
              technical enough for those used to such decisions, but it proved useful to me.
              http://www.wrox.com/ACON11.asp?WROXEMPTOKEN=87620ZUwNF3Eaw3YLdhXRpuVzK&ISBN=186100544X
              HTH
              Oskar
              

  • Reflection Performance / Best Practice

    Hi List
    Is reflection best practice in the followng situation, or should I head down the factory path? Having read http://forums.sun.com/thread.jspa?forumID=425&threadID=460054 I'm now wondering.
    I have a Web servlet application with a backend database. The servlet currently handles 8 different types of JSON data (there is one JSON data type for each table in the DB).
    Because JSON data is well structured, I have been able to write a simple handler, all using reflection, to dynamically invoke the Data Access Object and CRUD methods. So one class replaces 8 DAO's and 4 CRUD methods = 32 methods - this will grow as the application grows.
    Works brilliantly. It's also dynamic. I can add a new database table by simply subclassing a new DAO.
    Question is, is this best practice? Is there a better way? There are two sets of Class.forName(), newInstance(), getClass().getMethod(), invoke() ; one for getting the DAO and one for getting the CRUD method.....
    What is best practice here. Performance is important.
    Thanks, Len

    bocockli wrote:
    What is best practice here. Performance is important.I'm going to ignore the meat of your question (sorry, there are others who probably have better insights there) and focus on this point, because I think it's important.
    A best practice, when it comes to performance is: have clear, measurable goals.
    If your only performance-related goal is "it has to be fast", then you never know when you're done. You can always optimize some more. But you almost never need to.
    So you need to have a goal that can be verified. If your goal is "I need to be able to handle 100 update requests for Foo and 100 update requests for Bar and 100 read-only queries for Baz at the same time per second", then you have a definite goal and can check if you reached it (or how far away you are).
    If you don't have such a goal, then you'll be optimizing until the end of time and still won't be "done".

  • OWB Repository Performance, Best Practice

    Hi
    We are considering installing OWB repository in its own database, designed solely to the design repository to achieve maximum performance at the design center.
    Does anyone have knowledge of best practice in setting up the database to OWB repository? (db parameters, block size and so on).
    We are currently using Release 11.1.
    BR
    Klaus

    You can found all this informations in the documentation. Just here:
    http://download.oracle.com/docs/cd/B31080_01/doc/install.102/b28224/reqs01.htm#sthref48
    You will find all Initialization Parameters for the Runtime Instance and for the design instance.
    Success
    Nico

  • Performance best-practices?

    Does a program run slower if for each method invoked I declare and use a bunch of intermediate references/variables inside the method? or faster, if some of those references/variables were declared as members of the class that owns the method?

    Does a program run slower if for each method invoked I
    declare and use a bunch of intermediate
    references/variables inside the method? or faster, if
    some of those references/variables were declared as
    members of the class that owns the method?Theoretically, it would make thing run faster to declare everything on the class level, since the JVM would not have to allocate/deallocated memory for the temporary variables. However in practice, this method of development can actually be seriously detrimental to your physical health as
    1) You tear out your hair (if you have any left), blood pressure skyrockets, etc... as you try to maintain the code in the face of large numbers of static/global variables.
    2) The person who maintains this after you tracks you down and causes you seriously bodily injury.
    Unless the method-local variable initialization is expensive (i.e. database connections, large chunks of memory, etc...) , then the increase is negligible and not worth the fact that extracting local methods to class level in general makes your code:
    1) less robust/flexible
    2) not thread safe
    3) less cohesive
    4) harder to maintain
    For the certain items where it is expensive for allocation, then other data structures rather than class level instances can help out. (c.f. fly weight pattern, singleton pattern, object pooling). Be careful of haphazard and poorly considered scope changes. If you are experiencing performance problems or want to tune your code, chances are the bottleneck is NOT with method-local variables. As Donald Knuth says, premature optimization is the root of all evil.
    - N

  • Best practices Struts for tech. proj. leads

    baseBeans engineering won best training by readers of JDJ and published the first book on Struts called FastTrack to Struts.
    Upcoming class is live in NYC, on 5/2 from 7:30 AM to 1:00PM. We will cover db driven web site development, process, validation, tiles, multi row, J2EE security, DAO, development process, SQL tuning, etc.
    We will teach a project tech lead methods that will increase the productivity of his team and review best practices, so that they can benchmark their environment.
    Sign up now for $150, the price will be $450 soon as we get closer to the date (price goes up every few days). The web site to sign up on is baseBeans.net* .
    You will receive a lab/content CD when you sign up.
    Contact us for more details.
    ·     We preach and teach simple.
    ·     We use a very fast DAO DB Layer – with DAO side data cache
    ·     We use JSTL
    ·     We use a list backed Bean w/ DAO helper design for access to any native source and to switch out DAO.
    ·     We use J2EE security, container managed declarative authorization and authentication. (no code, works on any app. server).
    ·     Struts based Content Management System. A Struts menu entry like this:
    <Item name="About_Contacts"      title="About/Contacts"
    toolTip="About Us and Contact Info" page="/do/cmsPg?content=ABOUT" />
    passes to the action the parm of “about” which the DAO populates.
    You can peak at the source code at sourceforge.net/projects/basicportal or go to our site baseBeans.net. (16,000 downloads since Oct. 2002)
    Note that the baseBeans.net is using the Content Management System (SQL based) that we train on. (our own dog food)
    Note: We always offer money back on our public classes.
    Vic Cekvenich
    Project Recovery Specialist
    [email protected]
    800-314-3295
    <a href =”baseBeans.net”>Struts Training</a>
    ps:
    to keep on training, details, best practice, etc. sign up to this mail list:
    http://www.basebeans.net:8080/mailman/listinfo/mvc-programmers
    (1,000 + members)

    Hi,
    We use only Stateful release modes for application modules, defined in the action mappings in struts-config.xml exactly the same way as in your example. Stateful mode releases the module instance back to the pool and it can be reused by other sessions as well. However, all the code that uses the app modules and view objects, etc, must be written with the assumption that the module or the view object the code is operating on can be a different instance from the one in the previous request in the same session.
    The concept of BC4J is that this recycling of modules should be transparent for the users of the app modules, but this is not exactly the case. Some things are not passivated in the am's snapshots and are not activated in case of recycling, for example, custom view object properties or entries in the userData map (or at least were not in 9.0.5, I doubt this is changed in 10.1.2.) These are things that you have to manually passivate and activate if you use them to store some information that is relevant to a particular user session.
    All chances are that these strange things that you experience only occur in sessions that use recycled application modules, that is, there was passivation and subsequent activation of vo and am states. I have found it useful as a minimum to test the application with only 1 application module in the pool and at least 2 user sessions, constantly recycling this one am instance. Many of the problems that will surface in a real application usage only when there is a high load can be experienced in this artificial setup.

  • Best Practice / Solutions for using 11g DB+x86 or Small Computer to build iaas/paas?

    My customer wants to build their own iaas/paas using Oracle 11g DB, plus x86 or other small computer, running Linux or Solaris or Unix OS.
    Oracle Exadata is not feasible for them to use currently.
    Customer wants to know whether there are other customers have implemented their cloud solution based on these or not?
    If yes, would like to share the experience, presentation slides, best practices etc.
    Is there an Oracle email DL for asking this kind of question?
    Thanks,
    Boris

    Like Rick, I'm not aware of a specific "cloud implementors forum". Internally, Oracle has lots of material on implementing cloud, using any platform at all, although obviously we feel Engineered Systems are the most cost-effective solution for many customers. Are you interested in IaaS i.e. virtualised hardware, or PaaS i.e. DBaaS? They should not be confused, neither is required for the other, in fact, using IaaS to implement "DBaaS", as the OpenStack trove API attempts to do, is probably the most counter-productive way to go about it. Define the business-visible services you will be offering, and then design the most efficient means of supporting them. That way you gain from economies of scale, and set up appropriate management systems that address issues like patching, security, database virtualisation and so on.

  • What is the best practice to perform DB Backup on Sun Cluster using OSB

    I have a query on OSB 10.4.
    I want to configure OSB 10.4 on 2 Node Sun Cluster where the oracle database is running.
    When im performing DB backup, my DB backup job should not get failed if my node1 fails. What is the best practice to achieve this?

    Hi,
    Each Host that participates in an OSB administrative domain must also have some pre-configured way to resolve a host name to an IP address.Use DNS, NIS etc to do this.
    Specify cluster IP in OSB, so that OSB always looks for Cluster IP only instead of physical IPs of each node.
    Explanation :
    If it is 2-Node OR 4-Node, when Cluster software installed in these nodes we have to configure Cluster IP so that when one node fails Cluster IP will automatically move to the another node.
    This cluster IP we have to specify whether it is RMAN backup or Application JDBC connection. Failing to second node/another Node is the job of Cluster IP. So wherever we install cluster configuration we have to specify in all the failover places specify CLUSTER IP.
    Hope it helps..
    Thanks
    LaserSoft

  • Is there a Mac OS X manual/best practice/performance enhancement guide?

    Hi
    I just got all Mac'ed up and am pretty new to it. I've been a PC looser for ages, but at least I knew what I was doing with it! Although my Mac is super fast and efficient now, I am paranoid that, like all the PC's I've ever owned, this "new car smell" state will not last for ever, unless some maintenance is kept up. Is there any kind of manual or maybe a website out there that can tell me about best practice with Macs. Stuff like keeping the registry clean (like on a PC), the best way to remove applications (entirely!), managing memory for best performance, etc. I'd also like to know about stuff like how using non-Apple made/brand applications effects the integrity of the system. Basically just find out how Mac software works and how best to use it.
    Thanks in advance.

    Start with these:
    Switching from Windows to Mac OS X,
    Basic Tutorials on using a Mac,
    MacFixIt Tutorials,
    MacTips, and
    Switching to the Mac: The Missing Manual, Leopard Edition.
    For maintenance, see these:
    Macintosh OS X Routine Maintenance
    Mac OS X speed FAQ
    Maintaining OS X
    Additionally, *Texas Mac Man* recommends:
    Quick Assist.
    Welcome to the Switch To A Mac Guides, and
    A guide for switching to a Mac.

  • [SOLVED]xpra + winswitch - best practice due to /etc/X11/xorg.conf.d/*

    Running xpra causes some headaches.
    The application builds fine and installs just perfectly, the troubles comes from the default Xorg configuration is no longer stored in /etc/X11/xorg.conf, it's now stored and read from numerous files under /etc/X11/xorg.conf.d/ where just so happen to be my 5-nvidia.conf.
    Starting xpra via:
    xpra start :100
    Causes:
    X.Org X Server 1.14.3
    Release Date: 2013-09-12
    X Protocol Version 11, Revision 0
    Build Operating System: Linux 3.11.0-1-ARCH x86_64
    Current Operating System: Linux archie 3.11.2-1-ARCH #1 SMP PREEMPT Fri Sep 27 07:35:36 CEST 2013 x86_64
    Kernel command line: root=/dev/sda1 rw initrd=../initramfs-linux.img BOOT_IMAGE=../vmlinuz-linux
    Build Date: 13 September 2013 01:28:59PM
    Current version of pixman: 0.30.2
    Before reporting problems, check http://wiki.x.org
    to make sure that you have the latest version.
    Markers: (--) probed, (**) from config file, (==) default setting,
    (++) from command line, (!!) notice, (II) informational,
    (WW) warning, (EE) error, (NI) not implemented, (??) unknown.
    (++) Log file: "/home/torxed/.xpra/Xorg.:100.log", Time: Thu Oct 3 19:56:30 2013
    (++) Using config file: "/etc/xpra/xorg.conf"
    (==) Using config directory: "/etc/X11/xorg.conf.d"
    setversion 1.4 failed
    Initializing built-in extension Generic Event Extension
    Initializing built-in extension SHAPE
    Initializing built-in extension MIT-SHM
    Initializing built-in extension XInputExtension
    Initializing built-in extension XTEST
    Initializing built-in extension BIG-REQUESTS
    Initializing built-in extension SYNC
    Initializing built-in extension XKEYBOARD
    Initializing built-in extension XC-MISC
    Initializing built-in extension SECURITY
    Initializing built-in extension XINERAMA
    Initializing built-in extension XFIXES
    Initializing built-in extension RENDER
    Initializing built-in extension RANDR
    Initializing built-in extension COMPOSITE
    Initializing built-in extension DAMAGE
    Initializing built-in extension MIT-SCREEN-SAVER
    Initializing built-in extension DOUBLE-BUFFER
    Initializing built-in extension RECORD
    Initializing built-in extension DPMS
    Initializing built-in extension X-Resource
    Initializing built-in extension XVideo
    Initializing built-in extension XVideo-MotionCompensation
    Initializing built-in extension XFree86-VidModeExtension
    Initializing built-in extension XFree86-DGA
    Initializing built-in extension XFree86-DRI
    Initializing built-in extension DRI2
    Loading extension GLX
    (EE)
    Fatal server error:
    (EE) xf86OpenConsole: Cannot open /dev/tty0 (No such file or directory)
    (EE)
    (EE)
    Please consult the The X.Org Foundation support
    at http://wiki.x.org
    for help.
    (EE) Please also check the log file at "/home/torxed/.xpra/Xorg.:100.log" for additional information.
    (EE)
    (EE) Server terminated with error (1). Closing log file.
    2013-10-03 19:56:30,893
    2013-10-03 19:56:30,893 Xvfb command has terminated! xpra cannot continue
    2013-10-03 19:56:30,894
    2013-10-03 19:56:30,895 removing socket /home/torxed/.xpra/archie-100
    tl;dr: 5-nvidia.conf is trying to get loaded after xf86-video-dummy is loaded (which is the default for xpra).
    Now to the question: What's the best practice because there's no information about this issue, and i can't be the only one who's run in to this since the xorg.conf was removed. At the moment I move all my nvidia configs out of the way, start xpra and move them back in. This works for the better part as long as i execute xpra with my own scripts.
    But is there a better way (which doesn't perhaps include recreating xorg.conf since that appears to be something you should walk away from)?
    Last edited by Torxed (2013-10-04 10:29:35)

    Good thinking! (xpra is basically just screen but for x, or a X11 tunnel via SSH).
    After a really quick poking around and remembering that xpra really just uses xorg-server-xvfb and at the very bottom of /etc/xpra/xpra.conf there's a line that says:
    xvfb=xpra_Xdummy -dpi 96 -noreset -nolisten tcp +extension GLX +extension RANDR +extension RENDER -logfile ${HOME}/.xpra/Xorg.${DISPLAY}.log -config /etc/xpra/xorg.conf
    Appending -configdir appears to be working.
    Here are the full steps:
    [torxed@archie ~]$ sudo cp -r /etc/X11/xorg.conf.d /etc/xpra/
    [torxed@archie ~]$ sudo rm /etc/xpra/xorg.conf.d/*nvidia*
    [torxed@archie ~]$ sudo sed -i '$s/$/ -configdir \/etc\/xpra\/xpra.conf.d/' /etc/xpra/xpra.conf
    *nvidia* would have to match your graphic-driver-config that would cause xorg-server-xvbf to crash.
    I'll update the wiki and put a mention on the package about it because this either has to be standard or well known from the start because it really just breaks the package all together.
    Last edited by Torxed (2013-10-04 10:29:09)

  • Best practice to integrate the external(ERP or Database etc) eCommerce data in to CQ

    Hi Guys,
    I am refering to GEOMetrixx-Outdoors project for building eCommerce fucntionality in our project.
    Currently we are integrating with an ERP system to fetch the Product details.
    Now I need to store all the Product data from ERP system in to our CRX  under etc/commerce/products/<myproject> folder structure.
    Do I need to create a csv file structure as explained in the geometrixx-outdoors project  and place it exactly the way they have mentioned in the documentation? By doing this the csvimporter will import the data in to CRX and creates the Sling:folder and nt:unstructured nodes in to CRX?
    Please guide me  which is this best practice to integrate the external eCommerce data in to CQ system to build eCommerce projects?
    Are there any other best practices ?
    Your help in this regard is really appreciated.
    Thanks

    Hi Kresten,
    Thanks for your reply.
    I went through the eCommerce framework link which you sent.
    Can you get me few of the steps to utilise eCommerce framework to pull all the product information in to our CRX repository and also  how to synchronise between the ERP system and CRX data. Is that we have a scheduling mechanism to pull the data from our ERP system and synch it with CRX repository?
    Thanks

  • ASM on SAN datafile size best practice for performance?

    Is their a 'Best Practice' for datafile size for performance?
    In our current production, we have 25GB datafiles for all of our tablespaces in ASM on 10GR1, but was wondering what the difference would be if I used say 50GB datafiles? Is 25GB a kind of mid point so the data can be striped across multiple datafiles for better performance?

    We will be using Redhat Linux AS 4 update u on 64-bit AMD Opterons. The complete database will be on ASM...not the binarys though. All of our datafiles we have currently in our production system are all 25GB files. We will be using RMAN-->Veritas Tape backup and RMAN-->disk backup. I just didn't know if anybody out there was using smallfile tablespaces using 50GB datafiles or not. I can see that one of our tablespaces will prob be close to 4TB.

  • What is the best Practice to improve MDIS performance in setting up file aggregation and chunk size

    Hello Experts,
    in our project we have planned to do some parameter change to improve the MDIS performance and want to know the best practice in setting up file aggregation and chunk size when we importing large numbers of small files(one file contains one record and each file size would be 2 to 3KB) through automatic import process,
    below is the current setting in production:-
    Chunk Size=2000
    No. Of Chunks Processed In Parallel=40
    file aggregation-5
    Records Per Minute processed-37
    and we made the below setting in Development system:-
    Chunk Size=70000
    No. Of Chunks Processed In Parallel=40
    file aggregation-25
    Records Per Minute processed-111
    after making the above changes import process improved but we want to get expert opinion making these changes in production because there is huge number different between what is there in prod and what change we made in Dev.
    thanks in advance,
    Regards
    Ajay

    Hi Ajay,
    The SAP default values are as below
    Chunk Size=50000
    No of Chunks processed in parallel = 5
    File aggregation: Depends  largely on the data , if you have one or 2 records being sent at a time then it is better to cluster them together and send it at one shot , instead of sending the one record at a time.
    Records per minute Processed - Same as above
    Regards,
    Vag Vignesh Shenoy

  • Best practice to monitor 10gR3 OSB performance using JMX API?

    Hi guys,
    I need some advice on the best practice to monitor 10gR3 OSB performance using JMX API.
    Jus to show I have done my home work, I managed to get the JMX sample code from
    http://download.oracle.com/docs/cd/E13159_01/osb/docs10gr3/jmx_monitoring/example.html#wp1109828
    working.
    The following is the list of options I am think about:
    * Set up: I have a cluster of one 1 admin server with 2 managed servers, which managed server runs an instance of OSB
    * What I try to achieve:
    - use JMX API to collect OSB stats data periodically as in sample code above then save data as a record to a
         database table
    Options/ideas:
    1. Simplest approach: Run the modified version of JMX sample on the Admin Server to save stats data to database
    regularly. I can't see problems with this one ...
    2. Use WLI to schedule the Task of collecting stats data regularly. May be overkill if option 1 above is good for production
    3. Deploy a simple web app on Admin Server, say a simple servlet that displays a simple page to start/stop and configure
    data collection interval for the timer
    What approach would you experts recommend?
    BTW, the caveats os using JMX in http://download.oracle.com/docs/cd/E13159_01/osb/docs10gr3/jmx_monitoring/concepts.html#wp1095673
    says
         Oracle strongly discourages using this API in a concurrent manner with more than one thread or process. This is because a reset performed in
         one thread or process is not visible to another threads or processes. This caveat also applies to resets performed from the Monitoring Dashboard of
         the Oracle Service Bus Console, as such resets are not visible to this API.
    Under what scenario would I be breaking this rule? I am a little worried about its statement
         discourages using this API in a concurrent manner with more than one thread or process
    Thanks in advance,
    Sam

    Hi Manoj,
    Thanks for getting back. I am afraid configuring aggregation interval from Dashboard doesn't solve problem as I need to collect stats data of endpoint URI or in hourly or daily basis, then output to CSV files so line graphs can be drawn for chosen applications.
    Just for those who may be interested. It's not possible to use SQL to query database tables to extract OSB stats for a specified time period, say 9am - 5pm. I raised a support case already and the response I got back is 'No'.
    That means using JMX API will be the way to go :)
    Has anyone actually done this kind of OSB stats report and care to give some pointers?
    I am thinking of using 7 or 1 days as the aggregation interval set in Dashboard of OSB admin console then collects stats data using JMX(as described in previous link) hourly using WebLogic Server JMX Timer Service as described in
    http://download.oracle.com/docs/cd/E12840_01/wls/docs103/jmxinst/timer.html instead of Java's Timer class.
    Not sure if this is the best practice.
    Thanks,
    Regards,
    Sam

Maybe you are looking for

  • Web Start and Web Application

    I have a scenario as given below. I have a Central Server hosting a web application(JSPs, jars) in Jetty 6.0.0. I have many branch servers having Jetty 6.0.0 installed and the web application deployed. Now, I want to update the web application on all

  • Connecting older amp without HDMI to newer TV, Blu Ray, HD cable components with HDMI

    Hi - I have an older Yamaha RXV-600 receiver with no HMDI connectivity, plus a newer Sony HDTV, Samsung Blu Ray player and HD cable box, all of which have HDMI connectivity. Can someone suggest how I can "maximize" the connectivity between these comp

  • Apple ID's & ios 5

    What issues do you think I might have with my wife and I using the same Apple ID on both our phones with the release of ios 5?

  • How to prepare for the EBS R12 Supply Chain OCE Certification

    Hello everybody, I post this thread because I'm going to take the exams to get the certification as Oracle EBS R12 Supply Chain Certified Expert Consultant and I'd like to understand how to best prepare myself. Is there anyone who has already got thi

  • Purchasing Macbook Pro abroad

    Hi, I am currently living in South Korea and am looking to purchase a Macbook Pro 13" this month but am having a few troubles finding information that I need beacause of the language barrier. Anyone able to help please? 1) Will the 'Apple Cover' cove