Some thoughts / questions on createVirtualCopies() LR5

I've been playing a little with the createVirtualCopies funciton in LR5.
First adding the createVirtualCopies() function is great!
Thought / wise: could there be a similar function in the LrPhoto class, so that based on a photo a Virtual copy is created?
Then I'm not dependant upon any active selection.
Some questions:
Why is it a catalog namespace?
Now as a plug-in developer I have to deal with selected photos and I would love to create a Virtual copy on a particular photo object as I like.
And what happens if the user selects a different photo when the plug-in is running?
What is the return value?
At least it is not a properly LrPhoto object.
I did a little test:
local LrDialogs = import 'LrDialogs'
local LrTasks = import 'LrTasks'
local LrFunctionContext = import 'LrFunctionContext'
local LrApplication = import 'LrApplication'
Name          inspect
Purpose          module for listing a table
Source          https://github.com/kikito/inspect.lua
local inspect = require 'inspect'
LrFunctionContext.postAsyncTaskWithContext( "CreateVirtualCopyTest", function( context )
    LrDialogs.attachErrorDialogToFunctionContext( context )
    local catalog = LrApplication.activeCatalog()
local activePhoto = catalog:getTargetPhoto()
LrDialogs.message( inspect (activePhoto:getDevelopSettings()) )
if activePhoto ~= nil then
local virtualCopy = catalog:createVirtualCopies( "CreateVirtualCopyTest" )
LrDialogs.message( inspect (virtualCopy:getDevelopSettings()) )
else
LrDialogs.message( "No photo selected! Please select a photo" )
return
end
end )
Any thoughts?
(Example based on code of Rob Cole.)

dhmc05 wrote:
Any thoughts?
Poor design (sorry Adobe).
There are a number of new functions which bypass the catalog write methods (which is convenient) but are tied to the Lr user interface, e.g. library view mode, or develop module, which is bad - especially bad since there is no way to:
A. Tell which module / mode Lr is in.
B. Force which module / mode to be in.
C. Assure photos are targeted, even if being filtered, or are underlings in a stack..
I hope somebody who knows what they're doing with the SDK sorts this mess out before releasing Lr5 SDK.
Rob

Similar Messages

  • Nokia X/X2 -- Some thoughts and questions

    Moved from Symbian Belle, some thoughts:
    1. Email widget doesn't automatically refresh. Have to manually tap to load new mail despite specifying a sync. interval.
    2. Asking me to Sign in to a MS, Exchange, Nokia or Google account etc. to use the calendar. Can I not get a local one for which I can choose to Sync if desired, similar to Symbian?
    3. Does the X2 have a play via radio capability? I don't see the app but wanted to if the hardware capability is there?
    4. How do I know the phone is using Gps, Glonass or Beido? Specifications listed positioning capabilities.
    5. Don't see the mix radio app. Possibly not available in certain countries.
    6. Belle had option to automatically switch between GSM and 3G where available. X2 only allows a 3G only setting and GSM/CDMA.
    7. People and Phone apps need to be bundled together. While scrolling through contacts via phone app, if you accidently tap one, it starts dialing. Plus it doesn't show items as together and repeats contacts with more than one number.
    8. Didn't really like the Metro and Fastlane UIs. One reason was that the wallpaper was not changeable. So if you have gaps, then all you see is black.
    9. Glance screen doesn't seem customizable. Time appears on 2 lines, seems a bit odd.

    Reply to 9:
    Glance Screen on X v2 is almost amazing. http://developer.nokia.com/resources/library/nokia-x-ui/essentials.html
    A past Nokia X user, now a X user.

  • I have a Macbook Pro june 2011... I have 8GB ram but I only have 256mb VRAM... I've read some other questions about this and I realized... Why do I not have 560mb of VRAM since I have 8GB of RAM? Is there any way to get more VRAM to play games on steam?

    I have a Macbook Pro june 2011... I have 8GB ram but I only have 256mb VRAM...
    I've read some other questions about this and I realized... Why do I not have 560mb of VRAM since I have 8GB of RAM?
    Is there any way to get more VRAM to play games on steam?
    I've learned  by reading other topics that I can't upgrade my graphics card on my Macbook Pro because it's soldered into the motherboard or somthing like that, but please tell me if there is any way to get more video ram by chaning some setting or upgrading something else. I'd also like to know why I only have 256MB of VRAM when I have 8GB of RAM, since I have 8GB of RAM I thought I was supposed to have 560mb of VRAM...
    So the two questions are...
    Is there any way to upgrade my VRAM, so that I can play games on steam?
    Why do I only have 256MB VRAM when I have 8GB total RAM?
    Other Info:
    I have a quad core i7 Processor.
    My graphcics card is the AMD Radeon HD 6490M.
    I am also trying to play games on my BOOTCAMPed side of my mac, or my Windows 7 Professional side.
    THANK YOU SO MUCH IF YOU CAN REPLY,
    Dylan

    The only two items that a user can change on a MBP are the RAM and HDD (Retinas not included).  You have what the unit came with and the only way you will be able to change that is to purchase a MBP with superior graphics
    If you are very much into gaming, the I suggest A PC.  They are far superior for that type of application to a MBP.
    Ciao.

  • Some Thoughts On An OWB Performance/Testing Framework

    Hi all,
    I've been giving some thought recently to how we could build a performance tuning and testing framework around Oracle Warehouse Builder. Specifically, I'm looking at was in which we can use some of the performance tuning techniques described in Cary Millsap/Jeff Holt's book "Optimizing Oracle Performance" to profile and performance tune mappings and process flows, and to use some of the ideas put forward in Kent Graziano's Agile Methods in Data Warehousing paper http://www.rmoug.org/td2005pres/graziano.zip and Steven Feuernstein's utPLSQL project http://utplsql.sourceforge.net/ to provide an agile/test-driven way of developing mappings, process flows and modules. The aim of this is to ensure that the mappings we put together are as efficient as possible, work individually and together as expected, and are quick to develop and test.
    At the moment, most people's experience of performance tuning OWB mappings is firstly to see if it runs set-based rather than row-based, then perhaps to extract the main SQL statement and run an explain plan on it, then check to make sure indexes etc are being used ok. This involves a lot of manual work, doesn't factor in the data available from the wait interface, doesn't store the execution plans anywhere, and doesn't really scale out to encompass entire batches of mapping (process flows).
    For some background reading on Cary Millsap/Jeff Holt's approach to profiling and performance tuning, take a look at http://www.rittman.net/archives/000961.html and http://www.rittman.net/work_stuff/extended_sql_trace_and_tkprof.htm. Basically, this approach traces the SQL that is generated by a batch file (read: mapping) and generates a file that can be later used to replay the SQL commands used, the explain plans that relate to the SQL, details on what wait events occurred during execution, and provides at the end a profile listing that tells you where the majority of your time went during the batch. It's currently the "preferred" way of tuning applications as it focuses all the tuning effort on precisely the issues that are slowing your mappings down, rather than database-wide issues that might not be relevant to your mapping.
    For some background information on agile methods, take a look at Kent Graziano's paper, this one on test-driven development http://c2.com/cgi/wiki?TestDrivenDevelopment , this one http://martinfowler.com/articles/evodb.html on agile database development, and the sourceforge project for utPLSQL http://utplsql.sourceforge.net/. What this is all about is having a development methodology that builds in quality but is flexible and responsive to changes in customer requirements. The benefit of using utPLSQL (or any unit testing framework) is that you can automatically check your altered mappings to see that they still return logically correct data, meaning that you can make changes to your data model and mappings whilst still being sure that it'll still compile and run.
    Observations On The Current State of OWB Performance Tuning & Testing
    At present, when you build OWB mappings, there is no way (within the OWB GUI) to determine how "efficient" the mapping is. Often, when building the mapping against development data, the mapping executes quickly and yet when run against the full dataset, problems then occur. The mapping is built "in isolation" from its effect on the database and there is no handy tool for determining how efficient the SQL is.
    OWB doesn't come with any methodology or testing framework, and so apart from checking that the mapping has run, and that the number of rows inserted/updated/deleted looks correct, there is nothing really to tell you whether there are any "logical" errors. Also, there is no OWB methodology for integration testing, unit testing, or any other sort of testing, and we need to put one in place. Note - OWB does come with auditing, error reporting and so on, but there's no framework for guiding the user through a regime of unit testing, integration testing, system testing and so on, which I would imagine more complete developer GUIs come with. Certainly there's no built in ability to use testing frameworks such as utPLSQL, or a part of the application that let's you record whether a mapping has been tested, and changes the test status of mappings when you make changes to ones that they are dependent on.
    OWB is effectively a code generator, and this code runs against the Oracle database just like any other SQL or PL/SQL code. There is a whole world of information and techniques out there for tuning SQL and PL/SQL, and one particular methodology that we quite like is the Cary Millsap/Jeff Holt "Extended SQL Trace" approach that uses Oracle diagnostic events to find out exactly what went on during the running of a batch of SQL commands. We've been pretty successful using this approach to tune customer applications and batch jobs, and we'd like to use this, together with the "Method R" performance profiling methodology detailed in the book "Optimising Oracle Performance", as a way of tuning our generated mapping code.
    Whilst we want to build performance and quality into our code, we also don't want to overburden developers with an unwieldy development approach, because what we'll know will happen is that after a short amount of time, it won't get used. Given that we want this framework to be used for all mappings, it's got to be easy to use, cause minimal overhead, and have results that are easy to interpret. If at all possible, we'd like to use some of the ideas from agile methodologies such as eXtreme Programming, SCRUM and so on to build in quality but minimise paperwork.
    We also recognise that there are quite a few settings that can be changed at a session and instance level, that can have an effect on the performance of a mapping. Some of these include initialisation parameters that can change the amount of memory assigned to the instance and the amount of memory subsequently assigned to caches, sort areas and the like, preferences that can be set so that indexes are preferred over table scans, and other such "tweaks" to the Oracle instance we're working with. For reference, the version of Oracle we're going to use to both run our code and store our data is Oracle 10g 10.1.0.3 Enterprise Edition, running on Sun Solaris 64-bit.
    Some initial thoughts on how this could be accomplished
    - Put in place some method for automatically / easily generating explain plans for OWB mappings (issue - this is only relevant for mappings that are set based, and what about pre- and post- mapping triggers)
    - Put in place a method for starting and stopping an event 10046 extended SQL trace for a mapping
    - Put in place a way of detecting whether the explain plan / cost / timing for a mapping changes significantly
    - Put in place a way of tracing a collection of mappings, i.e. a process flow
    - The way of enabling tracing should either be built in by default, or easily added by the OWB developer. Ideally it should be simple to switch it on or off (perhaps levels of event 10046 tracing?)
    - Perhaps store trace results in a repository? reporting? exception reporting?
    at an instance level, come up with some stock recommendations for instance settings
    - identify the set of instance and session settings that are relevant for ETL jobs, and determine what effect changing them has on the ETL job
    - put in place a regime that records key instance indicators (STATSPACK / ASH) and allows reports to be run / exceptions to be reported
    - Incorporate any existing "performance best practices" for OWB development
    - define a lightweight regime for unit testing (as per agile methodologies) and a way of automating it (utPLSQL?) and of recording the results so we can check the status of dependent mappings easily
    other ideas around testing?
    Suggested Approach
    - For mapping tracing and generation of explain plans, a pre- and post-mapping trigger that turns extended SQL trace on and off, places the trace file in a predetermined spot, formats the trace file and dumps the output to repository tables.
    - For process flows, something that does the same at the start and end of the process. Issue - how might this conflict with mapping level tracing controls?
    - Within the mapping/process flow tracing repository, store the values of historic executions, have an exception report that tells you when a mapping execution time varies by a certain amount
    - get the standard set of preferred initialisation parameters for a DW, use these as the start point for the stock recommendations. Identify which ones have an effect on an ETL job.
    - identify the standard steps Oracle recommends for getting the best performance out of OWB (workstation RAM etc) - see OWB Performance Tips http://www.rittman.net/archives/001031.html and Optimizing Oracle Warehouse Builder Performance http://www.oracle.com/technology/products/warehouse/pdf/OWBPerformanceWP.pdf
    - Investigate what additional tuning options and advisers are available with 10g
    - Investigate the effect of system statistics & come up with recommendations.
    Further reading / resources:
    - Diagnosing Performance Problems Using Extended Trace" Cary Millsap
    http://otn.oracle.com/oramag/oracle/04-jan/o14tech_perf.html
    - "Performance Tuning With STATSPACK" Connie Dialeris and Graham Wood
    http://www.oracle.com/oramag/oracle/00-sep/index.html?o50tun.html
    - "Performance Tuning with Statspack, Part II" Connie Dialeris and Graham Wood
    http://otn.oracle.com/deploy/performance/pdf/statspack_tuning_otn_new.pdf
    - "Analyzing a Statspack Report: A Guide to the Detail Pages" Connie Dialeris and Graham Wood
    http://www.oracle.com/oramag/oracle/00-nov/index.html?o60tun_ol.html
    - "Why Isn't Oracle Using My Index?!" Jonathan Lewis
    http://www.dbazine.com/jlewis12.shtml
    - "Performance Tuning Enhancements in Oracle Database 10g" Oracle-Base.com
    http://www.oracle-base.com/articles/10g/PerformanceTuningEnhancements10g.php
    - Introduction to Method R and Hotsos Profiler (Cary Millsap, free reg. required)
    http://www.hotsos.com/downloads/registered/00000029.pdf
    - Exploring the Oracle Database 10g Wait Interface (Robin Schumacher)
    http://otn.oracle.com/pub/articles/schumacher_10gwait.html
    - Article referencing an OWB forum posting
    http://www.rittman.net/archives/001031.html
    - How do I inspect error logs in Warehouse Builder? - OWB Exchange tip
    http://www.oracle.com/technology/products/warehouse/pdf/Cases/case10.pdf
    - What is the fastest way to load data from files? - OWB exchange tip
    http://www.oracle.com/technology/products/warehouse/pdf/Cases/case1.pdf
    - Optimizing Oracle Warehouse Builder Performance - Oracle White Paper
    http://www.oracle.com/technology/products/warehouse/pdf/OWBPerformanceWP.pdf
    - OWB Advanced ETL topics - including sections on operating modes, partition exchange loading
    http://www.oracle.com/technology/products/warehouse/selfserv_edu/advanced_ETL.html
    - Niall Litchfield's Simple Profiler (a creative commons-licensed trace file profiler, based on Oracle Trace Analyzer, that displays the response time profile through HTMLDB. Perhaps could be used as the basis for the repository/reporting part of the project)
    http://www.niall.litchfield.dial.pipex.com/SimpleProfiler/SimpleProfiler.html
    - Welcome to the utPLSQL Project - a PL/SQL unit testing framework by Steven Feuernstein. Could be useful for automating the process of unit testing mappings.
    http://utplsql.sourceforge.net/
    Relevant postings from the OTN OWB Forum
    - Bulk Insert - Configuration Settings in OWB
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=291269&tstart=30&trange=15
    - Default Performance Parameters
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=213265&message=588419&q=706572666f726d616e6365#588419
    - Performance Improvements
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=270350&message=820365&q=706572666f726d616e6365#820365
    - Map Operator performance
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=238184&message=681817&q=706572666f726d616e6365#681817
    - Performance of mapping with FILTER
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=273221&message=830732&q=706572666f726d616e6365#830732
    - Poor mapping performance
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=275059&message=838812&q=706572666f726d616e6365#838812
    - Optimizing Mapping Performance With OWB
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=269552&message=815295&q=706572666f726d616e6365#815295
    - Performance of mapping with FILTER
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=273221&message=830732&q=706572666f726d616e6365#830732
    - Performance of the OWB-Repository
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=66271&message=66271&q=706572666f726d616e6365#66271
    - One large JOIN or many small ones?
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=202784&message=553503&q=706572666f726d616e6365#553503
    - NATIVE PL SQL with OWB9i
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=270273&message=818390&q=706572666f726d616e6365#818390
    Next Steps
    Although this is something that I'll be progressing with anyway, I'd appreciate any comment from existing OWB users as to how they currently perform performance tuning and testing. Whilst these are perhaps two distinct subject areas, they can be thought of as the core of an "OWB Best Practices" framework and I'd be prepared to write the results up as a freely downloadable whitepaper. With this in mind, does anyone have an existing best practices for tuning or testing, have they tried using SQL trace and TKPROF to profile mappings and process flows, or have you used a unit testing framework such as utPLSQL to automatically test the set of mappings that make up your project?
    Any feedback, add it to this forum posting or send directly through to me at [email protected]. I'll report back on a proposed approach in due course.

    Hi Mark,
    interesting post, but I think you may be focusing on the trees, and losing sight of the forest.
    Coincidentally, I've been giving quite a lot of thought lately to some aspects of your post. They relate to some new stuff I'm doing. Maybe I'll be able to answer in more detail later, but I do have a few preliminary thoughts.
    1. 'How efficient is the generated code' is a perennial topic. There are still some people who believe that a code generator like OWB cannot be in the same league as hand-crafted SQL. I answered that question quite definitely: "We carefully timed execution of full-size runs of both the original code and the OWB versions. Take it from me, the code that OWB generates is every bit as fast as the very best hand-crafted and fully tuned code that an expert programmer can produce."
    The link is http://www.donnapkelly.pwp.blueyonder.co.uk/generated_code.htm
    That said, it still behooves the developer to have a solid understanding of what the generated code will actually do, such as how it will take advantage of indexes, and so on. If not, the developer can create such monstrosities as lookups into an un-indexed field (I've seen that).
    2. The real issue is not how fast any particular generated mapping runs, but whether or not the system as a whole is fit for purpose. Most often, that means: does it fit within its batch update window? My technique is to dump the process flow into Microsoft Project, and then to add the timings for each process. That creates a Critical Path, and then I can visually inspect it for any bottleneck processes. I usually find that there are not more than one or two dogs. I'll concentrate on those, fix them, and re-do the flow timings. I would add this: the dogs I have seen, I have invariably replaced. They were just garbage, They did not need tuning at all - just scrapping.
    Gee, but this whole thing is minimum effort and real fast! I generally figure that it takes maybe a day or two (max) to soup up system performance to the point where it whizzes.
    Fact is, I don't really care whether there are a lot of sub-optimal processes. All I really care about is performance of the system as a whole. This technique seems to work for me. 'Course, it depends on architecting the thing properly in the first place. Otherwise, no amount of tuning of going to help worth a darn.
    Conversely (re. my note about replacing dogs) I do not think I have ever tuned a piece of OWB-generated code. Never found a need to. Not once. Not ever.
    That's not to say I do not recognise the value of playing with deployment configuration parameters. Obviously, I set auditing=none, and operating mode=set based, and sometimes, I play with a couple of different target environments to fool around with partitioning, for example. Nonetheless, if it is not a switch or a knob inside OWB, I do not touch it. This is in line with my dictat that you shall use no other tool than OWB to develop data warehouses. (And that includes all documentation!). (OK, I'll accept MS Project)
    Finally, you raise the concept of a 'testing framework'. This is a major part of what I am working on at the moment. This is a tough one. Clearly, the developer must unit test each mapping in a design-model-deploy-execute cycle, paying attention to both functionality and performance. When the developer is satisifed, that mapping will be marked as 'done' in the project workbook. Mappings will form part of a stream, executed as a process flow. Each process flow will usually terminate in a dimension, a fact, or an aggregate. Each process flow will be tested as an integrated whole. There will be test strategies devised, and test cases constructed. There will finally be system tests, to verify the validity of the system as a production-grade whole. (stuff like recovery/restart, late-arriving data, and so on)
    For me, I use EDM (TM). That's the methodology I created (and trademarked) twenty years ago: Evolutionary Development Methodology (TM). This is a spiral methodology based around prototyping cycles within Stage cycles within Release cycles. For OWB, a Stage would consist (say) of a Dimensional update. What I am trying to now is to graft this within a traditional waterfall methodology, and I am having the same difficulties I had when I tried to do it then.
    All suggestions on how to do that grafting gratefully received!
    To sum up, I 'm kinda at a loss as to why you want to go deep into OWB-generated code performance stuff. Jeepers, architect the thing right, and the code runs fast enough for anyone. I've worked on ultra-large OWB systems, including validating the largest data warehouse in the UK. I've never found any value in 'tuning' the code. What I'd like you to comment on is this: what will it buy you?
    Cheers,
    Donna
    http://www.donnapkelly.pwp.blueyonder.co.uk

  • Some generic  questions about JMF

    Hi, as new to JMF, I would like to ask some generic questions about JMF from this community
    Though including RTP stack, JMF is only compatible to old RTP draft, RFC1889. I know that Sun stopped developing/supporting JMF. What is the alternative solution for JAVA RTP implementation suitable for a serious SERVER project, not just a school-term project? Has anyone thought about that or may have already moved forward? Is that any commercial or more updated Java based RTP stack currently? I searched with google. Besides jrtp, a school project, and I could not find a nice one. Majority of RTPstack implementation is C/C++.
    My second question is if there is the performance analysis and comparison of JMF. I know that some client applications such as SIP-Communicator adopt JMF for media delivery. But it is for one user in one machine. I wonder if any server application has adopt JMF as media relaying layer.
    I appreciate any input from this community.

    erwin2009 wrote:
    Hi, as new to JMF, I would like to ask some generic questions about JMF from this communitySure/
    Though including RTP stack, JMF is only compatible to old RTP draft, RFC1889. I know that Sun stopped developing/supporting JMF. What is the alternative solution for JAVA RTP implementation suitable for a serious SERVER project, not just a school-term project? Has anyone thought about that or may have already moved forward? Is that any commercial or more updated Java based RTP stack currently? I searched with google. Besides jrtp, a school project, and I could not find a nice one. Majority of RTPstack implementation is C/C++. There is one active project out there, Freedom for Media in Java, that has RTP capabilities. It's obviously not from Sun, but it's the only other RTP-capable project I am aware of, other than FOBS4Java, I believe is the name of it.
    What they are suitable for is beyond my scope of knowledge, I only know of the 2 projects.
    My second question is if there is the performance analysis and comparison of JMF. I know that some client applications such as SIP-Communicator adopt JMF for media delivery. But it is for one user in one machine. I wonder if any server application has adopt JMF as media relaying layer. None that I have seen, except for various projects I've helped people with on this forum. But, if someone is asking for help on a forum with his/her server application, it probably doesn't meet the guidelines you just laid out ;-)
    I appreciate any input from this community.Sorry I don't have more to add to your inquires...

  • Pro Tools convert to Logic Newbie some simple question for logic users

    Hey guys, I am a recent convert from Pro Tools! and There were just some simple questions I thought some of you may be able to help me out with. I am using Logic 9.1.1
    1) After I set up a multi timbral software instrument (in this case MOTU Symphonic Library) how do I set up midi channels to trigger the individual sounds in the instrument. (in pro tools I used to create a midi track then send an output of that track to that particular soft synth)
    2) How do I set up a and forgive my Pro Tools vocab but basically I want to set up a reverb channel on an aux track have it receive on a bus so I can send other tracks to that particular bus.
    Thanks for the insight and really looking forward to finally make use of all this ram! lol

    1) I have a nord stage and would like record via midi. In pro tools i would usually need to create a midi track with the output set the the nord and an audo aux track to hear it What would I do in order to record midi from my nord stage?
    To record, just select a non-audio track, hit record and play the keyboard. The notes will get recorded into a region.
    If you want to play it back to the Nord Stage as well, then it makes sense to set the Nord up as a MIDI device in your templates, so you can just add the Nord to a track and not worry about what MIDI ports it's on etc. Open the environment, create a New -> Instrument, called it "Nord Stage", set the MIDI port and MIDI channels it's on, and sav your template.
    Now, when you start a song, to sequence the Nord, you just assign the Nord instrument to a track (by for instance, right clicking on a track and choosing your Nord instrument from the list of environment obects that appears).
    2) The multi instrument trick is quick and easy its great but How would I do that manually?
    Lots of ways, look into your key commands, ther are all kinds of "Create track with same instrument", "Create track with next MIDI channel" and so on. There are too many ways to do this to quickly talk about here. In short, all tracks have an environment object assigned to them, whether it's an audio object, or a MIDI instrument object, and this determines where the data on the track goes.
    3) How do you set up midi so that it doen'ts overwrite notes when recording over them?
    Not sure I understand you here...

  • Discussions: Some Thoughts From Abroad

    Hi all,
    I would like to add some thoughts on the points debate. I think we all recogonise that the majority of users of Apple Discussions are in the USA, I don't have a problem with that, it's a fact of life. However, the new points system does have a bias towards those replying from the same country as the OP due to time zone differences.
    I don't ever see the day when I can imbibe in the Lounge with the 4s and 5s. By the time I get to my computer each day, all the questions I am able to answer have already been answered unless there are a few insomniacs around, or I become one myself!
    I liked the old system of awarding points. Not just because I was able to get a few myself, but mainly because I was able to give recognition to someone who had helped me expand my knowledge of Macs, even when I was not the OP or after the question had been answered. I realise that this system was open to abuse, but only by a few surely. Limiting the amount of votes one could award each day according to your Level would address that.
    In closing I would just like to say a big thankyou to all who keep these Discussions going. Your work is appreciated.
    Cheers
    NikkiG

    737/3183
    Hi NikkiG,
    "I liked the old system of awarding points. Not just because I was able to get a few myself, but mainly because I was able to give recognition to someone who had helped me expand my knowledge of Macs, even when I was not the OP or after the question had been answered. I realise that this system was open to abuse, but only by a few surely. Limiting the amount of votes one could award each day according to your Level would address that."
    And this possibility of awarding others was shaping a very different spirit:
    A lot of tips could be found among "side-replies" that were on topic but coming later, after the OP was already satisfied with the urgent fix.
    It was also shaping a spirit of sharing knowledge in a deeper, more elaborate way. To me this was really about a community of people who shared this kind of enthusiasm for good knowledge, for sense and for general improvement.
    Now that we live in immediateness, the "game" is much more about Shooting First, the one short and "simplified" answer that would, yes, quick-fix, but also kill most of the "side-discussion" as considered almost off-topic...
    "By the time I get to my computer each day, all the questions I am able to answer have already been answered"
    Well, the general feeling, anyway, no matter what time is it in Australia, is more like
    "By the time I'm done reading the question as it appears on top of a forum, one or two "Link-to-TheXLab"-type of quick answer have already been copy-pasted within the same minute"!
    (not only Donald but some others who are better at self-censorship, who manage equally well in the new system or who don't speak much because they don't know enough about true reasons/motivations, have left a word or two here, in their own way, about your feeling)
    "A bias towards those replying from the same country as the OP due to time zone differences"
    I see what you mean,
    although for me European at the contrary there is not anymore this period of disabling the "20% chances" thing that I was feeling in the Old Discussions
    you still can give quick replies to some newbies from Japan/New Zealand/Australia, but I agree there may not be as many as those from the US...
    Your suggestions:
    I second and support them, but I'm not sure, now that the system works again smoothly with no slowing down
    - Thanks so much Mods!! -
    whether they are very high anymore on the list of priorities?
    This depends also on the "political" point of viewing them. If with Tuttle we consider that "many of the new features mainly enhance the forums' usefulness to Apple"
    (my main personal guess about this aspect, would be about the gain in Phone Technical Support Crew management)
    possibly there will be no more changes about this?
    Possibly even more immediateness in the future,
    after all we are seriously entering the "iPod world" now, although let's hope the good old times of computing being a matter of a few Cupertino geniuses and a bunch of truly enthusiastic pure Mac people, is not too far gone now?
    Let me finish with your words, as they are exactly those I'd have liked to write myself:
    In closing I would just like to say a big thankyou to all who keep these Discussions going. Your work is appreciated.
    Axl

  • Some quick questions before purchase of MBP

    Hey Guys/Gals,
    I've been a PC user all my life, but will now be making a permanent switch to Macs. I've been wanting to buy a MBP since they have been released, but have been hesitant (like many others I think...) about the issues of whine, extreme heat, warped screens etc...
    With the recent updates of the logic board and even battery, I think I'm ready to go for it. I just had some quick questions before I place my order:
    1. I could order my MBP from the computer store at my university. Do you guys think that will get me the newest possible MBP or will they maybe get me one that they have had in stock for ages? I really don't want to get stuck with an outdated logic board and/or battery. Is it safest just to order directly from the site?
    2. Although the issue with really hot MBPs seems to have gotten better, I am still quite shocked that people say that temperatures well above 70 Celsius are "within spec". I understand that Apple officially states this and that the Intel chip can handle up to 100 Celsius, but this seems crazy to me Basically, I want to ask if I am correct in my observations that the new Intel chips don't get much hotter than other chips around, but rather Apple has chosen to set the temperature mark that sets off the fans pretty hight (around the 70s if I remember correctly?)
    I would assume that in following their "mission" of creating beautifully designed computers, Apple also wants to have a really silent one. Now I'm all for a quiet machine, but I feel as though if fans were on either (a) all the time or (b) when the CPU reaches the high 50s lots of people would be happier.
    Anyway, it's just a thought. I noticed there are lots of experts around here, so I'm just interested to hear what you guys think about this approach that Apple seems to have taken. I don't think the heat will stop me from getting this amazing notebook, but it certainly is a little frustrating.
    Any responses are greatly appreciated.

    1. I could order my MBP from the computer store at my university. Do you guys think that will get me the newest possible MBP or will they maybe get me one that they have had in stock for ages? I really don't want to get stuck with an outdated logic board and/or battery. Is it safest just to order directly from the site?
    No matter where you take possession there's no assurance of when the machine was manufactured. There is really just as good a chance of getting a perfectly good computer regardless of place of purchase. If there is a problem it's covered under warranty. You can have the computer replaced within the first two weeks. Afterwards they will replace or repair. Without Applecare's extended warranty the computer is warranted for one year after purchase.
    2. Although the issue with really hot MBPs seems to have gotten better, I am still quite shocked that people say that temperatures well above 70 Celsius are "within spec". I understand that Apple officially states this and that the Intel chip can handle up to 100 Celsius, but this seems crazy to me Basically, I want to ask if I am correct in my observations that the new Intel chips don't get much hotter than other chips around, but rather Apple has chosen to set the temperature mark that sets off the fans pretty hight (around the 70s if I remember correctly?)
    You are wrong in your assumption. The Core Duo's maximum temperature is 125C before shutdown. Normal operating temperatures according to Intel's own documentation is 50-70C under average load. MBPs under heavy load will typically reach core temperatures over 80C. Then there's the added heat from the hard drive and GPU. The aluminum case acts as a partial heat radiator hence the case temperatures can get high. The MBP User Guide specifically states this and issues a warning about using the computer on one's lap.
    I would assume that in following their "mission" of creating beautifully designed computers, Apple also wants to have a really silent one. Now I'm all for a quiet machine, but I feel as though if fans were on either (a) all the time or (b) when the CPU reaches the high 50s lots of people would be happier.
    In fact the fans in the MBP are very quiet at least mine are. Unless you are in the hardware design business with intimate knowledge of using Core Duo chips in portable computers, then your observations, although certainly your opinion, are not based on anything factual.
    Anyway, it's just a thought. I noticed there are lots of experts around here, so I'm just interested to hear what you guys think about this approach that Apple seems to have taken. I don't think the heat will stop me from getting this amazing notebook, but it certainly is a little frustrating.
    More likely there are a lot of people around here who think they are experts. As the remark goes, "An 'expert' derives from ex, meaning 'has been', and spert, meaning 'drip under pressure.'"
    You'll read lots of opinions on these forums. Don't take for granted that everything you read is factual or truth. A lot of the opinions expressed are without any basis in fact.

  • SGA Size for 8.1.7.4 32 bit? , some Interview Questions

    Hi buddies,
    I got some interview questions, might be simple for geeks in DBA. I am in need of answers. Could anyone help me.
    Thanks,
    Raaj
    1) Does windows NT support direct I/O?
    Answer: Choose one of the answers that apply
    A: No, only AIO
    B: Yes, depending on hardware.
    C: Yes.
    D: No.
    2) Can you take a coldbackup from solaris and use it on windows NT?
    Answer: Choose one of the answers that apply
    A: Yes.
    B: Yes if RMAN backup performed from NT server.
    C: Yes, after running RMAN convert.
    D: No.
    3) All of the following will alter the number of checkpoints that occur in one hour on the database, except one. Which is it?
    Answer: Choose one of the answers that apply
    A: Decreasing tablespace size
    B: Decreasing size of redo log members
    C: Setting LOG_CHECKPOINT_INTERVAL greater than the size of the redo log file
    D: Setting LOG_CHECKPOINT_TIMEOUT to zero
    4) The DBA is attempting to back up the Oracle database control file. After
    issuing the ALTER DATABASE BACKUP CONTROLFILE TO TRACE command, where can the DBA find the backup control file creation materials Oracle created for him or her ?
    Answer: Choose one of the answers that apply
    A: USER_DUMP_DEST
    B: LOG_ARCHIVE_DEST
    C: CORE_DUMP_DEST
    D: BACKGROUND_DUMP_DEST
    5) What is the most important action a DBA must perform after changing the database from NOARCHIVELOG TO ARCHIVELOG?
    Answer: Choose one of the answers that apply
    A: Shutdown normal and restart the database
    B: Perform a full logical database backup
    C: Perform a full offline database backup
    D: Manually switch the log files
    6) Which of the following choices lists an ALTER USER option that can be executed by the user herself or himself?
    Answer: Choose one of the answers that apply
    A: DEFAULT TABLESPACE
    B: IDENTIFIED BY
    C: TEMPORARY TABLESPACE
    D: PROFILE
    7) You need to view the initialization parameter settings for your Oracle
    database. Which of the following choices does not identify a method
    you can use to obtain values set for your initialization parameters?
    Answer: Choose one of the answers that apply
    A: Issue SELECT * FROM DBA_PARAMETERS; from SQL*Plus
    B: Issue SELECT * FROM V$PARAMETER; from SQL*Plus
    C: Issue SHOW PARAMETERS from Server Manager
    D: Use OEM Instance Manager
    8) As a result of a media failure, the current online redo log group is corrupted, the database crashes, as the current online group is inaccessible. Which type of incomplete recovery are you most likely to perform ?
    Answer: Choose one of the answers that apply
    A: Change-based
    B: Time-based
    C: Recovery using a backup control file
    D: Cancel-based
    9) User SNOW executes the following statement: SELECT * FROM EMP. This
    statement executes successfully, and SNOW can see the output. Table
    EMP is owned by user REED. What object would be required in order for
    this scenario to happen ?
    Answer: Choose one of the answers that apply
    A: User SNOW would need the role to view table EMP.
    B: User SNOW would need the privileges to view table EMP.
    C: User SNOW would need a synonym for table EMP.
    D: User SNOW would need the password for table EMP.
    10) Which one of the following statements is true?
    Answer: Choose one of the answers that apply
    A: The request queue is common, and the response queue is different for all the dispatchers.
    B: The request queue and response queue are different for all the dispatchers.
    C: The request queue is different, and response queue is common for all the dispatchers.
    D: The request queue and response queue are common for all the dispatchers.
    11) What is the largest SGA size for 8.1.7.4 32 bit?
    Answer: Choose one of the answers that apply
    A: approximately 2GB
    B: approximately 3.5GB
    C: approximately 4GB
    D: approximately 8GB
    E: approximately 16GB
    12) The DBA is about to perform some administrative tasks. Specifying the
    OPTIMAL parameter has which of the following appropriate uses?
    Answer: Choose one of the answers that apply
    A: Limiting concurrent users
    B: Limiting concurrent transactions
    C: Limiting growth of rollback segments
    D: Limiting growth of tables
    13) If the DBA wants to find information about how often transactions are
    wrapping transaction information between multiple rollback segment
    extents, where would the DBA look to find that information?
    Answer: Choose one of the answers that apply
    A: DBA_ROLLBACK_SEGS
    B: V$ROLLSTAT
    C: V$ROLLNAME
    D: DBA_SEGMENTS
    14) You have 30 rollback segments in your database, for which
    TRANSACTIONS_PER_ ROLLBACK_SEGMENT is set to 49 and
    TRANSACTIONS is set to 1000. During periods of heavy usage, about how many rollback segments will be actively used by Oracle?
    Answer: Choose one of the answers that apply
    A: 50
    B: 60
    C: 20
    D: 30
    15) The DBA has a table created with the following statement:
    CREATE TABLE EMPL
    (EMPID NUMBER(10),
    LASTNAME VARCHAR2(40),
    RESUME LONG RAW);
    The DBA attempts to issue the following statement:
    ALTER TABLE EMPL
    ADD ( PERF_APPRAISE LONG);
    What happens?
    Answer: Choose one of the answers that apply
    A: The statement succeeds.
    B: The statement succeeds, but column is added as VARCHAR2.
    C: The statement fails.
    D: The statement adds a disabled constraint.
    16) The primary key of the EMP table has three columns, EMPID, LASTNAME,
    and FIRSTNAME. You issue the following SELECT statement:
    SELECT * FROM EMP WHERE LASTNAME = 'HARRIS' AND FIRSTNAME = 'BILLI'
    AND EMPID = '5069493';
    Where would you look to see if this query will use the index associated
    with the primary key?
    Answer: Choose one of the answers that apply
    A: DBA_IND_COLUMNS
    B: DBA_TAB_COLUMNS
    C: DBA_INDEXES
    D: DBA_CLU_COLUMNS
    17) You are configuring your index to be stored in a tablespace. Which of the
    following storage parameters are not appropriate for indexes?
    Answer: Choose one of the answers that apply
    A: OPTIMAL
    B: INITIAL
    C: PCTINCREASE
    D: NEXT
    18) You need to set up auditing in an order entry and product shipment
    application so that when the ORDER_STATUS column in the ORDERS
    table changes to ‘SHIPPED’, a record is placed in a special table associated
    with a part of the application that gives sales representatives a daily list
    of customers to call on a follow-up to make sure the customer is satisfied
    with the order. Which of the following choices represents the best way
    to perform this auditing?
    Answer: Choose one of the answers that apply
    A: Statement auditing
    B: Object auditing
    C: Audit by access
    D: Value-based auditing
    19) Information in the buffer cache is saved back to disk in each of the
    following situations except one. In which situation does this not occur?
    Answer: Choose one of the answers that apply
    A: When a time-out occurs
    B: When a log switch occurs
    C: When the shared pool is flushed
    D: When a checkpoint occurs
    20) In order to allow remote administration of users and tablespaces on an Oracle database, which of the following types of files must exist in the database?
    Answer: Choose one of the answers that apply
    A: Password file
    B: Initialization file
    C: Datafile
    D: Control file
    E: Nothing, SYSDBA privileges are not required for these actions.
    21) You are planning the storage requirements for your database. Which of the following is an effect of maintaining a high PCTFREE for a table?
    Answer: Choose one of the answers that apply
    A: Oracle will manage filling data blocks with new records more actively.
    B: Oracle will manage filling data blocks with new records less actively.
    C: Oracle will leave more space free in data blocks for existing records.
    D: Oracle will leave less space free in data blocks for existing records.
    22) You manage database access privileges with roles where possible.
    You have granted the SELECT_MY_TABLE role to another role, called
    EMP_DEVELOPER. To view information about other roles that may be
    granted to EMP_DEVELOPER, which of the following dictionary views
    are appropriate?
    Answer: Choose one of the answers that apply
    A: DBA_ROLE_PRIVS
    B: DBA_TAB_PRIVS
    C: USER_SYS_PRIVS
    D: ROLE_ROLE_PRIVS
    23) In order to set your SQL*Plus session so that your NLS_DATE_FORMAT
    information is altered in a specific way every time you log into Oracle,
    what method would be used?
    Answer: Choose one of the answers that apply
    A: Setting preferences in the appropriate menu option
    B: Creating an appropriate LOGIN.SQL file
    C: Issuing the ALTER USER statement
    D: Issuing the ALTER TABLE statement
    24) You create a sequence with the following statement:
    CREATE SEQUENCE MY_SEQ
    START WITH 394
    INCREMENT BY 12
    NOMINVALUE
    NOMAXVALUE
    NOCACHE
    NOCYCLE;
    Two users have already issued SQL statements to obtain NEXTVAL, and
    four more have issued SQL statements to obtain CURRVAL. If you issue a
    SQL statement to obtain the NEXTVAL, what will Oracle return?
    Answer: Choose one of the answers that apply
    A: 406
    B: 418
    C: 430
    D: 442

    1.-
    2.c
    3.a
    4.a
    5.c
    6.b
    7.a
    8.d
    9.b
    10.a -
    11.a
    12.c
    13.b
    14.d
    15.c
    16.a -
    17.a
    18.d
    19.c
    20.a
    21. -
    22.d
    23.b
    24.?
    hope it helps u.
    Thanks
    Kuljeet

  • HT5312 So I have a problem with iTunes not letting me download anything without first responding to some security questions which I don't remember setting up, how can fix it? Oh, and it won't let me reset the questions either!

    So I have a problem with iTunes not letting me download anything without first responding to some security questions which I don't remember setting up, how can fix it? Oh, and it won't let me reset the questions either!

    If you mean that you aren't getting the reset link, then from the page that you posted from :
    Note: The option to send an email to reset your security questions and answers will not be available if a rescue email address is not provided. You will need to contact iTunes Store support in order to do so. 
    You can contact iTunes Support in your country via this page : http://support.apple.com/kb/HT5699
    When they've been reset you can then use the steps half-way down the HT5312 page that you posted from to add a rescue email address for potential future use

  • I have some problem when I want to download app in iTunes some security question ask me about my visa when I right my visa number I got an errors and on top of the page told me go to iTunes support I do these things and now what should I do

    I have some problem when I want to download app in iTunes some security question ask me about my visa when I right my visa number I got an errors and on top of the page told me go to iTunes support I do these things and now what should I do

    Most of the people on these forums, including myself, are fellow users - you're not talking to iTunes Support here.
    You can contact iTunes Support via this link : http://www.apple.com/support/itunes/contact/ - click on Contact iTunes Store Support on the right-hand side of the page, then Purchases, Billing & Redemption

  • I got a new ipad and went buy music on iTunes and in is telling me that this is a new device on this account and wants me to answer some security questions I dont remember setting thes up and dont know the answers how do i reset the security questions

    i got a new ipad and went buy music on iTunes and in is telling me that this is a new device on this account and wants me to answer some security questions I dont remember setting thes up and dont know the answers how do i reset the security questions

    Click here and search the article for '2 out of 3' without the quotes; this generally involves either a message being sent to your rescue email address or contacting the iTunes Store staff directly.
    (74313)

  • Some basic questions on File Adapter

    Hello all,
    I have some basic questions on XI and File Adapter and hope you can help me. Any answer is appreciated.
    1. Can I use NFS transport protocol to poll a file from a machine in the network, which is not the XI? Or do I have to use FTP instead?
    2. If I understand it correctly - when using the FTP-File Adapter, XI has the role of a ftp client. I have to run a ftp server on my distant machine. XI connects to FTP-Server and polls the file.
    Can it also be configured the other way round? The scenario I think of would be: FTP client installed on distant machine, which connects to FTP-Server(XI) and loads up a file. So XI would act as FTP Server.
    I know this works, if I install a ftp Server on the computer my XI runs on, and use the NFS-File Adapter to observe the folder. But I want to know, if I need a second, independant ftp server for this.
    3. And last but not least: When do I need the active ftp mode instead of passive?
    Thanx a lot for your answers!
    Ilona

    > Hello all,
    > I have some basic questions on XI and File Adapter
    > and hope you can help me. Any answer is appreciated.
    >
    >
    > 1. Can I use NFS transport protocol to poll a file
    > from a machine in the network, which is not the XI?
    <b>yes</b>
    > Or do I have to use FTP instead?
    >
    <b>also you can do it</b>
    > 2. If I understand it correctly - when using the
    > FTP-File Adapter, XI has the role of a ftp client. I
    > have to run a ftp server on my distant machine. XI
    > connects to FTP-Server and polls the file.
    > Can it also be configured the other way round? The
    > scenario I think of would be: FTP client installed on
    > distant machine, which connects to FTP-Server(XI) and
    > loads up a file. So XI would act as FTP Server.
    > I know this works, if I install a ftp Server on the
    > computer my XI runs on, and use the NFS-File Adapter
    > to observe the folder. But I want to know, if I need
    > a second, independant ftp server for this.
    >
    <b>XI cannot act as FTP server, but it is always a client. When XI is reading (File sender adpater) when XIis writing than it is File Receiver adapter</b>
    > 3. And last but not least: When do I need the active
    > ftp mode instead of passive?
    >
    <b>It depends on your firewall configuration. The best and the fastests is active mode but not always available.</b>
    > Thanx a lot for your answers!
    > Ilona

  • Some generic questions on BPM of NW CE

    Hi Gurus
      i have over 6 years of ABAP business workflow experience and over 3 years of XI experience. Now i'm getting tough with this new stuff, BPM of NW CE. Here i have some generic questions of it:
    1. How can i trigger a BPM task while a document, like a sales order , has been created in the system? is there any processes to catch the event of R3 in NWDI?
    2. Can BPM consume RFC/BAPIs in R3?
    3. Is it possible to make message mapping in BPM? i found there is a 'mapping' choice in the context setting of NWDI whereas it is quite simple compared with PI.
    4. Can the BPM be integrated with webdynpro ABAP? or send the relevant tasks into R3 directly? Actually what i mean is , can i bypass java programing while dealing with BPM of CE, if possible.
    thanks

    Hi Stephen,
    In the first released version of SAP NetWeaver BPM it is not possible to catch events of R/3 directly. Anyway. In case you have the possibility to trigger a web service in that situation you could start off a specific (BPM) process instance that handles the situation.
    The same is basically true for the question regarding RFC: In the initial version you'd need to wrap the RFC into a web service in order to consume it.
    The mapping possibilities in SAP NetWeaver BPM are meant to perform data flow between different activities in a BPMN based process model. This way you could transform a message that came into the process in that way that if fits your needs in regards to the BPM process (data) context.
    Bypassing Java when using SAP NetWeaver BPM would be difficult as the supported UI technology is Web Dynpro for Java. In addition the possibility to add more flexibility to your mappings by defining custom functions is based on EJB, thus again Java.
    Best regards,
    Martin

  • Some interview Question?

    hello all, as this forum has many brilliant minds, i have some interview question, if you mind please let me know the answer....
    q1. default level at which validation accurs?
    q2. in which case property pallet display **** as a property value, what its meaning?
    q3. best way to ensure that item can not accept query criteria?
    q4. what is the use of key-other trigger?
    q5. what is the use of forms module validation unit. if it is set to form then data block pre-text-item trigger at point will raise?
    i tried but did not find any relevant answer.

    Hi yash...
    q1. default level at which validation accurs?Well, the default level is where ur trigger exist ( Form Level ,Block Level or Item Level )
    q2. in which case property pallet display **** as a property value, what its meaning?Conceal ; it means hiding displaying data and it is almost used in password text item.
    q3. best way to ensure that item can not accept query criteria?Well, actually i don't have any forms now but u could try it urself item property > query criteria (any related query allowed property ) >NO
    q4. what is the use of key-other trigger?A Key-Others trigger fires when an operator presses the associated key.
    A Key-Others trigger is associated with all keys that can have key triggers associated with them but are not currently defined by function key triggers (at any level).
    A Key-Others trigger overrides the default behavior of a Runform function key (unless one of the restrictions apply). When this occurs, however, Oracle Forms still displays the function key's default entry in the Keys screen.
    Hope this helps...
    Good Luck :)
    Regards,
    Amatu Allah.

Maybe you are looking for